WO2009157198A1 - 記録媒体、再生装置、記録装置、再生方法、記録方法、プログラム - Google Patents
記録媒体、再生装置、記録装置、再生方法、記録方法、プログラム Download PDFInfo
- Publication number
- WO2009157198A1 WO2009157198A1 PCT/JP2009/002905 JP2009002905W WO2009157198A1 WO 2009157198 A1 WO2009157198 A1 WO 2009157198A1 JP 2009002905 W JP2009002905 W JP 2009002905W WO 2009157198 A1 WO2009157198 A1 WO 2009157198A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- stream
- information
- subtitle
- playback
- display method
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims description 150
- 238000003860 storage Methods 0.000 claims description 29
- 230000008569 process Effects 0.000 claims description 25
- 238000012545 processing Methods 0.000 claims description 24
- 230000001360 synchronised effect Effects 0.000 claims description 10
- 238000001914 filtration Methods 0.000 claims description 6
- 238000000926 separation method Methods 0.000 claims 1
- 239000000872 buffer Substances 0.000 description 69
- 239000004065 semiconductor Substances 0.000 description 50
- 230000006870 function Effects 0.000 description 39
- 238000010586 diagram Methods 0.000 description 36
- 102100027667 Carboxy-terminal domain RNA polymerase II polypeptide A small phosphatase 2 Human genes 0.000 description 34
- 101000725947 Homo sapiens Carboxy-terminal domain RNA polymerase II polypeptide A small phosphatase 2 Proteins 0.000 description 34
- 238000009826 distribution Methods 0.000 description 25
- 238000007726 management method Methods 0.000 description 23
- 230000008859 change Effects 0.000 description 16
- 239000004973 liquid crystal related substance Substances 0.000 description 16
- 230000003287 optical effect Effects 0.000 description 14
- 238000012546 transfer Methods 0.000 description 14
- 239000000203 mixture Substances 0.000 description 13
- 239000008186 active pharmaceutical agent Substances 0.000 description 11
- 238000010187 selection method Methods 0.000 description 11
- 230000005540 biological transmission Effects 0.000 description 8
- 238000006243 chemical reaction Methods 0.000 description 8
- 230000009467 reduction Effects 0.000 description 7
- 238000011161 development Methods 0.000 description 6
- 230000018109 developmental process Effects 0.000 description 6
- 238000004519 manufacturing process Methods 0.000 description 6
- 238000001824 photoionisation detection Methods 0.000 description 6
- 230000003068 static effect Effects 0.000 description 6
- 238000004458 analytical method Methods 0.000 description 5
- 238000013461 design Methods 0.000 description 5
- 238000011069 regeneration method Methods 0.000 description 5
- 238000005516 engineering process Methods 0.000 description 4
- 239000000284 extract Substances 0.000 description 4
- 230000002093 peripheral effect Effects 0.000 description 4
- 230000008929 regeneration Effects 0.000 description 4
- 230000002123 temporal effect Effects 0.000 description 4
- 235000016496 Panda oleosa Nutrition 0.000 description 3
- 240000000220 Panda oleosa Species 0.000 description 3
- 238000009825 accumulation Methods 0.000 description 3
- 230000015572 biosynthetic process Effects 0.000 description 3
- 239000003086 colorant Substances 0.000 description 3
- 238000007906 compression Methods 0.000 description 3
- 230000006835 compression Effects 0.000 description 3
- 238000001514 detection method Methods 0.000 description 3
- 210000003128 head Anatomy 0.000 description 3
- 230000006872 improvement Effects 0.000 description 3
- 230000001172 regenerating effect Effects 0.000 description 3
- 206010047571 Visual impairment Diseases 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 238000005206 flow analysis Methods 0.000 description 2
- 230000002452 interceptive effect Effects 0.000 description 2
- 230000007246 mechanism Effects 0.000 description 2
- 238000005457 optimization Methods 0.000 description 2
- 230000008520 organization Effects 0.000 description 2
- 238000013468 resource allocation Methods 0.000 description 2
- 239000000758 substrate Substances 0.000 description 2
- 238000003786 synthesis reaction Methods 0.000 description 2
- LZDYZEGISBDSDP-UHFFFAOYSA-N 2-(1-ethylaziridin-1-ium-1-yl)ethanol Chemical compound OCC[N+]1(CC)CC1 LZDYZEGISBDSDP-UHFFFAOYSA-N 0.000 description 1
- TVZRAEYQIKYCPH-UHFFFAOYSA-N 3-(trimethylsilyl)propane-1-sulfonic acid Chemical compound C[Si](C)(C)CCCS(O)(=O)=O TVZRAEYQIKYCPH-UHFFFAOYSA-N 0.000 description 1
- 206010034719 Personality change Diseases 0.000 description 1
- 101100137868 Solanum lycopersicum PAD1 gene Proteins 0.000 description 1
- 206010053615 Thermal burn Diseases 0.000 description 1
- 230000001174 ascending effect Effects 0.000 description 1
- 230000008901 benefit Effects 0.000 description 1
- 230000033228 biological regulation Effects 0.000 description 1
- 230000003139 buffering effect Effects 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 238000012790 confirmation Methods 0.000 description 1
- 238000012937 correction Methods 0.000 description 1
- 238000005520 cutting process Methods 0.000 description 1
- 238000002296 dynamic light scattering Methods 0.000 description 1
- 238000009472 formulation Methods 0.000 description 1
- 239000012634 fragment Substances 0.000 description 1
- 238000010348 incorporation Methods 0.000 description 1
- 230000010354 integration Effects 0.000 description 1
- NUHSROFQTUXZQQ-UHFFFAOYSA-N isopentenyl diphosphate Chemical compound CC(=C)CCO[P@](O)(=O)OP(O)(O)=O NUHSROFQTUXZQQ-UHFFFAOYSA-N 0.000 description 1
- 230000009191 jumping Effects 0.000 description 1
- 230000033001 locomotion Effects 0.000 description 1
- 238000013507 mapping Methods 0.000 description 1
- 239000000463 material Substances 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000004806 packaging method and process Methods 0.000 description 1
- 229920003258 poly(methylsilmethylene) Polymers 0.000 description 1
- 238000013061 process characterization study Methods 0.000 description 1
- 239000011347 resin Substances 0.000 description 1
- 229920005989 resin Polymers 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 238000012216 screening Methods 0.000 description 1
- 230000011664 signaling Effects 0.000 description 1
- 239000007787 solid Substances 0.000 description 1
- 238000013519 translation Methods 0.000 description 1
- 238000002834 transmittance Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/10—Indexing; Addressing; Timing or synchronising; Measuring tape travel
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/156—Mixing image signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/172—Processing image signals image signals comprising non-image signal components, e.g. headers or format information
- H04N13/183—On-screen display [OSD] information, e.g. subtitles or menus
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/30—Image reproducers
- H04N13/356—Image reproducers having separate monoscopic and stereoscopic modes
- H04N13/359—Switching between monoscopic and stereoscopic modes
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/597—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/41—Structure of client; Structure of client peripherals
- H04N21/426—Internal components of the client ; Characteristics thereof
- H04N21/42646—Internal components of the client ; Characteristics thereof for reading from or writing on a non-volatile solid state storage medium, e.g. DVD, CD-ROM
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/47—End-user applications
- H04N21/488—Data services, e.g. news ticker
- H04N21/4884—Data services, e.g. news ticker for displaying subtitles
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/76—Television signal recording
- H04N5/84—Television signal recording using optical recording
- H04N5/85—Television signal recording using optical recording on discs or drums
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/79—Processing of colour television signals in connection with recording
- H04N9/80—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
- H04N9/82—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only
- H04N9/8205—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only involving the multiplexing of an additional signal and the colour video signal
- H04N9/8227—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only involving the multiplexing of an additional signal and the colour video signal the additional signal being at least another television signal
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B2220/00—Record carriers by type
- G11B2220/20—Disc-shaped record carriers
- G11B2220/21—Disc-shaped record carriers characterised in that the disc is of read-only, rewritable, or recordable type
- G11B2220/213—Read-only discs
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B2220/00—Record carriers by type
- G11B2220/20—Disc-shaped record carriers
- G11B2220/25—Disc-shaped record carriers characterised in that the disc is based on a specific recording technology
- G11B2220/2537—Optical discs
- G11B2220/2541—Blu-ray discs; Blue laser DVR discs
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/79—Processing of colour television signals in connection with recording
- H04N9/80—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
- H04N9/804—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components
- H04N9/806—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components with processing of the sound signal
- H04N9/8063—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components with processing of the sound signal using time division multiplex of the PCM audio and PCM video signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/79—Processing of colour television signals in connection with recording
- H04N9/80—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
- H04N9/82—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only
- H04N9/8205—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only involving the multiplexing of an additional signal and the colour video signal
Definitions
- the present invention is an invention belonging to the field of graphics subtitle display technology.
- “Graphics subtitles” is a technology that decodes graphics data compressed by so-called run-length encoding to display subtitles, and is used in BD-ROM playback devices, DVB-MHP, and DVD-Video. Graphics subtitles are generally provided for display by decoding a graphics subtitle stream.
- the graphics subtitle stream is an array of PES packets. Such PES packets include one storing graphics data, one storing pallet data, and one storing control data. It is length data, and is composed of a code value and a continuous length of the code value.
- Pallet data shows the correspondence between each code value and brightness and color difference.
- the control data includes a reference value indicating palette data to be used for the color conversion, information for designating a display area for writing graphics to the plane memory, and information on the display coordinates of the graphics in the plane memory, and the graphics based on the reference value Command display and color conversion.
- the age group of viewers who watch movie works using a playback device has a wide spread from elderly people to infants, and there are various variations in playback devices from low-priced models to high-end models. In the near future, it is said that a model that can switch between stereoscopic playback and planar playback will also appear. From the creators of movie works, according to the specs of these playback devices and the diversification of user groups There is a demand for realizing a unique caption display.
- control information incorporated in the graphics subtitle stream includes information for specifying the display area and display coordinates in the plane memory, the setting of the display coordinates in the control information and the range of the display area are changed.
- display effects such as scroll, wipe, cut-in, and cut-out can be realized.
- the type of control in such control information is limited to the display effect as described above, and there is no room for realizing a characteristic display according to the specifications of the playback device of the playback device and the diversification of the user layer. not exist.
- An object of the present invention is to provide a recording medium capable of realizing a change in subtitle display method according to the processing capability of the playback device and the user's age group while maintaining compatibility of the data structure of the graphics subtitle stream It is to be.
- a recording medium is a recording medium on which a subtitle stream to be reproduced in synchronization with a video stream and playlist information are recorded,
- the playlist information includes playback section information and a display method flag,
- the playback section information is information that defines in-time on the playback time axis of the video stream, and information that defines the playback section using information indicating the out time, and includes a stream information table,
- the display method flag indicates whether to enable control of the display method of captions in the playback section,
- the stream information table is characterized by instructing the playback apparatus which one of a plurality of subtitle streams should be selected when the display method is enabled in the playback section.
- the display method flag in the playlist information indicates whether or not the control of the subtitle display method in the playback section is enabled, and the stream information table in the playlist information enables the display method in the playback section.
- the content of the control information incorporated in the subtitle stream is the same as the conventional one.
- the playback device can select one of these various subtitle streams according to the configuration of the playback device.
- the subtitle display according to the specific display method can be executed by the playback device. Since the caption display is performed by the process of selecting the caption stream from a plurality of caption streams according to the effective display method, it is not necessary to cause fluctuations in the control information.
- the display method can be changed depending on whether stereoscopic playback is possible and how old the user is, so manufacturers can change the subtitle display method according to the configuration. Differentiate from other companies.
- FIG. 3 is a diagram illustrating a correspondence relationship between a physical unit of a BD-ROM and source packets that form one file extent.
- FIG. 3 is a diagram illustrating a plurality of numerical ranges that can be taken by a packet ID of a TS packet and a PES stream type of a TS packet having a packet ID in each numerical range in association with each other.
- positioning It is a figure which shows an example of the internal structure of the base view stream for stereoscopic vision, and the enhanced view stream. It is a figure which shows what kind of image
- It shows how the coordinates of the window in the graphics plane defined by window_horizontal_position and window_vertical_position and the coordinates of the graphics subtitle in the graphics plane defined by object_horizontal_position and object_vertical_position affect the stereoscopic caption.
- It is a figure which shows an example of a clip information file.
- It is a figure which shows the internal structure of an entry map table. Indicates registration of entry points using an entry map.
- It is a figure which shows the data structure of play list information.
- It is a figure which shows the internal structure of a subpath information table. This indicates what playback section is defined for the left view and right view.
- FIG. 1 It is a figure which shows the internal structure of a video stream number table.
- the internal structure of the subtitle stream information table in STN_table is shown.
- the PSR set 12 and the internal structure of the reproduction control engine 14 are shown.
- 3 is a diagram illustrating an internal configuration of a planar view control unit 41.
- the internal structure of the stereoscopic control unit 42 is shown. It is a figure which shows the internal structure of a graphics decoder. It is a flowchart which shows the process sequence of a playlist reproduction process. It is a flowchart which shows the reproduction
- FIG. 10 is a flowchart showing a procedure for setting PSR2 when a device state changes and when a stream change is requested. It is a flowchart which shows the selection procedure at the time of reproduction according to age. It is a flowchart which shows the process sequence of the selection procedure at the time of stereoscopic vision reproduction
- FIG. 1 is a diagram showing a form of usage of a recording medium and a playback device.
- a BD-ROM 101 and a playback device 102 which are examples of a recording medium constitute a home theater system together with a television 103, a liquid crystal shutter goggles 104, and a remote controller 100, and are used by a user.
- BD-ROM 101 supplies, for example, a movie work to the home theater system.
- the playback device 102 is connected to the television 103 and plays back the BD-ROM 101.
- the television 103 provides an interactive operation environment to the user by displaying a playback image of a movie work or displaying a menu or the like.
- the liquid crystal shutter goggles 104 are composed of a liquid crystal shutter and a control unit, and realize stereoscopic viewing using parallax in both eyes of the user.
- the liquid crystal shutter of the liquid crystal shutter goggles 104 is a shutter using a liquid crystal lens having a property that the light transmittance is changed by changing an applied voltage.
- the control unit of the liquid crystal shutter goggles 104 receives the synchronization signal for switching the output of the right-view image and the left-view image sent from the playback device, and in accordance with this synchronization signal, the first state and the second state Switch.
- FIG. 1B shows the first state.
- the first state is a state in which the applied voltage is adjusted so that the liquid crystal lens corresponding to the right view does not transmit light, and the applied voltage is adjusted so that the liquid crystal lens corresponding to the left view transmits light. In this state, the left-view image is used for viewing.
- FIG. 1 (c) shows the second state.
- the second state is a state in which the applied voltage is adjusted so that the liquid crystal lens corresponding to the right view transmits light, and the applied voltage is adjusted so that the liquid crystal lens corresponding to the left view does not transmit light.
- the liquid crystal shutter goggles can provide a right-view image for viewing.
- the short time interval may be a time enough to give an illusion that a person looks three-dimensional by the switching display described above.
- the remote controller 100 is a device that accepts an operation on a hierarchical GUI from a user. To accept such an operation, the remote controller 100 moves a menu key for calling a menu constituting the GUI and a focus of a GUI component constituting the menu. An arrow key, a determination key for performing a confirmation operation on a GUI component constituting the menu, a return key for returning a hierarchical menu to a higher level, and a numerical key are provided.
- FIG. 2 shows an internal configuration of an optical disc, more specifically a BD-ROM, which is an example of a recording medium described in the present embodiment.
- the first level shows a BD-ROM which is a multilayered optical disc
- the second level shows a spiral track extending on each recording layer extended in the horizontal direction. This spiral track is treated as one continuous recording area.
- the recording area includes a lead-in located on the innermost circumference, a lead-out located on the innermost circumference, a recording area of the first recording layer, a recording area of the second recording layer, and a recording area of the third recording layer existing therebetween. Consists of
- the third row shows the file system area in BD-ROM.
- the file system area is composed of a “volume management area” and a “logical address space”.
- the “volume management area” stores file system management information for handling the recording area of the first recording layer, the recording area of the second recording layer, and the recording area of the third recording layer as one continuous file system space. It is an area.
- Logical address space is an address space indicated by a logical block number (LBN) in which sectors continue. That is, the recording area of the first recording layer, the recording area of the second recording layer, and the recording area of the third recording layer in the second stage constitute one continuous logical address space.
- LBN logical block number
- the fourth row shows area allocation in the logical address space of the file system management area.
- a non-AV data recording area exists on the inner circumference side.
- An AV data recording area exists immediately after the non-AV data recording area.
- the fifth row shows extents recorded in these non-AV data recording area and AV data recording area.
- extents EXT, EXT, EXT... In the figure
- non-AV data recording area there are extents (EXT, EXT, EXT... In the figure) constituting files other than AV files.
- Fig. 3 is a diagram showing the application format of BD-ROM.
- BDMV directory is a directory in which data such as AV contents and management information handled by the BD-ROM is recorded. Under the BDMV directory, there are five subdirectories called “JAR directory”, “BDJO directory”, “PLAYLIST directory”, “CLIPINF directory”, and “STREAM directory”. In the BDMV directory, “index.bdmv” ”And“ MovieObject.bdmv ”are arranged.
- index.bdmv is management information relating to the entire BD-ROM, and after the disc is inserted into the playback device, the index.bdmv is read first so that the disc is uniquely recognized by the playback device.
- index.bdmv indicates the correspondence between the title numbers of a plurality of titles that can be reproduced on the BD-ROM and BD-J objects or movie objects that define individual titles.
- “MovieObject.bdmv” stores one or more movie objects.
- a movie object is a management object that defines the control procedure to be performed by the playback device in an operation mode (HDMV mode) mainly controlled by a command interpreter.
- HDMV mode operation mode
- One or more commands, GUI menu calls, and title calls can be used by the user.
- JAR directory is a directory in which JAR files corresponding to archive files are placed.
- An archive file is a file obtained by combining one or more class files, one or more data files, etc. into one.
- One or more class files, one or more data files, and the like can be combined into one by, for example, an archiver (not shown).
- Java (registered trademark) archive file will be described as an example of an archive file.
- BD-J mode mainly controlled by a Java virtual machine that is a byte code interpreter provided in the playback device
- the control procedure to be performed by the playback device is specified.
- a file storing a JAR file is identified by a 5-digit number zzzzz and an extension jar.
- the “BDJO directory” is a file that stores a management object (BDJ object) that defines the control procedure to be performed by the playback device in an operation mode (BD-J mode) mainly controlled by a Java virtual machine that is a bytecode interpreter.
- BDJ object management object
- a file storing a BDJ object is identified by a 5-digit number zzzzz and an extension bdjo.
- a file storing playlist information including main path information specifying a playback section for a base-view video stream and sub-path information specifying a playback section for an enhanced view video stream is arranged.
- the file storing the playlist information is identified by a 5-digit identification number “yyyyy” and an extension “mpls”.
- the base-view video stream is a video stream that constitutes a left view or a right view, and can realize planar display.
- a video stream that constitutes a right view or a left view and is not a base view video stream is referred to as an “enhanced view video stream”.
- the picture data constituting the enhanced view video stream is compression-encoded based on the frame correlation with the picture data constituting the base view video stream.
- MVC Multiview Video Coding
- the stream that makes up the base view and enhanced view is not limited to the video stream.
- the subtitle stream also forms a base view and an enhanced view.
- the base view video stream and the base view subtitle stream are collectively referred to as a “base view stream”.
- the enhanced view video stream and the enhanced view subtitle stream are collectively referred to as an “enhanced view stream”.
- CLIPINF directory is a directory in which files (clip information files) storing clip information are placed.
- the clip information file is identified by a 5-digit identification number “xxxxx” and an extension “clpi”. Inside the clip information file, there are entry maps corresponding to the left-view video stream and the right-view video stream.
- Extents that make up the files in the above directories are recorded in the non-AV data area.
- “STREAM directory” is a directory in which an AV clip file storing a stereoscopic video stream and an AV clip file storing a stereoscopic video stream are arranged.
- a file storing a planar video stream is identified by a 5-digit identification number “xxxxx” and an extension “m2ts”.
- a file storing a stereoscopic video stream is identified by a 5-digit identification number “xxxxx” and an extension “ilts”.
- the extents constituting the base view stream file stored in the STREAM directory and the extents constituting the enhanced view stream file to be stored in the STREAM directory are recorded in the AV data recording area.
- FIG. 4 shows the process by which each source packet constituting the base view stream and the enhanced view stream is written into the AV data area.
- the first level in the figure shows TS packets that constitute a base view stream or an enhanced view stream.
- TS_extra_header (hatched portion in the figure) as shown in the second row to become a source packet with a length of 192 bytes.
- This TS_extra_header includes Arrival_Time_Stamp indicating the decoder input time information of the TS packet.
- the source packets that make up the base view stream and the enhanced view stream form one or more “ATC sequences”.
- the “ATC sequence” is an array of source packets constituting the time axis of the ATS, and means that there is no discontinuity (no arrival time-base discontinutiy) in the Arrival_Time_Clock referenced by the Arrival_Time_Stamp.
- a source packet sequence having continuity in the Arrival_Time_Clock referenced by the Arrival_Time_Stamp is referred to as an “ATC sequence”.
- ATS is added to the head of the TS packet as follows, and indicates the transfer time to the decoder.
- the ATC sequence becomes an AV clip and is recorded on the recording layer with the file name xxxxx.m2ts.
- Such an AV clip is divided into one or more file extents and recorded in an area on each recording layer like a normal computer file.
- the third row shows the AV clip
- the fourth row schematically shows how the AV clip is recorded on each recording layer.
- Each file extent constituting the file in the fourth level has a data length equal to or larger than a predetermined size (this size is referred to as S_EXT).
- FIG. 5 is a diagram showing the correspondence between the physical unit of the BD-ROM and the source packets constituting one file extent.
- a plurality of sectors are formed in the AV file recording area of the BD-ROM.
- the source packets constituting the file extent are grouped every 32 and written into three consecutive sectors.
- Thirty-two source packets stored in three sectors are called "Aligned Unit", and writing to the BD-ROM is performed in units of Aligned Unit.
- error correction codes are added in units of 32 sectors, and an ECC block is configured.
- the playback device can obtain 32 complete source packets as long as the BD-ROM is accessed in units of Aligned Unit. The above is the process of writing the AV clip to the BD-ROM.
- FIG. 6A is a diagram in which a plurality of numerical ranges that can be taken by the packet ID (PID) of the TS packet and the PES stream types of the TS packet having the packet ID of each numerical range are associated with each other.
- PID packet ID
- TS packets having a packet ID of 0x0100 constitute a program map (Program_map), and TS packets having a packet ID of 0x1001 constitute a program clock reference (PCR).
- Program_map program map
- PCR program clock reference
- TS packets having a packet ID of 0x1011 constitute a base-view video stream
- Ox1012 TS packets constitute an enhanced-view video stream
- TS packets with packet IDs of 0x1100 to 0x111F constitute an audio stream.
- a TS packet having a packet ID of 0x1220 to x123F forms a base view subtitle stream.
- TS packets having packet IDs of 0x1240 to 0x125F constitute an enhanced view subtitle stream.
- the packet IDs of TS packets constituting a graphics subtitle stream for planar view that cannot be a base-view subtitle stream are in the numerical range of 0x1200 to 0x121F.
- TS packets that make up the video stream and TS packets that make up the subtitle stream are grouped into those that make up the base view and those that make up the enhanced view.
- FIG. 6B shows an example.
- the group of source packets that make up the base view is the source packet of the base view video stream (Video in the figure) to which the PID of 0x1011 is assigned, and the source of the audio stream to which the PID of 0x1100 is assigned.
- a packet (Audio in the figure) and a graphics subtitle stream source packet (PG in the figure) to which PIDs of 0x1220, 0x1221,0x1222,0x1223, 0x1224, 0x1225, and 0x1226 are assigned.
- the source packet group constituting the enhanced view includes the source packet of the enhanced view video stream (Video in the figure) to which the PID of 0x1012 is assigned, and the source packet of the audio stream to which the PID of 0x1101 is assigned (in the figure). Audio), 0x1240, 0x1241,0x1242, 0x1243, 0x1244, and 0x1245 PIDs of graphics subtitle streams to which source packets (PG in the figure) are configured.
- FIG. 7 is a diagram illustrating an example of an interleaved arrangement.
- the interleaved arrangement means that the extents that make up the base view and enhanced view are recorded with the regularity of "base view”, “enhanced view”, “base view”, “enhanced view” ... It is that you are.
- the first row shows the AV file
- the second row shows the extents EXT_L [i], EXT_L [i + 1], EXT_R [i], and EXT_R [i + 1] constituting the AV file.
- the third level shows the source packet sequence belonging to each extent
- the fourth level shows the sector sequence in the recording layer.
- variables i and i + 1 in parentheses indicate the number of extents to be reproduced.
- Dashed arrows h1, h2, h3, and h4 indicate the belonging relationship that the extents EXT_L [i] and EXT_L [i + 1] belong to either the base view stream or the enhanced view stream.
- the belonging relationship indicated by the arrows h1 and h2 it can be seen that the extents EXT_L [i] and EXT_L [i + 1] belong to the base view stream.
- the arrows h3 and h4 it can be seen that the extents EXT_R [i] and EXT_R [i + 1] belong to the enhanced view stream.
- SEXT_L and SEXT_R are alternately read out into two buffers, a right-view read buffer and a left-view read buffer, and provided to a video decoder in the playback apparatus.
- the sizes of SEXT_L and SEXT_R need to be determined in consideration of the time required for buffering the right-view read buffer and the left-view read buffer.
- the transfer rate to the read buffer for right view is Rmax1
- Right-view read buffer Rmax1 x "Time to fill left-view read buffer with jump”
- the capacity of the read buffer for right view must be determined so as to satisfy the relationship.
- jumping is synonymous with disc seeking. This is because the continuous area that can be secured for recording in the BD-ROM is limited, and the base view stream and the enhanced view stream are not necessarily recorded side by side, and may be recorded in a skipped area. It is.
- TS packet accumulation in the left-view read buffer is performed at a transfer rate of Rud-Rmax2. This means the difference between the output rate Rmax2 from the left-view read buffer and the input rate Rud to the left-view read buffer. Then, the time for filling the left-view read buffer becomes RB2 / (Rud-Rmax2).
- the size of the view read buffer RB1 is RB1 ⁇ Rmax1 ⁇ ⁇ 2 ⁇ Tjump + RB2 / (Rud-Rmax2) ⁇ become.
- the left view read buffer capacity RB2 is calculated.
- RB2 ⁇ Rmax2 ⁇ ⁇ 2 ⁇ Tjump + RB1 / (Rud-Rmax1) ⁇ become.
- the specific values of the memory size of the right-view read buffer and the left-view read buffer are 1.5 Mbytes or less.
- the extent sizes SEXT_R and SEXT_L are the right-view read buffer and the left-view read. It is set to the same size as the buffer size or approximately the same size. This completes the description of how the base view stream and the enhanced view stream are recorded. Next, the internal configuration of the base view stream and the enhanced view stream will be described.
- FIG. 8 is a diagram illustrating an example of an internal configuration of a base view stream and an enhanced view stream for stereoscopic viewing.
- the base view stream and the enhanced view stream include, for example, picture data.
- picture data There are a plurality of types of picture data, including picture data such as I picture, P picture, and B picture.
- An I picture is picture data for one screen.
- a P picture is picture data indicating a difference from a reference I picture.
- a B picture is picture data generated by a standard I picture and P picture.
- the second level in the figure shows the internal structure of the base view stream.
- This stream includes picture data of picture data I1, P2, Br3, Br4, P5, Br6, Br7, and P9.
- the picture data are decoded according to DTS (decoding time stamp: information indicating the start time of decoding by the decoder).
- the first level shows a left view image.
- the decoded picture data I1, P2, Br3, Br4, P5, Br6, Br7, P9 are played in the order of I1, Br3, Br4, P2, Br6, Br7, P5 according to PTS, and the left view The image will be played back.
- the fourth level shows the internal structure of the enhanced view stream.
- This secondary video stream includes picture data P1, P2, B3, B4, P5, B6, B7, and P8. These picture data are decoded according to DTS.
- the third row shows a right view image.
- the decoded picture data P1, P2, B3, B4, P5, B6, B7, and P8 are converted into P1, B3, B4, P2, and P2 according to PTS (presentation time stamp: information indicating video / audio presentation time). By reproducing in the order of B6, B7, and P5, the right-view image is reproduced.
- the fifth row shows how the state of the liquid crystal shutter goggles 104 is changed. As shown in the fifth row, the right-view liquid crystal shutter is closed when viewing the left-view video, and the left-view liquid crystal shutter is closed when viewing the right-view video.
- These primary video stream and secondary video stream are compressed by inter-picture predictive coding using inter-view redundancy in addition to inter-picture predictive coding using temporal redundancy.
- the pictures in the enhanced view stream are compressed with reference to the pictures at the same display time in the base view stream.
- the first P picture of the enhanced view stream refers to the I picture of the base view stream
- the B picture of the enhanced view stream refers to the Br picture of the base view stream
- the second P picture of the enhanced view stream is Referring to the P picture of the base view stream.
- FIG. 9 is a diagram showing what video is used for reproduction by switching the light transmission / shading of goggles according to the timing of FIG.
- the frame display period is 1/24 second, and if the translucency / shading of the right view and the left view in the goggles is changed every 1/48 seconds, the pictures of the right view and the left view are alternately displayed. Will appear.
- the left view image and the right view image shown in FIG. 9 schematically show that the orientation and position of the face of the person appearing in the image are slightly shifted between the left view image and the right view image. (Incidentally, deviations in the orientation and position of a person's face in FIGS. 9 and 10 are schematic.)
- FIG. 10 is a diagram showing a stereoscopic image formed by the afterimage reaction of the eyes.
- FIG. 11 is a diagram illustrating a configuration of a graphics subtitle stream.
- the first row in FIG. 11A shows a TS packet sequence constituting an AVClip.
- the second level shows a PES packet sequence that constitutes a graphics subtitle stream.
- the PES packet sequence in the second level is configured by extracting the payload from the TS packets having a predetermined PID among the TS packets in the first level and concatenating them.
- the third row shows the structure of the graphics subtitle stream.
- the graphics subtitle stream is composed of functional segments called PCS (Presentation, Composition, Segment), WDS (Window, Definition, Segment), PDS (Palette, Definition, Segment), ODS (Object_Definition_Segment), and END (END, End, Display, Set, Segment).
- PCS is called a screen composition segment
- WDS, PDS, ODS, and END are called definition segments.
- the correspondence between PES packets and functional segments is a one-to-one relationship or a one-to-many relationship. That is, the functional segment is converted into one PES packet and recorded on the BD-ROM, or fragmented, converted into a plurality of PES packets, and recorded on the BD-ROM.
- FIG. 11B is a diagram showing a PES packet obtained by converting a functional segment.
- the PES packet includes a “packet header” and a “payload”, and this payload corresponds to a functional segment entity.
- the packet header there are DTS and PTS corresponding to this functional segment.
- the DTS and PTS existing in the header of the PES packet in which the functional segment is stored are treated as the DTS and PTS of the functional segment.
- FIG. 12 is a diagram showing a logical structure composed of various types of functional segments.
- the functional segment is shown in the third row
- the Display Set is shown in the second row
- Epoch is shown in the first row.
- the second stage Display Set (abbreviated as DS) refers to a set of graphics constituting one screen of a plurality of functional segments constituting a graphics subtitle stream.
- the broken line kz1 in the figure indicates the attribution relationship to which DS the third-stage functional segment belongs. It can be seen that a series of functional segments called PCS-WDS-PDS-ODS-END constitutes one DS.
- the playback device can configure graphics for one screen by reading a plurality of functional segments constituting the DS from the BD-ROM.
- the first-stage Epoch refers to one period having memory management continuity on the playback time axis of the AVClip, and a data group assigned to this period.
- the memory assumed here is a graphics plane for storing graphics for one screen and an object buffer for storing graphics data in an expanded state. There is continuity in memory management for these, because the graphics plane and object buffer are not flushed during the period corresponding to this Epoch, and only within a certain rectangular area in the graphics plane. Erasing and redrawing are performed (* Flash here means clearing all stored contents of the plane and buffer).
- the vertical and horizontal sizes and positions of the rectangular area are fixed throughout the period corresponding to Epoch.
- Epoch can be said to be a unit on the playback time axis that can ensure synchronization of video-graphics.
- the change time point must be defined on the playback time axis, and after that change point, a new Epoch must be set. In this case, video-graphics synchronization is not guaranteed at the boundary between two Epochs.
- FIG. 13 is a diagram illustrating a relationship between a subtitle display position and Epoch.
- consideration is given to changing the position of the subtitle according to the pattern of each picture of the moving image.
- the two subtitles “Really”, “Sorry” are at the bottom of the screen, “From that”, “Three years have passed” Located on the upper side of the screen.
- Epoch1 the period in which subtitles appear in the lower margin
- the period in which subtitles appear in the upper margin is another Epoch2.
- Each of these two Epochs has its own subtitle drawing area.
- the lower margin of the screen is the subtitle drawing area (window 1).
- the upper margin of the screen becomes the subtitle drawing area (window 2).
- Epoch Start a series of Epoch Start, Acquisition Point, Normal Case constitutes the first stage Epoch.
- Epoch Start “Acquisition Point”, and “NormalDSCase” are types of DS.
- the order of Acquisition Point and Normal Case in this figure is only an example, and either may be the first.
- Epoch Start indicates the start of a new Epoch. Therefore, EpochpoStart includes all the functional segments necessary for the next screen composition. Epoch Start is placed at a position where it has been found to be cued, such as a chapter in a movie work.
- Acquisition Point is a Display Set that includes all functional segments necessary for the next screen composition, but not at the start of Epoch. If you cue from the Acquisition Point DS, you can reliably display graphics. In other words, Acquisition Point DS has the role of enabling screen composition from the middle of Epoch.
- the Acquisition Point Display Set is embedded at a position where it can be the cue point. Such positions include positions that can be specified by time search.
- the time search is an operation of accepting a time input of minutes and seconds from the user and cuing from the playback point corresponding to the time input. Since such time input is performed in rough units such as 10 minutes and 10 seconds, the reproduction position at intervals of 10 minutes and the reproduction position at intervals of 10 seconds are positions that can be designated by time search.
- the Acquisition Point at a position that can be designated by time search, it is possible to suitably perform graphics subtitle stream reproduction at time search.
- Normal ⁇ Case includes only the difference from the previous Display Set. For example, a subtitle of a certain DSv has the same content as the preceding DSu, but if the screen configuration is different from this preceding DSu, a DSS with only PCS and END is provided and this DSv is set as a DS of Normal Case . In this way, there is no need to provide overlapping ODS, which can contribute to capacity reduction in the BD-ROM. On the other hand, since the DS of Normal Case is only a difference, normal Case alone cannot be used for screen composition.
- Definition Segment is a functional segment that defines graphics subtitles. This graphics subtitle will be described below. Since the AVClip recorded on the BD-ROM has a high picture quality equivalent to that of a high-definition video, the resolution of graphics subtitles is also set to a high-definition size of 1920 x 1080 pixels. Since it has a resolution of 1920x1080, the BD-ROM can vividly reproduce the subtitle display for theater screenings, that is, the handwritten subtitle display. Graphics subtitles consist of a plurality of run-length data.
- the run-length data is data that represents a pixel string by a pixel code indicating a pixel value and a continuous length of the pixel value.
- the pixel code is an 8-bit value and takes a value of 1 to 255.
- any 256 colors can be selected from 16,777,216 full-color colors and set as the pixel color using this pixel code.
- graphics subtitles When displayed as subtitles, graphics subtitles must be drawn by arranging character strings on a transparent background.
- the definition of graphics subtitles by ODS has a data structure as shown in FIG. As shown in FIG.
- the ODS uniquely identifies “segment_type” indicating that it is an ODS, “segment_length” indicating the data length of the ODS, and a graphics subtitle corresponding to this ODS in the Epoch. It consists of “object_id”, “object_version_number” indicating the version of ODS in Epoch, “last_in_sequence_flag”, and continuous byte length data “object_data_fragment” which is a part or all of the graphics subtitles.
- PDS Picture Difinition Segment
- the palette data is data indicating a combination of a pixel code of 1 to 255 and a pixel value.
- the pixel value is composed of a red color difference component (Cr value), a blue color difference component (Cb value), a luminance component (Y value), and transparency (T value).
- Cr value red color difference component
- Cb value blue color difference component
- Y value luminance component
- T value transparency
- the PDS has “segment_type” indicating that it is a PDS, “segment_length” indicating the data length of the PDS, “pallet_id” uniquely identifying the palette included in this PDS, Epoch “Pallet_version_number” indicating the version of Epoch's PDS and information "pallet_entry” for each entry.
- “Pallet_entry” indicates a red color difference component (Cr value), a blue color difference component (Cb value), a luminance component Y value, and transparency (T value) in each entry.
- Window_definition_segment is a functional segment for defining a rectangular area of the graphics plane. It has already been described in Epoch that continuity occurs in memory management only when clearing and redrawing are performed within a rectangular area in the graphics plane. The rectangular area in this graphics plane is called “window” and is defined by this WDS.
- FIG. 15A shows the data structure of WDS. As shown in this figure, the WDS displays “window_id” that uniquely identifies a window in the graphics plane, “window_horizontal_position” that indicates the horizontal position of the upper left pixel in the graphics plane, and the vertical position of the upper left pixel in the graphics plane. “Window_vertical_position”, “window_width” indicating the horizontal width of the window in the graphics plane, and “window_height” indicating the vertical width in the graphics plane.
- window_horizontal_position The values that can be taken by window_horizontal_position, window_vertical_position, window_width, and window_height will be described.
- the coordinate system assumed by these is an internal area of the graphics plane, and the graphics plane has a two-dimensional size of vertical: video_height and horizontal: video_width.
- window_horizontal_position is the horizontal address of the upper left pixel in the graphics plane, it takes a value of 1 to video_width, and window_vertical_position takes the value of 1 to video_height because it is the vertical address of the upper left pixel in the graphics plane.
- window_width is the horizontal width of the window in the graphics plane, it takes a value of 1 to video_width-window_horizontal_position, and window_height has a value of 1 to video_height-window_vertical_position because it is the vertical width in the graphics plane.
- WDS window_horizontal_position, window_vertical_position, window_width, and window_height can specify for each Epoch where the window is to be placed and how large the window should be. Therefore, during a period when a picture belonging to a certain Epoch is displayed, it is possible to make an adjustment during authoring so that a window appears at a position corresponding to a margin on the picture so as not to obstruct the pattern in the picture. Thereby, it is possible to make the subtitle display by graphics easy to see. Since WDS can be defined for each Epoch, even if there is a temporal variation in the picture pattern, graphics can be displayed in an easy-to-read manner according to the variation. As a result, the quality of the movie work can be improved to the same level as subtitles are incorporated into the video body.
- END of Display Set Segment is an index indicating the end of transmission of Display Set, and is placed immediately after the last ODS in the functional segment in Display Set.
- the internal structure of this END of Display SetSegment consists of “segment_type” indicating that it is END of Display SetSegment and “segment_length” indicating the data length of the relevant functional segment. There are no elements. Therefore, illustration is abbreviate
- PCS is a functional segment that constitutes an interactive screen.
- the PCS has a data structure shown in FIG. As shown in this figure, the PCS has “segment_type”, “segment_length”, “composition_number”, “composition_state”, “pallet_update_flag”, “pallet_id”, “composition_object (1) to (m)” Consists of
- composition_number identifies the graphics update in Display Set using a number between 0 and 15. As for how to identify, if there is a graphics update from the beginning of the Epoch to this PCS, composition_number is set by a rule that it is incremented every time it passes through these graphics updates.
- Composition_state indicates whether the Display Set starting from this PCS is Normal Case, ACquisition Point, or Epoch Start.
- PalletOnly Displey Update is an update made by switching only the previous palette to a new one. If such an update is made in this PCS, this field is set to "1".
- PalletOnly Displey Update is an update made by switching only the palette from the previous Display Set to a new one. This field is set to "1" when such an update is made in this PCS.
- composition_object (1)... (N) is control information for realizing the screen configuration in the Display Set to which this PCS belongs.
- a broken line wd1 in FIG. 15B closes up the internal configuration of an arbitrary composition_object (i).
- composition_object (i) is composed of “object_id_ref”, “window_id_ref”, “object_cropped_flag”, “object_horizontal_position”, “object_vertical_position”, “cropping_rectangle information (1) (2) (n) ) ”.
- Object_id_ref is a reference value of the graphics subtitle identifier (object_id). This reference value means an identifier of a graphics subtitle to be used in realizing the screen configuration corresponding to composition_object (i).
- Window_id_ref is a reference value of the window identifier (window_id). This reference value indicates to which window a graphics subtitle should be displayed in realizing the screen configuration corresponding to composition_object (i).
- “Object_cropped_flag” is a flag for switching between displaying the graphics subtitles cropped in the object buffer or not displaying the graphics subtitles. When “1” is set, the graphics subtitles cropped in the object buffer are displayed. When “0” is set, the graphics subtitles are not displayed.
- Object_horizontal_position indicates the horizontal position of the upper left pixel of the graphics subtitle in the graphics plane.
- Object_vertical_position indicates the vertical position of the upper left pixel in the graphics plane.
- Croping_rectangle information (1) (2)... (N) is an information element that is valid when “object_cropped_flag” is set to 1.
- a broken line wd2 closes up the internal configuration of arbitrary cropping_rectangle information (i). As shown by the broken line, cropping_rectangle information (i) includes “object_cropping_horizontal_position”, “object_cropping_vertical_position”, “object_cropping_width”, and “object_cropping_height”.
- “Object_cropping_horizontal_position” indicates the horizontal position of the upper left pixel of the crop rectangle in the graphics plane.
- the crop rectangle is a frame for cutting out a part of the graphics subtitle, and corresponds to “Region” in the ETSI EN 300 300 743 standard.
- “Object_cropping_vertical_position” indicates the vertical position of the upper left pixel of the crop rectangle in the graphics plane.
- “Object_cropping_width” indicates the horizontal width of the crop rectangle in the graphics plane.
- “Object_cropping_height” indicates the vertical width of the crop rectangle in the graphics plane.
- FIG. 16 shows a description example of WDS and PCS belonging to Display ⁇ Set.
- FIG. 16 is a diagram illustrating a description example of the PCS in the display set.
- window_horizontal_position and window_vertical_position of the WDS indicate the window upper left coordinates LP1 in the graphics plane
- window_width and window_height indicate the horizontal and vertical widths of the window display frame.
- the object_cropping_horizontal_position and object_cropping_vertical_position of the crop information in FIG. 16A indicate the reference SDT of the crop range in the coordinate system with the upper left coordinate of the graphics subtitle in the object buffer as the origin. Then, the range indicated by object_cropping_width and object_cropping_height from the reference point (thick frame portion in the figure) becomes the cropping range.
- the cropped graphics subtitle is arranged in a broken line range cp1 with object_horizontal_position and object_vertical_position as reference points (upper left) in the coordinate system of the graphics plane. By doing so, “really” is written in the window in the graphics plane. As a result, the caption “Really” is combined with the moving image and displayed.
- FIG. 16B is a diagram showing a description example of the PCS in DS2.
- the description of WDS in this figure is the same as in FIG.
- the description of the crop information is different from that in FIG.
- the crop information object_cropping_horizontal_position and object_cropping_vertical_position in FIG. 16B indicate the upper left coordinates of “Komen” in the caption “Really sorry” on the object buffer. Show. By doing so, “I'm sorry” is written in the window in the graphics plane. As a result, the subtitle “I'm sorry” is combined with the moving image and displayed.
- Epoch is a period in which memory management continues on the playback time axis, and since Epoch is composed of one or more Display Sets, it becomes a problem how to assign Display Set to the playback time axis of AVClip.
- the playback time axis of AVClip refers to an assumed time axis for defining the decoding timing and playback timing of individual picture data constituting a video stream multiplexed on AVClip. On this playback time axis, the decode timing and playback timing are expressed with a time accuracy of 90 KHz.
- DTS and PTS added to PCS and ODS in Display ⁇ Set indicate the timing at which synchronous control should be realized on this playback time axis.
- Performing synchronous control using the DTS and PTS added to the PCS and ODS is the assignment of Display Set to the playback time axis.
- FIG. 17 is a diagram showing an AVClip playback time axis to which DSn is assigned.
- DTS DTS (DSn [PCS])
- PTS PTS (DSn [PCS])
- PTS PTS (DSn [PCS])
- the timing at which the first display is performed in DSn is also indicated in the PCS PTS value (PTS (DSn [PCS])). If the timing at which a desired picture of a video stream appears on the AVClip playback time axis matches PTS (DSn [PCS]), the initial display of DSn is synchronized with that video stream.
- PTS (DSn [PCS]) is a value obtained by adding the period required for ODS decoding (DECODEDURATION) to DTS (DSn [PCS]).
- a period mc1 in FIG. 17 indicates a period during which an arbitrary ODS (ODSm) belonging to DSn is decoded.
- the starting point of this decoding period is indicated by DTS (ODSn [ODSm]), and the end point of this decoding is indicated by PTS (ODSn [ODSm]).
- Epoch is defined by assigning the playback time axis as described above to all ODSs belonging to Epoch. This completes the description of the allocation to the playback time axis.
- FIG. 18 is a diagram showing the types of graphics subtitles defined by ODS.
- A is a graphics subtitle that is a subtitle of hiragana characters
- (b) is a graphics subtitle that is a subtitle of characters with a hiragana character.
- C is a graphics subtitle that forms an expanded character subtitle
- (d) and (e) are graphics subtitles that form a thick character subtitle.
- the visible angles are slightly different between those of (d) and (e).
- FIG. 19 shows how the subtitle stream constituting the base view and the subtitle stream constituting the enhanced view are decoded.
- the graphics decoder, graphics plane, and video plane for the base view are referred to as a basic graphics decoder, a basic graphics plane, and a basic video plane.
- the graphics decoder, graphics plane, and video plane for enhanced view are referred to as an extended graphics decoder, an extended graphics plane, and an extended video plane.
- the graphics subtitles obtained by decoding these source packets are stored in the object buffer in the basic graphics decoder. The Then, the graphics subtitles obtained by decoding these source packets are written in the basic graphics plane, and are used for synthesis with the pictures stored in the basic video plane.
- graphics subtitles obtained by decoding these source packets are stored in the object buffer in the extended graphics decoder. Then, graphics subtitles obtained by decoding these source packets are written in the extended graphics plane, and are used for synthesis with pictures stored in the extended video plane.
- the graphics subtitles representing the thick characters shown in FIG. 18 are written to the basic graphics plane and the extended graphics plane in a state shifted to the left and right, so that the graphics appear to appear before the video.
- FIG. 20 illustrates an example of a stereoscopic video that is played back when the playback of the base-view video stream and the enhanced-view video stream is synchronized with the playback of the base-view caption stream and the enhanced-view caption stream. Since the graphics are written in each of the basic graphics plane and the extended graphics plane, the subtitles appear stereoscopically in the stereoscopic video.
- FIG. 21 shows how the coordinates of the window in the graphics plane defined by window_horizontal_position and window_vertical_position and the coordinates of the graphics subtitle in the graphics plane defined by object_horizontal_position and object_vertical_position affect the stereoscopic caption.
- the interval between the position of the window and graphics subtitle in the left-view graphics plane and the position of the window and graphics subtitle in the right-view graphics plane is called “offset” in FIG. (A) shows a case where the offset is large and the interval between the left-view subtitle and the right-view subtitle is wide, and (b) shows that the offset is narrow, the left-view subtitle, the right-view subtitle, The case where the interval of is narrow is shown.
- the stereoscopic subtitle is closer to the foreground.
- the smaller the distance between the position of the window and graphics subtitle in the left-view graphics plane and the position of the window and graphics subtitle in the right-view graphics plane the smaller the stereoscopic subtitle is. Go far away.
- Subtitles include a text subtitle stream defined by a text subtitle stream defined by a text code and font, in addition to the graphics subtitle stream defined by the run-length graphics data described above. Although these have different data structures, they have the same usage as subtitles and are treated uniformly. Compared with the text subtitle stream, the graphics stream subtitle stream has the advantage of being able to appropriately express the thickness for stereoscopic viewing and the like, although the time required for authoring is large, and to improve the reproduction quality.
- FIG. 22 is a diagram illustrating an example of a clip information file.
- the clip information file is management information of the AV clip as shown in the figure, has a one-to-one correspondence with the AV clip, and includes a stream attribute table and an entry map table.
- the lead line zh1 shows a close-up of the internal structure of the stream attribute table.
- attribute information about each stream included in the AV clip is registered for each PID.
- the attribute information has different information for each base view stream and enhanced view stream.
- the entry map is a table showing the correspondence between the source packet number of an arbitrary source packet and the PTS on the STC time axis among the STC time axes specified by using a certain packet ID.
- the STC time axis is an MPEG2-TS time axis representing decoding time and display time.
- a group of one source packet in which there is no system time-base discontinuity of STC (System Time Clock) that is the system reference time of the AV stream is called an “STC sequence”.
- FIG. 23A shows the internal structure of the entry map table.
- a lead line eh1 shows a close-up of the internal structure of the entry map table.
- information including a pair of PTS and SPN is called an “entry point”.
- the entry point is information in which a combination of PTS and SPN is associated with a display method flag (is_angle_change flag) indicating whether or not decoding from the SPN is possible.
- a value incremented for each entry point with the head set to 0 is referred to as “entry point ID (hereinafter referred to as EP_ID)”.
- the playback device can specify the source packet position corresponding to an arbitrary point on the time axis of the video stream. For example, at the time of special playback such as fast forward / rewind, it is possible to efficiently perform processing without analyzing an AV clip by specifying, selecting and playing back an I picture registered in the entry map.
- An entry map is created for each video stream multiplexed in an AV clip and managed by PID.
- FIG. 24 shows entry point registration using an entry map.
- the first level shows a time axis defined by the STC sequence.
- the second level shows an entry map in clip information.
- the third level shows a source packet sequence constituting the STC sequence.
- the arrows te1, te2, te3, and te4 schematically show the correspondence between the playback points t1, t11, t21, and t31 on the STC time axis and the entry points
- the arrows sh1, sh2, sh3, and sh4 show the ATCsequece
- SPN n1, n11, n21, and n31 and entry points
- FIG. 25 is a diagram showing how entry maps corresponding to the left view and the right view are set.
- the source packet number in the STC sequence is described in the source packet number of each entry map in the entry map
- the PTS in the STC sequence is described in the PTS of each entry map in the entry map. That is done. It shows how the correspondence between the time axis source packet and the time axis is taken by the entry map.
- the fifth row is the left-view and right-view extents recorded in an interleaved manner, and is the same as shown in the previous figures.
- the extents of the left view and the right view that should be played back at the same playback time point on the time axis are recorded at different positions in the AV data recording area, but the entry maps associated with each are used.
- the source packet at the head of the left view extent and the right view extent is uniquely accessed using the PTS.
- FIG. 26 is a diagram showing the data structure of the playlist information.
- the playlist information includes reproduction attribute information, main path information, sub path information table, and extension data as indicated by the lead line mp1.
- a leader line mp3 shows a close-up of the internal structure of the reproduction attribute information.
- the reproduction attribute information includes “version number”, “reproduction type”, and “stereoscopic reproduction flag” of the standard on which the corresponding content is based.
- version number a version number such as BD-ROM application format version 2.00 can be stored.
- playback type it is possible to instruct the playback apparatus to play back “sequential” or “random / shuffle” which means that the play items included in the playlist are played back in order from the top.
- a lead line mp1 shows a close-up of the internal configuration of the main path information.
- MainPath is defined from a plurality of PlayItem information # 1,..., #N as indicated by an arrow mp1.
- PlayItem information defines one logical playback section that constitutes the MainPath.
- the structure of the PlayItem information is closed up by a lead line mp2.
- the PlayItem information includes “Clip_Information_file_name” indicating the file name of the playback section information of the AV clip to which the IN point and Out point of the playback section belong, “Clip_codec_identifier” indicating the encoding method of the AV clip, “Is_multi_angle” indicating whether or not the PlayItem constitutes a multi-angle, “connection_condition” indicating the connection state between this PlayItem (current PlayItem) and the previous PlayItem (previousPlayItem), and this PlayItem “Ref_to_STC_id [0]” uniquely indicating the STC_Sequence that is present, time information “In_time” indicating the start point of the playback section, time information “Out_time” indicating the end point of the playback section, and the user operation to be masked in this PlayItem “UO_mask_table” indicating whether or not “PlayItem_random_” indicating whether or not random access to the middle
- the playback path is composed of a set of time information “In_time” indicating the start point of the playback section and time information “Out_time” indicating the end point of the playback section. It consists of a set of “Out_time”.
- STN_table (STream Number_table) is a table that assigns a logical stream number to a combination of a stream entry including a packet ID and a stream attribute.
- the order of combinations of stream entries and stream attributes in the STN_table indicates the priority order of the corresponding streams.
- BaseView_indicator If BaseView_indicator is 0, BaseView is Left, and if it is 1, BaseView is Right.
- FIG. 27 shows the internal structure of the subpath information table.
- a lead line su1 shows a close-up of the internal configuration of the subpath information.
- the subpath information table includes a plurality of subpath information 1, 2, 3,. These subpath information is a plurality of instances derived from one class structure, and the internal configuration is common.
- a lead line su2 shows a close-up of a common internal configuration of Subpath information.
- each Subpath information includes SubPath_type indicating the type of the subpath and one or more sub play item information (... sub play item information # 1 to VOB # m ).
- a lead line su3 shows a close-up of the internal structure of SubPlayItem.
- the sub play item information includes “Clip_information_file_name”, “Clip_codec_identifier”, “ref_to_STC_id [0]”, “SubPlayItem_In_time”, “SubPlayItem_Out_time”, “sync_PlayItem_id”, and “sync_start_PTS_of_PlayItem”.
- the internal configuration of SubPlayItem will be described.
- “Clip_information_file_name” is information for uniquely specifying a SubClip corresponding to a SubPlayItem by describing the file name of the clip information.
- “Clip_codec_identifier” indicates the encoding method of the AV clip.
- SubPlayItem_In_time is information indicating the start point of SubPlayItem on the playback time axis of SubClip.
- SubPlayItem_Out_time is information indicating the end point of SubPlayItem on the playback time axis of SubClip.
- Sync_PlayItem_id is information that uniquely designates the PlayItems constituting the MainPath that should be synchronized with the SubPlayItem.
- SubPlayItem_In_time exists on the playback time axis of Play Item specified by this sync_PlayItem_id.
- Sync_start_PTS_of_PlayItem indicates where the start point of the SubPlayItem specified by SubPlayItem_In_time exists on the playback time axis of the Play Item specified by sync_PlayItem_id with a time accuracy of 45 KHz.
- FIG. 28 shows what playback sections are defined for the left view and the right view. This figure is drawn based on FIG. 14, and In_Time and Out_Time of PlayItem are drawn on the time axis of the second stage of this figure. In_Time and Out_Time of SubPlayItem are drawn on the time axis of the first row.
- the third to fifth stages are the same as the third to fifth stages in FIG.
- the left view and right view I pictures are at the same point in time.
- Left view and right view are associated with each other by play item information and sub play item information.
- FIG. 18 is a diagram showing an internal configuration of extension data in playlist information.
- a lead line et1 shows a close-up of the internal structure of the extension data.
- the extension data is composed of STN_table_extention corresponding to each of the play item information # 1 to #N.
- a lead line et2 shows a close-up of the internal configuration of STN_table_extention corresponding to PlayItem information # 1.
- STN_table_extention corresponding to PlayItem information # 1 includes a “video stream number table”.
- FIG. 29 shows the internal structure of the video stream number table. As indicated by the lead line mh1, the video stream number table is composed of a set of stream_entry and stream_attribute.
- Stream_entry is “ref_to_stream_PID_of_main_Clip” indicating a reference value for the PID of the PES packet constituting the primary video stream, “video_format” indicating a video display format such as NTSC, PAL, 1/24 seconds, 1 / 29.94 seconds, etc. Includes “frame_rate”.
- FIG. 30 shows the internal structure of the subtitle stream information table in STN_table.
- the subtitle stream information table in STN_table includes “display method information” and “N stream information”. Each of the n pieces of stream information is associated with each stream number, and includes stream_entry and stream_attribute.
- a lead line gh1 shows a close-up of the internal structure of the stream_entry.
- Stream_entry is set to either “ref_to_stream_PID_of_mainClip” or “ref_to_Sub_Path_id”, “ref_to_SubClip__entry_id”, or “ref_to_stream_PID_of_subClip”. “Ref_to_stream_PID_of_SubClip” indicates the PID for the subtitle stream when the subtitle stream corresponding to the stream number exists in the same AVClip as the video stream.
- Stream_attribute is composed of a “language attribute” and a “character attribute” as indicated by the lead line gh2.
- a lead line gh3 shows a close-up of the internal structure of the character attribute.
- the character attribute is composed of “character type” and “character size”.
- the lead line gh4 closes up the character type configuration.
- the character type can indicate whether the subtitle stream corresponding to this stream_attribute is a hiragana character or a character with a kana character.
- the character with a kana here means "shake a kana for normal display".
- “Katakana” may be set.
- the leader line gh5 is a close-up of the internal structure of the character size. As shown in the leader line gh5, the character size can indicate whether the subtitle character of the subtitle stream corresponding to stream_attribute is an enlarged character, a normal character, or a reduced character.
- FIG. 30B shows the internal structure of the display method information in the caption stream information table. As shown in the figure, when the display method information is set to “1”, it indicates that the display method control according to the user age setting in the playback apparatus is effective, and is set to “0”. This indicates that the display method control according to the user age setting in the playback device is invalid.
- FIG. 31 is a diagram showing an internal configuration of extension data in playlist information.
- a lead line et1 shows a close-up of the internal structure of the extension data.
- the extension data is composed of STN_table_extention corresponding to each of the play item information # 1 to #N.
- a lead line et2 shows a close-up of the internal configuration of STN_table_extention corresponding to PlayItem information # 1.
- STN_table_extention corresponding to PlayItem information # 1 includes a “video stream number table”.
- FIG. 32 shows the internal structure of the video stream number table.
- Stream_entry includes “ref_to_stream_PID_of_main_Clip” indicating a reference value for the PID of the PES packet constituting the primary video stream as indicated by the lead line vh1, and stream_attribute is “video_format” “frame_rate” as indicated by the lead line vh2. "including.
- the order of stream_entry in these tables means the priority of stream selection when the playback apparatus selects a stream. That is, the playback apparatus preferentially selects the entry having the highest order in the table.
- the packet ID of 0x1011 and the packet ID of 0x1012 are described in ref_to_stream_of_MainCLip.
- FIG. 33 is a diagram illustrating an internal configuration of a caption stream information table in STN_table_extension.
- FIG. 33A shows the overall structure of the caption stream information table.
- the subtitle stream information table is composed of display method information indicated by f3 and N pieces of stream information indicated by f4. Each of the n pieces of stream information is associated with each stream number, and includes “stream_entry”, “stream_attribute”, and “stereoscopic display information”.
- a lead line gh3 shows a close-up of the internal structure of the stream_entry.
- any one of “ref_to_stream_PID_of_mainClip”, “ref_to_Sub_Path_id”, “ref_to_SubClip__entry_id”, and “ref_to_stream_PID_of_subClip” is set.
- “Ref_to_stream_PID_of_mainClip” indicates the PID for the subtitle stream when the subtitle stream corresponding to the stream number exists in the same AVClip as the video stream.
- “Ref_to_Sub_Path_id” indicates the ID of the sub path referring to the subtitle stream when the subtitle stream corresponding to the stream number exists in an AVClip different from the video stream.
- Ref_to_SubClip__entry_id indicates the ID of the AVClip when the subtitle stream corresponding to the stream number exists in an AVClip different from the video stream.
- “Ref_to_stream_PID_of_SubClip” indicates the PID for the subtitle stream when the subtitle stream corresponding to the stream number exists in an AVClip different from the video stream.
- the lead line gh4 shows a close-up of the internal structure of the stereoscopic display information.
- the stereoscopic display information indicates whether the corresponding subtitle stream is for left view or right view, and other subtitle streams to be selected simultaneously with the subtitle stream corresponding to the stream number. Indicates the stream number of the subtitle stream.
- FIG. 33B shows the internal structure of the display method information in the caption stream information table.
- the display method information is set to “1” to indicate that the display method control according to the stereoscopic playback setting in the playback device is effective, and is set to “0”. This indicates that the display method control according to the stereoscopic playback setting in the playback device is invalid.
- the content in the present specification is a unit including playlist information managed by a certain title number and a video stream multiplexed on an AV clip referenced from the playlist information. Called. *
- the data structure of the subtitle stream information table in the present application is essential in that it indicates which one of a plurality of subtitle streams should be selected when the playback method of stereoscopic playback is enabled in the playback section.
- the stream information table only needs to be able to instruct a combination of two subtitle streams to be selected at the same time when the stereoscopic display method is enabled in the playback section, so that the stream number is associated with the packet ID of the TS packet corresponding thereto. Any data structure can be adopted as long as it is possible.
- a format may be used in which a packet identifier of a TS packet constituting a left-view caption stream and a packet identifier of a TS packet constituting a right-view caption stream are associated with a stream number. That is, the format may be such that the packet identifier 0x1220 of the left-view subtitle stream and the packet identifier 0x1240 of the right-view subtitle stream are associated with the stream number.
- FIG. 34 shows the internal structure of the playback device.
- the main components constituting the playback apparatus in this figure are a front end unit 105, a system LSI 106, a memory device 107, a back end unit 108, a nonvolatile memory 109, a host microcomputer 110, and a network I / F 111.
- the front end unit 105 is a data input source.
- the front end unit 105 includes, for example, a BD-ROM drive 112 and a local storage 113.
- the system LSI 106 is composed of logic elements and forms the core of the playback device.
- the device 17a and the combiner 17b are included in the system LSI 106.
- the system LSI may be composed of one chip or a plurality of chips.
- the memory device 107 is composed of an array of memory elements such as SDRAM.
- read buffer 1 read buffer 2
- basic video plane 6a extended video plane 6b
- basic graphics plane 8a extended graphics plane 8b
- static scenario memory 13 heap memory 21
- the dynamic scenario memory 23 will be included in the memory device 107.
- the back end unit 108 is a connection interface between the playback device and other devices.
- the back-end unit 108 includes, for example, the HDMI transmission / reception unit 10.
- the non-volatile memory 109 is a readable / writable recording medium, and is a medium that can retain the recorded contents even when power is not supplied, and is used for configuration information backup for displaying subtitles.
- the nonvolatile memory 109 for example, flash memory, FeRAM, or the like can be used.
- the host microcomputer 110 is the core of a playback device composed of an MPU, ROM, and RAM. Among the specific components shown in FIG. 35 described later, the BD-J platform 22 and the command interpreter 25 are included in the host microcomputer 110.
- the network interface 111 is for communicating with the outside of the playback device, and can access a server accessible via the Internet or a server connected via a local network. For example, it can be used to download additional BD-ROM content published on the Internet, or data communication can be performed with a server on the Internet specified by the content, enabling content playback using the network function.
- the BD-ROM additional content is content that does not exist in the original BD-ROM, and includes, for example, additional sub audio, subtitles, privilege video, and applications.
- the network interface 111 can be controlled from the BD-J platform, and additional content published on the Internet can be downloaded to the local storage 113.
- the front end unit 105 includes the BD-ROM drive 112 and the local storage 113, for example.
- the BD-ROM drive 112 includes, for example, a semiconductor laser (not shown), a collimator lens (not shown), a beam splitter (not shown), an objective lens (not shown), a condensing lens (not shown), light
- An optical head (not shown) having a detector (not shown) is provided.
- the light beam emitted from the semiconductor laser passes through the collimator lens, the beam splitter, and the objective lens, and is condensed on the information surface of the optical disk.
- the condensed light beam is reflected / diffracted on the optical disk, and is collected on the photodetector through the objective lens, the beam splitter, and the condenser lens.
- the generated signal corresponds to the data read from the BD-ROM according to the amount of light collected by the photodetector.
- the local storage 113 includes built-in media and removable media, and is used for storing downloaded additional content and data used by applications.
- the storage area for additional content is divided for each BD-ROM, and the area that an application can use to hold data is divided for each application.
- merge management information that describes how merged downloaded content and data on the BD-ROM are merged is also stored in the built-in and removable media.
- Build-in media is a writable recording medium such as a hard disk drive or memory built in the playback device.
- the removable media is, for example, a portable recording medium, and preferably a portable semiconductor memory card such as an SD card.
- a playback device has a slot (not shown) for mounting a removable medium and an interface (for example, a memory card) for reading the removable medium mounted in the slot.
- an interface for example, a memory card
- the removable media and the playback device are electrically connected and recorded in the semiconductor memory using an interface (for example, a memory card I / F). It is possible to read out the converted data into an electrical signal.
- FIG. 35 is a diagram showing a detailed configuration within the front end unit 105 to the host microcomputer 110. As shown in FIG.
- the read buffer 1 temporarily stores source packets that constitute extents that constitute the base view stream read from the BD-ROM drive 112, adjusts the transfer speed, and transfers the packets to the demultiplexer 4. And has the scale of “RB1” as described above.
- the read buffer 2 is a buffer for temporarily storing source packets constituting extents constituting the enhanced view stream read from the BD-ROM drive 112, adjusting the transfer speed, and transferring the packets to the demultiplexer 4. And has the scale of “RB2” as described above.
- the virtual file system 3 merges the additional content stored in the local storage with the content on the BD-ROM based on the merge management information downloaded to the local storage 113 together with the additional content.
- Virtual package The virtual package and the original BD-ROM can be referenced without distinction from the command interpreter, which is the HDMV mode operator, and the BD-J platform, which is the BD-J mode operator.
- the playback device performs playback control using both data on the BD-ROM and data on the local storage.
- the demultiplexer 4 includes a source packet depacketizer and a PID filter, receives a packet identifier instruction corresponding to a stream to be reproduced, and executes packet filtering based on the packet identifier. If the source packets that make up the base-view video stream and the source packets that make up the enhanced-view video stream are transferred via the read buffers 1 and 2, the TS packets that make up the base-view video stream and the enhanced-view video stream are made up The TS packet to be processed is separated from the extent and transferred to the video decoder 5a and the video decoder 5b.
- the basic video decoder 5a decodes the TS packet constituting the output from the demultiplexer 4 and writes a non-compressed picture into the basic video plane 6a.
- the extended video decoder 5b decodes the enhanced view video stream output from the demultiplexer 4, decodes the TS packet, and writes a non-compressed picture to the extended video plane 6b.
- the basic video plane 6a is a memory for storing uncompressed pictures that make up the base-view video stream.
- the extended video plane 6b is a memory for storing uncompressed pictures that make up the enhanced view video stream.
- the basic graphics decoder 7a decodes TS packets constituting the base-view subtitle stream output from the demultiplexer 4 and writes uncompressed graphics subtitles to the basic graphics plane 8a.
- the extended graphics decoder 7b decodes TS packets constituting the enhanced view subtitle stream output from the demultiplexer 4 and writes uncompressed graphics subtitles in the extended graphics plane 8b.
- the basic graphics plane 8a is a memory for storing uncompressed graphics subtitles obtained by decoding the base-view subtitle stream.
- the extended graphics plane 8b is a memory for storing uncompressed graphics subtitles obtained by decoding the enhanced view subtitle stream.
- the audio decoder 9 decodes the audio frame output from the demultiplexer 4 and outputs uncompressed audio data.
- the HDMI transmission / reception unit 10 includes, for example, an interface compliant with the HDMI standard (HDMI: High Definition Multimedia Interface), and performs transmission / reception so as to comply with the HDMI standard with a device (in this example, the television 103) connected to the playback device.
- the picture data stored in the basic video plane 6 a and the extended video plane 6 b and the uncompressed audio data decoded by the audio decoder 9 are transmitted to the television 103 via the HDMI interface 10.
- the television 103 holds information regarding whether it is compatible with stereoscopic display, information regarding resolution capable of planar display, information regarding resolution capable of stereoscopic display, and when there is a request from the playback device via the HDMI interface, The television 103 returns the requested required information (for example, information regarding whether or not stereoscopic display is supported, information regarding resolution capable of planar display, and information regarding resolution capable of stereoscopic display) to the playback device.
- information regarding whether or not the television 103 supports stereoscopic display can be acquired from the television 103 via the HDMI transmission / reception unit 10.
- a playback status / setting register (Player Status / Setting ⁇ ⁇ ⁇ Register) set 12 is a playback status register for storing the playback status of a playlist, a playback setting register for storing configuration information indicating the configuration of the playback device, and an arbitrary content used by the content. It is a collection of registers including general-purpose registers that can store information.
- the reproduction state of the playlist indicates a state such as which AV data is used in various AV data information described in the playlist and which position (time) of the playlist is being reproduced.
- the playback control engine 14 stores the contents in the PSR set 12.
- the value specified by the application can be stored or the stored value can be stored in the application according to the instruction from the command interpreter that is the HDMV mode operation subject or the Java platform that is the BD-J mode operation subject. It is possible to pass
- the static scenario memory 13 is a memory for storing current playlist information and current clip information.
- Current playlist information refers to information that is currently processed among multiple playlist information that can be accessed from a BD-ROM, a built-in media drive, or a removable media drive.
- Current clip information refers to information that is currently processed among a plurality of clip information that can be accessed from a BD-ROM, a built-in media drive, or a removable media drive.
- the playback control engine 14 executes an AV playback function and a playlist playback function in response to a function call from a command interpreter that is an HDMV mode operating entity and a Java platform that is an BD-J mode operating entity.
- the AV playback function is a group of functions followed from DVD players and CD players. Playback start, playback stop, pause, release of pause, release of still image function, fast forward with specified playback speed, playback speed Is a process such as rewind, audio switching, sub-video switching, angle switching, etc., designated as an immediate value.
- the playlist playback function refers to performing playback start and playback stop in accordance with current playlist information and current clip information constituting the current playlist in the AV playback function.
- the color conversion unit 15a converts the pixel code stored in the basic graphics plane 8a into pixel values such as Y, Cr, and Cb using a color lookup table.
- the color conversion unit 15b converts the pixel code stored in the extended graphics plane 8b into pixel values such as Y, Cr, and Cb using the color lookup table.
- the scaler 16a executes enlargement / reduction of graphics stored in the basic graphics plane 8a.
- the scaler 16b executes enlargement / reduction of the graphics stored in the extended graphics plane 8b.
- the synthesizer 17a synthesizes each pixel of uncompressed picture data stored in the basic video plane 6a and each pixel of graphics developed on the basic graphics plane 8a.
- the synthesizer 17b synthesizes each pixel of uncompressed picture data stored in the basic video plane 6b and each pixel of graphics developed in the extended graphics plane 8b.
- the text subtitle decoder 18 obtains the text subtitle by developing the text code constituting the text subtitle stream in the font, and the graphics plane. Write to 8a, b.
- the heap memory 21 is a stack area where system application byte codes, BD-J application byte codes, system parameters used by system applications, and application parameters used by BD-J applications are arranged.
- the BD-J platform 22 is a Java platform that is the main operation of the BD-J mode, and is fully equipped with Java2Micro_Edition (J2ME) Personal Basis Profile (PBP 1.0) and Globally Executable MHP specification (GEM1.0.2) for package media targets. It is implemented and includes a class loader and a bytecode interpreter.
- the class loader is one of the system applications, and loads the BD-J application by reading the bytecode from the class file existing in the JAR archive file and storing it in the heap memory.
- the byte code interpreter converts the byte code constituting the BD-J application stored in the heap memory 21 and the byte code constituting the system application into a native code, and causes the MPU 21 to execute it.
- the dynamic scenario memory 23 is a memory that stores a current dynamic scenario and is used for processing by a command interpreter that is an HDMV mode operation subject and a Java platform that is an BD-J mode operation subject.
- the current dynamic scenario refers to an index.bdmv, BD-J object, or movie object that is currently being executed among BD-ROM, built-in media, and removable media.
- the module manager 24 which is an example of the mode management module, holds Index.bdmv read from the BD-ROM, built-in media drive, and removable media drive, and performs mode management and branch control.
- the mode management by the module manager 24 is an assignment of modules, which command interpreter 40 and BD-J module execute a dynamic scenario.
- the command interpreter 25 which is an example of an HDMV module, is a DVD virtual player that is an HDMV mode operation subject, and an HDMV mode execution subject.
- the command interpreter which is the main operation in the HDMV mode, decodes and executes the navigation commands that make up the scenario program. Since navigation commands are described in a syntax similar to DVD-Video, DVD-Video-like playback control can be realized by executing such navigation commands.
- the UO detection module 26 detects a user operation performed on the remote control 100 or the front panel of the playback device 102, and outputs information indicating the user operation (hereinafter referred to as UO (User Operation)) to the mode management module 25. From the UO, only the UO appropriate for the mode in the current playback device is selected and transferred to the module that executes the mode. For example, if UOs such as up / down / left / right and activate are accepted during execution of HDMV mode, these UOs are output to the module in HDMV mode.
- UO User Operation
- FIG. 36 shows the internal configuration of the playback control engine 14 and the internal configuration of the PSR set 12.
- PSR2 is set to a value between 1 and 32 to indicate the subtitle current stream number.
- PSR4 is set to a value between 1 and 100 to indicate the current title number.
- ⁇ PSR5 indicates the current chapter number when set to a value between 1 and 999, and indicates that the chapter number is invalid in the playback device when set to 0xFFFF.
- PSR6 is set to a value between 0 and 999 to indicate the current playlist number.
- ⁇ PSR7 is set to a value between 0 and 255 to indicate the current play item number.
- ⁇ PSR8 is set to a value between 0 and OxFFFFFF to indicate the current playback point (current PTM) with a time accuracy of 45 KHz.
- current PTM current playback point
- PSR15 indicates the caption display capability in the playback device.
- PSR16 indicates the language setting of the playback device.
- the PSR set 12 includes a stereoscopic reproduction capability and a stereoscopic reproduction display method flag.
- the stereoscopic capability indicates whether or not the playback device has the ability to perform stereoscopic playback.
- the stereoscopic playback flag indicates whether or not the user intends to execute stereoscopic playback.
- FIG. 36B is a diagram showing an internal configuration of the regeneration control engine 14. As shown in the figure, the playback control engine 14 executes a playlist playback control 41 for performing playlist playback control when executing planar playback, and performs playlist playback control when stereoscopic viewing is set.
- the stereoscopic control unit 42 is configured.
- FIG. 37 is a diagram showing an internal configuration of the planar view control unit 41. As shown in the figure, the plan view control unit 41 includes a procedure control unit 43 and a PID setting unit 44.
- ⁇ Procedure control unit 43> When a change from one PlayItem information to another PlayItem information occurs or when an operation for switching the stream number is performed by the user, the procedure control unit 43 executes a predetermined stream selection procedure and sets the PSR2 Write the stream number of the new secondary video stream. Since the playback device plays back the secondary video stream according to the stream number written in PSR2, the secondary video stream is selected through the setting of PSR2.
- the stream selection procedure is executed because the STN_Table exists for each PlayItem information, so that a secondary video stream that was playable in one PlayItem information cannot be played in another PlayItem information. This is because it is possible.
- the procedure control unit 43 extracts the stream number corresponding to the user age in the PSR 13 from the STN_table in the current playitem information, and stores the stream number in the PSR 2 as the current stream number.
- the procedure control unit 43 determines that the user using the playback device 102 is a child, the procedure control unit 43 sets the packet identifier of the subtitle stream whose language type is “Hiragana” in the demultiplexer, and plays the playback Causes the transport stream referenced by the list to be demultiplexed. In this way, the graphics subtitles whose language type is “Hiragana” are output by the basic graphics decoder 7a.
- FIG. 38 shows the internal configuration of the stereoscopic control unit 42. As shown in the figure, the stereoscopic control unit 42 includes a procedure control unit 45 and a PID setting unit 46.
- ⁇ Procedure control unit 45> When the capability 2 and the display method flag indicate that stereoscopic playback is possible, the procedure control unit 45 extracts two stream numbers from STN_table_extension in Extension_Data, and stores the stream numbers as current stream numbers in PSR2. .
- a general-purpose register may be referred to, or the information that the playback apparatus 102 independently holds other than the PSR set 12 can be used as a determination material.
- the flat / stereoscopic display setting can be realized by a user operation using the remote controller 100 or an instruction from the BD-J application.
- the stereoscopic display flag indicates that stereoscopic display is not possible
- the video for the right view or the video for the left view the subtitle stream is selected. specify. It is not always necessary that the video and subtitles are stereoscopically displayed. Also, when selecting either the right view or the left view, priority is given to the smaller stream number.
- a part of a video stream for stereoscopic display in this example, a video for a left view or a right view
- One of the streams and a part of the subtitle stream can be used for flat display.
- the PID setting unit 46 extracts two PIDs corresponding to the two stream numbers stored in the PSR2 from the stream_entry in the STN_table_extension, sets them in the demultiplexer 4, and performs packet filtering based on this PID for the demultiplexer 4 Let me do it.
- FIG. 39 is a diagram showing an internal configuration of the graphics decoder.
- the graphics decoder 12 includes a coded data buffer 31, a peripheral circuit 36, a stream graphics processor 32, an object buffer 33, a composition buffer 34, and a graphics controller 35.
- the coded data buffer 31 is a buffer in which functional segments are stored together with DTS and PTS. Such a functional segment is obtained by removing the TS packet header and the PES packet header from each TS packet of the transport stream and sequentially arranging the payloads. Of the removed TS packet header and PES packet header, PTS / DTS is stored in association with the PES packet.
- the stream graphics processor 32 decodes the ODS, and writes the uncompressed graphics composed of the index colors obtained by the decoding into the object buffer 33 as graphics subtitles.
- Decoding by the stream graphics processor 32 is performed instantaneously, and the stream graphics processor 32 temporarily holds graphics subtitles by decoding.
- Decoding by the stream graphics processor 32 is performed instantaneously, but writing from the stream graphics processor 32 to the object buffer 33 does not end instantaneously. This is because, in the BD-ROM standard player model, writing to the object buffer 33 is performed at a transfer rate of 128 Mbps. Since the writing completion time to the object buffer 33 is indicated in the PTS of the END segment, processing for the next DS is waited until the time indicated in the PTS of this END segment elapses.
- Writing of graphics subtitles obtained by decoding each ODS starts at the DTS time associated with the ODS and ends by the decoding end time indicated in the PTS associated with the ODS.
- the object buffer 33 is a buffer in which graphics subtitles obtained by decoding by the stream graphics processor 32 are arranged.
- the object buffer 33 must be set to be twice or four times as large as the graphics plane. This is because, considering the case of realizing Scrolling, it is necessary to store graphics subtitles twice and four times as large as the graphics plane 8.
- the composition buffer 34 is a memory in which PCS and PDS are arranged. If there are two Display sets to be processed and the active periods of these PCS overlap, a plurality of PCSs to be processed are stored in the composition buffer 16.
- the graphics controller 35 decodes the PCS, and writes the graphics subtitles into the object buffer 33, reads the graphics subtitles from the object buffer 33, and displays the graphics subtitles according to the PCS decoding results.
- the display by the graphics controller 35 is executed at the time indicated by the PTS of the PES packet storing the PCS.
- the interval from the display of the graphics subtitles belonging to DSn to the display of the graphics subtitles belonging to DSn + 1 by the graphics controller 35 is as described above.
- the peripheral circuit 36 is wire logic that realizes transfer between the coded data buffer 31 and the stream graphics processor 32 and transfer between the coded data buffer 31 and the composition buffer 34.
- the ODS is transferred from the coded data buffer 31 to the stream graphics processor 32.
- the PCS and PDS are transferred to the composition buffer 34. This completes the description of the graphics decoder. Next, details of the playlist reproduction process will be described.
- FIG. 40 is a flowchart showing a processing procedure of playlist reproduction processing.
- step S1 the playlist information file is read, and the process in steps S2 to S5 is started.
- Step S2 is a determination as to whether or not the playback device has the capability.
- Step S ⁇ b> 3 is a determination as to whether or not the television to which the playback apparatus is connected has a stereoscopic playback processing capability.
- Step S4 is a determination of whether or not the display method flag in the reproduction attribute information of the current playlist is valid. If any of Steps S2 to S4 is determined as No, the process proceeds to Step S5, and play item reproduction based on STN_table in each play item information is executed.
- step S2 to step S4 If all of step S2 to step S4 are Yes, play item reproduction based on STN_table_extension in each play item information is executed in step S5.
- FIG. 41 is a flowchart showing a playback procedure of playlist information.
- step S51 the current PlayItem number is initialized to “1”, and the process proceeds to a loop of steps S52 to S62.
- This loop repeats the process of executing steps S52 to S60 for the current play item number and incrementing the current play item number (step S61) until the current play item number is final. Yes (Yes in step S62).
- Steps S52 to S60 are as follows.
- step S52 the current PlayItem.In_Time and the current PlayItem.Out_Time are converted into Start_SPN [i] and End_SPN [i] using the entry map corresponding to the packet ID of the base view stream.
- the enhanced view stream is selected, the current subtitle stream is selected (step S53), the current stream number of the selected stream is written in the PSR (step S54), and the SubPlayItem corresponding to the current stream number is specified (step S55).
- SubPlayItemIn_Time and SubPlayItemOut_Time specified using the entry map [j] corresponding to the packet ID [j] of the enhanced view stream are converted to Start_SPN [j] and End_SPN [j] (step S56).
- step S57 The extent belonging to the read range [i] for reading the TS packet [i] with the packet ID [i] from Start_SPN [i] to End_SPN [i] is specified (step S57), and the TS packet [i] with the packet ID [j]
- step S58 The extent belonging to the reading range for reading j] from Start_SPN [j] to End_SPN [j] is specified (step S58).
- step S59 the extents belonging to the reading ranges [i] and [j] are sorted in ascending order of addresses, and the extents belonging to the reading ranges [i] and [j] are continuously used by using the addresses sorted in step S60. Instruct the drive to read automatically.
- the above is the reproduction procedure based on STN_table and STN_table_extension.
- FIG. 42 (a) is a flowchart showing the procedure for setting PSR2 when the apparatus state changes.
- Step S11 is a determination as to whether or not the number of entries in the STN_table is 0. If it is 0, the value of PSR2 is maintained (step S13).
- Step S12 determines whether the number of entries in STN_table is larger than that in PSR2 and the condition (A) is true when the number of entries in STN_table is not zero.
- the condition (A) is that the playback apparatus has the ability to play back the audio stream specified by PSR2. If Step S12 is Yes, PSR2 is maintained (Step S14). If the value of PSR2 is larger than the entry number or the condition (A) is not satisfied, PSR2 is reset (step S15).
- FIG. 42B is a flowchart showing the procedure for setting PSR2 when the stream changes.
- the difference between this flowchart and FIG. 10A is that the notation of PSR2 in FIG.
- This X is a value based on User Operation.
- Step S20 in this flowchart determines whether the number of entries in the STN_table is larger than X and the condition (A) is true.
- the condition (A) is that the playback apparatus has the capability of playing back the audio stream specified by PSR2, and is determined by comparing PSR15 and Stream_coding_type of the audio stream. If X satisfies this condition, X is set in PSR2 (step S21).
- step S22 If X is larger than the entry number or the condition (A) is not satisfied, it is determined whether X is 0xFFFF (step S22). If it is not OxFFFF, the audio stream number that the user intends to select is considered invalid, so the value X based on the user operation is ignored and the setting value of PSR2 is maintained (step S24). If the setting value of PSR2 is 0xFFFF, PSR2 is set (step S23).
- FIG. 43 is a flowchart showing a selection procedure during reproduction by age.
- Steps S31 and S32 in this flowchart are a loop process that repeats the processes of steps S33 to S35 for each subtitle stream described in the STN_table.
- a subtitle stream to be processed in this loop processing is a subtitle stream i.
- step S33 it is determined whether the subtitle stream i is a graphics subtitle stream or a text subtitle stream. If the subtitle stream i is a graphics subtitle stream, the process proceeds to step S34.
- Step S34 is a determination of whether or not the graphics subtitle stream i satisfies the following (a) and (b).
- the playback device has the capability to play back the graphics subtitle stream i.
- the language attribute of the graphics subtitle stream i matches the language setting of the playback device.
- the condition (b) is made by determining whether or not PG_language_code in the STN_table matches PSR17.
- step S35 is a determination as to whether or not the text subtitle stream i satisfies (a) and (b).
- the playback device has the ability to expand and play back the text subtitle stream i in fonts.
- the language attribute of the text subtitle stream i matches the language setting of the playback device Whether or not the condition (a) is satisfied is determined based on whether or not the PSR 30 of the playback device indicates “with playback capability”. Whether or not the condition (b) is satisfied is determined by whether or not the textST_language_code of the STN_table matches the setting value of the PSR17. If the processes in steps S33 to S35 are repeated for all subtitle streams, the processes in steps S36 to S41 are executed.
- Step S36 is a determination as to whether or not there is a subtitle stream satisfying (a). If there is no subtitle stream, an invalid value (0xFFFF) is set to PSR2 in step S39 (step S38).
- Step S37 is a determination as to whether or not there is a subtitle stream that satisfies both (a) and (b). If there is a subtitle stream that satisfies both (a) and (b), the entry rank in the STN_table is the highest. The higher one is set to PSR2 (step S39).
- Step S41 sets PSR2 to the one with the highest entry ranking in the STN_table among the graphics subtitle stream satisfying only (a) and the text subtitle stream satisfying only (a).
- Step S40 is a determination of whether the user age in PSR13 means a young person. In the present embodiment, it is assumed that the age of 5 years or younger is set as a young person.
- the scaler enlarges the subtitle character.
- Step S44 determines whether the user age is a value indicating an elderly person. In this embodiment, it is assumed that 65 years old or older is determined as an elderly person. If Yes, in step S45, the stream number of the subtitle stream whose character attribute in stream_entry is an expanded character is set to PSR2.
- FIG. 44 is a flowchart illustrating a processing procedure of a selection procedure at the time of setting a stereoscopic view.
- step S46 the PID of the stream_entry corresponding to the stream number of PSR2 and the PID of the subtitle designated as “another subtitle” in the stereoscopic display information corresponding to the stream number of PSR2 are set in the demultiplexer. Enable packet filtering.
- step S47 the basic graphics decoder and the extended graphics decoder are activated to decode two TS packet sequences.
- FIG. 45 is a diagram showing an AV stream and play list information to be operated.
- FIG. 6A shows the internal structure of the AV stream, which is the same as that shown in FIG.
- FIG. 45 (b) shows the internal structure of the playlist information.
- the play item information 1 in this playlist information includes STN_table for assigning stream numbers to the four subtitle streams.
- the display method information is set such that display method control according to the user age is effective
- the stream information in this STN_table is a subtitle stream composed of TS packets to which a PID of 0x1220 is assigned, 0x1221
- stream numbers 1 to 4 are assigned.
- the language codes are set to Japanese, Japanese, Japanese, and English, respectively, and the character attributes are set to none, hiragana, expanded characters, and none.
- the Extension_Data of the playlist information in FIG. 45 (b) includes STN_table_extension that assigns stream numbers to the four subtitle streams.
- the display method information is set to enable display method control according to the stereoscopic playback setting
- the stream information in this STN_table_extension is a subtitle stream composed of TS packets to which a PID of 0x1225 is assigned.
- stream numbers 6 to 9 are assigned.
- Fig. 46 shows the content of subtitles.
- the run-length data stored in the 0x1221 TS packet represents the hiragana character “Let's start”.
- the run-length data stored in the 0x1222 TS packet represents the expanded character “Go Club”.
- the run-length data stored in the 0x1242 TS packet indicates the English text “This is stereogram”.
- the run-length data stored in the TS packet of 0x1243 is the English sentence “This is stereogram” and the display position is slightly different.
- a subtitle stream that is output so as to display subtitles of a normal size is multiplexed with the video stream. Yes.
- one of the two subtitle streams described above is appropriately specified, and the specified subtitle stream is output by the graphics decoder. In this way, it is not necessary to provide a function for enlarging subtitles in the playback apparatus, and the configuration of the playback apparatus can be simplified and manufactured at low cost.
- FIG. 47 shows what subtitles are displayed according to the configuration information “User age is 4 years old” in the playback apparatus.
- (A) is specific contents of the configuration information, and the user age in the PSR 13 is set to “4”, that is, 4 years old.
- the language setting in PSR16 is Japanese, the stereoscopic capability is set to “none”, and the stereoscopic playback flag is set to “off”.
- the subtitle “Come on now !” shown in FIG. 46 is combined with the moving image and displayed.
- FIG. 48 shows what subtitles are displayed according to the configuration information “user age is 70 years old” in the playback apparatus.
- (A) is specific contents of the configuration information, and the user age in the PSR 13 is set to “70”, that is, 70 years old.
- the language setting in PSR16 is Japanese, the stereoscopic capability is set to “none”, and the stereoscopic playback flag is set to off.
- the data is selectively output to the video decoder 5 by the demultiplexer 4.
- FIG. 49 shows what subtitles are displayed according to the configuration information “stereoscopic capability is on” and “stereoscopic playback flag is on” in the playback apparatus.
- (A) is specific contents of the configuration information, and is set in the configuration information that “stereoscopic capability is on” and “stereoscopic reproduction flag is on”.
- the playlist information is written as follows. That is, when a BD-ROM is loaded, additional content corresponding to the BD-ROM is acquired from the WWW server via the network in accordance with a request from the application. The acquired additional content includes the playlist information.
- a control unit that performs recording control writes the acquired playlist information in the local storage in accordance with a request from the application. By doing so, the virtual package can be constructed by combining the content recorded on the BD-ROM and the additional content recorded on the local storage.
- the disc root certificate identifier, the identifier of the organization that distributed the BD-ROM content, and the identifier of the BD-ROM are recorded in the BD-ROM, and the area in which the additional content is to be stored is the disc root
- the file path includes a certificate identifier, an organization identifier, and a BD-ROM identifier.
- the application performs writing by passing a file path that specifies an area in which additional content is to be stored to the control unit.
- the file path used for writing to the local storage has a directory name of 8 characters or less, and , Including the designation of the file name and extension in an 8.3 file system having a file name and an extension name of 3 characters or less.
- the playlist information is written as follows.
- the playback device 102 when the playback device 102 is supplied with playlist information by the on-demand manufacture service or the electronic sell-through service, the default directory and the MODEST directory are created under the root directory of the removable medium, and the MODEST directory is created. Create a BDMV directory under.
- the MODEST directory is a first MODEST directory
- the first MODEST directory is a MODEST directory that is created when the service is received for the first time.
- the control unit in the playback apparatus 102 creates a MODEST directory corresponding to the second or later service.
- This startup program is a program that is to be executed first when a recording medium is loaded in the playback device 102.
- the startup program displays a menu for accepting an operation for selecting a BDMV directory from the user on the playback device 102, and the root program is executed.
- the playback device 102 is caused to execute the change function.
- This route change function is a function for recognizing the MODEST directory to which the selected BDMV directory belongs as the root directory when the selection operation on the menu is performed by the user.
- the BD-J application may be, for example, an electronic commerce (EC) client application or an online online game. Further, various online services may be supplied to the user in cooperation with the search engine.
- EC electronic commerce
- various online services may be supplied to the user in cooperation with the search engine.
- the information indicating whether to specify right view or left view when the 3D display is not possible is configured to be included in the configuration information of the playback device 102, and when the 3D display is not supported, playback is performed.
- the stream for left view or right view (video stream and subtitle stream) may be designated according to information indicating which one is designated for left view set in the configuration information of the apparatus 102.
- the nonvolatile memory 109 can also be implemented using a writable recording medium in a local storage or a BD drive. In addition, even if the medium can be accessed via the network I / F, it can be used as the non-volatile memory 109 as long as the recorded contents can be held without depending on the power supply state of the playback device. is there.
- Pre-registration details User information such as the preferred size of the displayed subtitles may be stored and used as pre-registration information of the playback device 102, or information registered in the playback status register or general-purpose register included in the PSR set 12 May be used as configuration information.
- the content creator since the content creator does not need to prepare a plurality of subtitle streams of different sizes, it is possible to reduce the usage capacity of the recording medium and to reduce the content creation cost.
- the expanded character subtitle stream is stored separately from the normal character subtitle stream to execute the character expansion for the elderly.
- the normal character subtitle stream is By expanding the scaler, the subtitle character may be enlarged.
- the playback apparatus 102 may have a function for setting whether to enable or disable the automatic subtitle character change function. If the subtitle character automatic change function is valid, as shown in FIG. 10, if the subtitle character automatic change function is invalid, the user who uses the playback device 102 is a child. Even if it is determined that the subtitle stream is “normal”, the subtitle stream is set in the demultiplexer 4 and the transport stream referred to by the play list to be reproduced is demultiplexed.
- avorite setting For example, user information in which the user of the playback device 102 has registered information on age and preferred subtitles (such as preferring display of hiragana) in advance may be stored and used as configuration information of the playback device 102, or playback. Information registered in the status register or general-purpose register may be used as configuration information.
- Video stream for stereoscopic viewing Recording a video stream for left view and right view on a BD-ROM is merely an example. For each picture, a video stream representing a depth value for each pixel may be recorded on the BD-ROM as an enhanced view video stream for playback.
- the BD-J Extension includes various packages specialized to give the Java (TM) platform functionality beyond GEM [1.0.2].
- the packages supplied with BD-J Extension are as follows.
- ⁇ Org.bluray.media This package provides special features that should be added to Java (TM) Media FrameWork. Control over angle, audio, and subtitle selection is added to this package.
- TM Java
- ⁇ Org.bluray.ti This package includes API for mapping "services" to "titles” in GEM [1.0.2], a mechanism for querying title information from BD-ROM, and a mechanism for selecting new titles.
- Org.bluray.application This package includes an API for managing the lifetime of an application.
- Org.bluray.ui This package defines constants for key events specific to BD-ROM, and includes classes that realize synchronization with video playback.
- Org.bluray.vfs In order to seamlessly play back data regardless of the location of this data, this package can be used for content recorded on BD-ROM (on-disc content) and content on Local Storage that is not recorded on BD-ROM (off Provides a binding scheme for -disc content).
- the Binding Scheme associates content (AV clip, subtitle, BD-J application) on BD-ROM with related content on Local Storage.
- This Binding Scheme realizes seamless playback regardless of the location of the content.
- Java TM
- Java is used as the programming language of the virtual machine, but other than Java (TM), such as B-Shell used in UNIX (TM) OS, Perl Script, ECMA Script, etc. It may be a programming language.
- Multi drive In the above-described embodiment, a BD-ROM drive has been described as an example of a BD-ROM as an example of a recording medium and an example of specific means having a function of reading data from a BD-ROM.
- BD-ROM is merely an example, and even if the recording medium is an optical disk medium such as BD-R, BD-RE, DVD, or CD, data having the above-described data structure is stored in these recording media. In addition, if there is a drive device that reads these recording media, the operation described in the above embodiment is possible.
- the recording medium in each embodiment includes all package media such as an optical disk and a semiconductor memory card.
- the recording medium according to the present embodiment has been described by taking an example of an optical disc (for example, an existing readable optical disc such as a BD-ROM or DVD-ROM) on which necessary data is recorded.
- a terminal device having a function of writing 3D content including data necessary for carrying out the present invention distributed via broadcasting or a network to an optical disc (for example, the function described on the left may be incorporated in a playback device) It may be good or may be a device different from the playback device) and recorded on a writable optical disc (for example, an existing writable optical disc such as BD-RE, DVD-RAM) and the recorded optical disc
- a writable optical disc for example, an existing writable optical disc such as BD-RE, DVD-RAM
- the present invention can be implemented even if the recording medium is a removable medium (semiconductor memory card) such as an SD memory card, for example, besides the optical disk.
- a removable medium semiconductor memory card
- semiconductor memory card such as an SD memory card
- the read buffer 1 and the read buffer 2 are routed through an interface for reading data in the semiconductor memory card.
- the virtual file system 3 may be sent to the heap memory 21, the dynamic scenario memory 23, the static scenario memory 13, the text subtitle decoder 18, and the demultiplexer 4 from the virtual file system 3. .
- the playback device 102 and the semiconductor memory card are electrically connected via the memory card I / F.
- Data recorded on the semiconductor memory card is sent to the virtual file system 3 via the read buffer 1 and the read buffer 2 via the memory card I / F, and the heap memory 21 and dynamic scenario memory 23 from the virtual file system 3. What is necessary is just to comprise so that it may transfer to the static scenario memory 13, the text subtitle decoder 18, and the demultiplexer 4.
- a part of the data may be encrypted as necessary from the viewpoint of protecting the copyright and improving the confidentiality of the data.
- the encrypted data may be any of the data recorded on the BD-ROM.
- the data corresponds to the video stream and the audio stream. Or data corresponding to a subtitle stream or data corresponding to a stream including these.
- data for example, a device key
- a key necessary for decrypting the encrypted data in the BD-ROM is stored in advance in the playback device.
- the BD-ROM decrypts the data corresponding to the key necessary for decrypting the encrypted data (for example, the MKB (media key block) corresponding to the device key described above) and the encrypted data.
- Data obtained by encrypting the key itself (for example, the above-described device key and encrypted title key corresponding to the MKB) is recorded.
- the device key, the MKB, and the encrypted title key are paired, and can also be associated with an identifier (for example, a volume ID) written in an area that cannot be normally copied (area called BCA) on the BD-ROM. Has been. If this combination is not correct, the code cannot be decrypted.
- a key necessary for decryption (for example, a title key obtained by decrypting the encrypted title key based on the above-described device key, MKB, and volume ID) can be derived.
- the encrypted data can be decrypted using the necessary key.
- the playback is performed.
- the data is not played back. This is because the key (title key) necessary for decrypting the encrypted data is recorded on the BD-ROM with the key itself encrypted (encrypted title key), and a combination of MKB and device key. If is not correct, the key necessary for decryption cannot be derived.
- the playback apparatus is configured such that the video stream is decoded by the decoder using the title key, and the audio stream is decoded by the audio decoder.
- the BD-ROM has been described as an example of the recording medium.
- the recording medium is not necessarily limited to the BD-ROM.
- a readable / writable semiconductor memory for example, an SD card or the like is portable. Even when it is applied to a semiconductor memory card, it can be implemented.
- data corresponding to data recorded on the BD-ROM may be recorded on a semiconductor memory card using, for example, electronic distribution, and reproduced from the semiconductor memory card. Even when the necessary data is distributed using electronic distribution and the distributed data is recorded, some or all of the distributed data is encrypted and distributed as necessary. However, it is desirable to record the data necessary for the semiconductor memory while being encrypted. For example, an operation for recording data (distributed data) corresponding to the data described in the present embodiment in the semiconductor memory using electronic distribution will be described.
- the above-described operation may be configured such that the playback device described in the present embodiment can perform such an operation, or distribution data is stored in a semiconductor memory separately from the playback device of the present embodiment. It is also possible to use a form that is performed by a dedicated terminal device that performs this. Here, an example performed by the playback apparatus will be described. Further, an SD card will be described as an example of a recording destination semiconductor memory.
- the playback device When recording distribution data on an SD memory card inserted in a slot provided in the playback device, first, transmission of distribution data is requested to a distribution server (not shown) that stores the distribution data. At this time, the playback device uses the identification information for uniquely identifying the inserted SD memory card (for example, an identification number unique to each SD memory card, more specifically, for example, the serial number of the SD memory card) to the SD memory card. And the read identification information is transmitted to the distribution server together with the distribution request.
- a distribution server not shown
- the identification information for uniquely identifying the SD memory card corresponds to, for example, the volume ID described above.
- the distribution server encrypts necessary data (for example, a video stream, an audio stream, etc.) among the data to be distributed using a key (for example, a title key) necessary for decryption. Stored on the server.
- necessary data for example, a video stream, an audio stream, etc.
- a key for example, a title key
- the distribution server holds a secret key and is configured so that different public key information can be dynamically generated for each unique identification number of the semiconductor memory card.
- the distribution server is configured to be able to encrypt the key (title key) necessary for decrypting the encrypted data (that is, configured to generate an encrypted title key).
- the generated public key information includes, for example, information corresponding to the above-described MKB, volume ID, and encrypted title key.
- the encrypted data is, for example, a key necessary for decryption (for example, Based on the device key, MKB, and identification number unique to the semiconductor memory, a title key obtained by decrypting the encrypted title key) is obtained, and using the obtained key (title key) necessary for decryption, Encrypted data can be decrypted.
- the playback device records the received public key information and distribution data in the recording area of the semiconductor memory card inserted in the slot.
- the received public key information includes, for example, a public key body (for example, the above-described MKB and encrypted title key), signature information, a unique identification number of the semiconductor memory card, and a device list indicating information on a device to be invalidated. Yes.
- the signature information includes, for example, a hash value of public key information.
- This may be a device that is likely to be played illegally, such as a device key pre-recorded on the playback device, an identification number of the playback device, or an identification number of a decoder included in the playback device, or a component included in the device, or This is information for uniquely identifying a function (program).
- the following describes the playback of encrypted data among the distribution data recorded in the recording area of the semiconductor memory card.
- (1) Check whether the identification information unique to the semiconductor memory included in the public key information matches the unique identification number stored in advance in the semiconductor memory card. (2) The public key information calculated in the playback device. Check whether hash value and hash value included in signature information match (3) Check whether playback device that performs playback is capable of unauthorized playback based on information shown in device list included in public key information (For example, check whether the device key shown in the device list included in the public key information matches the device key stored in advance in the playback device) To do. These checks may be performed in any order.
- the identification information unique to the semiconductor memory included in the public key information does not match the unique identification number stored in advance in the semiconductor memory, and is calculated by the playback device. If the hash value of the key information and the hash value included in the signature information do not match or if it is determined that there is a possibility that the playback device that performs playback may be played back illegally, the playback device Control to prevent decryption of encrypted data.
- the identification number unique to the semiconductor memory the public key body included in the public key information
- the combination of the device key recorded in advance on the playback device is correct, and is obtained by decrypting the encrypted title key based on the key necessary for decryption (device key, MKB and identification number unique to the semiconductor memory)
- the encrypted data is decrypted using the title key. For example, when the encrypted data is a video stream or an audio stream, the video decoder decrypts the video stream by using the above-described key necessary for decryption (the title key obtained by decrypting the encrypted title key).
- the audio decoder decodes (decodes) the audio stream using the key necessary for the above-described decryption.
- the audio decoder decodes (decodes) the audio stream using the key necessary for the above-described decryption.
- information for identifying these is shown in the device list. If it is distributed, it is possible to prevent decryption using the public key information (public key body) when the playback device includes the ones shown in the device list. Even if the combination of the device identification number, the public key body included in the public key information, and the device key recorded in advance on the playback device is correct, it can be controlled so that the encrypted data cannot be decrypted. Use of distribution data on a simple device can be suppressed.
- the unique identifier of the semiconductor memory card recorded in advance on the semiconductor memory card is stored in a highly confidential recording area.
- a semiconductor memory card for example, an SD memory card as an example, the serial number of an SD memory card
- illegal copying can be easily performed.
- a different unique identification number is assigned to each of the plurality of semiconductor memory cards, but if the tampering is performed so that the unique identification numbers are the same, the determination of (1) above is made. This is because it makes no sense, and there is a possibility that illegal copies corresponding to the number of falsifications will be made.
- a recording area for recording highly confidential data such as a unique identifier of the semiconductor memory card is used as a recording area for storing normal data (the first area).
- a control circuit for accessing the second recording area Provided in a different recording area (referred to as a second recording area), a control circuit for accessing the second recording area, and a second recording area. Access to the access point is configured so that it can be accessed only through the control circuit.
- the data recorded in the second recording area is encrypted and recorded
- the control circuit includes, for example, a circuit for decrypting the encrypted data.
- the encryption is decrypted and the decrypted data is returned.
- the control circuit holds information on the storage location of the data recorded in the second recording area, and if there is a data access request, specifies the storage location of the corresponding data, and specifies the specified storage location It may be configured to return the data read from the.
- An application that operates on a playback device and requests to record on a semiconductor memory card using electronic distribution is used to transmit data recorded in the second recording area to the control circuit via the memory card I / F (eg, semiconductor).
- the control circuit that has received the request reads the data recorded in the second recording area and returns it to the application operating on the playback device. It is configured to request a distribution server for a required data distribution request together with a unique identification number of the semiconductor memory card, and record the public key information sent from the distribution server and the corresponding distribution data in the first recording area. That's fine.
- an application that operates on the playback device and requests recording to the semiconductor memory card using electronic distribution is used to record data (in the second recording area) to the control circuit via the memory card I / F.
- the application before issuing an access request to an identification number unique to a semiconductor memory, it is desirable to check in advance whether the application has been tampered with.
- the existing X A check using a digital certificate compliant with the 509 specification may be used.
- the application program shown in each embodiment can be created as follows. First, a software developer uses a programming language to write a source program that implements each flowchart and functional components. In this description, the software developer describes a source program that embodies each flowchart and functional components using a class structure, a variable, an array variable, and an external function call according to the syntax of the programming language.
- the described source program is given to the compiler as a file.
- the compiler translates these source programs to generate an object program.
- Translator translation consists of processes such as syntax analysis, optimization, resource allocation, and code generation.
- syntax analysis lexical analysis, syntax analysis, and semantic analysis of the source program are performed, and the source program is converted into an intermediate program.
- optimization operations such as basic block formation, control flow analysis, and data flow analysis are performed on the intermediate program.
- resource allocation in order to adapt to the instruction set of the target processor, a variable in the intermediate program is allocated to a register or memory of the processor of the target processor.
- code generation each intermediate instruction in the intermediate program is converted into a program code to obtain an object program.
- the object program generated here is composed of one or more program codes that cause a computer to execute the steps of the flowcharts shown in the embodiments and the individual procedures of the functional components.
- program codes such as a processor native code and a JAVA byte code.
- a call statement that calls the external function becomes a program code.
- a program code that realizes one step may belong to different object programs.
- each step of the flowchart may be realized by combining arithmetic operation instructions, logical operation instructions, branch instructions, and the like.
- the linker allocates these object programs and related library programs to a memory space, and combines them into one to generate a load module.
- the load module generated in this manner is premised on reading by a computer, and causes the computer to execute the processing procedures and the functional component processing procedures shown in each flowchart.
- Such a program may be recorded on a computer-readable recording medium and provided to the user.
- the system LSI is a package in which a bare chip is mounted on a high-density substrate and packaged.
- a system LSI that includes a plurality of bare chips mounted on a high-density substrate and packaged to give the bare chip an external structure like a single LSI is also included in system LSIs (such systems LSI is called a multichip module.)
- system LSIs are classified into QFP (Quad-Flood Array) and PGA (Pin-Grid Array).
- QFP is a system LSI with pins attached to the four sides of the package.
- the PGA is a system LSI with many pins attached to the entire bottom surface.
- pins serve as an interface with other circuits. Since pins in the system LSI have such an interface role, the system LSI plays a role as the core of the playback device 102 by connecting other circuits to these pins in the system LSI.
- Such a system LSI can be incorporated not only in the playback apparatus 102 but also in various devices that handle video playback, such as a TV, a game, a personal computer, a one-seg mobile phone, etc., and can broaden the application of the present invention.
- the system LSI architecture conforms to the Uniphier architecture.
- a system LSI that conforms to the Uniphier architecture consists of the following circuit blocks.
- ⁇ Data parallel processor DPP This is a SIMD type processor in which multiple element processors operate in the same way. By operating the arithmetic units incorporated in each element processor simultaneously with a single instruction, the decoding process for multiple pixels constituting a picture is performed in parallel. Plan
- Instruction parallel processor IPP This is a "Local Memory Controller” consisting of instruction RAM, instruction cache, data RAM, and data cache, "Processing Unit” consisting of instruction fetch unit, decoder, execution unit and register file, and Processing Unit part for parallel execution of multiple applications. It consists of a “Virtual Multi Processor Unit section” to be performed.
- MPU block This is a peripheral interface such as ARM core, external bus interface (Bus Control Unit: BCU), DMA controller, timer, vector interrupt controller, UART, GPIO (General Purpose Input Output), synchronous serial interface, etc. Consists of.
- -Stream I / O block This performs data input / output with drive devices, hard removable media drive devices, and SD memory card drive devices connected to the external bus via the USB interface or ATA Packet interface.
- ⁇ AVI / O block This is composed of audio input / output, video input / output, and OSD controller, and performs data input / output with TV and AV amplifier.
- Memory control block This is a block that realizes reading and writing of the SD-RAM connected via the external bus.
- the internal bus connection part that controls the internal connection between each block, the SD-RAM connected outside the system LSI It consists of an access control unit that transfers data to and from the RAM, and an access schedule unit that adjusts SD-RAM access requests from each block.
- the buses connecting circuit elements, ICs, LSIs, their peripheral circuits, external interfaces, etc. will be defined.
- connection lines, power supply lines, ground lines, clock signal lines, and the like will be defined.
- the circuit diagram is completed while adjusting the operation timing of each component in consideration of the specifications of the LSI, and making adjustments such as ensuring the necessary bandwidth for each component.
- Mounting design refers to where on the board the parts (circuit elements, ICs, and LSIs) on the circuit board created by circuit design are placed, or how the connection lines on the circuit board are placed on the board. This is a board layout creation operation for determining whether to perform wiring.
- the mounting design result is converted into CAM data and output to equipment such as an NC machine tool.
- NC machine tools perform SoC implementation and SiP implementation based on this CAM data.
- SoC (System on chip) mounting is a technology that burns multiple circuits on a single chip.
- SiP (System-in-Package) packaging is a technology that combines multiple chips into one package with resin or the like.
- the integrated circuit generated as described above may be called IC, LSI, super LSI, or ultra LSI depending on the degree of integration.
- the hardware configuration shown in each embodiment can be realized.
- the LUT is stored in the SRAM, and the contents of the SRAM disappear when the power is turned off.
- the LUT that realizes the hardware configuration shown in each embodiment is defined by the definition of the configuration information. Must be written to SRAM. *
- the user interface part, middleware part, and system LSI part of the playback device can be developed independently and in parallel, making development more efficient It becomes possible.
- There are various ways to cut each interface part For example, when the video decoder 5a, the video decoder 5b, the audio decoder 9, the color converter 15a, the color converter 15b, the synthesizer 17a, and the synthesizer 17b shown as included in the system LSI 106 are integrated into one chip, these are controlled.
- the middleware to be used and the interface part between these and the corresponding middleware are developed on the chip development side, and after completion, the chip is incorporated into the playback device, and the developed middleware and interface portion is the memory in the playback device. Incorporation into a storage unit such as the above enables development on the playback device side and development on the chip side to be performed in parallel, improving development efficiency.
- the portion configured by the system LSI is not limited to being configured only by the LSI, but may be configured by using a signal processing circuit corresponding to the function to be included in the system LSI. .
- a recording medium capable of dynamically switching between a flat display and a stereoscopic display, a reproducing apparatus and a reproducing method thereof, in particular, the movie industry involved in the production of AV contents, and the manufacture of equipment for processing the same.
- a reproducing apparatus and a reproducing method thereof in particular, the movie industry involved in the production of AV contents, and the manufacture of equipment for processing the same.
- it can be used as a BD-ROM disc and a BD-ROM player.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Human Computer Interaction (AREA)
- Television Signal Processing For Recording (AREA)
- Signal Processing For Digital Recording And Reproducing (AREA)
- Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
- Indexing, Searching, Synchronizing, And The Amount Of Synchronization Travel Of Record Carriers (AREA)
Abstract
Description
前記プレイリスト情報は、再生区間情報と、表示方式フラグとを含み、
前記再生区間情報は、ビデオストリームの再生時間軸におけるインタイムを示す情報、及び、アウトタイムを示す情報を用いて再生区間を定義する情報であり、ストリーム情報テーブルとを含み、
表示方式フラグは、再生区間における字幕の表示方式の制御を有効とするか否かを示し、
ストリーム情報テーブルは、再生区間において表示方式を有効とする場合、複数の字幕ストリームのうちどれを選択させるべきかを再生装置に指示する
ことを特徴としている。
図4は、ベースビューストリーム、エンハンスドビューストリームを構成する各ソースパケットがどのような過程を経てAVデータ領域に書き込まれるかを示す。本図の第1段目は、ベースビューストリーム又はエンハンスドビューストリームを構成するTSパケットを示す。
ライトビュー用リードバッファ=Rmax1×"ジャンプを伴いながらレフトビュー用リードバッファをフルにする時間"
という関係を満たすよう、ライトビュー用リードバッファの容量を定めねばならない。ここでジャンプとは、ディスクシークと同義である。何故なら、BD-ROMにおいて記録に確保できる連続領域は有限であり、ベースビューストリーム及びエンハンスドビューストリームは、必ずしも、隣合わせで記録されるとは限らず、飛び飛びの領域に記録されることも有り得るからである。
レフトビュー用リードバッファにデータを読み出すにあたっては、ライトビュービデオストリームからレフトビュービデオストリームへのジャンプ時間(Tjump)と、レフトビュービデオストリームからライトビュービデオストリームへのジャンプ時間(Tjump)とを考慮する必要があるので、
レフトビュー用リードバッファの蓄積には(2×Tjump+RB2/(Rud-Rmax2))という時間が必要になる。
RB1≧Rmax1×{2×Tjump+RB2/(Rud-Rmax2)}
になる。
同様の手順で、レフトビュー用リードバッファの容量RB2を求めると、
RB2≧Rmax2×{2×Tjump+RB1/(Rud-Rmax1)}
になる。
ライトビュー用リードバッファ,レフトビュー用リードバッファのメモリサイズの具体的な値としては、1.5Mbyte以下であり、本実施形態においてエクステントサイズSEXT_R、SEXT_Lは、このライトビュー用リードバッファ,レフトビュー用リードバッファのサイズと同じサイズか、またはこれにほぼ等しいサイズに設定されている。以上がベースビューストリーム、エンハンスドビューストリームの記録のされ方についての説明である。続いて、ベースビューストリーム及びエンハンスドビューストリームの内部構成について説明する。
『Object_Definition_Segment』は、グラフィクス字幕を定義する機能セグメントである。このグラフィクス字幕について以下説明する。BD-ROMに記録されているAVClipは、ハイビジョン並みの高画質をセールスポイントにしているため、グラフィクス字幕の解像度も、1920×1080画素という高精細な大きさに設定されている。1920×1080という解像度があるので、BD-ROMでは、劇場上映用の字幕の字体、つまり、手書きの味わい深い字体の字幕表示を鮮やかに再現できる。グラフィクス字幕は複数のランレングスデータからなる。ランレングスデータとは、ピクセル値を示すピクセルコードと、ピクセル値の連続長とにより、画素列を表現したデータである。ピクセルコードは、8ビットの値であり、1~255の値をとる。ランレングスデータでは、このピクセルコードによりフルカラーの16,777,216色から任意の256色を選んで画素の色として設定することができる。尚、字幕として表示される場合、グラフィクス字幕は、透明色の背景に、文字列を配置することで描画せねばならない。
ODSによるグラフィクス字幕の定義は、図14(a)に示すようなデータ構造をもってなされる。ODSは、図14(a)に示すように自身がODSであることを示す『segment_type』と、ODSのデータ長を示す『segment_length』と、EpochにおいてこのODSに対応するグラフィクス字幕を一意に識別する『object_id』と、EpochにおけるODSのバージョンを示す『object_version_number』と、『last_in_sequence_flag』と、グラフィクス字幕の一部又は全部である連続バイト長データ『object_data_fragment』とからなる。
<クリップ情報ファイル>
図22は、クリップ情報ファイルの一例を示す図である。クリップ情報ファイルは、本図に示すようにAVクリップの管理情報であり、AVクリップと1対1に対応し、ストリーム属性テーブルとエントリーマップテーブルから構成される。
図23(a)は、エントリーマップテーブルの内部構成を示す図である。引き出し線eh1は、エントリーマップテーブルの内部構成をクローズアップして示している。
図30は、STN_tableにおける字幕ストリーム情報テーブルの内部構成を示す。STN_tableにおける字幕ストリーム情報テーブルは、「表示方式情報」と、「N個のストリーム情報」とから構成される。n個のストリーム情報のそれぞれは、ストリーム番号のそれぞれに対応付けられており、stream_entryと、stream_attributeとから構成される。引き出し線gh1は、stream_entryの内部構成をクローズアップして示している。stream_entryには、『ref_to_stream_PID_of_mainClip』、又は、『ref_to_Sub_Path_id』、『ref_to_SubClip__entry_id』、『ref_to_stream_PID_of_subClip』のどちらが設定される。『ref_to_stream_PID_of_SubClip』は、ストリーム番号に対応する字幕ストリームが、ビデオストリームと同じAVClipに存在する場合に、その字幕ストリームについてのPIDを示す。
<再生装置における詳細構成>
これらフロントエンド部105~ホストマイコン110内の構成要素を更に詳細に説明する。図35は、フロントエンド部105~ホストマイコン110内の詳細な構成を示す図である。本図に示すように、フロントエンド部105~ホストマイコン110内には、リードバッファ1,2、仮想ファイルシステム3、デマルチプレクサ4、ビデオデコーダ5a,b、ビデオプレーン6a,b、グラフィクスデコーダ7a,b、グラフィクスプレーン8a,b、オーディオデコーダ9、HDMIインターフェイス10、再生状態/設定レジスタセット12、静的シナリオメモリ13、再生制御エンジン14、色変換部15a,b、スケーラ部16a,b、合成器17a,b、テキスト字幕デコーダ18、ヒープメモリ21、BD-Jプラットフォーム22、動的シナリオメモリ23、モード管理モジュール24、コマンドインタプリタ25、UO検知モジュール26といった構成要素が存在する。
動的シナリオメモリ23は、カレント動的シナリオを格納しておき、HDMVモードの動作主体であるコマンドインタプリタ、BD-Jモードの動作主体であるJavaプラットフォームによる処理に供されるメモリである。カレント動的シナリオとは、BD-ROMまたはビルドインメディア、リムーバブルメディアに記録されているIndex.bdmv、BD-Jオブジェクト、ムービーブジェクトのうち、現在実行対象になっているものをいう。
プロシージャ制御部43は、あるPlayItem情報から別のPlayItem情報への切り換わりが生じた場合、又はストリーム番号を切り換える旨の操作がユーザによりなされた場合、所定のストリーム選択プロシージャを実行して、PSR2に新たなセカンダリビデオストリームのストリーム番号を書き込む。再生装置は、PSR2に書き込まれたストリーム番号に応じて、セカンダリビデオストリームを再生するので、かかるPSR2の設定を通じて、セカンダリビデオストリームが選択されることになる。
PID設定部44は、PSR2に格納されているストリーム番号に対応するPIDを、STN_tableにおけるstream_entryから取り出して、デマルチプレクサ4に設定してこのPIDに基づくパケットフィルタリングを、デマルチプレクサ4に行わせる。
図38は、立体視制御部42の内部構成を示す。本図に示すように立体視制御部42は、プロシージャ制御部45、PID設定部46によって構成される。
プロシージャ制御部45は、ケーパビリティ2、表示方式フラグが、立体視再生可能を示す場合、Extension_DataにおけるSTN_table_extensionの中から、2つのストリーム番号を取り出して、そのストリーム番号をカレントストリーム番号としてPSR2に格納する。
PID設定部46は、PSR2に格納されている2つのストリーム番号に対応する2つのPIDを、STN_table_extensionにおけるstream_entryから取り出して、デマルチプレクサ4に設定してこのPIDに基づく、パケットフィルタリングを、デマルチプレクサ4に行わせる。
本フローチャートにおけるステップS20は、XよりSTN_tableのentry数が多く、尚且つ、条件(A)が真であるかを判定するものである。条件(A)とは、PSR2で特定されるオーディオストリームを再生する能力が再生装置に存在することであり、PSR15と、オーディオストリームのStream_coding_typeの比較で判定される。もしXがこの条件を満たすなら、PSR2にXを設定する(ステップS21)。
本フローチャートのステップS31、ステップS32は、STN_tableに記述されている字幕ストリームのそれぞれについて、ステップS33~ステップS35の処理を繰り返すループ処理になっている。本ループ処理において処理対象となる字幕ストリームを字幕ストリームiとする。ステップS33は、字幕ストリームiが、グラフィクス字幕ストリームであるか、テキスト字幕ストリームであるかの判定であり、もしグラフィクス字幕であるならステップS34に移行する。
(a)グラフィクス字幕ストリームiを再生する能力が再生装置に存在すること
(b)グラフィクス字幕ストリームiの言語属性が再生装置の言語設定と一致すること
この(b)の条件は、STN_tableにおけるPG_language_codeがPSR17と一致するか否かの判定でなされる。
一方ステップS35は、テキスト字幕ストリームiが(a)(b)を満たすかを否かの判定である。
(a)テキスト字幕ストリームiをフォントで展開して再生する能力が再生装置に存在すること
(b)テキスト字幕ストリームiの言語属性が再生装置の言語設定と一致すること
(a)の条件を具備しているかの判定は、再生装置のPSR30が"再生能力有"を示すかどうかでなされる。(b)の条件を具備しているかの判定は、STN_tableのtextST_language_codeがPSR17の設定値と一致しているかどうかでなされる。
以上のステップS33~ステップS35の処理が全ての字幕ストリームについて繰り返されれば、ステップS36~ステップS41の処理が実行される。
図44は、立体視設定時における選択手順の処理手順を示すフローチャートである。
図48は、再生装置における"ユーザ年齢が70歳"というコンフィグレーション情報に応じて、どのような字幕が表示されるかを示す。(a)は、コンフィグレーション情報の具体的な内容であり、PSR13におけるユーザ年齢が"70"つまり70歳に設定されている。PSR16における言語設定は日本語であり、立体視ケーパビリティは"無し"に、立体視再生フラグはオフに設定されている。かかるコンフィグレーション情報の再生装置によって、先の図のプレイリスト情報が読み出されれば、上述したような処理手順を実行することにより、(b)に示すように、ストリーム番号=3番の字幕ストリームが選択の対象になる。その結果、PSR2には、ストリーム番号=3が設定されることになる。そうすると、STN_tableにおいて、ストリーム番号=3に対応付けられている、0x1222が、(c)に示すようにPSR2に設定されるから、0x1222のPIDが付与されたTSパケットが、(d)に示すようにデマルチプレクサ4によって選択的にビデオデコーダ5に出力されることになる。結果として、(e)に示すように、図46に示した"囲碁サロン"という字幕が、動画像に合成されて表示されることになる。
図49は、再生装置における"立体視ケーパビリティがオン""立体視再生フラグがオン"というコンフィグレーション情報に応じて、どのような字幕が表示されるかを示す。(a)は、コンフィグレーション情報の具体的な内容であり、"立体視ケーパビリティがオン""立体視再生フラグがオン"というコンフィグレーション情報に設定されている。PSR16における言語設定は英語である。かかるコンフィグレーション情報の再生装置によって、先の図のプレイリスト情報が読み出されれば、上述したような処理手順を実行することにより、(b)に示すように、ストリーム番号=8、9番の字幕ストリームが選択の対象になる。その結果、PSR2には、ストリーム番号=8、9が設定されることになる。そうすると、STN_tableにおいて、ストリーム番号=8、9に対応付けられている、0x1226、0x1241がデマルチプレクサ4に設定されるから、0x1226、0x1241のPIDが付与されたTSパケットが、選択的にビデオデコーダ5に出力されることになる。結果として、(d)に示すように、図36に示した"This is streogram"という字幕が、動画像に合成されて表示されることになる。
以上、本願の出願時点において、出願人が知り得る最良の実施形態について説明したが、以下に示す技術的トピックについては、更なる改良や変更実施を加えることができる。各実施形態に示した通り実施するか、これらの改良・変更を施すか否かは、何れも任意的であり、実施する者の主観によることは留意されたい。
(記録装置としての実施)
再生装置102は、ビルドインメディアドライブ、リムーバブルメディアを含むローカルストレージを具備していて、これらへの書き込みを想定した構成になっているので、本願明細書に記載された再生装置は、記録装置としての機能を兼備しているといえる。再生装置102が記録装置として機能する場合、以下の2つの態様によって、プレイリスト情報の書き込みを実行する。
i)再生装置102が仮想パッケージを再生する機能をもつ場合、プレイリスト情報の書き込みを以下のように行う。つまり、BD-ROMが装填された際、アプリケーションからの要求に従い、前記BD-ROMに対応する追加コンテンツを、ネットワークを介して、WWWサーバから取得する。取得された追加コンテンツは上記プレイリスト情報を含む。再生装置102において、記録制御を行う制御部は、前記アプリケーションからの要求に従い、取得したプレイリスト情報をローカルストレージに書き込む。こうすることで、BD-ROMに記録されたコンテンツと、前記ローカルストレージに記録された追加コンテンツとを組み合わせることで、前記仮想パッケージを構築することが可能になる。
ii)再生装置102がオンデマンドマニュファクチャサービス又は電子的セルスルーサービス(MODEST)の供給を受ける機能をもつ場合、プレイリスト情報の書き込みを以下のように行う。
(Java(TM)アプリケーション)
BD-Jアプリケーションは、例えば電子商取引(EC(Electronic Commerce))のクライアントアプリケーションであってもよいし、ネット対戦型のオンラインゲームであってもよい。更に、検索エンジンと連携して、様々なオンラインサービスを、ユーザに供給するものでもよい。
(ライトビュー用、あるいは、レフトビュー用、どちらを指定するかを示す情報の置き場所)
また、上述の例では立体表示に対応していない場合、ライトビュー用、あるいは、レフトビュー用、どちらを指定するかを示す情報をプレイリストに含ませるような構成を説明したがこれに限定される必要はない。立体表示できない場合、ライトビュー用、あるいは、レフトビュー用、どちらを指定するかを示す情報は、再生装置102のコンフィグレーション情報に含まれるように構成し、立体表示に対応していない場合、再生装置102のコンフィグレーション情報に設定されたレフトビュー用、どちらを指定するかを示す情報に従って、レフトビュー用またはライトビュー用のストリーム(映像ストリームおよび字幕ストリーム)を指定するようにしても良い。
(不揮発性メモリのバリエーション)
不揮発性メモリ109は、ローカルストレージ、BDドライブにおける書き込み可能な記録媒体を利用しても実施可能である。また、ネットワークI/Fを経由してアクセス可能な媒体であっても、再生装置の電源供給状態に依存せず記録内容を保持できるのであれば、不揮発性メモリ109として利用しても実施可能である。
表示される字幕の好ましいサイズなどの利用者情報を再生装置102の事前登録情報として記憶し、使用しても良いし、PSRセット12に含まれる再生状態レジスタ、あるいは、汎用レジスタに登録された情報をコンフィグレーション情報として使用しても良い。
上記実施形態では、拡大文字の字幕ストリームを、通常文字の字幕ストリームとは別に格納することで、年配者向けの文字拡大を実行したが、ユーザが年配者でされば、通常文字の字幕ストリームをスケーラが拡大することで、字幕文字の拡大を実行してもよい。
(字幕の自動拡大・縮小機能の有効、無効を設定)
字幕の自動拡大・縮小機能の有効、無効を設定する機能があっても良い。もし、字幕の自動拡大・縮小機能が有効であった場合には、拡大を実行するが、字幕の自動拡大・縮小機能が無効であった場合には、再生装置102を使用している利用者を年配者だと判断した場合であっても、字幕のサイズを変更させない。
(自動変更機能の有効、無効設定)
再生装置102に字幕文字の自動変更機能の有効、無効を設定する機能があっても良い。もし、字幕文字の自動変更機能が有効であった場合には、図10の通りだが、字幕文字の自動変更機能が無効であった場合には、再生装置102を使用している利用者を幼少者だと判断した場合であっても、言語タイプが"通常"である字幕ストリームをデマルチプレクサ4に設定し、前記再生するプレイリストが参照しているトランスポートストリームの多重分離を行わせる。
(お気に入り設定)
例えば再生装置102の利用者が事前に年齢、好ましい字幕(平仮名の表示を好むなど)の情報を登録した利用者情報を再生装置102のコンフィグレーション情報として記憶し、使用しても良いし、再生状態レジスタ、あるいは、汎用レジスタに登録された情報をコンフィグレーション情報として使用しても良い。
(立体視のためのビデオストリーム)
レフトビュー用、ライトビュー用のビデオストリームをBD-ROMに記録しておくというのは、一例に過ぎない。ピクチャ毎に、画素毎の奥行き値を表すビデオストリームをエンハンスドビュービデオストリームとしてBD-ROMに記録しておいて、再生に供してもよい。
アプリケーション実行装置の実施にあたっては、以下のBD-J Extensionをアプリケーション実行装置に実装するのが望ましい。BD-J Extensionは、GEM[1.0.2]を越えた機能を、Java(TM)プラットフォームに与えるために特化された、様々なパッケージを含んでいる。BD-J Extensionにて供給されるパッケージには、以下のものがある。
・org.bluray.media
このパッケージは、Java(TM) Media FrameWorkに追加すべき、特殊機能を提供する。アングル、音声、字幕の選択についての制御が、このパッケージに追加される。
・org.bluray.ti
このパッケージは、GEM[1.0.2]における"サービス"を"タイトル"にマップして動作するためのAPIや、BD-ROMからタイトル情報を問い合わせる機構や新たなタイトルを選択する機構を含む。
・org.bluray.application
このパッケージは、アプリケーションの生存区間を管理するためのAPIを含む。また、アプリケーションを実行させるにあたってのシグナリングに必要な情報を問い合わせるAPIを含む。
・org.bluray.ui
このパッケージは、BD-ROMに特化されたキーイベントのための定数を定義し、映像再生との同期を実現するようなクラスを含む。
・org.bluray.vfs
このパッケージは、データの所在に拘らず、データをシームレスに再生するため、BD-ROMに記録されたコンテンツ(on-discコンテンツ)と、BD-ROMに記録されていないLocal Storage上のコンテンツ(off-discコンテンツ)とをバインドする機構(Binding Scheme)を提供する。
(プログラミング言語の適用範囲)
上記実施形態では、仮想マシンのプログラミング言語としてJava(TM)を利用したが、Java(TM)ではなく、UNIX(TM) OSなどで使われているB-Shellや、Perl Script、ECMA Scriptなど他のプログラミング言語であっても良い。
(マルチドライブ化)
上記実施形態では、記録媒体の一例としてBD-ROM、BD-ROMからデータを読み出す機能を有する具体的な手段の一例としてBD-ROMドライブを例に挙げて説明をした。しかしながら、BD-ROMは単なる一例であり、記録媒体としてBD-R、BD-RE、DVD、CDなどの光ディスク媒体であっても、これらの記録媒体に上述したデータ構造を有するデータが格納されていること、これらの記録媒体を読み取るドライブ装置があれば、上述の実施の形態で説明した動作が可能である。
装填されたBD-ROMを再生装置において再生する場合、例えばBD-ROM内の暗号化タイトルキー、MKBと対になっている(または対応する)デバイスキーが再生装置内になければ、暗号化されたデータは再生がなされない。何故ならば、暗号化されたデータの解読に必要な鍵(タイトルキー)は、鍵自体が暗号化されて(暗号化タイトルキー)BD-ROM上に記録されており、MKBとデバイスキーの組み合わせが正しくなければ、暗号の解読に必要な鍵を導き出すことができないからである。
本実施の形態では記録媒体としてBD-ROMを例に説明をしたが、BD-ROMに必ずしも限定されるのではなく、例えば、読込み/書込み可能な半導体メモリー(例えばSDカードなどの可搬性を有する半導体メモリーカード)に適用した場合においても、実施が可能である。
例えば電子配信を利用して、本実施の形態で説明をしたデータに相応するデータ(配信データ)を半導体メモリーに記録する動作について説明をする。
生成される公開鍵情報は例えば上述のMKB、ボリュームID及び暗号化タイトルキーに相当する情報を含む。暗号化されたデータは例えば半導体メモリー固有の識別番号、後述する公開鍵情報に含まれる公開鍵本体、および再生装置に予め記録されたデバイスキーの組み合わせが正しければ、暗号解読に必要な鍵(例えばデバイスキー、MKB及び半導体メモリー固有の識別番号を元に、暗号化タイトルキーを復号して得られるタイトルキー)が得られ、この得られた暗号解読に必要な鍵(タイトルキー)を用いて、暗号化されたデータの解読ができるものである。
次に、再生装置は、受信した公開鍵情報と配信データをスロットに挿入した半導体メモリーカードの記録領域に記録する。
(1) 公開鍵情報に含まれる半導体メモリー固有の識別情報と半導体メモリーカードに予め記憶されている固有の識別番号とが一致するかどうかのチェック
(2) 再生装置内で算出した公開鍵情報のハッシュ値と署名情報に含まれるハッシュ値が一致するかのチェック
(3) 公開鍵情報に含まれるデバイスリストに示される情報に基づいて、再生を行う再生装置が不正な再生が可能かどうかのチェック(例えば公開鍵情報に含まれるデバイスリストに示されるデバイスキーと、再生装置に予め記憶されたデバイスキーが一致するかどうかのチェック)
を行なう。これらのチェックを行なう順番どのような順序で行なってもよい。
また、公開鍵情報に含まれる半導体メモリーカードの固有の識別情報と半導体メモリーカードに予め記憶されている固有の識別番号とが一致し、かつ再生装置内で算出した公開鍵情報のハッシュ値と署名情報に含まれるハッシュ値が一致し、かつ再生を行う再生装置が不正に再生される可能性がないと判断したのであれば、半導体メモリー固有の識別番号、公開鍵情報に含まれる公開鍵本体、および再生装置に予め記録されたデバイスキーの組み合わせが正しいと判断し、暗号解読に必要な鍵(デバイスキー、MKB及び半導体メモリー固有の識別番号を元に、暗号化タイトルキーを復号して得られるタイトルキー)を用いて、暗号化されたデータの解読を行なう。
例えば暗号化されたデータがビデオストリーム、オーディオストリームである場合、ビデオデコーダは上述の暗号解読に必要な鍵(暗号化タイトルキーを復号して得られるタイトルキー)を利用してビデオストリームを復号し(デコードし)、オーディオデコーダは、上述の暗号解読に必要な鍵を利用してオーディオストリームを復号する(デコードする)。
このように構成をすることにより、電子配信時において不正利用される可能性がある再生装置、部品、機能(プログラム)などが分っている場合、これらを識別するための情報をデバイスリストに示して、配信するようにすれば、再生装置側がデバイスリストに示されているものを含むような場合には公開鍵情報(公開鍵本体)を用いた復号を抑止できるようにできるため、半導体メモリー固有の識別番号、公開鍵情報に含まれる公開鍵本体、および再生装置に予め記録されたデバイスキーの組み合わせが、たとえ正しくても、暗号化されたデータの解読がなされないように制御できるため、不正な装置上での配信データの利用を抑止することが可能となる。
また半導体メモリーカードに予め記録されている半導体メモリーカードの固有の識別子は秘匿性の高い記録領域に格納するような構成を採用するのが望ましい。何故ならば、半導体メモリーカードに予め記録されている固有の識別番号(例えばSDメモリーカードを例にすればSDメモリーカードのシリアル番号等)は改竄がなされると、違法コピーが容易になされてしまう。何故ならば複数の半導体メモリーカードには、それぞれ異なる固有の識別番号が割り当てられているが、この固有の識別番号が同一となるように改竄がなされてしまえば、上述の(1)の判定が意味を成さなくなり、改竄がなされた数に相当する違法コピーがなされてしまう可能性があるからである。
再生装置上で動作するアプリケーションで、電子配信を利用して半導体メモリーカードに記録する要求するアプリケーションは、メモリーカードI/Fを介して制御回路へ第2の記録領域に記録されたデータ(例えば半導体メモリ固有の識別番号)へのアクセス要求を発行すると、要求を受けた制御回路は第2の記録領域に記録されたデータを読み出して再生装置上で動作するアプリケーションへ返す。この半導体メモリーカードの固有の識別番号とともに必要なデータの配信要求を配信サーバに要求し、配信サーバから送られる公開鍵情報、および対応する配信データを第1の記録領域に記録するように構成すればよい。
(プログラムの実施形態)
各実施形態に示したアプリケーションプログラムは、以下のようにして作ることができる。先ず初めに、ソフトウェア開発者は、プログラミング言語を用いて、各フローチャートや、機能的な構成要素を実現するようなソースプログラムを記述する。この記述にあたって、ソフトウェア開発者は、プログラミング言語の構文に従い、クラス構造体や変数、配列変数、外部関数のコールを用いて、各フローチャートや、機能的な構成要素を具現するソースプログラムを記述する。
(システムLSIの単体実施)
システムLSIとは、高密度基板上にベアチップを実装し、パッケージングしたものをいう。複数個のベアチップを高密度基板上に実装し、パッケージングすることにより、あたかも1つのLSIのような外形構造を複数個のベアチップに持たせたものも、システムLSIに含まれる(このようなシステムLSIは、マルチチップモジュールと呼ばれる。)。
これは、複数の要素プロセッサが同一動作するSIMD型プロセッサであり、各要素プロセッサに内蔵されている演算器を、1つの命令で同時動作させることで、ピクチャを構成する複数画素に対するデコード処理の並列化を図る。
これは、命令RAM、命令キャッシュ、データRAM、データキャッシュからなる「Local Memory Controller」、命令フェッチ部、デコーダ、実行ユニット、レジスタファイルからなる「Processing Unit部」、複数アプリケーションの並列実行をProcessing Unit部に行わせる「Virtual Multi Processor Unit部」で構成される。
これは、ARMコア、外部バスインターフェイス(Bus Control Unit:BCU)、DMAコントローラ、タイマー、ベクタ割込コントローラといった周辺回路、UART、GPIO(General Purpose Input Output)、同期シリアルインターフェイスなどの周辺インターフェイスで構成される。
これは、USBインターフェイスやATA Packetインターフェイスを介して、外部バス上に接続されたドライブ装置、ハードリムーバブルメディアドライブ装置、SDメモリカードドライブ装置とのデータ入出力を行う。
これは、オーディオ入出力、ビデオ入出力、OSDコントローラで構成され、テレビ、AVアンプとのデータ入出力を行う。
これは、外部バスを介して接続されたSD-RAMの読み書きを実現するブロックであり、各ブロック間の内部接続を制御する内部バス接続部、システムLSI外部に接続されたSD-RAMとのデータ転送を行うアクセス制御部、各ブロックからのSD-RAMのアクセス要求を調整するアクセススケジュール部からなる。
3 仮想ファイルシステム
4 デマルチプレクサ
5a,b ビデオデコーダ
6a,b、ビデオプレーン
7a,b グラフィクスデコーダ
8a,b グラフィクスプレーン
9 オーディオデコーダ
10 HDMIインターフェイス
12 再生状態/設定レジスタセット
13 静的シナリオメモリ
14 再生制御エンジン
15a,b 色変換
16a,b スケーラ部
17a,b 合成器
18 テキスト字幕デコーダ
21 ヒープメモリ
22 BD-Jプラットフォーム
23 動的シナリオメモリ
24 モード管理モジュール
25 コマンドインタプリタ
26 UO検知モジュール
101 BD-ROM
102 再生装置
103 テレビ
104 ゴーグル
105 フロントエンド部
106 システムLSI
107 メモリデバイス
108 バックエンド部
109 不揮発性メモリ
110 ホストマイコン
111 ネットワークI/F
112 BD-ROMドライブ
113 ローカルストレージ
Claims (17)
- ビデオストリームと同期して再生すべき字幕ストリームと、プレイリスト情報とが記録された記録媒体であって、
前記プレイリスト情報は、再生区間情報と、表示方式フラグとを含み、
前記再生区間情報は、ビデオストリームの再生時間軸におけるインタイムを示す情報、及び、アウトタイムを示す情報を用いて再生区間を定義する情報であり、ストリーム情報テーブルとを含み、
表示方式フラグは、再生区間における字幕の表示方式の制御を有効とするか否かを示し、
ストリーム情報テーブルは、再生区間において表示方式を有効とする場合、複数の字幕ストリームのうちどれを選択させるべきかを再生装置に指示する
ことを特徴とする記録媒体。 - 前記字幕の表示方式には、ライトビュー用の字幕、レフトビュー用の字幕を個別に表示することで、立体視を実現する立体視の表示方式があり、
ストリーム情報テーブルは、再生区間において立体視の表示方式を有効とする場合、同時に選択すべき2つの字幕ストリームの組合せを再生装置に指示する
ことを特徴とする請求項1記載の記録媒体。 - 前記字幕の表示方式には、ユーザの年齢に応じた表示方式があり、
前記ストリーム番号テーブルは、複数のエントリーからなるエントリー列を含み、各エントリーには、字幕ストリームを構成するパケットのパケット識別子と、個々の字幕ストリームの文字属性とが記述されており、
ストリーム情報テーブルは、年齢に応じた表示方式を有効とする場合、所定の文字属性をもつ字幕ストリームを再生装置に指示する
ことを特徴とする請求項1記載の記録媒体。 - 所定の文字属性をもつ字幕ストリームとは、平仮名表記の字幕ストリーム、振り仮名付きの字幕ストリームである
ことを特徴とする請求項3記載の記録媒体。 - 前記所定の文字属性をもつ字幕ストリームとは、拡大文字によって現された字幕ストリームである
ことを特徴とする請求項3記載の記録媒体。 - 記録媒体に記録されている複数の字幕ストリームのうち1つを選んで、ビデオストリームと同期して再生する再生装置であって、
記録媒体には、ビデオストリームと同期して再生すべき字幕ストリームと、プレイリスト情報とが記録されており、
前記プレイリスト情報は、再生区間情報と、表示方式フラグとを含み、
前記再生区間情報は、ビデオストリームの再生時間軸におけるインタイムを示す情報、及び、アウトタイムを示す情報を用いて再生区間を定義する情報であり、ストリーム情報テーブルとを含み、
表示方式フラグは、再生区間における字幕の表示方式の制御を有効とするか否かを示し、
ストリーム情報テーブルは、再生区間において表示方式を有効とする場合、複数の字幕ストリームのうちどれを選択させるべきかを再生装置に指示し、
装置におけるコンフィグレーション情報を格納するコンフィグレーション格納部と、
コンフィグレーション情報において、所定の表示方式の制御を有効にするかが設定されているかどうかの判定を行い、有効とする場合、複数の字幕ストリームのうち、当該表示方式に適合するものを、ビデオストリームと同期すべき字幕ストリームとして選択する選択手段と、
選択されたグラフィクス字幕ストリームと、ビデオストリームとを再生する再生手段と
を備えることを特徴とする再生装置。 - 前記字幕の表示方式には、ライトビュー用の字幕、レフトビュー用の字幕を個別に表示することで、立体視を実現する立体視方式があり、
ストリーム情報テーブルは、再生区間において立体視方式を有効とする場合、同時に選択すべき2つの字幕ストリームの組合せを示し、
前記表示方式に適合する字幕ストリームとは、
ストリーム情報テーブルによって指示されている2つの字幕ストリームの組合せである、ことを特徴とする請求項6記載の再生装置。 - 前記再生装置は、多重分離部を備え、
再生手段は、
2つの字幕ストリームのうち、一方をデコードしてレフトビュー字幕を得るグラフィクスデコーダ、他方をデコードしてライトビュー字幕を得るグラフィクスデコーダを含み、
前記多重分離部はパケットフィルタリングを実行するものであり、ビデオストリーム及び複数の字幕ストリームを構成する複数のTSパケットのうち、選択手段から指示されたパケット識別子をもつものを、ビデオデコーダ、グラフィクスデコーダに出力し、
前記立体視の表示方式を有効とする場合における選択手段による選択は、2つの字幕ストリームのそれぞれに対応するパケット識別子を、ストリーム情報テーブルから取得して、当該2つのパケット識別子に基づくパケットフィルタリングを多重分離部に命じることでなされる
ことを特徴とする請求項7記載の再生装置。 - 前記再生装置のコンフィグレーション格納部における表示方式が、平面表示を意味している場合、
立体視再生をするために用いる2つの字幕ストリームのうち、いずれか一方のみを再生手段に再生させることにより、字幕の平面表示を行う
ことを特徴とする請求項7記載の再生装置。 - 前記再生装置は、表示装置との接続が可能であり、
再生装置は、さらに、
表示装置が立体表示できるかどうかに関する情報を表示装置から取得する取得手段を備え、
前記取得した情報から前記表示装置が立体表示できないと判断した場合、立体表示用の字幕ストリームの一部および映像ストリームの一部を用いて平面表示を行い、
前記立体表示は、前記取得した情報から前記表示装置が立体表示できると判断した場合に実行される
ように構成した請求項7記載の再生装置。 - 前記再生装置のコンフィグレーション格納部は、年齢情報を含み、
表示方式には、ユーザの年齢に応じた表示方式があり、
年齢に応じた表示方式は、
前記コンフィグレーション格納部に設定された年齢情報が、年少者又は年配者向けの年齢を示す場合、字幕ストリームに対応する字幕を拡大することでなされる
ことを特徴とする請求項6記載の再生装置。 - 前記再生装置のコンフィグレーション格納部は、字幕の表示サイズを含み、
年齢に応じた表示方式は、前記コンフィグレーション格納部に設定されたサイズにまで、字幕サイズを拡大することでなされる
ことを特徴とする請求項6記載の再生装置。 - 前記選択手段は、年齢に応じた表示方式を有効とする場合、所定の文字属性をもつ字幕ストリームを選択し、
所定の文字属性をもつ字幕ストリームとは、
平仮名表記の字幕ストリーム、又は、振り仮名付きの字幕ストリームである、ことを特徴する請求項6記載の再生装置。 - ビデオストリームと同期して再生すべき字幕ストリームと、プレイリスト情報とを記録媒体に記録する記録装置であって、
前記プレイリスト情報は、再生区間情報と、表示方式フラグとを含み、
前記再生区間情報は、ビデオストリームの再生時間軸におけるインタイムを示す情報、及び、アウトタイムを示す情報を用いて再生区間を定義する情報であり、ストリーム情報テーブルとを含み、
表示方式フラグは、再生区間における字幕の表示方式の制御を有効とするか否かを示し、
ストリーム情報テーブルは、再生区間において表示方式を有効とする場合、複数の字幕ストリームのうちどれを選択させるべきかを再生装置に指示する
ことを特徴とする記録装置。 - 記録媒体に記録されている複数の字幕ストリームのうち1つを選んで、ビデオストリームと同期して再生する処理をコンピュータ上で実行する再生方法であって、
記録媒体には、ビデオストリームと同期して再生すべき字幕ストリームと、プレイリスト情報とが記録されており、
前記プレイリスト情報は、再生区間情報と、表示方式フラグとを含み、
前記再生区間情報は、ビデオストリームの再生時間軸におけるインタイムを示す情報、及び、アウトタイムを示す情報を用いて再生区間を定義する情報であり、ストリーム情報テーブルとを含み、
表示方式フラグは、再生区間における字幕の表示方式の制御を有効とするか否かを示し、
ストリーム情報テーブルは、再生区間において表示方式を有効とする場合、複数の字幕ストリームのうちどれを選択させるべきかを指示し、
コンピュータにおけるコンフィグレーション情報を格納するコンフィグレーション格納部と、
コンフィグレーション情報において、所定の表示方式の制御を有効にするかが設定されているかどうかの判定を行い、有効とする場合、複数の字幕ストリームのうち、当該表示方式に適合するものを、ビデオストリームと同期すべき字幕ストリームとして選択する選択ステップと、
選択されたグラフィクス字幕ストリームと、ビデオストリームとを再生する再生ステップと
を備えることを特徴とする再生方法。 - ビデオストリームと同期して再生すべき字幕ストリームと、プレイリスト情報とを記録媒体に記録する記録方法であって、
前記プレイリスト情報は、再生区間情報と、表示方式フラグとを含み、
前記再生区間情報は、ビデオストリームの再生時間軸におけるインタイムを示す情報、及び、アウトタイムを示す情報を用いて再生区間を定義する情報であり、ストリーム情報テーブルとを含み、
表示方式フラグは、再生区間における字幕の表示方式の制御を有効とするか否かを示し、
ストリーム情報テーブルは、再生区間において表示方式を有効とする場合、複数の字幕ストリームのうちどれを選択させるべきかを再生装置に指示する
ことを特徴とする記録方法。 - 記録媒体に記録されている複数の字幕ストリームのうち1つを選んで、ビデオストリームと同期して再生する処理をコンピュータに実行させるプログラムであって、
記録媒体には、ビデオストリームと同期して再生すべき字幕ストリームと、プレイリスト情報とが記録されており、
前記プレイリスト情報は、再生区間情報と、表示方式フラグとを含み、
前記再生区間情報は、ビデオストリームの再生時間軸におけるインタイムを示す情報、及び、アウトタイムを示す情報を用いて再生区間を定義する情報であり、ストリーム情報テーブルとを含み、
表示方式フラグは、再生区間における字幕の表示方式の制御を有効とするか否かを示し、
ストリーム情報テーブルは、再生区間において表示方式を有効とする場合、複数の字幕ストリームのうちどれを選択させるべきかを指示し、
コンピュータにおけるコンフィグレーション情報を格納するコンフィグレーション格納部と、
コンフィグレーション情報において、所定の表示方式の制御を有効にするかが設定されているかどうかの判定を行い、有効とする場合、複数の字幕ストリームのうち、当該表示方式に適合するものを、ビデオストリームと同期すべき字幕ストリームとして選択する選択ステップと、
選択されたグラフィクス字幕ストリームと、ビデオストリームとを再生する再生ステップと
をコンピュータに実行させることを特徴とするプログラム。
Priority Applications (6)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN200980123473.8A CN102067591B (zh) | 2008-06-26 | 2009-06-24 | 再现装置、记录装置、再现方法及记录方法 |
EP09769908.6A EP2293553B1 (en) | 2008-06-26 | 2009-06-24 | Recording medium, reproducing device, recording device, reproducing method, recording method, and program |
ES09769908T ES2428921T3 (es) | 2008-06-26 | 2009-06-24 | Medio de grabación, dispositivo de reproducción, dispositivo de grabación, procedimiento de reproducción, procedimiento de grabación y programa |
MX2010013448A MX2010013448A (es) | 2008-06-26 | 2009-06-24 | Medio de grabacion, dispositivo de reproduccion, dispositivo de grabacion, metodo de reproduccion, metodo de grabacion y programa. |
RU2010150618/07A RU2518189C2 (ru) | 2008-06-26 | 2009-06-24 | Носитель записи, устройство воспроизведения, устройство записи, способ воспроизведения, способ записи и программа |
JP2010517765A JP5065488B2 (ja) | 2008-06-26 | 2009-06-24 | 再生装置、再生方法、再生プログラム |
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2008-166807 | 2008-06-26 | ||
JP2008166807 | 2008-06-26 | ||
US10961808P | 2008-10-30 | 2008-10-30 | |
US61/109,618 | 2008-10-30 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2009157198A1 true WO2009157198A1 (ja) | 2009-12-30 |
Family
ID=41444270
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2009/002905 WO2009157198A1 (ja) | 2008-06-26 | 2009-06-24 | 記録媒体、再生装置、記録装置、再生方法、記録方法、プログラム |
Country Status (8)
Country | Link |
---|---|
US (2) | US8265453B2 (ja) |
EP (1) | EP2293553B1 (ja) |
JP (2) | JP5065488B2 (ja) |
CN (1) | CN102067591B (ja) |
ES (1) | ES2428921T3 (ja) |
MX (1) | MX2010013448A (ja) |
RU (1) | RU2518189C2 (ja) |
WO (1) | WO2009157198A1 (ja) |
Cited By (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR20110106149A (ko) * | 2010-03-22 | 2011-09-28 | 엘지전자 주식회사 | 영상표시장치 및 그 동작방법 |
JP2012023488A (ja) * | 2010-07-13 | 2012-02-02 | Ntt Docomo Inc | 画像処理装置、画像処理方法、表示装置及びプログラム |
CN102379123A (zh) * | 2010-02-12 | 2012-03-14 | 索尼公司 | 信息处理设备、信息处理方法、重放设备、重放方法、程序和记录介质 |
JP2013026644A (ja) * | 2011-07-15 | 2013-02-04 | Hitachi Consumer Electronics Co Ltd | 受信装置、受信方法および送受信方法 |
JP2013527678A (ja) * | 2010-04-14 | 2013-06-27 | サムスン エレクトロニクス カンパニー リミテッド | デジタル字幕放送のための放送ビットストリームを生成する方法及びその装置、デジタル字幕放送のための放送ビットストリームを受信する方法及びその装置 |
JP2013534097A (ja) * | 2010-06-18 | 2013-08-29 | サムスン エレクトロニクス カンパニー リミテッド | 字幕サービスを含むデジタル放送サービスを提供する方法及びその装置 |
JP2013545344A (ja) * | 2010-10-13 | 2013-12-19 | エレクトロニクス アンド テレコミュニケーションズ リサーチ インスチチュート | ステレオスコピック映像情報の送信方法及び装置 |
AU2010248580B2 (en) * | 2009-05-12 | 2015-05-28 | Sony Corporation | Data structure, recording medium, reproducing device, reproducing method, program, and program storage medium |
WO2016039025A1 (ja) * | 2014-09-08 | 2016-03-17 | ソニー株式会社 | 情報処理装置、情報記録媒体、および情報処理方法、並びにプログラム |
WO2018066382A1 (ja) * | 2016-10-06 | 2018-04-12 | ソニー株式会社 | 再生装置および再生方法、記録装置および記録方法、再生記録装置および再生記録方法、並びにプログラム |
TWI853240B (zh) | 2022-05-04 | 2024-08-21 | 台灣大哥大股份有限公司 | 提供視角切換之影音播放系統及方法 |
Families Citing this family (37)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8335425B2 (en) * | 2008-11-18 | 2012-12-18 | Panasonic Corporation | Playback apparatus, playback method, and program for performing stereoscopic playback |
US10257493B2 (en) * | 2009-01-20 | 2019-04-09 | Koninklijke Philips N.V. | Transferring of 3D image data |
WO2010093115A2 (en) * | 2009-02-12 | 2010-08-19 | Lg Electronics Inc. | Broadcast receiver and 3d subtitle data processing method thereof |
KR20110129903A (ko) * | 2009-02-18 | 2011-12-02 | 코닌클리케 필립스 일렉트로닉스 엔.브이. | 3d 시청자 메타데이터의 전송 |
JP4985807B2 (ja) * | 2009-04-15 | 2012-07-25 | ソニー株式会社 | 再生装置および再生方法 |
US8437603B2 (en) * | 2009-05-25 | 2013-05-07 | Panasonic Corporation | Recording medium, reproduction device, integrated circuit, reproduction method, and program |
US20100303437A1 (en) * | 2009-05-26 | 2010-12-02 | Panasonic Corporation | Recording medium, playback device, integrated circuit, playback method, and program |
WO2010147311A2 (en) | 2009-06-16 | 2010-12-23 | Lg Electronics Inc. | 3d display device and selective image display method thereof |
JP5521486B2 (ja) * | 2009-06-29 | 2014-06-11 | ソニー株式会社 | 立体画像データ送信装置および立体画像データ送信方法 |
CN102474603B (zh) * | 2009-07-04 | 2015-04-22 | 杜比实验室特许公司 | 帧兼容三维传输中全分辨率图形、菜单和字幕的支持 |
EP2525579B1 (en) * | 2010-01-11 | 2019-03-06 | LG Electronics Inc. | Broadcasting receiver and method for displaying 3d images |
JP5533163B2 (ja) * | 2010-04-08 | 2014-06-25 | ソニー株式会社 | 情報処理装置、情報処理システム、および情報処理方法、並びにプログラム |
KR101789636B1 (ko) * | 2010-04-30 | 2017-10-25 | 엘지전자 주식회사 | 이미지 처리 방법 및 장치 |
JP2011239169A (ja) * | 2010-05-10 | 2011-11-24 | Sony Corp | 立体画像データ送信装置、立体画像データ送信方法、立体画像データ受信装置および立体画像データ受信方法 |
US8771064B2 (en) | 2010-05-26 | 2014-07-08 | Aristocrat Technologies Australia Pty Limited | Gaming system and a method of gaming |
CA2799704C (en) * | 2010-05-30 | 2016-12-06 | Jongyeul Suh | Method and apparatus for processing and receiving digital broadcast signal for 3-dimensional subtitle |
JP5637750B2 (ja) | 2010-06-30 | 2014-12-10 | 日立コンシューマエレクトロニクス株式会社 | 記録装置/方法/媒体、再生装置/方法 |
JP5527727B2 (ja) * | 2010-08-06 | 2014-06-25 | 日立コンシューマエレクトロニクス株式会社 | 映像表示システム及び表示装置 |
JP2012044407A (ja) * | 2010-08-18 | 2012-03-01 | Sony Corp | 画像処理装置および方法、並びにプログラム |
GB2485140A (en) * | 2010-10-26 | 2012-05-09 | Sony Corp | A Method and Apparatus For Inserting Object Data into a Stereoscopic Image |
GB2488746B (en) * | 2010-12-23 | 2016-10-26 | Samsung Electronics Co Ltd | Improvements to subtitles for three dimensional video transmission |
US8566870B2 (en) * | 2011-03-31 | 2013-10-22 | Sony Corporation | Higher speed high definition television (HDTV) channel and external input surfing |
US8988512B2 (en) * | 2011-04-14 | 2015-03-24 | Mediatek Inc. | Method for adjusting playback of multimedia content according to detection result of user status and related apparatus thereof |
JP6307213B2 (ja) * | 2012-05-14 | 2018-04-04 | サターン ライセンシング エルエルシーSaturn Licensing LLC | 画像処理装置、画像処理方法およびプログラム |
JP6140416B2 (ja) * | 2012-10-10 | 2017-05-31 | サターン ライセンシング エルエルシーSaturn Licensing LLC | 情報処理装置、情報処理方法およびプログラム |
KR20140049832A (ko) * | 2012-10-18 | 2014-04-28 | 삼성전자주식회사 | 블루레이 디스크와 이를 재생하기 위한 블루레이 디스크 재생 장치 및 그 자막 표시 방법 |
EP2866435A1 (en) * | 2013-10-22 | 2015-04-29 | TP Vision Holding B.V. | Consumer device and method for adapting the duration of a temporary control status related to the display of a user interface element |
CN111212251B (zh) * | 2014-09-10 | 2022-05-27 | 松下电器(美国)知识产权公司 | 再现装置以及再现方法 |
CN111899770B (zh) * | 2014-09-12 | 2022-07-08 | 松下电器(美国)知识产权公司 | 非暂时性计算机可读介质、再现装置以及再现方法 |
JP2016081553A (ja) * | 2014-10-17 | 2016-05-16 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカPanasonic Intellectual Property Corporation of America | 記録媒体、再生方法、および再生装置 |
CN104681023A (zh) * | 2015-02-15 | 2015-06-03 | 联想(北京)有限公司 | 一种信息处理方法及电子设备 |
CN106980579B (zh) | 2016-09-30 | 2020-08-14 | 阿里巴巴集团控股有限公司 | 一种图片加载方法及装置 |
CN106570096B (zh) * | 2016-10-31 | 2019-02-22 | 腾讯科技(深圳)有限公司 | 图片排序方法和装置 |
CN107168892A (zh) * | 2017-03-29 | 2017-09-15 | 联想(北京)有限公司 | 一种数据的写入方法及装置 |
CN108924599A (zh) | 2018-06-29 | 2018-11-30 | 北京优酷科技有限公司 | 视频字幕显示方法及装置 |
CN109151507B (zh) * | 2018-08-08 | 2021-06-11 | 武汉市风奥科技股份有限公司 | 视频播放系统及方法 |
CN114500974B (zh) * | 2020-07-17 | 2024-01-30 | 深圳市瑞立视多媒体科技有限公司 | 基于虚幻引擎的字幕实现方法、设备及存储介质 |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH10208446A (ja) | 1996-11-22 | 1998-08-07 | Toshiba Corp | 多種言語記録媒体 |
JP2001333335A (ja) * | 2000-05-24 | 2001-11-30 | Minolta Co Ltd | 映像情報機器 |
JP2003100055A (ja) * | 2001-06-04 | 2003-04-04 | Matsushita Electric Ind Co Ltd | 記録装置、記録媒体、再生装置、プログラム、方法 |
WO2004082297A1 (ja) | 2003-03-11 | 2004-09-23 | Seijiro Tomita | 立体映像表示装置 |
JP2005252989A (ja) * | 2004-03-08 | 2005-09-15 | Kazunari Era | 立体視化パラメータ記憶媒体、立体視化パラメータ埋込装置、立体視画像再生装置、立体視化パラメータ埋込プログラム、立体視画像再生プログラム、及びビデオコンテンツデータの流通方法 |
Family Cites Families (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
SG82587A1 (en) * | 1997-10-21 | 2001-08-21 | Sony Corp | Recording apparatus, recording method, playback apparatus, playback method, recording/playback apparatus, recording/playback method, presentation medium and recording medium |
DE69841745D1 (de) * | 1997-10-21 | 2010-08-12 | Sony Corp | Aufzeichnungsgerät, Aufzeichnungsverfahren, Aufzeichnungsmedium |
EP2261908A3 (en) * | 2001-06-04 | 2012-10-17 | Panasonic Corporation | Recording apparatus and method, playback apparatus and method and recording medium |
JPWO2003092303A1 (ja) * | 2002-04-25 | 2005-09-08 | シャープ株式会社 | マルチメディア情報生成装置およびマルチメディア情報再生装置 |
JP2004274125A (ja) * | 2003-03-05 | 2004-09-30 | Sony Corp | 画像処理装置および方法 |
WO2005010882A1 (en) * | 2003-07-24 | 2005-02-03 | Lg Electronics Inc. | Recording medium having a data structure for managing reproduction of text subtitle data recorded thereon and recording and reproducing methods and apparatuses |
BRPI0409003A (pt) * | 2003-11-10 | 2006-03-28 | Matsushita Electric Ind Co Ltd | meio de gravação, aparelho de reprodução, programa, método de reprodução, circuito integrado do sistema |
KR100739680B1 (ko) * | 2004-02-21 | 2007-07-13 | 삼성전자주식회사 | 스타일 정보를 포함하는 텍스트 기반 서브타이틀을 기록한저장 매체, 재생 장치 및 그 재생 방법 |
WO2005124779A1 (ja) * | 2004-06-18 | 2005-12-29 | Matsushita Electric Industrial Co., Ltd. | 再生装置、プログラム、再生方法 |
EP1775726A4 (en) * | 2004-07-22 | 2009-05-27 | Panasonic Corp | PLAYING DEVICE, PLAY PROCESS, PROGRAM AND COMPUTER READABLE RECORDING MEDIUM |
JP4339206B2 (ja) * | 2004-08-19 | 2009-10-07 | ソニー株式会社 | 再生装置、再生方法および再生プログラム、ならびに、記録媒体 |
CN101853680B (zh) * | 2004-12-01 | 2013-02-27 | 松下电器产业株式会社 | 再现装置和再现方法 |
JP4879480B2 (ja) * | 2004-12-02 | 2012-02-22 | ソニー株式会社 | 再生装置、再生方法および再生プログラム、記録媒体、ならびに、データ構造体 |
KR101377736B1 (ko) * | 2006-10-11 | 2014-03-24 | 코닌클리케 필립스 엔.브이. | 3차원 그래픽 데이터의 생성 |
JP2009135686A (ja) * | 2007-11-29 | 2009-06-18 | Mitsubishi Electric Corp | 立体映像記録方法、立体映像記録媒体、立体映像再生方法、立体映像記録装置、立体映像再生装置 |
-
2009
- 2009-06-24 MX MX2010013448A patent/MX2010013448A/es active IP Right Grant
- 2009-06-24 RU RU2010150618/07A patent/RU2518189C2/ru not_active IP Right Cessation
- 2009-06-24 EP EP09769908.6A patent/EP2293553B1/en active Active
- 2009-06-24 ES ES09769908T patent/ES2428921T3/es active Active
- 2009-06-24 CN CN200980123473.8A patent/CN102067591B/zh not_active Expired - Fee Related
- 2009-06-24 US US12/490,649 patent/US8265453B2/en not_active Expired - Fee Related
- 2009-06-24 WO PCT/JP2009/002905 patent/WO2009157198A1/ja active Application Filing
- 2009-06-24 JP JP2010517765A patent/JP5065488B2/ja not_active Expired - Fee Related
-
2012
- 2012-06-27 JP JP2012143845A patent/JP5400932B2/ja active Active
- 2012-07-31 US US13/562,442 patent/US8509593B2/en not_active Expired - Fee Related
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH10208446A (ja) | 1996-11-22 | 1998-08-07 | Toshiba Corp | 多種言語記録媒体 |
JP2001333335A (ja) * | 2000-05-24 | 2001-11-30 | Minolta Co Ltd | 映像情報機器 |
JP2003100055A (ja) * | 2001-06-04 | 2003-04-04 | Matsushita Electric Ind Co Ltd | 記録装置、記録媒体、再生装置、プログラム、方法 |
WO2004082297A1 (ja) | 2003-03-11 | 2004-09-23 | Seijiro Tomita | 立体映像表示装置 |
JP2005252989A (ja) * | 2004-03-08 | 2005-09-15 | Kazunari Era | 立体視化パラメータ記憶媒体、立体視化パラメータ埋込装置、立体視画像再生装置、立体視化パラメータ埋込プログラム、立体視画像再生プログラム、及びビデオコンテンツデータの流通方法 |
Non-Patent Citations (1)
Title |
---|
See also references of EP2293553A4 |
Cited By (28)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
AU2010248580B2 (en) * | 2009-05-12 | 2015-05-28 | Sony Corporation | Data structure, recording medium, reproducing device, reproducing method, program, and program storage medium |
CN103763541A (zh) * | 2010-02-12 | 2014-04-30 | 索尼公司 | 信息处理设备、信息处理方法、重放设备、重放方法、程序和记录介质 |
US9025933B2 (en) | 2010-02-12 | 2015-05-05 | Sony Corporation | Information processing device, information processing method, playback device, playback method, program and recording medium |
CN103763541B (zh) * | 2010-02-12 | 2015-11-25 | 索尼公司 | 信息处理设备、信息处理方法、重放设备、重放方法、程序和记录介质 |
JP2013511789A (ja) * | 2010-02-12 | 2013-04-04 | ソニー株式会社 | 情報処理方法、再生装置、および記録媒体 |
JP2013070385A (ja) * | 2010-02-12 | 2013-04-18 | Sony Corp | 再生装置、記録媒体、および情報処理方法 |
JP2013081181A (ja) * | 2010-02-12 | 2013-05-02 | Sony Corp | 再生装置、記録媒体、および情報処理方法 |
JP2013081180A (ja) * | 2010-02-12 | 2013-05-02 | Sony Corp | 再生装置、記録媒体、および情報処理方法 |
CN102379123B (zh) * | 2010-02-12 | 2015-05-13 | 索尼公司 | 信息处理设备、信息处理方法、重放设备和重放方法 |
JP2013138467A (ja) * | 2010-02-12 | 2013-07-11 | Sony Corp | 記録媒体および情報処理方法 |
JP2013157992A (ja) * | 2010-02-12 | 2013-08-15 | Sony Corp | 再生装置、記録媒体、および情報処理方法 |
JP2013158003A (ja) * | 2010-02-12 | 2013-08-15 | Sony Corp | 再生装置、記録媒体、および情報処理方法 |
CN102379123A (zh) * | 2010-02-12 | 2012-03-14 | 索尼公司 | 信息处理设备、信息处理方法、重放设备、重放方法、程序和记录介质 |
KR101638541B1 (ko) * | 2010-03-22 | 2016-07-11 | 엘지전자 주식회사 | 영상표시장치 및 그 동작방법 |
KR20110106149A (ko) * | 2010-03-22 | 2011-09-28 | 엘지전자 주식회사 | 영상표시장치 및 그 동작방법 |
US9313442B2 (en) | 2010-04-14 | 2016-04-12 | Samsung Electronics Co., Ltd. | Method and apparatus for generating a broadcast bit stream for digital broadcasting with captions, and method and apparatus for receiving a broadcast bit stream for digital broadcasting with captions |
JP2013527678A (ja) * | 2010-04-14 | 2013-06-27 | サムスン エレクトロニクス カンパニー リミテッド | デジタル字幕放送のための放送ビットストリームを生成する方法及びその装置、デジタル字幕放送のための放送ビットストリームを受信する方法及びその装置 |
JP2013534097A (ja) * | 2010-06-18 | 2013-08-29 | サムスン エレクトロニクス カンパニー リミテッド | 字幕サービスを含むデジタル放送サービスを提供する方法及びその装置 |
JP2012023488A (ja) * | 2010-07-13 | 2012-02-02 | Ntt Docomo Inc | 画像処理装置、画像処理方法、表示装置及びプログラム |
JP2013545344A (ja) * | 2010-10-13 | 2013-12-19 | エレクトロニクス アンド テレコミュニケーションズ リサーチ インスチチュート | ステレオスコピック映像情報の送信方法及び装置 |
JP2013026644A (ja) * | 2011-07-15 | 2013-02-04 | Hitachi Consumer Electronics Co Ltd | 受信装置、受信方法および送受信方法 |
JPWO2016039025A1 (ja) * | 2014-09-08 | 2017-06-01 | ソニー株式会社 | 情報処理装置、情報記録媒体、および情報処理方法、並びにプログラム |
WO2016039025A1 (ja) * | 2014-09-08 | 2016-03-17 | ソニー株式会社 | 情報処理装置、情報記録媒体、および情報処理方法、並びにプログラム |
JP2017139052A (ja) * | 2014-09-08 | 2017-08-10 | ソニー株式会社 | 情報処理装置、情報記録媒体、および情報処理方法、並びにプログラム |
WO2018066382A1 (ja) * | 2016-10-06 | 2018-04-12 | ソニー株式会社 | 再生装置および再生方法、記録装置および記録方法、再生記録装置および再生記録方法、並びにプログラム |
JPWO2018066382A1 (ja) * | 2016-10-06 | 2019-07-18 | ソニー株式会社 | 再生装置および再生方法、記録装置および記録方法、再生記録装置および再生記録方法、並びにプログラム |
US11398077B2 (en) | 2016-10-06 | 2022-07-26 | Sony Corporation | Reproducing device, reproducing method, recording device, recording method, reproducing/recording device, and reproducing/recording method for display of a 3D model arranged in a space recreated by a free visual point moving image stream |
TWI853240B (zh) | 2022-05-04 | 2024-08-21 | 台灣大哥大股份有限公司 | 提供視角切換之影音播放系統及方法 |
Also Published As
Publication number | Publication date |
---|---|
CN102067591B (zh) | 2014-03-19 |
JP5065488B2 (ja) | 2012-10-31 |
JP5400932B2 (ja) | 2014-01-29 |
US20120294585A1 (en) | 2012-11-22 |
US8265453B2 (en) | 2012-09-11 |
JPWO2009157198A1 (ja) | 2011-12-08 |
MX2010013448A (es) | 2010-12-22 |
ES2428921T3 (es) | 2013-11-12 |
US8509593B2 (en) | 2013-08-13 |
EP2293553A4 (en) | 2011-12-28 |
JP2012213220A (ja) | 2012-11-01 |
EP2293553B1 (en) | 2013-09-11 |
CN102067591A (zh) | 2011-05-18 |
EP2293553A1 (en) | 2011-03-09 |
RU2518189C2 (ru) | 2014-06-10 |
US20090324202A1 (en) | 2009-12-31 |
RU2010150618A (ru) | 2012-08-10 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP5400932B2 (ja) | 再生装置、再生方法 | |
JP5291026B2 (ja) | 3d映像を再生する再生装置、および配信装置 | |
JP4923162B2 (ja) | 受信装置、受信方法 | |
WO2010038409A1 (ja) | 再生装置、記録媒体、及び集積回路 | |
WO2010095411A1 (ja) | 記録媒体、再生装置、集積回路 | |
WO2010095410A1 (ja) | 記録媒体、再生装置、集積回路 | |
WO2010095382A1 (ja) | 記録媒体、再生装置、集積回路 | |
WO2010095381A1 (ja) | 記録媒体、再生装置、集積回路 | |
WO2009157159A1 (ja) | 記録媒体、再生装置、集積回路、再生方法、プログラム |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 200980123473.8 Country of ref document: CN |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 09769908 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2010517765 Country of ref document: JP |
|
WWE | Wipo information: entry into national phase |
Ref document number: MX/A/2010/013448 Country of ref document: MX Ref document number: 2009769908 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 4711/KOLNP/2010 Country of ref document: IN |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2010150618 Country of ref document: RU |