WO2011001859A1 - 立体画像データ送信装置および立体画像データ受信装置 - Google Patents
立体画像データ送信装置および立体画像データ受信装置 Download PDFInfo
- Publication number
- WO2011001859A1 WO2011001859A1 PCT/JP2010/060587 JP2010060587W WO2011001859A1 WO 2011001859 A1 WO2011001859 A1 WO 2011001859A1 JP 2010060587 W JP2010060587 W JP 2010060587W WO 2011001859 A1 WO2011001859 A1 WO 2011001859A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- information
- image data
- disparity
- data
- eye image
- Prior art date
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/172—Processing image signals image signals comprising non-image signal components, e.g. headers or format information
- H04N13/183—On-screen display [OSD] information, e.g. subtitles or menus
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/597—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/235—Processing of additional data, e.g. scrambling of additional data or processing content descriptors
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/236—Assembling of a multiplex stream, e.g. transport stream, by combining a video stream with other content or additional data, e.g. inserting a URL [Uniform Resource Locator] into a video stream, multiplexing software data into a video stream; Remultiplexing of multiplex streams; Insertion of stuffing bits into the multiplex stream, e.g. to obtain a constant bit-rate; Assembling of a packetised elementary stream
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/236—Assembling of a multiplex stream, e.g. transport stream, by combining a video stream with other content or additional data, e.g. inserting a URL [Uniform Resource Locator] into a video stream, multiplexing software data into a video stream; Remultiplexing of multiplex streams; Insertion of stuffing bits into the multiplex stream, e.g. to obtain a constant bit-rate; Assembling of a packetised elementary stream
- H04N21/23614—Multiplexing of additional data and video streams
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/434—Disassembling of a multiplex stream, e.g. demultiplexing audio and video streams, extraction of additional data from a video stream; Remultiplexing of multiplex streams; Extraction or processing of SI; Disassembling of packetised elementary stream
- H04N21/4348—Demultiplexing of additional data and video streams
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/47—End-user applications
- H04N21/488—Data services, e.g. news ticker
- H04N21/4884—Data services, e.g. news ticker for displaying subtitles
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/81—Monomedia components thereof
- H04N21/816—Monomedia components thereof involving special video data, e.g 3D video
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N2213/00—Details of stereoscopic systems
- H04N2213/003—Aspects relating to the "2D+depth" image format
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N2213/00—Details of stereoscopic systems
- H04N2213/005—Aspects relating to the "3D+depth" image format
Definitions
- the present invention relates to a stereoscopic image data transmission device and a stereoscopic image data reception device, and more particularly to a stereoscopic image data transmission device that can satisfactorily display superimposed information such as closed caption information, subtitle information, graphics information, and text information. .
- Patent Document 1 proposes a transmission method using a television broadcast radio wave of stereoscopic image data.
- stereoscopic image data including left-eye image data and right-eye image data is transmitted, and stereoscopic image display using binocular parallax is performed in the television receiver.
- FIG. 108 shows the relationship between the display position of the left and right images of an object (object) on the screen and the playback position of the stereoscopic image in stereoscopic image display using binocular parallax.
- object object
- FIG. 108 shows the relationship between the display position of the left and right images of an object (object) on the screen and the playback position of the stereoscopic image in stereoscopic image display using binocular parallax.
- the right and left line of sight intersects in front of the screen surface, so that the stereoscopic image is reproduced.
- the position is in front of the screen surface.
- DPa represents a horizontal disparity vector related to the object A.
- the right and left lines of sight intersect on the screen surface. It becomes on the surface.
- the left image Lc is shifted to the left and the right image Rc is shifted to the right, the right and left lines of sight intersect at the back of the screen surface.
- the playback position is behind the screen.
- DPc represents a horizontal disparity vector related to the object C.
- Superimposition information to be superimposed on an image for example, closed caption information, subtitle information, graphics information, text information, etc., is not only two-dimensional spatial, but also three-dimensional sense of depth in conjunction with stereoscopic image display. Expected to be rendered.
- subtitles that are closed caption information or subtitle information are superimposed on an image (overlay display)
- the viewer will not be able to display it in front of the closest object (object) in the perspective.
- other graphics information or text information is superimposed on an image, it is expected that parallax adjustment is performed according to the perspective of each object in the image to maintain the consistency of perspective.
- An object of the present invention is to maintain perspective consistency with each object in an image in displaying superimposed information such as closed caption information, subtitle information, graphics information, and text information.
- the concept of this invention is A stereoscopic image data output unit for outputting stereoscopic image data including left-eye image data and right-eye image data;
- a superimposition information data output unit for outputting superimposition information data to be superimposed on an image based on the left eye image data and the right eye image data;
- a parallax information output unit for outputting parallax information for shifting the superimposition information to be superimposed on the image based on the left-eye image data and the right-eye image data and providing parallax;
- the payload portion includes the stereoscopic image data output from the stereoscopic image data output unit, and is output from the superimposition information data output from the superimposition information data output unit and the parallax information output unit in the user data area of the header portion.
- a transmission unit for transmitting a video elementary stream including disparity information The data of the superimposition information and the disparity information included in the user data area of the header part of the video elementary stream are in a stereoscopic image data transmission device that can be identified by user data identification information.
- the stereoscopic image data output unit outputs stereoscopic image data including left-eye image data and right-eye image data. Further, the superimposition information data output unit outputs superimposition information data to be superimposed on the image based on the left eye image data and the right eye image data.
- the superimposition information means information to be superimposed on the image, such as closed caption information for displaying a caption and subtitle information.
- the parallax information output unit outputs parallax information for adding parallax by shifting the superimposition information to be superimposed on the image based on the left eye image data and the right eye image data.
- a video elementary stream including the above-described stereoscopic image data in the payload portion and including the above-described superimposition information data and the above-described parallax information in the user data area of the header portion is transmitted by the transmission unit.
- the superimposition information data and the disparity information transmitted by being included in the user data area of the header portion of the video elementary stream can be identified by the user data identification information.
- stereoscopic image data including left-eye image data and right-eye image data for displaying a stereoscopic image is transmitted by being included in the payload portion of the video elementary stream.
- the superimposition information data and the disparity information are included and transmitted in the user data area of the header portion of the mental stream in an identifiable state. Therefore, on the receiving side, in addition to acquiring stereoscopic image data from this video elementary stream, it is possible to acquire superimposition information data and disparity information.
- the superimposition information to be shifted based on the parallax information is specified from among the superimposition information superimposed on the image based on the left eye image data and the superimposition information superimposed on the image based on the right eye image data.
- Shift target designation information may be added. With this shift target designation information, only the superimposition information superimposed on the left eye image is shifted, only the superimposition information superimposed on the right eye image is shifted, or the superposition superimposed on both the left eye image and the right eye image Information can be shifted.
- the disparity information may be disparity information that is sequentially updated in each frame of a predetermined number of frame periods in which the superimposition information is displayed.
- the parallax since the parallax is given to the superimposition information by the parallax information that is sequentially updated at each frame in the predetermined number of frame periods in which the superimposition information is displayed on the reception side, the parallax to be given to the superposition information It becomes possible to change dynamically in conjunction with the change of.
- the disparity information that is sequentially updated in each frame includes disparity information of the first frame in a predetermined number of frame periods, and offset information of the second and subsequent frames with respect to disparity information of the previous frame. You may make it consist of. In this case, the data amount of the parallax information can be suppressed.
- the disparity information may be disparity information that is used in common for each frame in a predetermined number of frame periods in which superimposition information is displayed. In this case, the data amount of the parallax information can be significantly suppressed.
- the disparity information is disparity information that is commonly used in each frame of a predetermined number of frame periods in which the superimposition information is displayed, or a predetermined number of frame periods in which the superimposition information is displayed.
- the disparity information is sequentially updated in each frame, and flag information indicating whether the disparity information is commonly used in each frame or the disparity information sequentially updated in each frame is added to the disparity information. May be.
- disparity information that is commonly used in each frame or disparity information that is sequentially updated in each frame can be selectively transmitted according to, for example, image contents.
- disparity information that is sequentially updated in each frame is transmitted, and on the receiving side, the disparity added to the superimposition information is dynamically changed in conjunction with the change of the image content. Can do.
- the data amount of the disparity information can be suppressed by transmitting disparity information that is commonly used in each frame.
- the payload portion includes stereoscopic image data including left-eye image data and right-eye image data, and superimposition information data to be superimposed on the image by the left-eye image data and the right-eye image data in the user data area of the header portion;
- the data of the superimposition information included in the user data area of the header part including parallax information for shifting the superimposition information to be superimposed on the image based on the left eye image data and the right eye image data to give parallax
- a receiving unit that receives a video elementary stream in which the disparity information can be identified by user data identification information;
- a stereoscopic image data acquisition unit that acquires the stereoscopic image data from the payload portion of the video elementary stream received by the reception unit;
- a superimposition information data acquisition unit that acquires data of the superimposition information from a user data area of a header portion of the video elementary stream received by the reception unit;
- a disparity information acquisition unit that acquires the disparity information from a user data area of a header
- the video elementary stream is received by the receiving unit.
- This video elementary stream includes a stereoscopic image including left eye image data and right eye image data in a payload portion. Further, in this video elementary stream, the data of superimposition information and the disparity information are included in the user data area of the header portion so as to be identifiable by the user data identification information.
- the stereoscopic image data acquisition unit acquires the stereoscopic image data from the payload portion of the video elementary stream. Further, the superimposition information data acquisition unit acquires the superimposition information data from the user data area of the header portion of the video elementary stream. In addition, the disparity information acquisition unit acquires disparity information from the user data area in the header portion of the video elementary stream. As described above, the superimposition information data and the parallax information are included in the user data area so as to be identifiable by the user data identification information. Therefore, the parallax information can be favorably acquired from the user data area together with the superimposition information data based on the user data identification information.
- the image data processing unit uses the left eye image data and the right eye image data, the superimposition information data, and the parallax information, and gives the parallax to the same superimposition information to be superimposed on the left eye image and the right eye image. Data of the left eye image on which the superimposition information is superimposed and data of the right eye image on which the superposition information is superimposed are obtained.
- disparity information suitable for superimposition information (closed caption information, subtitle information, etc.) superimposed on the left eye image and the right eye image. Can be granted. Therefore, in the display of superimposition information, perspective consistency with each object in the image can be maintained in an optimum state.
- a stereoscopic image data output unit for outputting stereoscopic image data including left-eye image data and right-eye image data
- a parallax information output unit for outputting parallax information for shifting the superimposition information to be superimposed on the image based on the left-eye image data and the right-eye image data and providing parallax
- a transmission unit that transmits stereoscopic image data output from the stereoscopic image data output unit and parallax information output from the parallax information output unit;
- shift target designating information for designating superimposition information to be shifted based on the parallax information among superimposition information to be superimposed on the image based on the left eye image data and superimposition information superimposed on the image based on the right eye image data. Is added to the stereoscopic image data transmission device.
- the stereoscopic image data output unit outputs stereoscopic image data including left-eye image data and right-eye image data.
- the parallax information output unit outputs parallax information for adding parallax by shifting the superimposition information to be superimposed on the image based on the left eye image data and the right eye image data.
- the superimposition information means information to be superimposed on the image, such as closed caption information for displaying a caption and subtitle information.
- the transmission unit transmits the stereoscopic image data output from the stereoscopic image data output unit and the parallax information output from the parallax information output unit.
- the parallax information transmitted together with the stereoscopic image data includes the superimposition information to be shifted based on the parallax information among the superimposition information superimposed on the image based on the left eye image data and the superimposition information superimposed on the image based on the right eye image data. Shift target designation information for designating is added.
- parallax information is transmitted together with left-eye image data and right-eye image data for displaying a stereoscopic image. Therefore, on the receiving side, appropriate disparity can be given to the same superimposition information (closed caption information, subtitle information, etc.) superimposed on the left eye image and the right eye image using the disparity information. Therefore, in the display of superimposition information, the perspective consistency with each object in the image can be maintained in an optimum state.
- superimposition information closed caption information, subtitle information, etc.
- shift target designation information is added to the received parallax information. Therefore, on the receiving side, only the superimposition information to be superimposed on the left eye image is shifted, only the superimposition information to be superimposed on the right eye image is shifted, or the left eye image and the right eye are shifted by this shift target designation information. It is possible to shift superimposition information superimposed on both images.
- Another concept of the present invention is Stereoscopic image data including left-eye image data and right-eye image data, and parallax information for giving parallax by shifting superimposition information to be superimposed on an image based on the left-eye image data and the right-eye image data are received.
- parallax information For giving parallax by shifting superimposition information to be superimposed on an image based on the left-eye image data and the right-eye image data are received.
- shift target designating information for designating superimposition information to be shifted based on the parallax information among superimposition information to be superimposed on the image based on the left eye image data and superimposition information superimposed on the image based on the right eye image data are received.
- the stereoscopic image data receiving apparatus further includes an image data processing unit for obtaining data.
- the receiving unit receives a stereoscopic image including left-eye image data and right-eye image data, and parallax information.
- This parallax information is for giving parallax by shifting the superimposition information superimposed on the image by the left eye image data and the right eye image data.
- Shift information specifying information for superimposing information to be shifted based on the parallax information out of the superimposition information to be superimposed on the image based on the left eye image data and the superimposition information superimposed on the image based on the right eye image data is added to the parallax information.
- the left eye image data, the right eye image data, and the parallax information are used by the image data processing unit, the parallax is given to the same superimposition information to be superimposed on the left eye image and the right eye image, and the superimposition information is superimposed.
- Data of the left eye image and data of the right eye image on which the superimposition information is superimposed are obtained. In this case, based on the shift target designation information, only the superimposition information superimposed on the left eye image is shifted, only the superimposition information superimposed on the right eye image is shifted, or both the left eye image and the right eye image The superimposition information to be superimposed on is shifted.
- disparity information (with shift target designation information) suitable for the superimposition information is added to the superimposition information (closed caption information, subtitle information, etc.) superimposed on the left eye image and the right eye image.
- the superimposition information closed caption information, subtitle information, etc.
- an appropriate parallax can be given. Therefore, in the display of superimposition information, perspective consistency with each object in the image can be maintained in an optimum state.
- a stereoscopic image data output unit for outputting stereoscopic image data including left-eye image data and right-eye image data
- a parallax information output unit for outputting parallax information for shifting the superimposition information to be superimposed on the image based on the left-eye image data and the right-eye image data and providing parallax
- a transmission unit that transmits stereoscopic image data output from the stereoscopic image data output unit and parallax information output from the parallax information output unit
- the disparity information is in a stereoscopic image data transmission device that is disparity information that is sequentially updated in each frame of a predetermined number of frame periods in which the superimposition information is displayed.
- the stereoscopic image data output unit outputs stereoscopic image data including left-eye image data and right-eye image data.
- the parallax information output unit outputs parallax information for adding parallax by shifting the superimposition information to be superimposed on the image based on the left eye image data and the right eye image data.
- the superimposition information means information to be superimposed on the image, such as closed caption information for displaying a caption and subtitle information.
- the transmission unit transmits the stereoscopic image data output from the stereoscopic image data output unit and the parallax information output from the parallax information output unit.
- the disparity information transmitted together with the stereoscopic image data is disparity information that is sequentially updated in each frame of a predetermined number of frame periods in which the superimposition information is displayed.
- parallax information is transmitted together with left-eye image data and right-eye image data for displaying a stereoscopic image. Therefore, on the receiving side, appropriate disparity can be given to the same superimposition information (closed caption information, subtitle information, etc.) superimposed on the left eye image and the right eye image using the disparity information. Therefore, in the display of superimposition information, the perspective consistency with each object in the image can be maintained in an optimum state.
- superimposition information closed caption information, subtitle information, etc.
- the received disparity information is disparity information that is sequentially updated in each frame in a predetermined number of frame periods in which superimposition information is displayed. Therefore, on the reception side, parallax can be given to the superimposition information by the parallax information that is sequentially updated in each frame of a predetermined number of frame periods in which the superimposition information is displayed. Therefore, the parallax to be added to the superimposition information can be dynamically changed in conjunction with the change of the image content.
- the disparity information sequentially updated in each frame includes disparity information of the first frame in a predetermined number of frame periods and offset information of the second and subsequent frames relative to the disparity information of the previous frame. You may do it. In this case, the data amount of the parallax information can be suppressed.
- Another concept of the present invention is Stereoscopic image data including left-eye image data and right-eye image data, and parallax information for giving parallax by shifting superimposition information to be superimposed on an image based on the left-eye image data and the right-eye image data are received.
- the disparity information is disparity information that is sequentially updated in each frame of a predetermined number of frame periods in which the superimposition information is displayed,
- the left eye image data and the right eye image data included in the stereoscopic image data received by the receiving unit and the parallax information received by the receiving unit are used to superimpose the left eye image data and the right eye image.
- a stereoscopic image data receiving apparatus further comprising an image data processing unit that gives parallax to the same superimposition information and obtains data of a left eye image on which the superimposition information is superimposed and data of a right eye image on which the superimposition information is superimposed It is in.
- the receiving unit receives a stereoscopic image including left-eye image data and right-eye image data, and parallax information.
- This parallax information is for giving parallax by shifting the superimposition information superimposed on the image by the left eye image data and the right eye image data.
- the disparity information is disparity information that is sequentially updated in each frame in a predetermined number of frame periods in which superimposition information is displayed.
- the left eye image data, the right eye image data, and the parallax information are used by the image data processing unit, the parallax is given to the same superimposition information to be superimposed on the left eye image and the right eye image, and the superimposition information is superimposed. Data of the left eye image and data of the right eye image on which the superimposition information is superimposed are obtained.
- disparity information (with shift target designation information) suitable for the superimposition information is added to the superimposition information (closed caption information, subtitle information, etc.) superimposed on the left eye image and the right eye image.
- an appropriate parallax can be given. Therefore, in the display of superimposition information, perspective consistency with each object in the image can be maintained in an optimum state.
- parallax can be given to superimposition information by the parallax information sequentially updated in each frame of a predetermined number of frame periods in which superimposition information is displayed. Therefore, the parallax to be added to the superimposition information can be dynamically changed in conjunction with the change of the image content.
- a stereoscopic image data output unit for outputting stereoscopic image data including left-eye image data and right-eye image data
- a parallax information output unit for outputting parallax information for shifting the superimposition information to be superimposed on the image based on the left-eye image data and the right-eye image data and providing parallax
- a transmission unit that transmits stereoscopic image data output from the stereoscopic image data output unit and parallax information output from the parallax information output unit
- the disparity information is disparity information used in common for each frame of a predetermined number of frame periods in which the superimposition information is displayed,
- the disparity information is in the stereoscopic image data transmission apparatus to which information indicating whether the disparity information is commonly used in each frame or the disparity information sequentially updated in each frame is added.
- the stereoscopic image data output unit outputs stereoscopic image data including left-eye image data and right-eye image data.
- the parallax information output unit outputs parallax information for adding parallax by shifting the superimposition information to be superimposed on the image based on the left eye image data and the right eye image data.
- the superimposition information means information to be superimposed on the image, such as closed caption information for displaying a caption and subtitle information.
- the transmission unit transmits the stereoscopic image data output from the stereoscopic image data output unit and the parallax information output from the parallax information output unit.
- the disparity information transmitted together with the stereoscopic image data is disparity information that is commonly used in each frame of a predetermined number of frame periods in which the superimposition information is displayed.
- parallax information is transmitted together with left-eye image data and right-eye image data for displaying a stereoscopic image. Therefore, on the receiving side, appropriate disparity can be given to the same superimposition information (closed caption information, subtitle information, etc.) superimposed on the left eye image and the right eye image using the disparity information. Therefore, in the display of superimposition information, the perspective consistency with each object in the image can be maintained in an optimum state.
- superimposition information closed caption information, subtitle information, etc.
- the received disparity information is disparity information that is used in common for each frame in a predetermined number of frame periods in which superimposition information is displayed. Therefore, it is possible to greatly reduce the data amount of the parallax information.
- Another concept of the present invention is Stereoscopic image data including left-eye image data and right-eye image data, and parallax information for giving parallax by shifting superimposition information to be superimposed on an image based on the left-eye image data and the right-eye image data are received.
- the disparity information is disparity information used in common for each frame of a predetermined number of frame periods in which the superimposition information is displayed,
- the left eye image data and the right eye image data included in the stereoscopic image data received by the receiving unit and the parallax information received by the receiving unit are used to superimpose the left eye image data and the right eye image.
- a stereoscopic image data receiving apparatus further comprising an image data processing unit that gives parallax to the same superimposition information and obtains data of a left eye image on which the superimposition information is superimposed and data of a right eye image on which the superimposition information is superimposed It is in.
- the receiving unit receives a stereoscopic image including left-eye image data and right-eye image data, and parallax information.
- This parallax information is for giving parallax by shifting the superimposition information superimposed on the image by the left eye image data and the right eye image data.
- This disparity information is used as disparity information used in common for each frame in a predetermined number of frame periods in which superimposition information is displayed.
- the left eye image data, the right eye image data, and the parallax information are used by the image data processing unit, the parallax is given to the same superimposition information to be superimposed on the left eye image and the right eye image, and the superimposition information is superimposed. Data of the left eye image and data of the right eye image on which the superimposition information is superimposed are obtained.
- disparity information (with shift target designation information) suitable for the superimposition information is added to the superimposition information (closed caption information, subtitle information, etc.) superimposed on the left eye image and the right eye image.
- the disparity information is commonly used in each frame of a predetermined number of frame periods in which the superimposition information is displayed. Therefore, the processing load can be reduced and the capacity of the memory for storing the parallax information can be suppressed.
- a stereoscopic image data output unit for outputting stereoscopic image data including left-eye image data and right-eye image data;
- a parallax information output unit for outputting parallax information for shifting the superimposition information to be superimposed on the image based on the left-eye image data and the right-eye image data and providing parallax;
- a transmission unit that transmits stereoscopic image data output from the stereoscopic image data output unit and parallax information output from the parallax information output unit;
- the disparity information is disparity information used in common for each frame of a predetermined number of frame periods in which the superimposition information is displayed, or in each frame of a predetermined number of frame periods in which the superimposition information is displayed. Disparity information that is updated sequentially, In the stereoscopic image data transmitting apparatus, flag information indicating whether the disparity information is commonly used in each frame or the disparity information sequentially updated in each frame is added to the disparity information. .
- the stereoscopic image data output unit outputs stereoscopic image data including left-eye image data and right-eye image data.
- the parallax information output unit outputs parallax information for adding parallax by shifting the superimposition information to be superimposed on the image based on the left eye image data and the right eye image data.
- the superimposition information means information to be superimposed on the image, such as closed caption information for displaying a caption and subtitle information.
- the transmission unit transmits the stereoscopic image data output from the stereoscopic image data output unit and the parallax information output from the parallax information output unit.
- the disparity information transmitted together with the stereoscopic image data is disparity information used in common for each frame in a predetermined number of frame periods in which the superimposition information is displayed, or a predetermined number of frames in which the superimposition information is displayed.
- the disparity information is sequentially updated in each frame of the period. Then, flag information indicating whether the disparity information is commonly used in each frame or the disparity information sequentially updated in each frame is added to the disparity information.
- parallax information is transmitted together with left-eye image data and right-eye image data for displaying a stereoscopic image. Therefore, on the receiving side, appropriate disparity can be given to the same superimposition information (closed caption information, subtitle information, etc.) superimposed on the left eye image and the right eye image using the disparity information. Therefore, in the display of superimposition information, the perspective consistency with each object in the image can be maintained in an optimum state.
- superimposition information closed caption information, subtitle information, etc.
- disparity information that is commonly used in each frame or disparity information that is sequentially updated in each frame for example, according to the image content.
- disparity information that is sequentially updated in each frame is transmitted, and on the receiving side, the disparity added to the superimposition information is dynamically changed in conjunction with the change of the image content. Can do.
- the data amount of the disparity information can be suppressed by transmitting disparity information that is commonly used in each frame.
- Another concept of the present invention is Stereoscopic image data including left-eye image data and right-eye image data, and parallax information for giving parallax by shifting superimposition information to be superimposed on an image based on the left-eye image data and the right-eye image data are received.
- the disparity information is disparity information used in common for each frame of a predetermined number of frame periods in which the superimposition information is displayed, or in each frame of a predetermined number of frame periods in which the superimposition information is displayed.
- Disparity information that is updated sequentially In the disparity information, flag information indicating whether the disparity information is commonly used in each frame or the disparity information sequentially updated in each frame is added.
- the stereoscopic image data receiving apparatus further includes an image data processing unit for obtaining the image data.
- the receiving unit receives a stereoscopic image including left-eye image data and right-eye image data, and parallax information.
- This parallax information is for giving parallax by shifting the superimposition information superimposed on the image by the left eye image data and the right eye image data.
- the disparity information is disparity information that is commonly used in each frame of a predetermined number of frame periods in which the superimposition information is displayed, or is sequentially updated in each frame of the predetermined number of frame periods in which the superimposition information is displayed. This disparity information is added with flag information indicating whether the disparity information is commonly used in each frame or the disparity information that is sequentially updated in each frame.
- the left eye image data, the right eye image data, and the parallax information are used by the image data processing unit, the parallax is given to the same superimposition information to be superimposed on the left eye image and the right eye image, and the superimposition information is superimposed.
- Data of the left eye image and data of the right eye image on which the superimposition information is superimposed are obtained.
- appropriate disparity is obtained by using disparity information suitable for superimposition information (closed caption information, subtitle information, etc.) superimposed on the left eye image and the right eye image. Can be granted. Therefore, in the display of superimposition information, perspective consistency with each object in the image can be maintained in an optimum state.
- the disparity information that is commonly used in each frame or the disparity information that is sequentially updated in each frame is received. Which is received can be determined from the flag information.
- the received disparity information is disparity information that is sequentially updated in each frame, it is possible to dynamically change the disparity added to the superimposition information in conjunction with the change in the image content. Further, when the received disparity information is disparity information that is commonly used in each frame, the processing load can be reduced, and the capacity of the memory that stores the disparity information can be suppressed.
- the disparity information set suitable for the superimposition information for the superimposition information (closed caption information, subtitle information, graphics information, text information, etc.) superimposed on the left eye image and the right eye image, Appropriate parallax can be given, and the consistency of perspective between each object in the image can be maintained in an optimum state in displaying superimposition information.
- the case (2) it is a diagram illustrating that a set (PackedDisparity ⁇ Set) of a plurality of disparity information sets corresponding to image data of each fixed period of image data is transmitted together. It is a figure which shows an example of the correlation information (Component_Linked_Info) of component element ID and a disparity vector. It is a figure which shows an example of the parallax information set produced in case (3) (both the general-purpose parallax information set and the parallax information set for specific service use are sent).
- case (3) it is a figure which shows that the collection (PackedDisparity
- H.264 2 is a diagram illustrating a configuration example of user data in each of H.264 AVC and VC-1. It is a figure which shows the structural example of the disparity linkage descriptor (Disparity_Linkage_Descriptor) which associates disparity sequence ID (Disparity_Sequence_id) and a service.
- FIG. 1 shows a configuration example of a stereoscopic image display system 10 as an embodiment.
- the stereoscopic image display system 10 includes a broadcasting station 100, a set top box (STB) 200, and a television receiver 300.
- STB set top box
- the set-top box 200 and the television receiver 300 are HDMI (High Definition Multimedia interface) cable 400.
- the set top box 200 is provided with an HDMI terminal 202.
- the television receiver 300 is provided with an HDMI terminal 302.
- One end of the HDMI cable 400 is connected to the HDMI terminal 202 of the set top box 200, and the other end of the HDMI cable 400 is connected to the HDMI terminal 302 of the television receiver 300.
- the broadcast station 100 transmits bit stream data on a broadcast wave.
- the bit stream data includes stereoscopic image data including left-eye image data and right-eye image data, audio data, superimposition information data, and disparity information (disparity vector).
- the superimposition information data is closed caption data, subtitle data, graphics data, text data, and the like.
- FIG. 2 shows a configuration example of the transmission data generation unit 110 that generates the above-described bit stream data in the broadcast station 100.
- This configuration example is an example in which a disparity vector is transmitted as numerical information.
- the transmission data generation unit 110 includes cameras 111L and 111R, a video framing unit 112, a video encoder 113, a parallax vector detection unit 114, and a parallax vector encoder 115.
- the transmission data generation unit 110 includes a microphone 116, an audio encoder 117, a subtitle / graphics generation unit 118, a subtitle / graphics encoder 119, a text generation unit 120, a text encoder 121, and a multiplexer 122. ing. In this embodiment, it is assumed that the text generation unit 120 also serves as a closed caption data generation unit.
- the closed caption data is text data for displaying closed caption captions.
- the camera 111L captures a left eye image and obtains left eye image data for stereoscopic image display.
- the camera 111R captures the right eye image and obtains right eye image data for stereoscopic image display.
- the video framing unit 112 processes the left eye image data obtained by the camera 111L and the right eye image data obtained by the camera 111R into a state corresponding to the transmission method.
- Example of transmission method of stereoscopic image data Here, the following first to third methods are listed as transmission methods for stereoscopic image data (3D image data), but other transmission methods may be used.
- 3D image data 3D image data
- the first transmission method is the “Top & Bottom” method, as shown in FIG. 4A, in which the left eye image data is transmitted in the first half in the vertical direction and the left eye in the second half in the vertical direction.
- This is a method for transmitting data of each line of image data.
- the vertical resolution is halved with respect to the original signal.
- the second transmission method is the “Side By Side” method, as shown in FIG. 4B, pixel data of the left eye image data is transmitted in the first half in the horizontal direction, and right eye image data in the second half in the horizontal direction.
- This is a method for transmitting pixel data.
- the pixel data in the horizontal direction is thinned out to 1/2.
- the horizontal resolution is halved.
- the third transmission method is a “Frame Sequential” method, in which left-eye image data and right-eye image data are sequentially switched and transmitted for each field as shown in FIG.
- the video encoder 113 performs encoding such as MPEG4-AVC, MPEG2, VC-1, etc. on the stereoscopic image data processed by the video framing unit 112 to obtain encoded video data.
- the video encoder 113 includes a stream formatter 113a in the subsequent stage.
- the stream formatter 113a generates a video elementary stream including encoded video data in the payload portion.
- the disparity vector detection unit 114 detects a disparity vector that is disparity information of the other of the left eye image and the right eye image at a predetermined position in the image based on the left eye image data and the right eye image data.
- the predetermined position in the image is all pixel positions, a representative position of each area composed of a plurality of pixels, or a superimposition information, here, a representative position of an area where graphic information or text information is superimposed.
- Disposity vector detection A detection example of a disparity vector will be described. Here, an example in which the parallax vector of the right eye image with respect to the left eye image is detected will be described. As shown in FIG. 5, the left eye image is a detected image, and the right eye image is a reference image. In this example, the disparity vectors at the positions (xi, yi) and (xj, yj) are detected.
- a case where a disparity vector at the position of (xi, yi) is detected will be described as an example.
- an 8 ⁇ 8 or 16 ⁇ 16 pixel block (parallax detection block) Bi is set in the left eye image with the pixel at the position (xi, yi) at the upper left. Then, a pixel block matching the pixel block Bi is searched in the right eye image.
- a search range centered on the position of (xi, yi) is set in the right eye image, and each pixel in the search range is sequentially set as a pixel of interest, for example, 8 ⁇ 8 as in the pixel block Bi described above.
- 16 ⁇ 16 comparison blocks are sequentially set.
- the sum of the absolute differences for each corresponding pixel is obtained.
- the pixel block Bi when the pixel value of the pixel block Bi is L (x, y) and the pixel value of the comparison block is R (x, y), the pixel block Bi, a certain comparison block, The sum of absolute differences between the two is represented by ⁇
- the disparity vector at the position (xi, yi) is detected as (xi′ ⁇ xi, yi′ ⁇ yi).
- an 8 ⁇ 8 or 16 ⁇ 16 pixel block in the left-eye image with the pixel at the position (xj, yj) at the upper left Bj is set and detected in the same process.
- FIG. 7A shows an example of the parallax vector VV detected by the parallax vector detection unit 114 at a predetermined position in the image.
- FIG. 7B if the left eye image (detected image) is shifted by the parallax vector VV at a predetermined position in the image, it means that it overlaps with the right eye image (reference image). .
- the disparity vector encoder 115 generates an elementary stream of disparity vectors including the disparity vector detected by the disparity vector detection unit 114.
- the elementary stream of disparity vectors includes the following contents. That is, ID (ID_Block), vertical position information (Vertical_Position), horizontal position information (Horizontal_Position), and disparity vector (View_Vector) are set as one set. Then, this one set is repeated for N blocks, which is the number of parallax detection blocks.
- FIG. 8 shows the transmission content of the disparity vector.
- the disparity vector includes a vertical direction component (View_Vector_Vertical) and a horizontal direction component (View_Vector_Horizontal).
- the vertical and horizontal positions of the parallax detection block are offset values in the vertical and horizontal directions from the upper left origin of the image to the upper left pixel of the block.
- the reason for attaching the ID of the parallax detection block to the transmission of each parallax vector is to enable a link with a superimposition information pattern such as closed caption information, subtitle information, graphics information, text information, etc. to be superimposed and displayed on the image. It is.
- horizontal and vertical disparity vectors are 16-bit information.
- the number of bits of the disparity vector is not limited to 16 bits.
- An appropriate number of bits can be taken depending on the display size or the size of parallax to be applied, such as 8 bits described later.
- the transmission contents include the IDs of the disparity detection blocks A to F as shown in FIG. 9 (b). , Vertical and horizontal position information, and disparity vectors.
- ID2 indicates the ID of the disparity detection block A
- Ha, Va indicates the vertical and horizontal position information of the disparity detection block A
- the disparity vector a Indicates a disparity vector of the disparity detection block A.
- timing for detecting and transmitting a disparity vector will be described.
- this timing for example, the following first to fourth examples can be considered.
- the disparity vector is transmitted in units of pictures.
- This picture unit is the finest unit for transmitting a disparity vector.
- FIG. 10B it is synchronized with a video scene. In this case, the disparity vector is transmitted in scene units.
- the display is synchronized with the display start timing of the subtitle information, graphics information, text information, and the like displayed superimposed on the image.
- the microphone 116 detects sound corresponding to the images photographed by the cameras 111 ⁇ / b> L and 111 ⁇ / b> R, and obtains sound data.
- the audio encoder 117 performs MPEG-2Audio on the audio data obtained by the microphone 116. Encoding such as AAC is performed to generate an audio elementary stream.
- the subtitle / graphics generating unit 118 generates subtitle information and graphics information data (subtitle data, graphics data) to be superimposed on the image.
- the subtitle information is, for example, a caption.
- the graphics information is, for example, a logo.
- the subtitle data and graphics data are bitmap data.
- the subtitle data and graphics data are added with idling offset information indicating the superimposed position on the image.
- This idling offset information indicates, for example, offset values in the vertical and horizontal directions from the upper left origin of the image to the upper left pixel of the superimposed position of the subtitle information and graphics information.
- the standard for transmitting caption data as bitmap data is standardized and operated as DVB_Subtitling in DVB, which is a European digital broadcasting standard.
- the subtitle / graphic encoder 119 inputs the subtitle information generated by the subtitle / graphics generation unit 118 and the data of the graphics information (subtitle data, graphics data). Then, the subtitle / graphic encoder 119 generates an elementary stream including these data in the payload portion.
- the text generator 120 generates text information data (text data) to be superimposed on the image.
- the text information is, for example, an electronic program guide or text broadcast content.
- idling offset information indicating the superimposed position on the image is added to the text data.
- This idling offset information indicates, for example, offset values in the vertical and horizontal directions from the upper left origin of the image to the upper left pixel of the text information overlapping position.
- Examples of text data transmission include EPG used as a program reservation and CC_data (Closed Caption) of the American digital terrestrial standard ATSC.
- the text encoder 121 inputs the text data generated by the text generator 120. Then, the text encoder 121 generates an elementary stream including these data in the payload portion.
- the multiplexer 122 multiplexes the packetized elementary streams output from the encoders 113, 115, 117, 119, and 121.
- the multiplexer 122 outputs bit stream data (transport stream) BSD as transmission data.
- the operation of the transmission data generation unit 110 shown in FIG. 2 will be briefly described.
- the camera 111L captures a left eye image.
- Left-eye image data for stereoscopic image display obtained by the camera 111L is supplied to the video framing unit 112.
- the camera 111R captures a right eye image.
- the right eye image data for stereoscopic image display obtained by the camera 111R is supplied to the video framing unit 112.
- the left-eye image data and the right-eye image data are processed into a state corresponding to the transmission method to obtain stereoscopic image data (see FIGS. 4A to 4C).
- the stereoscopic image data obtained by the video framing unit 112 is supplied to the video encoder 113.
- the stereoscopic image data is encoded by MPEG4-AVC, MPEG2, VC-1, or the like, and a video elementary stream including the encoded video data is generated. This video elementary stream is supplied to the multiplexer 122.
- the left eye image data and right eye image data obtained by the cameras 111L and 111R are supplied to the parallax vector detection unit 114 through the video framing unit 112.
- a disparity detection block is set at a predetermined position in the image based on the left eye image data and the right eye image data, and is the other disparity information for one of the left eye image and the right eye image. A disparity vector is detected.
- the disparity vector at a predetermined position in the image detected by the disparity vector detection unit 114 is supplied to the disparity vector encoder 115.
- the ID of the parallax detection block, the vertical position information of the parallax detection block, the horizontal position information of the parallax detection block, and the parallax vector are passed as one set.
- the disparity vector encoder 115 generates a disparity vector elementary stream including disparity vector transmission contents (see FIG. 8). This disparity vector elementary stream is supplied to the multiplexer 122.
- the microphone 116 detects sound corresponding to the images photographed by the cameras 111L and 111R. Audio data obtained by the microphone 116 is supplied to the audio encoder 117. In this audio encoder 117, MPEG-2Audio is processed for audio data. Encoding such as AAC is performed, and an audio elementary stream including encoded audio data is generated. This audio elementary stream is supplied to the multiplexer 122.
- the subtitle / graphics generation unit 118 generates subtitle information and graphics information data (subtitle data, graphics data) to be superimposed on the image.
- This data (bitmap data) is supplied to the subtitle / graphic encoder 119.
- the subtitle / graphics data is added with idling offset information indicating the superimposed position on the image.
- the graphics data is subjected to predetermined encoding, and an elementary stream including the encoded data is generated. This elementary stream is supplied to the multiplexer 122.
- the text generator 120 generates text information data (text data) to be superimposed on the image.
- This text data is supplied to the text encoder 121.
- idling offset information indicating the superimposed position on the image is added to the text data.
- the text encoder 121 the text data is subjected to predetermined encoding, and an elementary stream including the encoded data is generated. This elementary stream is supplied to the multiplexer 122.
- the multiplexer 122 multiplexes the elementary stream packets supplied from the encoders to obtain bit stream data (transport stream) BSD as transmission data.
- FIG. 12 shows an example of each data stream multiplexed in the transmission data generation unit 110 shown in FIG. This example shows a case where a disparity vector is detected and transmitted in units of video scenes (see FIG. 10B).
- Each stream packet is given a time stamp for synchronous display, and the reception side can control the superimposition timing of subtitle information, graphics information, text information, and the like on the image. .
- “Other configuration examples of the transmission data generator" 2 is configured to transmit disparity vector transmission content (see FIG. 8) to the receiving side as an independent elementary stream.
- the transmission content of the disparity vector is embedded in another stream for transmission.
- the transmission content of the disparity vector is embedded and transmitted as user data in a video stream.
- the transmission content of a disparity vector is embedded and transmitted in a subtitle, graphics, or text stream.
- FIG. 13 shows a configuration example of the transmission data generation unit 110A.
- This example is also an example in which a disparity vector is transmitted as numerical information.
- This transmission data generation unit 110A is configured to transmit the transmission content of the disparity vector by embedding it in the video stream as user data.
- parts corresponding to those in FIG. 2 are denoted by the same reference numerals, and detailed description thereof is omitted.
- the disparity vector at a predetermined position in the image detected by the disparity vector detection 114 is supplied to the stream formatter 113a in the video encoder 113.
- the ID of the parallax detection block, the vertical position information of the parallax detection block, the horizontal position information of the parallax detection block, and the parallax vector are passed as one set.
- the transmission content of the disparity vector (see FIG. 8) is embedded as user data in the video stream.
- the transmission data generation unit 110A shown in FIG. 13 is otherwise configured in the same manner as the transmission data generation unit 110 shown in FIG.
- FIG. 14 shows an example of an image data stream, a subtitle or graphics data stream, and a text data stream multiplexed in the transmission data generation unit 110A shown in FIG.
- the disparity vector (disparity information) is embedded in the image data stream and transmitted.
- FIG. 15 illustrates a configuration example of the transmission data generation unit 110B. This example is also an example in which a disparity vector is transmitted as numerical information.
- the transmission data generation unit 110B is configured to transmit the transmission content of the disparity vector by embedding it in a subtitle or graphics data stream.
- parts corresponding to those in FIG. 2 are denoted by the same reference numerals, and detailed description thereof is omitted.
- the disparity vector at a predetermined position in the image detected by the disparity vector detection 114 is supplied to the stream formatter 119a in the subtitle / graphics encoder 119.
- the ID of the parallax detection block, the vertical position information of the parallax detection block, the horizontal position information of the parallax detection block, and the parallax vector are passed as one set.
- disparity vector transmission contents are embedded in a subtitle or graphics data stream.
- the other parts of the transmission data generation unit 110B shown in FIG. 15 are configured in the same manner as the transmission data generation unit 110 shown in FIG. 2 and operate in the same manner.
- FIG. 16 shows an example of an image data stream, a subtitle or graphics data stream, and a text data stream multiplexed in the transmission data generation unit 110B shown in FIG.
- the disparity vector (disparity information) is embedded and transmitted in a subtitle or graphics data stream.
- the transmission data generation units 110, 110A, and 110B shown in FIGS. 2, 13, and 15 described above transmit parallax vectors as numerical information (see FIG. 8).
- the transmission side reflects the parallax information in advance on the data of superimposition information (for example, subtitle information, graphics information, text information, etc.) to be superimposed on the image, and transmits it. Is also possible.
- the transmission side when reflecting in the graphics information data, the transmission side generates graphics data corresponding to both the left-eye graphics information to be superimposed on the left-eye image and the right-eye graphics information to be superimposed on the right-eye image.
- the left eye graphics information and the right eye graphics information are the same graphics information.
- the display position in the image is shifted in the horizontal direction by, for example, the horizontal component of the parallax vector corresponding to the display position of the right-eye graphics information with respect to the left-eye graphics information.
- parallax vector a parallax vector detected at a plurality of positions in the image and corresponding to the superimposed position is used. Further, for example, as the parallax vector, a parallax vector at a position that is recognized closest in terms of perspective among parallax vectors detected at a plurality of positions in the image is used. Although the detailed description is omitted, the same applies to the case where the parallax information is reflected in the data of the subtitle information or the graphics information.
- FIG. 17A shows the superimposed positions of the left-eye graphics information and the right-eye graphics information when the transmission method is the above-described first transmission method (“Top ⁇ & ⁇ Bottom” method).
- These left-eye graphics information and right-eye graphics information are the same information.
- the right eye graphics information RGI superimposed on the right eye image IR is shifted in the horizontal direction by the horizontal component VVT of the parallax vector. It is said that.
- graphics data is generated so that the graphics information LGI and RGI are superimposed on the images IL and IR.
- the viewer can observe the graphics information LGI and RGI together with the images IL and IR with parallax, and can recognize the perspective in the graphics information.
- the graphics data of each graphics information LGI, RGI is generated as single area data as shown in FIG.
- the data other than the graphics information LGI and RGI may be generated as transparent data.
- the graphics data of each graphics information LGI, RGI is generated as data of another area as shown in FIG.
- FIG. 19A shows the superimposed positions of left-eye graphics information and right-eye graphics information when the transmission method is the above-described second transmission method (“SideSBy Side” method).
- These left-eye graphics information and right-eye graphics information are the same information.
- the right eye graphics information RGI superimposed on the right eye image IR is shifted in the horizontal direction by the horizontal component VVT of the parallax vector. It is said that. IT is an idling offset value.
- graphics data is generated so that the graphics information LGI and RGI are superimposed on the images IL and IR. Accordingly, as shown in FIG. 19B, the viewer can observe the graphics information LGI and RGI together with the images IL and IR with parallax, and can recognize the perspective in the graphics information.
- the graphics data of each graphics information LGI and RGI is generated as single area data as shown in FIG.
- the data other than the graphics information LGI and RGI may be generated as transparent data.
- FIG. 21 shows a configuration example of the transmission data generation unit 110C.
- the transmission data generation unit 110C is configured to transmit parallax information in data of superimposition information such as closed caption information, subtitle information, graphics information, text information, and the like.
- portions corresponding to those in FIG. 2 are denoted by the same reference numerals, and detailed description thereof is omitted.
- a subtitle / graphics processing unit 124 is inserted between the subtitle / graphics generation unit 118 and the subtitle / graphic encoder 119.
- a text processing unit 125 is inserted between the text generation unit 120 and the text encoder 121. The disparity vector at a predetermined position in the image detected by the disparity vector detection unit 114 is supplied to the subtitle / graphics processing unit 124 and the text processing unit 125.
- the subtitle / graphics processing unit 124 generates left eye and right eye subtitles and graphics information LGI and RGI data to be superimposed on the left eye image IL and the right eye image IR. In this case, it is generated based on the subtitle data or graphics data generated by the subtitle / graphics generating unit 118.
- the left-eye and right-eye subtitle information and graphics information are the same information. However, the superposition position in the image is shifted in the horizontal direction by the horizontal component VVT of the disparity vector, for example, with respect to the subtitle information or graphics information of the left eye. (See FIG. 17 (a) and FIG. 19 (a)).
- the subtitle data and graphics data generated in this way by the subtitle / graphics processing unit 124 are supplied to the subtitle / graphics encoder 119. Note that idling offset information indicating a superimposed position on an image is added to the subtitle data and graphics data.
- the subtitle / graphic encoder 119 generates an elementary stream of subtitle data and graphics data generated by the subtitle / graphics processing unit 124.
- the text processing unit 125 based on the text data generated by the text generation unit 120, the left-eye text information data superimposed on the left-eye image and the right-eye text information superimposed on the right-eye image. Data is generated.
- the left-eye text information and the right-eye text information are the same text information, but the superimposed position in the image is, for example, the left-eye text information, and the right-eye text information is the horizontal component of the disparity vector. Only VVT is shifted in the horizontal direction.
- the text data generated by the text processing unit 125 is supplied to the text encoder 121.
- idling offset information indicating the superimposed position on the image is added to the text data.
- the text encoder 121 generates an elementary stream of text data generated by the text processing unit.
- the transmission data generation unit 110C shown in FIG. 21 is otherwise configured in the same manner as the transmission data generation unit 110 shown in FIG.
- the disparity vector detection unit 114 uses a disparity vector (disparity) at a predetermined position in the image based on the left eye image data and the right eye image data. Information) is detected.
- the transmission data generation units 110, 110A, and 110B are configured to transmit the disparity information at a predetermined position in the image detected by the disparity vector detection unit 114 as it is to the reception side.
- a disparity information set including disparity information belonging to is selected based on, for example, the spatial density of the disparity vector requested by the receiving side, the transmission band, or the like.
- disparity information set a general-purpose disparity information set and a disparity information set for a specific service application can be considered.
- specific service applications include closed captions and subtitles.
- the following cases are conceivable as the disparity information set transmitted to the receiving side. (1) Send only a general-purpose disparity information set (when not related to a specific service) (2) Send only a general-purpose disparity information set (when related to a specific service) (3) Send both a general-purpose disparity information set and a specific service use disparity information set. (4) Send only a specific service use disparity information set.
- FIG. 22 shows a configuration example of the transmission data generation unit 110D.
- This transmission data generation unit 110D is configured to create a predetermined set of disparity information sets corresponding to any of the above cases (1) to (4) and transmit the predetermined set of disparity information sets to the reception side. ing.
- parts corresponding to those in FIG. 2 are denoted by the same reference numerals, and detailed description thereof is omitted as appropriate.
- the transmission data generation unit 110D includes a data extraction unit (archive unit) 130, a parallax information set creation unit 131, a video encoder 113, and an audio encoder 117.
- the transmission data generating unit 110D includes a subtitle / graphics generating unit 118, a subtitle / graphics encoder 119, a text generating unit 120, a text encoder 121, and a multiplexer 122.
- a data recording medium 130a is detachably attached to the data extraction unit 130, for example.
- audio data and parallax vectors are recorded in association with stereo image data including left eye image data and right eye image data.
- the data extraction unit 130 extracts and outputs stereoscopic image data, audio data, disparity vectors, and the like from the data recording medium 130a.
- the data recording medium 130a is a disk-shaped recording medium, a semiconductor memory, or the like.
- the stereoscopic image data recorded on the data recording medium 130a corresponds to the stereoscopic image data obtained by the video framing unit 112 of the transmission data generation unit 110 shown in FIG.
- the parallax vector recorded on the data recording medium 130a is, for example, a parallax vector for each pixel (pixel) constituting the image.
- the parallax information set creation unit 131 has a viewer function.
- the disparity information set creation unit 131 performs a downsizing process on the disparity vector output from the data extraction unit 130, that is, the disparity vector for each pixel, thereby generating a disparity vector belonging to a predetermined region.
- FIG. 23 shows an example of data in the relative depth direction given as the luminance value of each pixel (pixel).
- the data in the relative depth direction can be handled as a disparity vector for each pixel by a predetermined conversion.
- the luminance value of the person portion is high. This means that the value of the parallax vector of the person portion is large, and therefore, in stereoscopic image display, this means that the person portion is perceived as being raised.
- the luminance value of the background portion is low. This means that the value of the parallax vector in the background portion is small, and therefore, in stereoscopic image display, this means that the background portion is perceived as a sunken state.
- FIG. 24 shows an example of a disparity vector for each block.
- the block corresponds to an upper layer of pixels (picture elements) located at the lowermost layer.
- This block is configured by dividing an image (picture) region into a predetermined size in the horizontal direction and the vertical direction.
- the disparity vector of each block is obtained, for example, by selecting the disparity vector having the largest value from the disparity vectors of all pixels (pixels) existing in the block.
- the disparity vector of each block is indicated by an arrow, and the length of the arrow corresponds to the magnitude of the disparity vector.
- FIG. 25 shows an example of the downsizing process performed by the disparity information set creation unit 131.
- the disparity information set creation unit 134 obtains a disparity vector for each block using a disparity vector for each pixel (pixel).
- a block corresponds to an upper layer of pixels located at the lowest layer, and is configured by dividing an image (picture) region into a predetermined size in the horizontal direction and the vertical direction.
- the disparity vector of each block is obtained, for example, by selecting the disparity vector having the largest value from the disparity vectors of all the pixels (pixels) existing in the block.
- the disparity information set creation unit 131 uses a disparity vector for each block as illustrated in FIG.
- the disparity vector for each block is obtained.
- a group is an upper layer of a block, and is obtained by grouping a plurality of adjacent blocks together.
- each group is composed of four blocks bounded by a broken line frame.
- the disparity vector of each group is obtained, for example, by selecting the disparity vector having the largest value from the disparity vectors of all blocks in the group.
- the disparity information set creation unit 131 obtains a disparity vector for each partition (Partition) using the disparity vector for each group as shown in FIG.
- the partition is an upper layer of the group and is obtained by grouping a plurality of adjacent groups together.
- each partition is composed of two groups bounded by a broken line frame.
- the disparity vector of each partition is obtained, for example, by selecting the disparity vector having the largest value from the disparity vectors of all groups in the partition.
- the disparity information set creation unit 131 obtains the disparity vector of the entire picture (entire image) located in the highest layer using the disparity vector for each partition.
- the entire picture includes four regions that are bounded by a broken line frame.
- the disparity vector for the entire picture is obtained, for example, by selecting the disparity vector having the largest value from the disparity vectors for all regions included in the entire picture.
- information on the position (illustrated by “+”) of the original pixel (pixel) from which the disparity vector was obtained can be obtained and used as additional information of the disparity vector. . This is the same when obtaining the above-described disparity vectors of blocks, groups, and partitions.
- the disparity information set creation unit 131 performs the downsizing process on the disparity vector for each pixel (pixel) located in the lowest layer, and the disparity of each region in each layer of the block, group, partition, and entire picture A vector can be obtained.
- the four-layer disparity vectors of the block, group, partition, and entire picture are obtained in addition to the pixel (pixel) layer.
- the method of cutting the area of the hierarchy and the number of areas are not limited to this.
- the disparity information set creation unit 131 creates a disparity information set including a disparity vector for each region having a predetermined size by the above-described downsizing process.
- Information indicating the screen area to which each piece of parallax information belongs is added to the parallax information set. For example, information indicating the number of screen divisions such as 0 division, 4 divisions, 9 divisions, and 16 divisions is added as information indicating the screen area. Further, as information indicating the screen area, for example, information indicating the size of the area (block size) such as 16 ⁇ 16, 32 ⁇ 32, and 64 ⁇ 64 is added.
- the screen area to which each piece of parallax information belongs can be grasped from the information indicating the number of divisions of the screen or the information indicating the size of the area.
- the information indicating the number of screen divisions and the information indicating the size of the area are information indicating the screen area to which each parallax information belongs.
- the disparity information set creation unit 131 creates a disparity information set for a specific service use as necessary. For example, a disparity information set for closed captions, subtitles, graphics for a specific application (widget), or the like is created as necessary. In this case, the disparity vectors belonging to one or more screen areas corresponding to one or more display areas of superimposition information (closed caption information, subtitle information, graphics information, etc.) by the above-described downsizing process. Is created.
- FIG. 26A shows a display example of subtitles and application graphics on an image.
- FIG. 26B shows the perspective of images, subtitles, and application graphics.
- the subtitle is displayed in the center area of the lower half of the screen, but it is desired that the subtitle is recognized further forward than the closest position of the image.
- the disparity information set creation unit 131 creates a disparity vector for giving disparity to the subtitle from the display area of the subtitle and the disparity vector of the corresponding image area.
- the disparity information set creation unit 131 creates a disparity vector for giving disparity to the application graphics from the display region of the application graphics and the disparity vector of the corresponding image region.
- the disparity information set creation unit 131 has a viewer function as described above.
- the parallax information set creation unit 131 actually displays a stereoscopic image as shown in FIG. Therefore, the parallax information set creation unit 131 is supplied with stereoscopic image data (left-eye image data and right-eye image data) output from the data extraction unit 130.
- the disparity information set creation unit 131 includes subtitle data or graphics data generated by the subtitle / graphics generation unit 118, and text data (including closed caption data) generated by the text generation unit 120. Supplied.
- a disparity sequence ID (Disparity_Sequence_id) is added to each set of disparity information sets created by the disparity information set creation unit 131 and transmitted together with stereoscopic image data.
- the disparity sequence ID is identification information indicating the type of superimposition information that should use the disparity information included in the disparity information set.
- the disparity sequence ID identifies whether the disparity information set is a general-purpose disparity information set or a disparity information set for a specific service application. Further, this identification information identifies which parallax information set for a specific service application is a parallax information set for a specific service application.
- Case (1) is a case where only the general-purpose disparity information set is sent as described above. In this case (1), only a general-purpose disparity information set is created. In this case (1), the disparity information included in the disparity information set is not related to the specific service.
- FIG. 28 shows an example of a disparity information set created in this case (1).
- the disparity information set includes disparity vectors (horizontal disparity vectors) P0 to P8 belonging to each screen area obtained by dividing the screen into nine.
- the disparity vectors P0 to P8 are obtained by the above-described downsizing process using, for example, the disparity vector for each pixel (pixel).
- an arrow indicates a disparity vector, and the length of the arrow corresponds to the magnitude of the disparity vector.
- This disparity information set is a general-purpose disparity information set. Therefore, the identification information “Diparity_Sequence_id” added to the disparity information set is “0” indicating that it is a general-purpose disparity information set as shown in FIG.
- the general-purpose disparity information set is a disparity information set for each predetermined unit of image data, for example, for each picture of encoded video.
- the image data is divided every fixed period, and a set of plural disparity information sets (Packed DisparityparSet) corresponding to the image data of each fixed period is transmitted prior to the transmission of the image data of each fixed period.
- Sent. Therefore, as shown in FIG. 29, an offset (Offset) is added to the disparity information set for each predetermined unit as time information indicating the timing of using the disparity information set.
- the offset added to each disparity information set indicates the time from the usage time of the first disparity information set indicated by a time stamp (TimeStamp).
- Case (2) is a case where only the general-purpose disparity information set is sent as described above. In this case (2), only a general-purpose disparity information set is created. In this case (2), the disparity information included in the disparity information set is related to the specific service.
- FIG. 30 shows an example of a disparity information set created in this case (2).
- the disparity information set includes disparity vectors (horizontal disparity vectors) P0 to P8 belonging to each screen area obtained by dividing the screen into nine.
- the disparity vectors P0 to P8 are obtained by the above-described downsizing process using, for example, the disparity vector for each pixel (pixel).
- an arrow indicates a parallax vector, and the length of the arrow corresponds to the magnitude of the parallax vector.
- This disparity information set is a general-purpose disparity information set. Therefore, the identification information “Diparity_Sequence_id” added to the disparity information set is “0” indicating that it is a general-purpose disparity information set as shown in FIG.
- the general-purpose disparity information set is a disparity information set for each predetermined unit of image data, for example, for each picture of encoded video.
- the image data is divided every fixed period, and a set of plural disparity information sets (Packed DisparityparSet) corresponding to the image data of each fixed period is transmitted prior to the transmission of the image data of each fixed period.
- Sent. Therefore, as shown in FIG. 31, an offset (Offset) is added to the disparity information set for each predetermined unit as time information indicating the timing of using the disparity information set.
- the offset added to each disparity information set indicates the time from the usage time of the first disparity information set indicated by a time stamp (TimeStamp).
- the disparity information included in the disparity information set is related to the specific service.
- FIG. 31 shows an example in which the disparity vector P7 and the closed caption data DF0 are associated with each other. Note that the closed caption information by the closed caption data DF0 is displayed in the screen area to which the disparity vector P7 belongs, as shown in FIG. In this case, the component element ID (Component_Element ID) indicating the screen area for displaying the closed caption information is associated with the disparity vector P7.
- Component_Element ID Component_Element ID
- FIG. 32 shows an example of an association table (Component_Linked_Info) between component element IDs and disparity vectors in that case.
- the target position of the disparity vector is specified by the partition position ID (Partition_Position_ID).
- Partition_Position_ID indicates the number of the screen area among the plurality of screen areas specified by the number of divisions or the block size.
- FIG. 31 illustrates an example in which one piece of disparity information is associated with one piece of closed caption data.
- a plurality of pieces of disparity information may be associated with one closed caption data.
- any disparity information is used to give disparity to the closed caption information based on the closed caption data. Can be selected and used.
- the disparity vectors P6 and P7 are associated with the closed caption data DF1, and further, the disparity vectors P7 and P8 are associated with the closed caption data DF2.
- the component element ID indicating the screen area for displaying the closed caption information based on the closed caption data DF1 is associated with the disparity vectors P6 and P7.
- the component element ID indicating the screen area for displaying the closed caption information based on the closed caption data DF2 is associated with the disparity vectors P7 and P8.
- FIG. 34 shows an association table (Component_Linked_Info) between component element IDs and disparity vectors in that case.
- the disparity vector is specified by a partition position ID (Partition_Position_ID).
- Partition_Position_ID indicates the number of the screen area among the plurality of screen areas specified by the number of divisions or the block size.
- Case (3) is a case where both the general-purpose disparity information set and the specific service-use disparity information set are sent as described above.
- this case (3) both a general-purpose disparity information set and a specific-service-use disparity information set are created.
- the identification information (Disparity_Sequence_id) added to the general-purpose disparity information set is “0”.
- the identification information (Disparity_Sequence_id) added to the disparity information set for specific service use is a value other than “0”.
- the association of the identification information (Disparity_Sequence_id) with the service is performed, for example, by inserting the information into the user data as described above. Alternatively, this association is performed using, for example, a disparity linkage descriptor (Disparity_Linkage_Descriptor) as described above.
- FIG. 35 shows an example of a disparity information set created in this case (3).
- the general-purpose disparity information set includes disparity vectors (horizontal disparity vectors) P0 to P8 belonging to each screen area obtained by dividing the screen into nine.
- the disparity vectors P0 to P8 are obtained by the above-described downsizing process using, for example, the disparity vector for each pixel (pixel).
- an arrow indicates a disparity vector, and the length of the arrow corresponds to the magnitude of the disparity vector.
- a disparity information set for closed captions is created as a disparity information set for specific service applications.
- the disparity information set for closed caption includes a disparity vector CC1 corresponding to the closed caption information “Caption1” and a disparity vector CC2 corresponding to the closed caption information “Caption2”.
- the disparity information set is a disparity information set for each predetermined unit of image data, for example, for each picture of encoded video.
- the image data is divided every fixed period, and a set of plural disparity information sets (Packed DisparityparSet) corresponding to the image data of each fixed period is transmitted prior to the transmission of the image data of each fixed period.
- Sent. Therefore, as shown in FIG. 36, an offset (Offset) is added to the disparity information set for each predetermined unit as time information indicating the timing of using the disparity information set.
- the offset added to each disparity information set indicates the time from the usage time of the first disparity information set indicated by a time stamp (TimeStamp).
- the disparity information set for specific service use is a disparity information set for closed captions.
- FIG. 35 illustrates an example in which the disparity vector CC1 and the closed caption data DF1 are associated with each other, and the disparity vector CC2 and the closed caption data DF2 are associated with each other.
- the component element ID (Component_Element ID) indicating the screen area on which the closed caption information is displayed is associated with the disparity vectors CC1 and CC2.
- FIG. 37 shows an association table (Component_Linked_Info) between component element IDs and disparity vectors in that case.
- the target position partition position ID Partition_Position ID
- This partition position ID indicates the number of the screen area among the plurality of screen areas specified by the number of divisions or the block size.
- Case (4) is a case in which only a disparity information set for a specific service application is sent as described above. In this case (4), only the disparity information set for specific service use is created.
- the identification information (Disparity_Sequence_id) added to the disparity information set for specific service use is a value other than “0”.
- the association of the identification information (Disparity_Sequence_id) with the service is performed by inserting the information into the user data as described above. Alternatively, this association is performed using, for example, a disparity linkage descriptor (Disparity_Linkage_Descriptor) as described above.
- FIG. 38 shows an example of the disparity information set created in this case (4).
- a disparity information set for closed caption is created as a disparity information set for specific service use.
- the disparity information set for closed caption includes a disparity vector CC1 corresponding to the closed caption information “Caption1” and a disparity vector CC2 corresponding to the closed caption information “Caption2”.
- the disparity information set is a disparity information set for each predetermined unit of image data, for example, for each picture of encoded video.
- the image data is divided every fixed period, and a set of plural disparity information sets (Packed DisparityparSet) corresponding to the image data of each fixed period is transmitted prior to the transmission of the image data of each fixed period.
- Sent. Therefore, as shown in FIG. 39, an offset (Offset) is added to the disparity information set for each predetermined unit as time information indicating the timing of using the disparity information set.
- the offset added to each disparity information set indicates the time from the usage time of the first disparity information set indicated by a time stamp (TimeStamp).
- the disparity information set for specific service use is a disparity information set for closed captions.
- FIG. 39 illustrates an example in which the disparity vector CC1 and the closed caption data DF1 are associated with each other, and the disparity vector CC2 and the closed caption data DF2 are associated with each other.
- the component element ID (Component_Element ID) indicating the screen area where the closed caption information is displayed is associated with the disparity vectors CC1 and CC2 (see FIG. 37).
- the component element ID is a closed caption window ID (Window ID) as an association table of component element IDs and disparity vectors.
- Window ID a closed caption window ID
- the component element ID is the region ID (Region_id) of the DVB subtitle.
- subtitle segments (SubtitleSegment) and disparity vector data (Disparity Data) are associated using elementary PID (Elementary PID).
- page ID (Page_id) in the subtitle segment is associated with the disparity vector data identification information (Disparity_Sequence_id).
- region ID Region ID is associated with the disparity vector (Disparity_Vector_horizontal).
- the video encoder 113 performs encoding such as MPEG4-AVC, MPEG2, or VC-1 on the stereoscopic image data supplied from the data extracting unit 130 to obtain encoded video data.
- the video encoder 113 generates a video elementary stream including encoded video data in the payload portion by the stream formatter 113a provided in the subsequent stage.
- the predetermined set of disparity information sets created by the disparity information set creation unit 131 is supplied to the stream formatter 113 a in the video encoder 113.
- the stream formatter 113a embeds a predetermined set of disparity information sets as user data in a video stream.
- FIG. 41 schematically shows an example of the structure of a video elementary stream (Video Elementary Stream).
- a sequence header portion including parameters in sequence units is arranged at the head.
- a picture header including parameters and user data in units of pictures is arranged.
- a payload portion including picture data is arranged.
- the picture header part and the payload part are repeatedly arranged.
- FIG. 42 shows a configuration example of user data.
- Fig.42 (a) has shown the structure of the user data in case an encoding system is MPEG2.
- FIG. 42B shows that the encoding method is H.264.
- the structure of user data in the case of H.264 AVC (MPEG4-AVC) is shown.
- FIG. 42 (c) shows the configuration of user data when the encoding method is VC-1.
- the audio encoder 117 performs encoding such as MPEG-2Audio ⁇ ⁇ ⁇ AAC on the audio data supplied from the data extraction unit 130 to generate an audio elementary stream.
- the multiplexer 122 multiplexes the packetized elementary streams output from the encoders 113, 117, 119 and 121.
- the multiplexer 122 outputs bit stream data (transport stream) BSD as transmission data.
- the transmission data generation unit 110D shown in FIG. 22 is otherwise configured in the same manner as the transmission data generation unit 110 shown in FIG.
- the operation of the transmission data generation unit 110D shown in FIG. 22 will be briefly described.
- the stereoscopic image data output from the data extraction unit 130 is supplied to the video encoder 113.
- the stereoscopic image data is encoded by MPEG4-AVC, MPEG2, VC-1, or the like, and a video elementary stream including the encoded video data is generated. This video elementary stream is supplied to the multiplexer 122.
- the disparity vector for each pixel (pixel) output from the data extraction unit 130 is supplied to the disparity information set creation unit 131.
- a downsizing process or the like is performed on the disparity vector to create a predetermined disparity information set to be transmitted (cases (1) to (4)).
- the predetermined set of disparity information sets created by the disparity information set creation unit 131 is supplied to the stream formatter 113a in the video encoder 113.
- a predetermined set of disparity information sets are embedded as user data in a video stream.
- the audio data output from the data extraction unit 130 is supplied to the audio encoder 117.
- MPEG-2Audio is processed for audio data. Encoding such as AAC is performed, and an audio elementary stream including encoded audio data is generated. This audio elementary stream is supplied to the multiplexer 122.
- an elementary stream including subtitle data or encoded data of graphics data is supplied from the subtitle / graphic encoder 119 to the multiplexer 122.
- an elementary stream including encoded data of text data is supplied to the multiplexer 122 from the text encoder 121.
- the multiplexer 122 then multiplexes the elementary stream packets supplied from the encoders to obtain bit stream data (transport stream) BSD as transmission data.
- the association between the service component and the disparity information in each associated stream is performed using a disparity linkage descriptor (Disparity_Linkage_Descriptor). Further, the disparity linkage descriptor also associates disparity information set identification information (Disparity_Sequence_id) in each stream with a service.
- the disparity linkage descriptor is arranged in a header area of the multiplexed stream or a descriptor table such as a PMT in the multiplexed stream.
- FIG. 43 shows a configuration example of the disparity linkage descriptor in that case.
- FIG. 44 shows the contents of each piece of information in the configuration shown in FIG.
- “Descriptor_tag” is 8-bit data indicating the type of descriptor. Here, it is “0xD0”, which indicates that it is a disparity linkage descriptor.
- “Descriptor_length” is 8-bit data indicating the length (size) of this information.
- “Number_of_Linked_Streams” is 5-bit data indicating the number of elementary streams associated with each other by defining the relationship between the two streams from the viewpoint of disparity information data (Disparity Data).
- “ElementaryPID_of_disparity” is 13-bit data indicating a PID (program ID) of an elementary stream (PES) including disparity information data (DisparityData).
- “ElementaryPID_of_Service_Component” is 13-bit data indicating a PID (program ID) of an elementary stream (PES) including a service component such as a subtitle, caption, text, or graphics. With these two PIDs, disparity information and service components are associated for each associated elementary stream.
- “Number_of_linked_sequences” is 5-bit data indicating the number of disparity information sets (Disparity information) associated with each of the associated elementary streams.
- “Service_id” is 32-bit data indicating a unique word for service identification. For example, “0x47413934” is allocated for ATSC.
- “Data_Type” is 8-bit data indicating the data type in the service identified by “Service_id”. For example, “0x03” indicates closed caption data.
- Disparity sequence ID “Disparity_Sequence_id” and service page ID “Service_page_id” are arranged by the number indicated by “number_of_linked_sequences”. “Disparity_Sequence_id” is 8-bit data that is identification information for identifying each service. “Service_page_id” is 8-bit data specifying a service such as a subtitle page. Thereby, a predetermined number of disparity sequence IDs and service page IDs are associated with predetermined “Service_id (i)” and “Data_Type (i)”.
- FIG. 45A shows an example of the PID in the case where the disparity information set is inserted into the user data area of the image data stream and sent to the reception side, like the transmission data generation unit 110D in FIG.
- the disparity linkage descriptor (Disparity_Linkage_Descriptor) is arranged in the PMT in the multiplexed stream (TS).
- the PID of the video (image data) elementary stream (PES) including disparity information data (Disparity Data) included in the multiplexed stream (TS) is “PID_0”.
- the PID of the audio elementary stream (PES) included in the multiplexed stream (TS) is “PID_1”.
- the PID of the subtitle elementary stream (PES) included in the multiplexed stream (TS) is “PID_2”.
- FIG. 45 (b) generates a disparity vector elementary stream (PES) including a disparity information set, and multiplexes this independent stream with other streams, like a transmission data generation unit 110E of FIG.
- PES disparity vector elementary stream
- FIG. 45 (b) generates a disparity vector elementary stream (PES) including a disparity information set, and multiplexes this independent stream with other streams, like a transmission data generation unit 110E of FIG.
- PID disparity vector elementary stream
- TS multiplexed stream
- the PID of the video (image data) elementary stream (PES) included in the multiplexed stream (TS) is “PID_0”.
- the PID of the audio elementary stream (PES) included in the multiplexed stream (TS) is “PID_1”.
- the PID of the subtitle elementary stream (PES) included in the multiplexed stream (TS) is “PID_2”.
- the PID of the elementary stream (PES) of the disparity vector included in the multiplexed stream (TS) is “PID — 3”.
- the disparity information set includes a disparity vector “Region1” corresponding to the subtitle information “Subtitle1” and a disparity vector “Region2” corresponding to the subtitle information “Subtitle2”.
- the disparity information set includes a disparity vector “CC1” corresponding to the closed caption information “Caption1” and a disparity vector “CC2” corresponding to the closed caption information “Caption2”.
- FIG. 47 shows a case where the disparity information set is associated with a subtitle in the case of FIG. 45 (a) described above.
- “ElementaryPID_of_Service_Component” is set to the PID (Subtitle PID) of the subtitle elementary stream.
- “ElementaryPID_of_disparity” is set to the PID (Video (PID) of the video elementary stream.
- FIG. 48 illustrates a case where the disparity information set is associated with the subtitle in the case of FIG. 45 (b) described above.
- “ElementaryPID_of_Service_Component” is set to the PID (Subtitle PID) of the subtitle elementary stream.
- “ElementaryPID_of_disparity” is set to the PID (DisparityparPID) of the disparity vector elementary stream.
- FIG. 49 shows a case where the disparity information set is associated with the closed caption in the case of FIG. 45A described above (however, the closed caption data is inserted into the user data area of the image data stream and sent to the reception side). Shows the case.
- “Service_id” indicates ATSC
- “Data_Type” closed caption (CC).
- “ElementaryPID_of_Service_Component” is set to the PID (Video PID) of the video elementary stream.
- “ElementaryPID_of_disparity” is also set to the PID (Video PID) of the video elementary stream.
- FIG. 50 shows the case where the disparity information set is associated with the closed caption in the case of FIG. 45B described above (however, the closed caption data is inserted into the user data area of the image data stream and sent to the receiving side). Shows the case.
- “Service_id” indicates ATSC
- “Data_Type” closed caption (CC).
- “ElementaryPID_of_Service_Component” is set to the PID (Video PID) of the video elementary stream.
- “ElementaryPID_of_disparity” is set to the PID (DisparityparPID) of the disparity vector elementary stream.
- FIG. 51 illustrates an example of a detailed configuration of user data including disparity information (Disparity_Information). This example is an example when the encoding method is MPEG2.
- the data identifier after the start code (user_data_start_code) is a disparity information identifier (Disparity_Information_identifier)
- disparity information (Disparity_Information) is arranged as the subsequent data body.
- FIG. 52 shows a configuration of disparity information (Disparity_Information), in which disparity information data (Disparity_Information_Data) is arranged.
- This disparity information data can have the same information as the disparity linkage descriptor (Disparity_Linkage_Descriptor) in FIG.
- the disparity information data includes DSL (Disparity_Information_Link_Information) whose configuration is shown in FIG.
- the disparity information data includes DHI (Disparity_Header_Information) whose configuration is shown in FIGS. 54 and 55.
- the disparity information data includes DDS (Disparity Data Set) whose configuration is shown in FIG. 57 to 60 show the contents of each information in the configuration shown in FIGS. 53 to 56.
- This DSL is information for associating a disparity sequence ID (Disparity_Sequence_id) with a service.
- “Disparity_Info_Length” is 16-bit data indicating the length (size) of this information.
- Displayity_Info_Type is 2-bit data indicating the type of this information, and is “0x0” here, indicating that this information is “DSL”.
- This DHI also includes information associating each service component with a partition position ID (Partition_Position_ID) for specifying a disparity vector to be used for the component.
- Partition_Position_ID partition position ID
- Disition_Info_Length is 16-bit data indicating the length (size) of this information.
- Disarity_Info_Type is 2-bit data indicating the type of this information, and is “0x1” here, indicating that this information is “DHI”.
- “Views_offset” sets the offset expansion to “second “view”, for example, 1-bit data indicating whether to apply only to the right eye image or “both views”, that is, to each of the left eye image and the right eye image. If it is “1”, “both It shows that it applies to “views”. When it is “0”, it indicates that it is applied to “secondview”.
- “Disparity_precision” is 2-bit data indicating how much superimposition information is shifted with respect to the disparity vector. In the case of “0”, it indicates that the shift is performed by an amount corresponding to 1 ⁇ 2 of the parallax vector. In the case of “1”, it indicates that the shift is performed by an amount corresponding to the disparity vector.
- Distality_Sequence_id is 8-bit data which is identification information for identifying each service.
- Service_Linked_flag is 1-bit data indicating whether or not there is a disparity vector (disparity information) associated with the service in the DSL described above. “1” indicates that there is a disparity vector associated with the service. On the other hand, “0” indicates that there is no disparity vector associated with the service.
- “Target_Resolution_Type” is 2-bit data specifying the resolution of the disparity information set encoded for the target video. “0x0” indicates that the resolution is 1920 ⁇ 1080. “0x1” indicates that the resolution is 1440 ⁇ 1080. “0x2” indicates that the resolution is 1280 ⁇ 720.
- Block_Size is 2-bit data specifying the block size. This “Block_Size” is information indicating the size of the area constituting the information indicating the screen area. “0x0” indicates that the block size is 16 pixels ⁇ 16 pixels. “0x1” indicates that the block size is 32 pixels ⁇ 32 pixels. “0x2” indicates that the block size is 64 pixels ⁇ 64 pixels. Further, “0x3” indicates that the block size is 128 pixels ⁇ 128 pixels.
- Link_source indicates whether the disparity information set (source) is a general-purpose disparity information set or a disparity information set for a specific service application. “1” indicates that the disparity information set is a general-purpose disparity information set having a disparity vector of each screen area obtained by dividing the screen. “0” indicates that the disparity information set is a disparity information set for a specific service application having a disparity vector corresponding to the service component element.
- Partition indicates a screen division pattern. This “Partition” is information indicating the number of screen divisions constituting the information indicating the screen area.
- the image (picture) region is not divided, and one disparity vector (disparity information) representing this image (picture) region is transmitted. It shows that.
- the image (picture) region is divided into four, and four disparity vectors (disparity information) representing each of the divided regions are transmitted. Indicates. In this case, the disparity vectors in the upper left region are arranged first and the disparity vectors in the lower right region are arranged last, in the order of 0 to 3 in FIG. 61 (b).
- the image (picture) region is divided into nine, and nine disparity vectors (disparity information) representing these divided regions are transmitted. Indicates. In this case, the disparity vectors in the upper left region are arranged first and the disparity vectors in the lower right region are arranged last (in the order of 0 to 8 in FIG. 61C). In the case of “0x3”, as shown in FIG. 61 (d), the image (picture) region is divided into 16 parts, and 16 disparity vectors (disparity information) representing each of the divided regions are transmitted. Indicates. In this case, the disparity vectors in the upper left region are arranged first and the disparity vectors in the lower right region are arranged last (in the order of 0 to 15 in FIG. 61 (d)).
- 0x7 indicates that the image (picture) region is divided into a plurality of blocks, and a plurality of disparity vectors (disparity information) representing the respective divided regions are transmitted.
- the disparity vectors in the upper left area are the first and the disparity vectors in the lower right area are the last.
- “Number_of_Component_Elements” is 13-bit data indicating the number of screen areas specified by the above “Partition” or “Block_Size” or the number of component elements in the associated service.
- Component_Element is 8-bit data indicating a component element ID.
- the component element ID is, for example, a closed caption window ID (Window ID) or a DVB subtitle region ID (region_idga).
- Element_Linkage_Length is 8-bit data indicating the length (size) of information.
- Numberer_of_mutiple_link indicates the number of disparity vectors associated with the component element ID.
- Partition_Position_id is 13-bit information specifying a disparity vector. This “Partition_Position_id” indicates the number of the screen area among the plurality of screen areas specified by the above “Partition” or “Block_Size”.
- This DDS includes information on disparity vectors included in each disparity information set.
- “Disparity_Info_Length” is 16-bit data indicating the length (size) of this information.
- “Disparity_Info_Type” is 2-bit data indicating the type of this information, and is “0x2” here, indicating that this information is “DID”.
- “Disparity_Sequence_id” is 8-bit data for identifying whether the disparity information set is a general purpose disparity information set or a disparity information set for a specific service application. is there.
- “Number_of_Component_Elements” is 13-bit data indicating the number of screen areas specified by the above “Partition” or “Block_Size” or the number of component elements in the associated service. A value of P disparity vectors included in a disparity information set used in a certain period of video data, for example, each frame within a 15-frame period is arranged.
- “Disparity_sets_in_period” is 8-bit data indicating the number of disparity vector (disparity information) sets for providing an offset included in a certain period of video data.
- Offset_Frame_In_disparity_set is time information indicating the timing of using each disparity information set. This time information indicates the time (number of frames) from the use time of the first disparity information set indicated by the time stamp (Time Stamp).
- Distal_Vector_Horozontal (i) indicates the value of the disparity vector in the horizontal direction, which is the value of the i-th disparity vector.
- the example in which the information for associating the disparity sequence ID (Disparity_Sequence_id) with the service is arranged as DSL in the disparity information (Disparity_Information) as user data is shown.
- a disparity linkage descriptor Dispose_Linkage_Descriptor
- the disparity linkage descriptor DLD is arranged, for example, in a PMT table of a multiplexed stream (bit stream data BSD).
- identification information indicating the type of superimposition information that should use disparity information is provided together with left-eye image data and right-eye image data for displaying a stereoscopic image.
- a predetermined set of disparity information sets added is transmitted. Therefore, on the receiving side, the same superimposition information (closed caption information, subtitle information, graphics information, text information, etc.) superimposed on the left-eye image and right-eye image, and disparity information that matches the type of superimposition information Appropriate parallax can be given using the set.
- the disparity information set creation unit 131 performs downsizing processing on the disparity vector for each pixel (pixel) in the lowest layer, and As described above, the parallax vector is obtained. However, the parallax vector of each area of each hierarchy is recorded on the data recording medium 130a, and the parallax information set creation unit 131 may use it. "Other configuration examples of the transmission data generator"
- the transmission data generation unit 110D illustrated in FIG. 22 is configured to insert a predetermined set of disparity information sets created by the disparity information set creation unit 131 into an image data stream and send the image data stream to the reception side.
- a configuration in which a predetermined set of disparity information sets created by the disparity information set creation unit 131 is inserted into a data stream other than the image data stream and sent to the receiving side can also be employed.
- FIG. 62 shows a configuration example of the transmission data generation unit 110E in that case.
- portions corresponding to those in FIG. 22 are denoted by the same reference numerals, and detailed description thereof is omitted.
- the transmission data generation unit 110E includes a disparity vector encoder 115, similar to the transmission data generation unit 110 illustrated in FIG.
- the predetermined set of disparity information sets created by the disparity information set creation unit 131 is sent to the disparity vector encoder 115.
- the disparity vector encoder 115 generates a disparity vector elementary stream (disparity vector stream) including a predetermined set of disparity information sets. Then, this disparity vector stream is supplied to the multiplexer 122.
- the multiplexer 122 multiplexes the disparity vector stream together with other data streams to generate bit stream data BSD.
- this transmission data generation unit 110E Although detailed description is omitted, in this transmission data generation unit 110E, other configurations and operations are the same as those of the transmission data generation unit 110D shown in FIG. In this transmission data generation unit 110E, the same effect as that of the transmission data generation unit 110D shown in FIG. 22 can be obtained.
- either or both of a general-purpose disparity information set and a specific service use disparity information set can be transmitted as the disparity information set.
- disparity information for limited use such as closed caption information or subtitle information
- it may be possible to transmit disparity information for limited use such as closed caption information or subtitle information, and to have a simple data structure that can be easily linked to existing broadcasting standards (ATSC, CEA, etc.).
- FIG. 63 shows a configuration example of the transmission data generation unit 110F in that case.
- the transmission data generation unit 110F is configured to create disparity information whose use is limited to closed caption information and transmit the disparity information to the reception side.
- portions corresponding to those in FIG. 2 are denoted by the same reference numerals, and detailed description thereof will be omitted as appropriate.
- the transmission data generation unit 110F includes a data extraction unit (archive unit) 130, a parallax information generation unit 131, and a CC (Closed) Caption) encoder 133, video encoder 113, audio encoder 117, and multiplexer 122.
- a data recording medium 130a is detachably attached to the data extraction unit 130, for example.
- audio data and parallax vectors (parallax information) are recorded in association with stereoscopic image data including left-eye image data and right-eye image data.
- the data extraction unit 130 extracts and outputs stereoscopic image data, audio data, disparity vectors, and the like from the data recording medium 130a.
- the data recording medium 130a is a disk-shaped recording medium, a semiconductor memory, or the like.
- the stereoscopic image data recorded on the data recording medium 130a corresponds to the stereoscopic image data obtained by the video framing unit 112 of the transmission data generation unit 110 shown in FIG.
- the parallax vector recorded on the data recording medium 130a is, for example, a parallax vector for each pixel (pixel) constituting the image.
- CC encoder 133 is an encoder compliant with CEA-708, and outputs CC data (closed caption information data) for displaying closed caption captions. In this case, the CC encoder 133 sequentially outputs CC data of each closed caption information displayed in time series.
- the disparity information creation unit 132 performs downsizing processing on the disparity vector output from the data extraction unit 130, that is, the disparity vector for each pixel (pixel), and each window included in the CC data output from the CC encoder 133 described above.
- Disparity information (disparity vector) associated with the ID (WindowID) is output.
- This parallax information includes shift target designation information for designating closed caption information to be shifted based on this parallax information, among the closed caption information superimposed on the left eye image and the closed caption information superimposed on the right eye image. It has been added.
- the disparity information creating unit 132 associates disparity information to be used in a predetermined number of frame periods in which the closed caption information is displayed in association with the CC data of each closed caption information output from the CC encoder 133. Output.
- the disparity information is, for example, disparity information that is commonly used in each frame in a predetermined number of frame periods in which closed caption information is displayed, or disparity information that is sequentially updated in each frame.
- flag information indicating whether the disparity information is commonly used in each frame or the disparity information that is sequentially updated in each frame is added to the disparity information.
- the disparity information set creation unit 132 creates disparity information that is sequentially used in each frame when the motion of the image is large, and disparity information that is commonly used in each frame when the motion of the image is small. Create
- the disparity information creating unit 132 is configured with disparity information sequentially updated in each frame, for example, with disparity information of the first frame and offset information with respect to disparity information of the second and subsequent frames.
- disparity information sequentially updated in each frame, for example, with disparity information of the first frame and offset information with respect to disparity information of the second and subsequent frames.
- the video encoder 113 performs encoding such as MPEG4-AVC, MPEG2, or VC-1 on the stereoscopic image data supplied from the data extracting unit 130 to obtain encoded video data. Further, the video encoder 113 generates a video elementary stream including encoded video data in the payload portion by a stream formatter 113a provided in the subsequent stage.
- the CC data output from the CC encoder 133 and the disparity information generated by the disparity information generating unit 131 are supplied to the stream formatter 113a in the video encoder 113.
- the stream formatter 113a embeds CC data and disparity information as user data in a video elementary stream. That is, stereoscopic image data is included in the payload portion of the video elementary stream, and CC data and disparity information are included in the user data area of the header portion.
- the video elementary stream has a sequence header portion including parameters in sequence units at the head. Subsequent to the sequence header portion, a picture header including parameters and user data in units of pictures is arranged. Following this picture header portion, a payload portion including picture data is arranged. Hereinafter, the picture header part and the payload part are repeatedly arranged.
- CC data and the disparity information described above are embedded in, for example, a user data area in the picture header portion.
- CC data and disparity information embedded as user data can be identified by the user data type code. That is, the user data type code added to the disparity information is different from the user data type code added to the CC data.
- the audio encoder 117 performs encoding such as MPEG-2Audio ⁇ ⁇ ⁇ AAC on the audio data supplied from the data extraction unit 130 to generate an audio elementary stream.
- the multiplexer 122 multiplexes the packetized elementary streams output from the encoders 113 and 117.
- the multiplexer 122 outputs bit stream data (transport stream) BSD as transmission data.
- the stereoscopic image data output from the data extraction unit 130 is supplied to the video encoder 113.
- the stereoscopic image data is encoded by MPEG4-AVC, MPEG2, VC-1, or the like, and a video elementary stream including the encoded video data is generated. This video elementary stream is supplied to the multiplexer 122.
- the CC encoder 133 outputs CC data (closed caption information data) for displaying closed caption captions.
- the CC encoder 133 sequentially outputs CC data of each closed caption information displayed in time series.
- the disparity vector for each pixel (pixel) output from the data extracting unit 130 is supplied to the disparity information creating unit 132.
- the disparity vector is subjected to downsizing processing or the like, and disparity information (disparity) associated with each window ID (WindowID) included in the CC data output from the CC encoder 133 described above. Vector) is output.
- CC data output from the CC encoder 133 and the disparity information created by the disparity information creating unit 132 are supplied to the stream formatter 113a of the video encoder 113.
- CC data and disparity information are embeddably identifiable by user data identification information (user data type code) in the user data area of the header part of the video elementary stream.
- the audio data output from the data extraction unit 130 is supplied to the audio encoder 117.
- MPEG-2Audio is processed for audio data. Encoding such as AAC is performed, and an audio elementary stream including encoded audio data is generated.
- This audio elementary stream is supplied to the multiplexer 122.
- the multiplexer 122 the elementary stream packets supplied from the encoders are multiplexed to obtain bit stream data BSD as transmission data.
- FIG. 64 shows an example.
- FIG. 65 shows the contents of each piece of information in the configuration shown in FIG. This example is an example when the encoding method is MPEG2.
- a 32-bit code “0x47413934” for identifying the broadcast standard “ATSC” is arranged after the 32-bit start code (user_data_start_code).
- an 8-bit user data type code (user_data_type_code) is arranged.
- disparity information (caption_disparity_data) is arranged as the subsequent data body.
- FIG. 66 shows an example.
- 67 and 68 show the contents of each piece of information in the configuration shown in FIG.
- “Service_number” is information for identifying a closed caption channel. This “service_number” is associated with “caption service” of CEA-708.
- a closed caption has a plurality of channels such as English and Japanese, but disparity information for each channel is arranged.
- both the closed caption information superimposed on the left eye image and the closed caption information superimposed on the right eye image are shifted in opposite directions in the horizontal direction. Show.
- the disparity information (disparity) is an even value
- the closed caption information to be superimposed on the left eye image is shifted to the right by “disparity / 2” and is a negative value when the disparity information is a positive value. Is shifted to the left by “disparity / 2”.
- the closed caption information to be superimposed on the right eye image is shifted to the left by “disparity / 2” and is a negative value when the disparity information (disparity) is a positive value. In some cases, it is shifted to the right by “disparity / 2”.
- the closed caption information to be superimposed on the left eye image is shifted to the right by “(disparity + 1) / 2” when the disparity information (disparity) is a positive value. If it is a value, it is shifted to the left by “(disparity + 1) / 2”. Further, the closed caption information to be superimposed on the right eye image is shifted to the left by “(disparity ⁇ 1) / 2” when the disparity information (disparity) is an odd value and is a positive value, When it is a negative value, it is shifted to the right by “(disparity ⁇ 1) / 2”.
- “Number_of_caption_windows” is 3-bit information indicating the number of caption windows (CaptionWindow) associated with disparity information (disparity). The maximum number of caption windows is eight. Disparity information (disparity) is arranged for each caption window. “Caption_window_id” is 3-bit information indicating “caption window [0..7]” of CEA-708.
- “Disparity” is 8-bit information indicating the value of the disparity information of the first frame in the display frame period, and takes a value in the range of ⁇ 128 to +127.
- the disparity information of the second and subsequent frames is given as offset information with respect to the disparity information of the previous frame.
- “Number_of_frame_set” is 8-bit information indicating how many units are necessary when the number of frames after the second in the display frame period is divided into 11 frames. For example, when the display frame period is 30 frames, the number of second and subsequent frames is 29, and the number of units indicated by “number_of_frame_set” is “3”.
- stereoscopic image data including left-eye image data and right-eye image data for displaying a stereoscopic image is included in the payload portion of the video elementary stream and transmitted.
- disparity information for adding disparity to CC data and closed caption information based on the CC data in the user data area of the header portion of the video elementary stream can be identified by user data identification information (user_data_type_code) Included and sent.
- the disparity information includes the left-eye closed caption information superimposed on the left-eye image and the right-eye closed caption information superimposed on the right-eye image.
- Shift target designation information (select_view_shift) for designating closed caption information to be shifted based on is added. For this reason, it is possible to arbitrarily shift only the left-eye closed caption information, shift only the right-eye closed caption information, or both of them according to the shift target designation information.
- Flag information temporary_extension_flag indicating which one is added is added to the disparity information. Therefore, disparity information that is commonly used in each frame or disparity information that is sequentially updated in each frame can be selectively transmitted according to, for example, image contents.
- disparity information that is sequentially updated in each frame is transmitted, and on the receiving side, the disparity added to the superimposition information is dynamically changed in conjunction with the change of the image content. Can do.
- the data amount of the disparity information can be suppressed by transmitting disparity information that is commonly used in each frame.
- the disparity information sequentially updated in each frame includes the disparity information of the first frame in a predetermined number of frame periods and the previous frames of the second and subsequent frames. It consists of offset information for disparity information. Therefore, the data amount of parallax information can be suppressed.
- FIG. 69 shows a calculation example of the data amount (data size) in the configuration of the disparity information (caption_disparity_data) shown in FIG. (1-1) is an example of calculation when one piece of disparity information is commonly used in each frame of the closed caption information display period in each of the eight caption windows.
- the data amount of the disparity information (caption_disparity_data) is 18 bytes.
- (1-2) is a case where there is one caption window or the same disparity information is used in all caption windows, and one disparity is provided for each frame in the closed caption information display period. It is an example of calculation when information is used in common. In this case, the data amount of the disparity information (caption_disparity_data) is 4 bytes.
- (2) is a case where disparity information in which closed caption information is sequentially updated in each frame of the display period is used in each of the eight caption windows.
- the display period is 5 seconds (150 frames).
- the data amount of the disparity information (caption_disparity_data) is 362 bytes.
- disparity information whose use is limited to closed caption information is illustrated in FIG. 63.
- a transmission data generation unit that transmits disparity information whose use is limited to other superimposition information such as subtitle information can be similarly configured.
- the set-top box 200 receives bit stream data (transport stream) transmitted from the broadcast station 100 on a broadcast wave.
- the bit stream data includes stereoscopic image data including left eye image data and right eye image data, audio data, superimposition information data, and further disparity information (disparity vector).
- the superimposition information data is, for example, subtitle data, graphics data, text data (including closed caption data), and the like.
- the set top box 200 has a bit stream processing unit 201.
- the bit stream processing unit 201 extracts stereoscopic image data, audio data, superimposition information data, a disparity vector, and the like from the bit stream data.
- the bit stream processing unit 201 generates left-eye image data and right-eye image data on which superimposition information is superimposed, using stereoscopic image data, superimposition information data (subtitle data, graphics data, text data), and the like.
- left eye superimposition information and right eye superimposition information to be superimposed on the left eye image and the right eye image are generated based on the parallax vector and the superimposition information data.
- the left eye superposition information and the right eye superposition information are the same superposition information.
- the superposition position in the image is shifted in the horizontal direction by the horizontal component of the parallax vector, for example, with respect to the left eye superimposition information.
- FIG. 70 (a) shows the superimposed positions of left-eye graphics information and right-eye graphics information when the transmission method is the above-described second transmission method (“Side By Side” method).
- the right-eye graphics information RGI superimposed on the right-eye image IR is set at a position shifted in the horizontal direction by the horizontal component VVT of the parallax vector.
- IT is an idling offset value.
- bit stream processing unit 201 graphics data is generated such that the graphics information LGI and RGI are superimposed on the images IL and IR as shown in FIG.
- the bit stream processing unit 201 combines the generated left eye graphics data and right eye graphics data with the stereoscopic image data (left eye image data and right eye image data) extracted from the bit stream data, and performs processing Later stereo image data is acquired.
- this stereoscopic image data as shown in FIG. 70 (b), the viewer can observe each graphics information LGI, RGI together with each image IL, IR with parallax, and the graphics information has a sense of perspective. It becomes possible to recognize.
- FIG. 71A shows a state in which graphics images based on graphics data extracted from bit stream data are superimposed on the images IL and IR as they are.
- FIG. 71B the viewer observes the left half of the graphics information together with the left eye image IL and the right half of the graphics information together with the right eye image IR. Therefore, the graphics information cannot be recognized correctly.
- FIG. 70 shows the case of graphics information, but the same applies to other superimposition information (closed caption information, subtitle information, text information, etc.). That is, when the parallax vector is transmitted as numerical information, left eye superimposition information and right eye superimposition information to be superimposed on the left eye image and the right eye image are generated based on the parallax vector and the superimposition information data data. .
- the left eye superposition information and the right eye superposition information are the same superposition information.
- the superposition position in the image is shifted in the horizontal direction by, for example, the horizontal component of the parallax vector in the right eye superposition information with respect to the left eye superposition information, for example.
- disparity vectors As the disparity vector, it is conceivable to use the disparity vector at the closest recognized position in terms of perspective among the disparity vectors detected at a plurality of positions in the image.
- 72 (a), (b), (c), and (d) show disparity vectors (View Vector) at three object positions at times T0, T1, T2, and T3, respectively.
- the disparity vector VV0-1 at the position (H0, V0) corresponding to the object 1 is the maximum disparity vector MaxVV (T0).
- the disparity vector VV1-1 at the position (H1, V1) corresponding to the object 1 is the maximum disparity vector MaxVV (T1).
- the parallax vector VV2-2 at the position (H2, V2) corresponding to the object 2 is the maximum parallax vector MaxVV (T2).
- the parallax vector VV3-0 at the position (H3, V3) corresponding to the object 1 is the maximum parallax vector MaxVV (T3).
- Superimposition information can be displayed in front of an object in a close image.
- FIG. 73 (a) shows a display example of captions (for example, closed caption information and subtitle information) on an image.
- captions are superimposed on an image composed of a background and a foreground object.
- FIG. 73B shows the perspective of the background, the foreground object, and the caption, and indicates that the caption is recognized as being closest.
- FIG. 74A shows a display example of captions (for example, closed caption information and subtitle information) on the same image as FIG. 73A.
- FIG. 74 (b) shows left eye subtitle information LGI and right eye subtitle information RGI for displaying a caption.
- FIG. 74 (c) shows that disparity is given to each subtitle information LGI and RGI in order that the subtitle is recognized as being closest.
- a disparity vector detected at a plurality of positions in an image (including a disparity vector of each area of each layer included in a disparity information set) corresponding to the superimposed position should be used. Can be considered.
- FIG. 75A shows graphic information based on graphic data extracted from bit stream data and text information based on text data extracted from bit stream data.
- FIG. 75 (b) shows a state in which the left eye graphics information LGI and the left eye text information LTI are superimposed on the left eye image.
- the superimposition position thereof is regulated by the idling offset value (IT-0) in the horizontal direction.
- the left eye text information LTI has its superposition position regulated in the horizontal direction by an idling offset value (IT-1).
- FIG. 75 (c) shows a state in which the right eye graphics information RGI and the right eye text information RTI are superimposed on the right eye image.
- the superimposition position is restricted by the idling offset value (IT-0) in the horizontal direction, and the horizontal component VVT-0 of the parallax vector corresponding to this superposition position is also set to the left eye. It is shifted from the superimposed position of the graphics information LGI.
- the right-eye text information RTI has its superposition position restricted by an idling offset value (IT-1) in the horizontal direction, and the left-eye text corresponding to the horizontal component VVT-1 of the parallax vector corresponding to this superposition position. It is shifted from the superimposed position of the information LTI.
- the graphic information and text information superimposed position superimposed on the right eye image are shifted with reference to the same graphics information and text information superimposed position superimposed on the left eye image. ing. That is, in the example of FIG. 75, processing for shifting only the superimposed position of graphics information and text information superimposed on the right eye image is performed. However, it is also conceivable to perform processing so as to shift the overlapping positions of both the left eye image and the right eye image.
- FIG. 76 (a) shows graphic information based on graphic data extracted from the bit stream data and text information based on text data extracted from the bit stream data.
- FIG. 76B shows the superimposed position of the graphics information GI and the text information TI in the two-dimensional display. In this case, the superimposed position of text information is (x1, y1), and the superimposed position of graphics information is (x2, y2).
- FIG. 76 (c) shows a state in which the left eye graphics information LGI and the left eye text information LTI are superimposed on the left eye image.
- the left-eye graphics information LGI is shifted to the right by the number of pixels of the disparity vector D1 corresponding to the graphics information with respect to the superimposed position of the graphics information GI in the two-dimensional display.
- the left-eye text information LTI is shifted to the right by the number of pixels of the disparity vector D0 corresponding to the text information with respect to the superimposed position of the text information LT in the two-dimensional display.
- FIG. 76 (d) shows a state where the right eye graphics information RGI and the right eye text information RTI are superimposed on the right eye image.
- the right-eye graphics information RGI is shifted leftward by the number of pixels of the disparity vector D1 corresponding to the graphics information with respect to the superimposed position of the graphics information GI in the two-dimensional display.
- the right eye text information RTI is shifted leftward by the number of pixels of the disparity vector D0 corresponding to the text information with respect to the superimposed position of the text information LT in the two-dimensional display.
- 76 (c) and 76 (d) are examples in which the offset development is applied to each of the left eye image and the right eye image, and is set so as to be shifted by the number of pixels of the parallax vector. .
- the expansion of the offset is indicated by “Views_offset”, and how much the superimposition information is shifted with respect to the disparity vector is indicated by “Disparity_precision”.
- FIG. 77 (a) shows closed caption information C1 (“Caption1”) and C2 (“Caption2”) of two caption windows extracted from the bitstream data.
- FIG. 77 (b) shows the superimposed position of the closed caption information C1 and C2 in the two-dimensional display.
- the superimposed position of the closed caption information C1 is (x1, y1)
- the superimposed position of the closed caption information C2 is (x2, y2).
- FIG. 77 (c) shows a state in which the left-eye closed caption information LC1 and LC2 are superimposed on the left-eye image.
- the left-eye closed caption information LC1 is shifted to the right by the number of pixels of the disparity vector D0 corresponding to the closed caption information with respect to the superimposed position of the closed caption information C1 in the two-dimensional display.
- the left-eye closed caption information LC2 is shifted to the right by the number of pixels of the disparity vector D1 corresponding to the closed caption information with respect to the closed caption information C2 in the two-dimensional display.
- FIG. 77 (d) shows a state in which the right-eye closed caption information RC1 and RC2 are superimposed on the right-eye image.
- the left eye closed caption information RC1 is shifted leftward by the number of pixels of the disparity vector D0 corresponding to the closed caption information with respect to the superimposed position of the closed caption information C1 in the two-dimensional display.
- the right-eye closed caption information RC2 is shifted leftward by the number of pixels of the disparity vector D1 corresponding to the closed caption information with respect to the closed caption information C2 in the two-dimensional display.
- 77 (c) and 77 (d) are examples in which the offset development is applied to each of the left eye image and the right eye image and is set so as to be shifted by the number of pixels of the parallax vector. .
- the expansion of the offset is indicated by “select_view_shift”, for example, and how much the superimposition information is shifted with respect to the disparity vector is indicated by “offset_precision”.
- FIG. 78 (a) shows that objects A, B, and C exist in the image and, for example, text information indicating annotations of the objects is superimposed on positions near the objects.
- FIG. 78B shows the positions of the objects A, B, and C, the disparity vector list indicating the correspondence between the disparity vectors at the positions, the disparity vectors, and the annotations of the objects A, B, and C. It shows that it is used when parallax is given to the text information shown. For example, the text information “Text” is superimposed in the vicinity of the object A, but the disparity vector at the position (Ha, Va) of the object A is between the left-eye text information and the right-eye text information. Parallax corresponding to VV-a is given. The same applies to the text information superimposed in the vicinity of the B and C objects.
- FIG. 75 and 76 show a case where the superimposition information is graphics information and text information.
- FIG. 77 shows a case where the superimposition information is closed caption information.
- FIG. 78 shows a case where the superimposition information is text information.
- the superimposition information data extracted from the bitstream data includes data of left eye superimposition information and right eye superimposition information that are given disparity by a disparity vector.
- the bit stream processing unit 201 simply synthesizes the superimposition information data extracted from the bit stream data with the stereoscopic image data (left eye image data, right eye image data) extracted from the bit stream data, The processed stereoscopic image data is acquired.
- processing such as conversion of character code into bitmap data is necessary.
- FIG. 79 shows a configuration example of the set top box 200.
- the set top box 200 includes a bit stream processing unit 201, an HDMI terminal 202, an antenna terminal 203, a digital tuner 204, a video signal processing circuit 205, an HDMI transmission unit 206, and an audio signal processing circuit 207. ing.
- the set top box 200 includes a CPU 211, a flash ROM 212, a DRAM 213, an internal bus 214, a remote control receiving unit 215, and a remote control transmitter 216.
- the antenna terminal 203 is a terminal for inputting a television broadcast signal received by a receiving antenna (not shown).
- the digital tuner 204 processes the television broadcast signal input to the antenna terminal 203 and outputs predetermined bit stream data (transport stream) corresponding to the user's selected channel.
- the bit stream processing unit 201 extracts stereoscopic image data (left-eye image data, right-eye image data), audio data, superimposition information data, disparity information (disparity vector), and the like from the bit stream data.
- the superimposition information data is subtitle data, graphics data, text data (including closed caption data), and the like.
- the bit stream processing unit 201 synthesizes superimposition information (closed caption information, subtitle information, graphics information, text information, etc.) data with the stereoscopic image data, and acquires display stereoscopic image data. To do.
- the bit stream processing unit 201 outputs audio data. The detailed configuration of the bit stream processing unit 201 will be described later.
- the video signal processing circuit 205 performs image quality adjustment processing on the stereoscopic image data output from the bit stream processing unit 201 as necessary, and supplies the processed stereoscopic image data to the HDMI transmission unit 206.
- the audio signal processing circuit 207 performs sound quality adjustment processing or the like on the audio data output from the bit stream processing unit 201 as necessary, and supplies the processed audio data to the HDMI transmission unit 206.
- the HDMI transmission unit 206 transmits baseband image (video) and audio data from the HDMI terminal 202 by communication conforming to HDMI. In this case, since transmission is performed using the HDMI TMDS channel, image and audio data are packed and output from the HDMI transmission unit 206 to the HDMI terminal 202. Details of the HDMI transmission unit 206 will be described later.
- the CPU 211 controls the operation of each part of the set top box 200.
- the flash ROM 212 stores control software and data.
- the DRAM 213 constitutes a work area for the CPU 211.
- the CPU 211 develops software and data read from the flash ROM 212 on the DRAM 213 to activate the software, and controls each part of the set top box 200.
- the remote control receiving unit 215 receives the remote control signal (remote control code) transmitted from the remote control transmitter 216 and supplies it to the CPU 211.
- the CPU 211 controls each part of the set top box 200 based on the remote control code.
- the CPU 211, flash ROM 212 and DRAM 213 are connected to the internal bus 214.
- a television broadcast signal input to the antenna terminal 203 is supplied to the digital tuner 204.
- the digital tuner 204 processes the television broadcast signal and outputs predetermined bit stream data (transport stream) corresponding to the user's selected channel.
- the bit stream data output from the digital tuner 204 is supplied to the bit stream processing unit 201.
- the bit stream processing unit 201 extracts stereoscopic image data (left-eye image data, right-eye image data), audio data, graphics data, text data, disparity vectors, and the like from the bit stream data.
- data of superimposition information (closed caption information, subtitle information, graphics information, text information, and the like) is combined with the stereoscopic image data to generate stereoscopic image data for display.
- the display stereoscopic image data generated by the bit stream processing unit 201 is supplied to the HDMI transmission unit 206 after image quality adjustment processing or the like is performed as necessary by the video signal processing circuit 205. Also, the audio data obtained by the bit stream processing unit 201 is supplied to the HDMI transmission unit 206 after the audio signal processing circuit 207 performs sound quality adjustment processing or the like as necessary.
- the stereoscopic image data and audio data supplied to the HDMI transmission unit 206 are transmitted from the HDMI terminal 202 to the HDMI cable 400 via the HDMI TMDS channel.
- FIG. 80 shows a configuration example of the bit stream processing unit 201.
- the bit stream processing unit 201 has a configuration corresponding to the transmission data generation units 110 and 110E shown in FIGS.
- the bit stream processing unit 201 includes a demultiplexer 220, a video decoder 221, a subtitle / graphics decoder 222, a text decoder 223, an audio decoder 224, and a disparity vector decoder 225.
- the bit stream processing unit 201 includes a stereoscopic image subtitle / graphics generating unit 226, a stereoscopic image text generating unit 227, a video superimposing unit 228, and a multi-channel speaker control unit 229.
- the demultiplexer 220 extracts video, audio, disparity vectors, subtitles, graphics, and text packets from the bit stream data BSD, and sends them to each decoder.
- the video decoder 221 performs processing reverse to that of the video encoder 113 of the transmission data generation unit 110 described above. That is, the video decoder 221 reconstructs a video elementary stream from the video packet extracted by the demultiplexer 220, performs decoding processing, and generates stereoscopic image data including left-eye image data and right-eye image data. Get.
- the transmission method of the stereoscopic image data is, for example, the above-described first transmission method (“Top & Bottom” method), the second transmission method (“Side By Side” method), and the third transmission method (“Frame Sequential ”method) (see FIGS. 4A to 4C).
- the subtitle / graphics decoder 222 performs the reverse process of the subtitle / graphics encoder 119 of the transmission data generation unit 110 described above. That is, the subtitle / graphics decoder 222 reconstructs a subtitle / graphics elementary stream from the subtitle / graphics packet extracted by the demultiplexer 220. Then, the subtitle / graphics decoder 222 further performs a decoding process to obtain subtitle data and graphics data.
- the text decoder 223 performs processing reverse to that of the text encoder 121 of the transmission data generation unit 110 described above. That is, the text decoder 223 reconstructs a text elementary stream from the text packet extracted by the demultiplexer 220 and performs a decoding process to obtain text data (including closed caption data).
- the audio decoder 224 performs processing opposite to that of the audio encoder 117 of the transmission data generation unit 110 described above. That is, the audio decoder 224 reconstructs an audio elementary stream from the audio packet extracted by the demultiplexer 220, performs decoding processing, and obtains audio data.
- the disparity vector decoder 225 performs processing opposite to that of the disparity vector encoder 115 of the transmission data generation unit 110 described above. That is, the disparity vector decoder 225 reconstructs an elementary stream of disparity vectors from the disparity vector packets extracted by the demultiplexer 220, performs decoding processing, and obtains disparity vectors at predetermined positions in the image.
- the stereoscopic image subtitle / graphics generating unit 226 generates left-eye and right-eye subtitle information and graphics information to be superimposed on the left-eye image and the right-eye image, respectively. This generation process is performed based on the subtitle data and graphics data obtained by the decoder 222 and the disparity vector obtained by the decoder 225. In this case, the left-eye and left-eye subtitle information and graphics information are the same information. However, the superimposed position in the image is shifted in the horizontal direction by, for example, the horizontal component of the disparity vector in the right-eye title information and graphics information with respect to the left-eye subtitle information and graphics information, for example. . Then, the stereoscopic image subtitle / graphics generating unit 226 outputs the generated left-eye and left-eye subtitle information and graphics information data (bitmap data).
- the stereoscopic image text generating unit 227 Based on the text data obtained by the decoder 223 and the parallax vector obtained by the decoder 225, the stereoscopic image text generating unit 227 superimposes the left-eye text information and the right-eye image to be superimposed on the left-eye image and the right-eye image, respectively. Generate text information.
- the left-eye text information and the right-eye text information are the same text information, but the superimposed position in the image is, for example, the left-eye text information, and the right-eye text information is the horizontal component of the disparity vector. Only to be shifted horizontally. Then, the stereoscopic image text generation unit 227 outputs data (bitmap data) of the generated left eye text information and right eye text information.
- the video superimposing unit 228 superimposes the data generated by the generating units 226 and 227 on the stereoscopic image data (left-eye image data and right-eye image data) obtained by the video decoder 221 to display stereoscopic image data for display. Get Vout. Note that superimposition of superimposition information data on stereoscopic image data (left-eye image data, right-eye image data) is started by a time stamp of the system layer.
- the multi-channel speaker control unit 229 gives the sound data obtained by the audio decoder 224 a process for generating sound data of a multi-channel speaker for realizing 5.1ch surround, for example, and predetermined sound field characteristics. Apply processing.
- the multi-channel speaker control unit 229 controls the output of the multi-channel speaker based on the disparity vector obtained by the decoder 225.
- FIG. 81 shows an example of speaker output control when the parallax vector VV1 is larger on the left video object toward the television display.
- the Rear Left speaker volume of the multi-channel speaker is increased, the Front Left speaker volume is medium, and the Front Right and Rear Right speaker volumes are reduced.
- the disparity vector of the video content stereoscopic image data
- the viewer can experience a stereoscopic effect comprehensively.
- bit stream processing unit 201 The operation of the bit stream processing unit 201 shown in FIG.
- the bit stream data BSD output from the digital tuner 204 (see FIG. 79) is supplied to the demultiplexer 220.
- the demultiplexer 220 video, audio, disparity vectors, subtitles and graphics, and text TS packets are extracted from the bit stream data BSD and supplied to each decoder.
- a video elementary stream is reconstructed from the video packet extracted by the demultiplexer 220, and further, decoding processing is performed, so that stereoscopic image data including left eye image data and right eye image data is obtained. can get.
- the stereoscopic image data is supplied to the video superimposing unit 228.
- the disparity vector decoder 225 reconstructs the disparity vector elementary stream from the disparity vector packet extracted by the demultiplexer 220 and further performs decoding processing to obtain a disparity vector at a predetermined position in the image. (See FIG. 8, FIG. 29, etc.).
- the subtitle / graphics decoder 222 reconstructs a subtitle / graphics elementary stream from the subtitle / graphics packets extracted by the demultiplexer 220.
- the subtitle / graphics decoder 222 further performs a decoding process on the subtitle and graphics elementary streams to obtain subtitle data and graphics data.
- the subtitle data and graphics data are supplied to the stereoscopic image subtitle / graphics generating unit 226.
- the stereoscopic image subtitle / graphics generating unit 226 is also supplied with the disparity vector obtained by the disparity vector decoder 225.
- the stereoscopic image subtitle / graphics generating unit 226 generates left eye and right eye subtitle information and graphics information data to be superimposed on the left eye image and the right eye image, respectively. This generation process is performed based on the subtitle data and graphics data obtained by the decoder 222 and the disparity vector obtained by the decoder 225.
- the right-eye subtitle information and graphics information are shifted in the horizontal direction by the horizontal component of the disparity vector. Is done.
- the generated left-eye and right-eye subtitle information and graphics information data are output.
- the text decoder 223 reconstructs a text elementary stream from the text TS packet extracted by the demultiplexer 220, and further performs a decoding process to obtain text data.
- This text data is supplied to the stereoscopic image text generator 227.
- the stereoscopic image text generation unit 227 is also supplied with the disparity vector obtained by the disparity vector decoder 225.
- this stereoscopic image text generation unit 227 based on the text data obtained by the decoder 223 and the disparity vector obtained by the decoder 225, left-eye text information to be superimposed on the left-eye image and the right-eye image, Eye text information is generated.
- the left-eye text information and the right-eye text information are the same text information, but the superimposed position in the image is, for example, the left-eye text information, and the right-eye text information is the horizontal component of the disparity vector. Only to be shifted horizontally.
- data (bitmap data) of the generated left eye text information and right eye text information is output.
- the disparity information set that matches the superimposition information is determined based on the identification information (Disparity_Sequence_id) added to each disparity information set. Can be used. That is, with respect to superimposition information such as closed caption information and subtitle information, an appropriate disparity can be given using an associated disparity vector.
- the video superimposing unit 228 is supplied with data output from the subtitle / graphics generating unit 226 and the text generating unit 227. Is done.
- the video superimposing unit 228 superimposes the data generated by the subtitle / graphics generating unit 226 and the text generating unit 227 on the stereoscopic image data (left-eye image data, right-eye image data), and displays stereoscopic image data for display. Vout is obtained.
- the display stereoscopic image data Vout is supplied as transmission image data to the HDMI transmission unit 206 (see FIG. 79) via the video signal processing circuit 205.
- the audio decoder 224 reconstructs an audio elementary stream from the audio TS packet extracted by the demultiplexer 220, and further performs a decoding process to obtain audio data.
- This audio data is supplied to the multi-channel speaker control unit 229.
- processing for generating multi-channel speaker audio data for realizing 5.1ch surround, processing for giving predetermined sound field characteristics, and the like are performed on the audio data.
- the parallax vector obtained by the parallax vector decoder 225 is also supplied to the multi-channel speaker control unit 229.
- the multi-channel speaker control unit 229 controls the output of the multi-channel speaker based on the parallax vector.
- the multichannel audio data obtained by the multichannel speaker control unit 229 is supplied as transmission audio data to the HDMI transmission unit 206 (see FIG. 79) via the audio signal processing circuit 207.
- the image data stream, the subtitle or graphics data stream, the text data stream, and the disparity vector (FIG. 8, FIG. A disparity vector stream including a reference) is sent.
- the disparity vectors for each predetermined unit corresponding to each fixed period are collected in accordance with the start timing of each fixed period such as GOP (Group Of Pictures) of an encoded video, I (Intra picture) picture, or scene. Will be sent.
- GOP Group Of Pictures
- I Intra picture
- the predetermined unit for example, a picture (frame) unit or a unit of an integer multiple of a picture can be considered.
- parallax based on a corresponding disparity vector is given to the superimposition information for each predetermined unit in the superimposition period of the superimposition information.
- the arrows attached to the subtitle (graphics) data stream and text data stream portions in FIG. 82 represent the timing at which disparity is given to the superimposition information for each predetermined unit.
- the superimposition information is selected from the disparity vectors for the superimposition period.
- Parallax is given by the parallax vector (parallax information).
- the predetermined disparity vector is, for example, disparity information indicating the maximum disparity among the disparity information for the overlapping period.
- the arrows attached to the subtitle (graphics) data stream and text data stream portions in FIG. 83 represent the timing at which disparity is added to the superimposition information.
- parallax is given to the superimposition information at the beginning of the superposition period, and thereafter, the superposition information to which the parallax is given is used as superposition information to be superimposed on the left eye image and the right eye image.
- image data, closed caption data, and graphics data are sent from the broadcasting station 100 (see FIG. 1).
- the disparity information set (see FIGS. 29, 31, 33, and 36) is sent as a disparity vector stream.
- a disparity information set for each predetermined unit corresponding to each predetermined period is sent together in accordance with the start timing of each predetermined period such as a GOP of encoded video, an I picture, or a scene.
- the predetermined unit for example, a picture (frame) unit or a unit of an integer multiple of a picture can be considered.
- parallax based on a corresponding disparity vector is given to the superimposition information for each predetermined unit in the superimposition period of the superimposition information.
- the arrows attached to the closed caption data and graphics data portions in FIG. 84 represent the timing at which the parallax is added to the superposition information for each predetermined unit.
- the bit stream processing unit 201A shown in FIG. 85 has a configuration corresponding to the transmission data generation units 110A and 110D shown in FIGS. 13 and 22 described above.
- portions corresponding to those in FIG. 80 are denoted by the same reference numerals, and detailed description thereof is omitted.
- the bit stream processing unit 201A is provided with a disparity vector extracting unit 231 instead of the disparity vector decoder 225 of the bit stream processing unit 201 shown in FIG.
- the disparity vector extracting unit 231 extracts a disparity vector embedded in the user data area from a video stream obtained through the video decoder 221.
- the disparity vector extracting unit 231 supplies the extracted disparity vector to the stereoscopic image subtitle / graphics generating unit 226, the stereoscopic image text generating unit 227, and the multi-channel speaker control unit 229.
- bit stream processing unit 201A shown in FIG. 85 are configured in the same manner as the bit stream processing unit 201 shown in FIG. 80 and operate in the same manner.
- an image data stream, a subtitle or graphics data stream, and a text data stream are sent.
- the disparity vector (see FIG. 8, FIG. 29, etc.) is sent embedded in the image data stream.
- a disparity vector corresponding to the predetermined unit is embedded in the image data stream and sent.
- a parallax based on a corresponding disparity vector (information information) is given to the superimposition information for each predetermined unit in the superimposition period of the superimposition information.
- the arrows attached to the subtitle (graphics) data stream and text data stream portions in FIG. 86 represent the timing at which disparity is given to the superimposition information for each predetermined unit.
- an image data stream, a subtitle or graphics data stream, and a text data stream are sent from the broadcasting station 100 (see FIG. 1).
- the disparity vector (see FIG. 8, FIG. 29, etc.) is sent embedded in the image data stream.
- the disparity vectors for each predetermined unit corresponding to each fixed period are sent together in accordance with the start timing of each fixed period such as GOP of encoded video, I picture, or scene.
- the predetermined unit for example, a picture (frame) unit or a unit of an integer multiple of a picture can be considered.
- parallax based on a corresponding disparity vector is given to the superimposition information for each predetermined unit in the superimposition period of the superimposition information.
- the arrows attached to the subtitle (graphics) data stream and text data stream portions in FIG. 87 represent the timing at which disparity is given to the superposition information for each predetermined unit.
- the superimposition information is selected from the disparity vectors for the superimposition period.
- Parallax is given by the parallax vector (parallax information).
- the predetermined disparity vector is, for example, the predetermined disparity information is disparity information indicating the maximum disparity among the disparity information for the overlapping period.
- the arrows attached to the subtitle (graphics) data stream and text data stream portions in FIG. 88 represent the timing at which disparity is added to the superimposition information.
- parallax is given to the superimposition information at the beginning of the superposition period, and thereafter, the superposition information to which the parallax is given is used as superposition information to be superimposed on the left eye image and the right eye image.
- image data, closed caption data, and graphics data are sent from the broadcasting station 100 (see FIG. 1).
- the disparity information set (see FIGS. 29, 31, 33, and 36) is sent embedded in the image data stream.
- a disparity information set for each predetermined unit corresponding to each predetermined period is sent together in accordance with the start timing of each predetermined period such as a GOP of encoded video, an I picture, or a scene.
- the predetermined unit for example, a picture (frame) unit or a unit of an integer multiple of a picture can be considered.
- parallax based on a corresponding disparity vector is given to the superimposition information for each predetermined unit in the superimposition period of the superimposition information.
- the arrows attached to the closed caption data and graphics data portions in FIG. 89 represent the timing at which the parallax is given to the superposition information for each predetermined unit.
- bit stream processing unit 201B illustrated in FIG. 90 is configured to correspond to the transmission data generation unit 110B illustrated in FIG. 15 described above.
- parts corresponding to those in FIG. 80 are given the same reference numerals, and detailed descriptions thereof are omitted.
- the bit stream processing unit 201B is provided with a disparity vector extracting unit 232 instead of the disparity vector decoder 225 of the bit stream processing unit 201 shown in FIG.
- the disparity vector extracting unit 232 extracts a disparity vector embedded in a subtitle or graphics stream obtained through the subtitle / graphics decoder 222.
- the disparity vector extracting unit 232 supplies the extracted disparity vector to the stereoscopic image subtitle / graphics generating unit 226, the stereoscopic image text generating unit 227, and the multi-channel speaker control unit 229.
- bit stream processing unit 201B shown in FIG. 90 are configured in the same manner as the bit stream processing unit 201 shown in FIG. 80 and operate in the same manner. 90.
- the addition of parallax to the superimposition information in the bit stream processing unit 201B in FIG. 90 is the same as the parallax addition to the superposition information in the bit stream processing unit 201A in FIG. 85 described above (FIGS. 86 to 89). reference).
- the bit stream processing unit 201C illustrated in FIG. 91 has a configuration corresponding to the transmission data generation unit 110C illustrated in FIG. 21 described above. 91, parts corresponding to those in FIG. 80 are given the same reference numerals, and detailed descriptions thereof are omitted.
- the bit stream processing unit 201C is obtained by removing the parallax vector decoder 225, the stereoscopic image subtitle / graphics generating unit 226, and the stereoscopic image text generating unit 227 from the bit stream processing unit 201 shown in FIG.
- the disparity vector is reflected in advance in the data of subtitle information, graphics information, and text information.
- the transmitted subtitle data and graphics data include left eye subtitle information and graphics information data to be superimposed on the left eye image, and right eye subtitle information and graphics to be superimposed on the right eye image. Contains informational data.
- the transmitted text data includes left-eye text information data superimposed on the left-eye image and right-eye text information data superimposed on the right-eye image. Accordingly, the disparity vector decoder 225, the stereoscopic image subtitle / graphics generating unit 226, and the stereoscopic image text generating unit 227 are not necessary.
- the text data obtained by the text decoder 223 is code data (character code), it is necessary to convert it into bitmap data. This processing is performed at the final stage of the text decoder 223 or at the input stage of the video superimposing unit 228, for example.
- the bit stream processing unit 201D illustrated in FIG. 92 has a configuration corresponding to the transmission data generation unit 110F illustrated in FIG. 63 described above.
- portions corresponding to those in FIG. 80 are denoted by the same reference numerals, and detailed description thereof is omitted as appropriate.
- the bit stream processing unit 201D includes a demultiplexer 220, a video decoder 221, an audio decoder 224, a video superimposing unit 228, and a multi-channel speaker control unit 229.
- the bit stream processing unit 201D includes a parallax information extraction unit 235, a CC decoder 233, and a stereoscopic image CC generation unit 234.
- CC data and disparity information include user data type code (user data) in the user data area of the header portion of the video elementary stream output from the video encoder 113 of the transmission data generation unit 110F illustrated in FIG. It is embedded so that it can be identified by (identification information).
- the CC decoder 233 extracts CC data from the video elementary stream obtained through the video decoder 221. Further, the CC decoder 233 obtains closed caption information (caption character code) for each caption window (Caption Window), and control data for the superimposed position and display time from the CC data. Then, the CC decoder 233 supplies the closed caption information and the control data of the superimposed position and the display time to the stereoscopic image CC generation unit 234.
- the disparity information extracting unit 235 extracts disparity information from the video elementary stream obtained through the video decoder 221.
- This disparity information is associated with closed caption data (caption character code) for each caption window (Caption Window) acquired by the CC decoder 233 described above.
- This parallax information includes shift target designation information for designating closed caption information to be shifted based on this parallax information, among the closed caption information superimposed on the left eye image and the closed caption information superimposed on the right eye image. It has been added.
- the disparity information extracting unit 235 extracts disparity information used in the display frame period of the closed caption information based on the data in correspondence with the closed caption data for each caption window acquired by the CC decoder 233.
- This disparity information is, for example, disparity information that is commonly used in each frame in the display frame period of the closed caption information, or disparity information that is sequentially updated in each frame.
- the disparity information is added with flag information indicating whether the disparity information is commonly used in each frame in the display frame period or the disparity information that is sequentially updated in each frame in the display frame period.
- the disparity information extracting unit 235 extracts disparity information of the first frame and offset information with respect to disparity information of the second and subsequent frames as disparity information sequentially updated in each frame.
- the second and subsequent parallax information is offset information with respect to the parallax information of the previous frame.
- the disparity information extracting unit 235 extracts, for example, the disparity information of the first frame as disparity information that is commonly used in each frame.
- the stereoscopic image CC generating unit 234 includes, for each caption window (Caption ⁇ Window), left-eye closed caption information (caption) and right-eye closed caption information (caption) superimposed on the left-eye image and right-eye image, respectively. Generate data. This generation process is performed based on the closed caption data and superposition position control data obtained by the CC decoder 233 and the disparity information (disparity vector) extracted by the disparity information extracting unit 235.
- the stereoscopic image CC generating unit 234 performs a shift process for adding parallax to either or both of the left-eye closed caption information and the right-eye closed caption information based on the shift target designation information. . Further, the stereoscopic image CC generating unit 234 determines whether the disparity information extracted by the disparity information extracting unit 235 is the disparity information that is commonly used in each frame or the disparity information that is sequentially updated in each frame. Thus, disparity is given to the closed caption information as follows.
- the stereoscopic image CC generating unit 234 uses the closed caption information superimposed on the left eye image and the right eye image based on the common disparity information. Add parallax.
- the stereoscopic image CC generation unit 234 converts the disparity information updated for each frame into the closed caption information superimposed on the left eye image and the right eye image. Based on this, parallax is given.
- the disparity information that is sequentially updated in each frame includes, for example, disparity information of the first frame and offset information for the disparity information of the previous frame in the second and subsequent frames.
- first frame parallax is given to the closed caption information superimposed on the left eye image and the right eye image based on the parallax information of the first frame.
- second and subsequent frames the closed caption information superimposed on the left-eye image and the right-eye image is further given disparity by the offset information with reference to the disparity provision state of the previous frame.
- FIG. 93 shows a process of adding parallax to closed caption information superimposed on the left eye image and the right eye image based on the parallax information sequentially updated in each frame.
- a shift process is performed based on the parallax information DP0 of the first frame, and parallax is given to the left eye closed caption information LC and right eye closed caption information RC.
- the information LC and RC are further shifted by “+1” which is offset information (Differential) from the state of the frame (FrameDN).
- next frame the information LC and RC are further shifted by “+1” which is offset information from the state of the frame (Frame N + 1).
- the offset information is “0”, the shift state of the information LC, RC is kept the same as that of the frame (Frame N + 2).
- the information LC and RC are further shifted by “+1” which is offset information from the state of the frame (Frame N + 3).
- the video superimposing unit 228 performs the closed captioning of the left eye and the right eye obtained by the stereoscopic image CC generating unit 234 with respect to the stereoscopic image data (left eye image data and right eye image data) obtained by the video decoder 221.
- the information data is superimposed to obtain display stereoscopic image data Vout.
- superimposition of closed caption information data on stereoscopic image data is started by a time stamp of the system layer. Further, the superimposition duration time is controlled based on the display time control data for the closed caption information.
- bit stream processing unit 201D shown in FIG. 92 are configured in the same manner as the bit stream processing unit 201 shown in FIG.
- the bit stream data BSD output from the digital tuner 204 (see FIG. 79) is supplied to the demultiplexer 220.
- the demultiplexer 220 video and audio TS packets are extracted from the bit stream data BSD and supplied to each decoder.
- the video decoder 221 a video elementary stream is reconstructed from the video packet extracted by the demultiplexer 220, and further, decoding processing is performed, so that stereoscopic image data including left eye image data and right eye image data is obtained. can get.
- the stereoscopic image data is supplied to the video superimposing unit 228.
- the video video elementary stream reconstructed by the video decoder 221 is supplied to the CC decoder 233.
- the CC decoder 233 extracts CC data from the video elementary stream. Then, the CC decoder 233 obtains closed caption information (caption character code) for each caption window (Caption ⁇ Window), and control data for the superimposed position and display time from the CC data.
- the closed caption information and the superposition position and display time control data are supplied to the stereoscopic image CC generator 234.
- the video video elementary stream reconstructed by the video decoder 221 is supplied to the disparity information extracting unit 235.
- the disparity information extracting unit 235 extracts disparity information from the video elementary stream. This disparity information is associated with closed caption data (caption character code) for each caption window (Caption Window) acquired by the CC decoder 233 described above.
- the parallax information is supplied to the stereoscopic image CC generation unit 234.
- the stereoscopic image CC generating unit 2344 for each caption window (Caption Window), left-eye closed caption information (caption) and right-eye closed caption information (caption) to be superimposed on the left-eye image and the right-eye image, respectively. Data is generated. This generation process is performed based on the closed caption data and superposition position control data obtained by the CC decoder 233 and the disparity information (disparity vector) extracted by the disparity information extracting unit 235.
- the stereoscopic image CC generation unit 234 provides parallax to either or both of the left-eye closed caption information and the right-eye closed caption information based on the shift target designation information included in the parallax information. Shift processing is performed.
- the disparity information extracted by the disparity information extracting unit 235 is disparity information commonly used in each frame, the common caption information superimposed on the left eye image and the right eye image is included in the common caption information. Parallax is given based on the parallax information.
- the disparity information extracted by the disparity information extracting unit 235 is disparity information that is sequentially updated in each frame, the disparity information is updated for each frame into closed caption information superimposed on the left eye image and the right eye image. Parallax is given based on the obtained parallax information (see FIG. 93).
- the data of the closed caption information supplied from the CC generating unit 234 for stereoscopic images is superimposed on the stereoscopic image data (left-eye image data and right-eye image data) obtained by the video decoder 221.
- display stereoscopic image data Vout is obtained.
- the audio decoder 224 reconstructs an audio elementary stream from the audio TS packet extracted by the demultiplexer 220, and further performs a decoding process to obtain audio data.
- This audio data is supplied to the multi-channel speaker control unit 229.
- processing for generating multi-channel speaker audio data for realizing 5.1ch surround, processing for giving predetermined sound field characteristics, and the like are performed on the audio data.
- the multi-channel speaker control unit 229 can obtain the output of the multi-channel speaker.
- stereoscopic image data can be acquired from the payload portion of the video elementary stream, and CC data and disparity information can be acquired from the user data area of the header portion.
- the CC data and the disparity information are included in the user data area so as to be identifiable by the user data identification information (user_data_type_code). Therefore, the parallax information can be favorably acquired from the user data area together with the CC data based on the user data identification information.
- bit stream processing unit 201D illustrated in FIG. 92 only the left-eye closed caption information is shifted based on the shift target designation information (select_view_shift) added to the disparity information, and the right-eye closed caption information is displayed. Can only be shifted, or both can be shifted. That is, disparity reflecting the intention of the transmission side can be given to the left-eye closed caption information and the right-eye closed caption information.
- disparity information that is commonly used in each frame in the closed caption information display period or disparity information that is sequentially updated in each frame is received. Can be determined by flag information (temporal_extension_flag).
- the disparity added to the closed caption information can be dynamically changed in conjunction with the change in the image content.
- the processing load can be reduced, and the capacity of the memory that stores the disparity information can be suppressed.
- bit stream processing unit 201D illustrated in FIG. 92 an example is shown in which disparity information whose use is limited to closed caption information is received. Although a detailed description is omitted, for example, a bit stream processing unit that receives and processes disparity information whose use is limited to other superimposition information such as subtitle information can be similarly configured.
- FIG. 94 shows another configuration example of the set top box 200A.
- the set top box 200A includes a bit stream processing unit 201D, an HDMI terminal 202, an antenna terminal 203, a digital tuner 204, a video / graphic processing circuit 205A, an HDMI transmission unit 206, and an audio signal processing circuit 207. is doing.
- the set top box 200A includes a CPU 211, a flash ROM 212, a DRAM 213, an internal bus 214, a remote control receiver 215, a remote control transmitter 216, a network terminal 217, and an Ethernet interface 218.
- Ethernet is a registered trademark.
- the bit stream processing unit 201D extracts stereoscopic image data (left eye image data, right eye image data), audio data, superimposition information data, disparity information (disparity vector), and the like from the bit stream data BSD obtained by the digital tuner 204.
- the superimposition information data is subtitle data.
- This bit stream processing unit 201D is configured in substantially the same manner as the bit stream processing unit 201A shown in FIG.
- the bit stream processing unit 201D includes a demultiplexer 220A, a video decoder 221, a subtitle decoder 222A, and an audio decoder 224.
- the bit stream processing unit 201D includes a stereoscopic image subtitle generating unit 226A, a video superimposing unit 228, and a disparity vector extracting unit 231.
- the demultiplexer 220A extracts video, audio, and subtitle packets from the bit stream data BSD, and sends them to each decoder.
- the video decoder 221 reconstructs a video elementary stream from the video packet extracted by the demultiplexer 220A, performs decoding processing, and obtains stereoscopic image data including left-eye image data and right-eye image data.
- the subtitle decoder 222A reconstructs a subtitle elementary stream from the subtitle packet extracted by the demultiplexer 220A.
- the subtitle decoder 222A further performs a decoding process to obtain subtitle data.
- the audio decoder 224 reconstructs an audio elementary stream from the audio packet extracted by the demultiplexer 220A, performs a decoding process, obtains audio data, and outputs the audio data to the outside of the bit stream processing unit 201.
- the disparity vector extracting unit 231 extracts a disparity vector (disparity information set) embedded in the user data area from the video stream obtained through the video decoder 221.
- the disparity vector extracting unit 231 outputs the extracted disparity vector to the stereoscopic image subtitle generating unit 226 and outputs the disparity vector to the outside of the bit stream processing unit 201D.
- the stereoscopic image subtitle generating unit 226A generates left eye and right eye subtitle information to be superimposed on the left eye image and the right eye image, respectively. This generation process is performed based on the subtitle data obtained by the subtitle decoder 222A and the disparity vector (disparity information set) supplied from the disparity vector extracting unit 231. In this case, the left-eye and left-eye subtitle information is the same information. However, the superimposed position in the image is shifted in the horizontal direction by the horizontal component of the parallax vector, for example, with respect to the left-eye subtitle information. Then, the stereoscopic image subtitle generating unit 226A outputs left eye and left eye subtitle information data (bitmap data).
- the video superimposing unit 228 superimposes the left-eye and left-eye subtitle information data generated by the subtitle generating unit 226A on the stereoscopic image data (left-eye image data and right-eye image data) obtained by the video decoder 221. Then, stereoscopic image data for display is obtained. Note that superimposition of superimposition information data on stereoscopic image data (left-eye image data, right-eye image data) is started by a time stamp of the system layer.
- the video superimposing unit 228 outputs the stereoscopic image data for display to the outside of the bit stream processing unit 201D.
- the Ethernet interface 228 is connected to a network (not shown) such as the Internet via the network terminal 227.
- the Ethernet interface 228 can also acquire various information, for example, information such as a weather forecast and a stock price displayed in graphics using a widget according to a user operation.
- the Ethernet interface 228 is connected to the internal bus 214.
- the video / graphic processing circuit 205A performs image quality adjustment processing on the stereoscopic image data for display output from the bit stream processing unit 201D as necessary.
- the video / graphic processing circuit 205A superimposes data of graphics information by, for example, a widget on display stereoscopic image data output from the bit stream processing unit 201D.
- the video / graphic processing circuit 205A is supplied with a disparity vector (disparity information set) output from the bit stream processing unit 201D described above. Based on this disparity vector (disparity information set), the video / graphic processing circuit 205A gives disparity to the same graphics information to be superimposed on the left eye image and the right eye image. Thereby, as the same graphics information to be superimposed on the left eye image and the right eye image, it is possible to use information on which parallax adjustment is performed according to the perspective of each object in the image. The perspective consistency between each object in the image is maintained.
- the video / graphic processing circuit 205 ⁇ / b> A supplies the processed stereoscopic image data to the HDMI transmission unit 206.
- the audio signal processing circuit 207 performs sound quality adjustment processing or the like on the audio data output from the bit stream processing unit 201D as necessary, and supplies the processed audio data to the HDMI transmission unit 206.
- the HDMI transmission unit 206 transmits baseband image (video) and audio data from the HDMI terminal 202 by communication conforming to HDMI. In this case, the HDMI transmission unit 206 packs image and audio data and outputs them to the HDMI terminal 202 for transmission via the HDMI TMDS channel.
- the rest of the set top box 200A shown in FIG. 94 is configured in the same manner as the set top box 200 shown in FIG.
- a television broadcast signal input to the antenna terminal 203 is supplied to the digital tuner 204.
- the television broadcast signal is processed, and predetermined bit stream data BSD corresponding to the user's selected channel is obtained.
- the bit stream data BSD is supplied to the bit stream processing unit 201D.
- the bit stream processing unit 201D extracts stereoscopic image data (left eye image data, right eye image data), audio data, subtitle data, a disparity vector (disparity information set), and the like from the bit stream data. Further, in the bit stream processing unit 201D, the subtitle information data is combined with the stereoscopic image data to generate display stereoscopic image data.
- the bit stream processing unit 201D gives disparity to the same subtitle information to be superimposed on the left eye image and the right eye image based on the disparity vector.
- the same subtitle information superimposed on the left eye image and the right eye image can be used with the parallax adjusted according to the perspective of each object in the image. The perspective consistency between each object in the image is maintained.
- the stereoscopic image data for display generated by the bit stream processing unit 201D is subjected to image quality adjustment processing as necessary by the video / graphic processing circuit 205A. Also, in the video / graphic processing circuit 205A, graphics information data by a widget is combined with display stereoscopic image data generated by the bit stream processing unit 201D as necessary.
- disparity is given to the same graphics information to be superimposed on the left eye image and the right eye image based on the disparity vector (disparity information set) supplied from the bit stream processing unit 201D. Is done.
- the same graphics information to be superimposed on the left eye image and the right eye image it is possible to use information on which parallax adjustment is performed according to the perspective of each object in the image. The perspective consistency between each object in the image is maintained.
- the processed stereoscopic image data obtained from the video / graphic processing circuit 205A is supplied to the HDMI transmission unit 206.
- the audio data obtained by the bit stream processing unit 201D is supplied to the HDMI transmission unit 206 after the audio signal processing circuit 207 performs a sound quality adjustment process or the like as necessary.
- the stereoscopic image data and audio data supplied to the HDMI transmission unit 206 are transmitted from the HDMI terminal 202 to the HDMI cable 400 via the HDMI TMDS channel.
- the left eye image and the right Parallax is given to the same subtitle information superimposed on the eye image.
- the video / graphic processing circuit 205A gives disparity to the same graphics information superimposed on the left eye image and the right eye image based on the disparity vector (disparity information set). Therefore, not only the subtitle information sent from the broadcasting station but also the graphics information generated in the set top box 200A maintains the perspective consistency with each object in the image in the display. You can do that.
- FIG. 95 shows an example of superimposing subtitle information and graphics information on the left eye image and right eye image in the set top box 200A shown in FIG. FIG. 95 (a) shows a left eye image.
- the subtitle information “Subtitle 1” is superimposed at the position (x1, y1)
- the subtitle information “Subtitle 1” is superimposed at the position (x2, y2).
- the graphics information “Graphics 1” is superimposed at the position (x3, y3).
- FIG. 95B shows a right eye image.
- the subtitle information “Subtitle” “1” is superimposed at a position shifted by offset 1 (Offset 1) based on the corresponding disparity vector with respect to the superimposed position on the left eye image.
- the subtitle information “Subtitle 2” is superimposed at a position shifted by offset 2 (Offset 2) based on the corresponding disparity vector with respect to the superimposed position on the left eye image. Is done.
- the graphics information “Graphics 1” is superimposed at a position shifted by offset 3 (Offset 3) with respect to the superimposed position on the left eye image.
- bitstream processing unit 201D handles subtitle information and the video / graphic processing unit 205A handles graphic information. it can.
- the television receiver 300 receives stereoscopic image data sent from the set top box 200 via the HDMI cable 400.
- the television receiver 300 includes a 3D signal processing unit 301.
- the 3D signal processing unit 301 performs processing (decoding processing) corresponding to the transmission method on the stereoscopic image data to generate left-eye image data and right-eye image data. That is, the 3D signal processing unit 301 performs a process reverse to that of the video framing unit 112 in the transmission data generation units 110, 110A, 110B, and 110C shown in FIGS. 2, 13, 15, and 21. Then, the 3D signal processing unit 301 acquires the left eye image data and the right eye image data constituting the stereoscopic image data.
- FIG. 96 illustrates a configuration example of the television receiver 300.
- the television receiver 300 includes a 3D signal processing unit 301, an HDMI terminal 302, an HDMI receiving unit 303, an antenna terminal 304, a digital tuner 305, and a bit stream processing unit 306.
- the television receiver 300 includes a video / graphic processing circuit 307, a panel drive circuit 308, a display panel 309, an audio signal processing circuit 310, an audio amplification circuit 311, and a speaker 312.
- the television receiver 300 includes a CPU 321, a flash ROM 322, a DRAM 323, an internal bus 324, a remote control receiving unit 325, and a remote control transmitter 326.
- the antenna terminal 304 is a terminal for inputting a television broadcast signal received by a receiving antenna (not shown).
- the digital tuner 305 processes the television broadcast signal input to the antenna terminal 304 and outputs predetermined bit stream data (transport stream) corresponding to the user's selected channel.
- the bit stream processing unit 306 has the same configuration as the bit stream processing unit 201 of the set top box 200 shown in FIG.
- the bit stream processing unit 306 extracts stereoscopic image data (left eye image data, right eye image data), audio data, superimposition information data, disparity vectors (disparity information), and the like from the bit stream data.
- the superimposition information data is subtitle data, graphics data, text data (including closed caption data), and the like.
- the bit stream processing unit 306 combines the superimposition information data with the stereoscopic image data, and acquires display stereoscopic image data.
- the bit stream processing unit 306 outputs audio data.
- the HDMI receiving unit 303 receives uncompressed image data and audio data supplied to the HDMI terminal 302 via the HDMI cable 400 by communication conforming to HDMI.
- the HDMI receiving unit 303 has a version of, for example, HDMI 1.4, and can handle stereoscopic image data. Details of the HDMI receiving unit 303 will be described later.
- the 3D signal processing unit 301 performs decoding processing on the stereoscopic image data received by the HDMI receiving unit 303 or obtained by the bit stream processing unit 306 to generate left eye image data and right eye image data. To do. In this case, the 3D signal processing unit 301 performs decoding processing corresponding to the transmission method (see FIG. 4) on the stereoscopic image data obtained by the bit stream processing unit 306. Further, the 3D signal processing unit 301 performs a decoding process corresponding to a TMDS transmission data structure described later on the stereoscopic image data received by the HDMI receiving unit 303.
- the video / graphic processing circuit 307 generates image data for displaying a stereoscopic image based on the left-eye image data and the right-eye image data generated by the 3D signal processing unit 301.
- the video / graphic processing circuit 307 performs image quality adjustment processing on the image data as necessary. Further, the video / graphic processing circuit 307 synthesizes superimposition information data such as a menu and a program guide with the image data as necessary.
- the panel drive circuit 308 drives the display panel 309 based on the image data output from the video / graphic processing circuit 307.
- the display panel 309 is, for example, an LCD (Liquid Crystal Display), a PDP (Plasma). DisplayPanel) etc.
- the audio signal processing circuit 310 performs necessary processing such as D / A conversion on the audio data received by the HDMI receiving unit 303 or obtained by the bit stream processing unit 306.
- the audio amplification circuit 311 amplifies the audio signal output from the audio signal processing circuit 310 and supplies the amplified audio signal to the speaker 312.
- the CPU 321 controls the operation of each unit of the television receiver 300.
- the flash ROM 322 stores control software and data.
- the DRAM 323 constitutes a work area for the CPU 321.
- the CPU 321 develops software and data read from the flash ROM 322 on the DRAM 323 to activate the software, and controls each unit of the television receiver 300.
- the remote control receiving unit 325 receives the remote control signal (remote control code) transmitted from the remote control transmitter 326 and supplies it to the CPU 321.
- the CPU 321 controls each part of the television receiver 300 based on the remote control code.
- the CPU 321, flash ROM 322, and DRAM 323 are connected to the internal bus 324.
- the HDMI receiving unit 303 receives stereoscopic image data and audio data transmitted from the set top box 200 connected to the HDMI terminal 302 via the HDMI cable 400.
- the stereoscopic image data received by the HDMI receiving unit 303 is supplied to the 3D signal processing unit 301.
- the audio data received by the HDMI receiving unit 303 is supplied to the audio signal processing circuit 310.
- the TV broadcast signal input to the antenna terminal 304 is supplied to the digital tuner 305.
- the digital tuner 305 processes the television broadcast signal and outputs predetermined bit stream data (transport stream) corresponding to the user's selected channel.
- the bit stream data output from the digital tuner 305 is supplied to the bit stream processing unit 306.
- the bit stream processing unit 306 extracts stereoscopic image data (left eye image data, right eye image data), audio data, superimposition information data, disparity vectors (disparity information), and the like from the bit stream data. Further, the bit stream processing unit 306 synthesizes superimposition information (closed caption information, subtitle information, graphics information, text information) data with the stereoscopic image data to generate display stereoscopic image data.
- the display stereoscopic image data generated by the bit stream processing unit 306 is supplied to the 3D signal processing unit 301.
- the audio data obtained by the bit stream processing unit 306 is supplied to the audio signal processing circuit 310.
- the 3D signal processing unit 301 performs decoding processing on the stereoscopic image data received by the HDMI receiving unit 303 or obtained by the bit stream processing unit 306 to generate left eye image data and right eye image data. Is done.
- the left eye image data and right eye image data are supplied to the video / graphic processing circuit 307.
- image data for displaying a stereoscopic image is generated based on the left eye image data and the right eye image data, and image quality adjustment processing and superimposition information data synthesis processing are performed as necessary. Is also done.
- the image data obtained by the video / graphic processing circuit 307 is supplied to the panel drive circuit 308. Therefore, a stereoscopic image is displayed on the display panel 309.
- the left eye image based on the left eye image data and the right eye image based on the right eye image data are alternately displayed on the display panel 309 in a time division manner.
- the viewer can see only the left eye image with the left eye and the right eye with the right eye by wearing shutter glasses that alternately open the left eye shutter and the right eye shutter in synchronization with the display on the display panel 309. Only images can be seen, and stereoscopic images can be perceived.
- the audio signal processing circuit 310 necessary processing such as D / A conversion is performed on the audio data received by the HDMI receiving unit 303 or obtained by the bit stream processing unit 306.
- the audio data is amplified by the audio amplification circuit 311 and then supplied to the speaker 312. Therefore, sound corresponding to the display image on the display panel 309 is output from the speaker 312.
- FIG. 97 illustrates a configuration example of the HDMI transmission unit (HDMI source) 206 of the set top box 200 and the HDMI reception unit (HDMI sink) 303 of the television receiver 300 in the stereoscopic image display system 10 of FIG.
- the HDMI transmission unit 206 transmits a differential signal corresponding to pixel data of an uncompressed image for one screen in an effective image section (hereinafter, also referred to as an active video section as appropriate) using a plurality of channels.
- the effective image section is a section obtained by removing the horizontal blanking section and the vertical blanking section from the section from one vertical synchronization signal to the next vertical synchronization signal.
- the HDMI transmission unit 206 receives the differential signals corresponding to at least audio data, control data, other auxiliary data, etc. associated with the image on a plurality of channels in the horizontal blanking interval or the vertical blanking interval. Transmit to the unit 303 in one direction.
- the transmission channels of the HDMI system including the HDMI transmission unit 206 and the HDMI reception unit 303 include the following transmission channels. That is, three TMDS channels # 0 to ## as transmission channels for serially transmitting pixel data and audio data in one direction in synchronization with the pixel clock from the HDMI transmission unit 206 to the HDMI reception unit 303. There are two. There is also a TMDS clock channel as a transmission channel for transmitting a pixel clock.
- the HDMI transmission unit 206 includes an HDMI transmitter 81.
- the transmitter 81 converts, for example, pixel data of an uncompressed image into a corresponding differential signal, and is connected via the HDMI cable 400 with three TMDS channels # 0, # 1, and # 2 that are a plurality of channels. Serial transmission in one direction to the HDMI receiving unit 303.
- the transmitter 81 converts audio data accompanying uncompressed images, further necessary control data and other auxiliary data, etc. into corresponding differential signals, and converts them into three TMDS channels # 0, # 1, #. 2 serially transmits to the HDMI receiving unit 303 in one direction.
- the transmitter 81 transmits the pixel clock synchronized with the pixel data transmitted through the three TMDS channels # 0, # 1, and # 2 to the HDMI receiving unit 303 connected via the HDMI cable 400 using the TMDS clock channel. Send.
- the HDMI receiving unit 303 receives a differential signal corresponding to the pixel data transmitted from the HDMI transmitting unit 206 in one direction through a plurality of channels in the active video section. Further, the HDMI receiving unit 303 receives differential signals corresponding to audio data and control data transmitted in one direction from the HDMI transmitting unit 206 through a plurality of channels in a horizontal blanking interval or a vertical blanking interval. Receive.
- the HDMI receiving unit 303 has an HDMI receiver 82.
- This HDMI receiver 82 uses TMDS channels # 0, # 1, and # 2 to transmit a differential signal corresponding to pixel data and a difference corresponding to audio data and control data transmitted from the HDMI transmission unit 206 in one direction. Receive a motion signal. In this case, reception is performed in synchronization with the pixel clock transmitted from the HDMI transmission unit 206 via the TMDS clock channel.
- the transmission channels of the HDMI system include transmission channels called DDC (Display Data Channel) 83 and CEC line 84 in addition to the above-described TMDS channels # 0 to # 2 and the TMDS clock channel.
- the DDC 83 includes two signal lines (not shown) included in the HDMI cable 400.
- the DDC 83 is used by the HDMI transmitting unit 206 to read E-EDID (Enhanced Extended Display Identification Data) from the HDMI receiving unit 303.
- E-EDID Enhanced Extended Display Identification Data
- the HDMI receiving unit 303 has an EDID ROM (Read Only Memory) 85 that stores E-EDID, which is performance information related to its performance (Configuration / capability), in addition to the HDMI receiver 81. .
- E-EDID ROM Read Only Memory
- the HDMI transmission unit 206 reads the E-EDID from the HDMI reception unit 303 connected via the HDMI cable 400 via the DDC 83.
- the HDMI transmission unit 206 sends the read E-EDID to the CPU 211.
- the CPU 211 stores this E-EDID in the flash ROM 212 or the DRAM 213.
- the CPU 211 can recognize the performance setting of the HDMI receiving unit 303 based on the E-EDID. For example, the CPU 211 recognizes whether or not the television receiver 300 having the HDMI receiving unit 303 can handle stereoscopic image data, and if so, what kind of TMDS transmission data structure can be supported.
- the CEC line 84 is made up of one signal line (not shown) included in the HDMI cable 400, and is used for bidirectional communication of control data between the HDMI transmission unit 206 and the HDMI reception unit 303.
- the CEC line 84 constitutes a control data line.
- the HDMI cable 400 includes a line (HPD line) 86 connected to a pin called HPD (Hot Plug Detect).
- HPD line 86 is also used as a HEAC-line constituting a bidirectional communication path.
- the HDMI cable 400 includes a line (power line) 87 used for supplying power from the source device to the sink device.
- the HDMI cable 400 includes a utility line 88.
- the utility line 88 is also used as a HEAC + line constituting a bidirectional communication path.
- FIG. 98 shows a configuration example of the HDMI transmitter 81 and the HDMI receiver 82 of FIG.
- the HDMI transmitter 81 has three encoder / serializers 81A, 81B, and 81C corresponding to the three TMDS channels # 0, # 1, and # 2, respectively.
- Each of the encoders / serializers 81A, 81B, and 81C encodes the image data, auxiliary data, and control data supplied thereto, converts the parallel data into serial data, and transmits the data by a differential signal.
- the B component is supplied to the encoder / serializer 81A
- the G component is supplied to the encoder / serializer 81B
- the R component is supplied to the encoder / serializer 81C. Supplied.
- auxiliary data there are audio data and control packets.
- the control packets are supplied to, for example, the encoder / serializer 81A, and the audio data is supplied to the encoders / serializers 81B and 81C.
- the control data includes a 1-bit vertical synchronization signal (VSYNC), a 1-bit horizontal synchronization signal (HSYNC), and 1-bit control bits CTL0, CTL1, CTL2, and CTL3.
- the vertical synchronization signal and the horizontal synchronization signal are supplied to the encoder / serializer 81A.
- the control bits CTL0 and CTL1 are supplied to the encoder / serializer 81B, and the control bits CTL2 and CTL3 are supplied to the encoder / serializer 81C.
- the encoder / serializer 81A transmits the B component of the image data, the vertical synchronization signal and the horizontal synchronization signal, and auxiliary data supplied thereto in a time division manner. That is, the encoder / serializer 81A converts the B component of the image data supplied thereto into 8-bit parallel data that is a fixed number of bits. Further, the encoder / serializer 81A encodes the parallel data, converts it into serial data, and transmits it through the TMDS channel # 0.
- the encoder / serializer 81A encodes 2-bit parallel data of the vertical synchronization signal and horizontal synchronization signal supplied thereto, converts the data into serial data, and transmits the serial data through the TMDS channel # 0. Furthermore, the encoder / serializer 81A converts the auxiliary data supplied thereto into parallel data in units of 4 bits. Then, the encoder / serializer 81A encodes the parallel data, converts it into serial data, and transmits it through the TMDS channel # 0.
- Encoder / serializer 81B transmits the G component of image data, control bits CTL0 and CTL1, and auxiliary data supplied thereto in a time-sharing manner. That is, the encoder / serializer 81B sets the G component of the image data supplied thereto as parallel data in units of 8 bits, which is a fixed number of bits. Further, the encoder / serializer 81B encodes the parallel data, converts it into serial data, and transmits it through the TMDS channel # 1.
- the encoder / serializer 81B encodes the 2-bit parallel data of the control bits CTL0 and CTL1 supplied thereto, converts the data into serial data, and transmits it through the TMDS channel # 1. Furthermore, the encoder / serializer 81B converts the auxiliary data supplied thereto into parallel data in units of 4 bits. Then, the encoder / serializer 81B encodes the parallel data, converts it into serial data, and transmits it through the TMDS channel # 1.
- the encoder / serializer 81C transmits the R component of the image data, control bits CTL2 and CTL3, and auxiliary data supplied thereto in a time division manner. That is, the encoder / serializer 81C sets the R component of the image data supplied thereto as parallel data in units of 8 bits, which is a fixed number of bits. Further, the encoder / serializer 81C encodes the parallel data, converts it into serial data, and transmits it through the TMDS channel # 2.
- the encoder / serializer 81C encodes the 2-bit parallel data of the control bits CTL2 and CTL3 supplied thereto, converts it into serial data, and transmits it through the TMDS channel # 2. Furthermore, the encoder / serializer 81C converts the auxiliary data supplied thereto into parallel data in units of 4 bits. Then, the encoder / serializer 81C encodes the parallel data, converts it into serial data, and transmits it through the TMDS channel # 2.
- the HDMI receiver 82 has three recovery / decoders 82A, 82B, and 82C corresponding to the three TMDS channels # 0, # 1, and # 2, respectively. Then, each of the recovery / decoders 82A, 82B, and 82C receives image data, auxiliary data, and control data transmitted as differential signals through the TMDS channels # 0, # 1, and # 2. Further, each of the recovery / decoders 82A, 82B, and 82C converts the image data, auxiliary data, and control data from serial data to parallel data, and further decodes and outputs them.
- the recovery / decoder 82A receives the B component of image data, the vertical synchronization signal, the horizontal synchronization signal, and the auxiliary data that are transmitted as differential signals through the TMDS channel # 0. Then, the recovery / decoder 82A converts the B component of the image data, the vertical synchronization signal, the horizontal synchronization signal, and the auxiliary data from serial data to parallel data, and decodes and outputs them.
- the recovery / decoder 82B receives the G component of the image data, the control bits CTL0 and CTL1, and the auxiliary data transmitted by the differential signal through the TMDS channel # 1. Then, the recovery / decoder 82B converts the G component of the image data, the control bits CTL0 and CTL1, and the auxiliary data from serial data to parallel data, and decodes and outputs them.
- the recovery / decoder 82C receives the R component of the image data, the control bits CTL2 and CTL3, and the auxiliary data transmitted as a differential signal through the TMDS channel # 2. Then, the recovery / decoder 82C converts the R component of the image data, the control bits CTL2 and CTL3, and the auxiliary data from serial data to parallel data, and decodes and outputs them.
- FIG. 99 shows an example of the structure of TMDS transmission data.
- FIG. 99 shows sections of various transmission data when image data of horizontal ⁇ vertical 1920 pixels ⁇ 1080 lines is transmitted in TMDS channels # 0, # 1, and # 2.
- a video field in which transmission data is transmitted through the three TMDS channels # 0, # 1, and # 2 of HDMI.
- Video Data period Video Data period
- data island period data island period
- control period control period
- the video field period is a period from the rising edge (active edge) of a certain vertical synchronizing signal to the rising edge of the next vertical synchronizing signal.
- This video field period is divided into a horizontal blanking period (horizontal blanking), a vertical blanking period (verticalblanking), and an active video period (Active Video).
- This active video section is a section obtained by removing the horizontal blanking period and the vertical blanking period from the video field section.
- the video data section is assigned to the active video section.
- 1920 pixels (pixels) ⁇ 1080 lines of effective pixel (Active pixel) data constituting uncompressed image data for one screen is transmitted.
- Data island section and control section are assigned to horizontal blanking period and vertical blanking period.
- auxiliary data (Auxiliary data) is transmitted. That is, the data island period is assigned to a part of the horizontal blanking period and the vertical blanking period.
- audio data packets which are data not related to control, of auxiliary data are transmitted.
- the control section is assigned to other parts of the horizontal blanking period and the vertical blanking period.
- this control period for example, vertical synchronization signals, horizontal synchronization signals, control packets, and the like, which are data related to control, of auxiliary data are transmitted.
- FIG. 100 shows an example of the pin arrangement of the HDMI terminal.
- the pin arrangement shown in FIG. 100 is called type A (type-A).
- TMDS Data # i + and TMDS Data # i ⁇ which are differential signals of TMDS channel #i, are transmitted by two lines which are differential lines. These two lines consist of pins assigned TMDS Data # i + (pin numbers 1, 4, and 7) and pins assigned TMDS Data # i- (pin numbers 3, 6, and 7). 9 pin).
- the CEC line 84 through which the CEC signal, which is control data, is transmitted is connected to a pin whose pin number is 13.
- a line for transmitting an SDA (Serial Data) signal such as E-EDID is connected to a pin having a pin number of 16.
- a line through which an SCL (Serial Clock) signal, which is a clock signal used for synchronization during transmission and reception of the SDA signal, is connected to a pin having a pin number of 15.
- the above-described DDC 83 includes a line for transmitting the SDA signal and a line for transmitting the SCL signal.
- the HPD line (HEAC-line) 86 for the source device to detect the connection of the sink device is connected to a pin having a pin number of 19.
- the utility line (HEAC + line) 88 is connected to a pin having a pin number of 14.
- the line 87 for supplying power is connected to a pin having a pin number of 18.
- the HDMI transmission unit 206 receives the E-EDID from the HDMI reception unit 303 connected via the HDMI cable 400 via the DDC 83 in response to a request from the CPU 211 (see FIG. 79), for example. Read out. Then, based on this E-EDID, the CPU 211 recognizes the setting of the performance of the HDMI receiving unit 303, for example, whether or not stereoscopic image data can be handled.
- FIG. 101 shows an example of the data structure of E-EDID.
- This E-EDID is composed of a basic block and an extended block.
- data defined in the E-EDID1.3 standard represented by “E-EDID1.3 Basic Structure” is arranged at the head.
- timing information for maintaining compatibility with the conventional EDID represented by “Preferred timing” is arranged.
- timing information different from “Preferredtiming” for maintaining compatibility with the conventional EDID represented by “2nd timing” is arranged.
- the basic block has “Monitor” following “2nd timing”.
- Information indicating the name of the display device represented by “NAME” is arranged, followed by “Monitor” in the basic block.
- “Short Video Descriptor” is placed at the beginning of the extension block. This is information indicating the displayable image size (resolution), the frame rate, and whether it is interlaced or progressive. Subsequently, “Short Audio Descriptor” is arranged. This is information such as a reproducible audio codec system, a sampling frequency, a cutoff band, and the number of codec bits. Subsequently, information on left and right speakers represented by “Speaker Allocation” is arranged.
- “Speaker Allocation” is followed by data uniquely defined for each manufacturer represented by “Vender Specific”. Subsequently, timing information for maintaining compatibility with the conventional EDID represented by “3rd timing” is arranged in the extension block. In the extended block, subsequently, timing information for maintaining compatibility with the conventional EDID represented by “4th timing” is arranged.
- FIG. 102 shows a Vender Specific area (HDMI (Vendor Specific DataBlock) data structure example.
- Vender In the Specific area, a 0th block to an Nth block which are 1-byte blocks are provided.
- information indicating the number “0x000C03” registered for HDMI (R) represented by “24bit IEEE Registration Identifier (0x000C03) LSB first” is arranged.
- information indicating the physical address of the 24-bit sink device represented by “A”, “B”, “C”, and “D” is arranged.
- a flag indicating a function supported by the sink device which is represented by “Supports-AI”, is arranged.
- information specifying the number of bits per pixel represented by “DC-48 bit”, “DC-36 bit”, and “DC-30 bit” is arranged.
- the sink device represented by “DVI-Dual” is dual DVI (Digital A flag indicating whether it corresponds to (VisualInterface) is arranged.
- the seventh block information indicating the maximum frequency of the TMDS pixel clock represented by “Max-TMDS-Clock” is arranged.
- a flag indicating the presence / absence of video and audio delay information represented by “Latency” is arranged in the sixth and seventh bits of the eighth block.
- a flag indicating whether or not the additional HDMI video format (3D, 4k ⁇ 2k) represented by “HDMI_Video_present” can be handled is arranged in the fifth bit of the eighth block.
- the delay time data of progressive video represented by “Video Latency” is arranged in the ninth block, and the audio accompanying the progressive video represented by “Audio Latency” is arranged in the tenth block.
- the delay time data is arranged.
- delay time data of interlaced video represented by “Interlaced Video Latency” is arranged in the eleventh block.
- audio delay time data associated with interlaced video represented by “Interlaced Audio Latency” is arranged.
- a flag indicating whether or not 3D image data represented by “3D_present” can be handled is arranged in the seventh bit of the thirteenth block.
- the 7th to 5th bits of the 14th block in addition to the mandatory 3D data structure represented by “HDMI_VIC_LEN”, which is arranged after the 15th block (not shown), it can be handled.
- the block size information indicating Also, in the 4th to 0th bits of the 14th block, there is block size information indicating a 4k ⁇ 2k video format that can be handled after the 15th block (not shown) represented by “HDMI_3D_LEN”. Has been placed.
- FIG. 103 shows a 3D video format (Frame Packing) 3D video format, which is one of the TMDS transmission data structures of stereoscopic image data.
- This 3D video format is a format for transmitting progressive left-eye (L) and right-eye (R) image data as stereoscopic image data.
- image data of 1920 ⁇ 1080p and 1080 ⁇ 720p pixel format is transmitted as image data of the left eye (L) and right eye (R).
- FIG. 93 shows an example in which the left eye (L) image data and the right eye (R) image data are 1920 lines ⁇ 1080 pixels, respectively.
- transmission data is generated in units of a video field section including a horizontal blanking period (Hblank), a vertical blanking period (Vblank), and an active video section (Hactive ⁇ Vactive) divided by a vertical synchronization signal. Is done.
- the active video section has two active video areas (Active video) and one active space area (Activespace) between them. Left eye (L) image data is arranged in the first active video area, and right eye (R) image data is arranged in the second active video area.
- FIG. 104 shows a line alternative (Line) which is one of TMDS transmission data structures of stereoscopic image data.
- An alternative) 3D video format (3D VideoFormat) is shown.
- This 3D video format is a format for transmitting progressive left-eye (L) and right-eye (R) image data as stereoscopic image data.
- image data of a 1920 ⁇ 1080p pixel format is transmitted as image data of the left eye (L) and right eye (R).
- transmission is performed in units of video field sections including a horizontal blanking period (Hblank), a vertical blanking period (2 ⁇ Vblank), and an active video section ((Hactive ⁇ 2Vactive)), which are separated by a vertical synchronization signal.
- Hblank horizontal blanking period
- 2 ⁇ Vblank vertical blanking period
- active video section ((Hactive ⁇ 2Vactive)
- FIG. 105 shows a side-by-side (Full) 3D video format (3D VideoFormat) which is one of the TMDS transmission data structures of stereoscopic image data.
- This 3D video format is a format for transmitting progressive left-eye (L) and right-eye (R) image data as stereoscopic image data.
- image data of a 1920 ⁇ 1080p pixel format is transmitted as image data of the left eye (L) and right eye (R).
- 3D video format transmission is performed in units of a video field section including a horizontal blanking period (2 ⁇ Hblank), a vertical blanking period (Vblank), and an active video section ((2Hactive ⁇ Vactive)) separated by a vertical synchronization signal.
- a horizontal blanking period (2 ⁇ Hblank) a horizontal blanking period
- Vblank a vertical blanking period
- active video section ((2Hactive ⁇ Vactive)) separated by a vertical synchronization signal.
- left eye (L) image data is arranged in the first half of the horizontal direction
- R right eye
- HDMI 1.4 defines a 3D video format as a TMDS transmission data structure of stereoscopic image data in addition to the 3D video format shown in FIGS. 103 to 105 described above.
- a frame packing (frame packing for interlaced format) method for example, a field alternative method, a side-bay-side (half) method, and the like.
- the same superimposition information (superimposed on the left eye image and the right eye image) based on the parallax information on the other of the left eye image and the right eye image (Parllax is given to closed caption information, subtitle information, graphics information, text information, and the like. Therefore, as the same superimposition information superimposed on the left eye image and the right eye image, information on which parallax adjustment is performed according to the perspective of each object (object) in the image can be used. In this case, it is possible to maintain perspective consistency with each object in the image.
- FIG. 106 shows an example of a method for associating a disparity vector with a component element ID.
- the component element ID of the closed caption information is a window ID (window_id).
- the component element ID of the subtitle information is a region ID (region_id).
- a partition position ID having a maximum of 13 bits indicating the screen area to which the disparity vector belongs is associated with the component element ID.
- connect_flag 1-bit connect flag
- the connect flag added to the first partition position ID is “1”, which indicates that there is a further associated partition position ID.
- the connect flag added to the second partition position ID is set to “0”, which indicates that there is no associated partition position ID.
- the stereoscopic image display system 10 includes the broadcasting station 100, the set top box 200, and the television receiver 300.
- the television receiver 300 includes a bit stream processing unit 201 that functions in the same manner as the bit stream processing unit 201 in the set-top box 200, as shown in FIG. Therefore, as shown in FIG. 107, a stereoscopic image display system 10A including a broadcasting station 100 and a television receiver 300 is also conceivable.
- the set-top box 200 and the television receiver 300 are connected by an HDMI digital interface.
- the present invention can be applied even when these are connected by a digital interface similar to the HDMI digital interface (including wireless as well as wired).
- the present invention can be applied to a stereoscopic image display system or the like that superimposes information such as closed caption information, subtitle information, graphics information, text information and the like on an image.
- Data recording medium 131 ... Parallax information set creation unit, 132 ... Parallax information creation unit, 133 ... CC encoder, 200, 200A ... set Top box, 201, 201A, 201B, 201C, 201D ... bit stream processing unit, 202 ... HDMI terminal, 203 ... antenna terminal, 204 ... digital tuner, 205 ... video signal processing circuit, 205A ... Video / graphic processing unit, 206 ... HDMI transmission unit, 207 ... Audio signal processing circuit, 211 ... CPU, 212 ... Flash ROM, 213 ... DRAM, 214 ... Internal bus, 215, remote control receiver, 216, remote control transmitter, 220, 220A, demultiplexer 221...
- Video decoder 222... Subtitle / Graphics decoder, 222 A... 223... Text decoder, 224.
- digital tuner 306 ... bit stream processing unit, 307 ... video / graphic processing circuit, 308 ... panel drive circuit, 309 ... Display panel, 310 ... Audio signal processing circuit, 311 ... Audio amplification circuit, 312 ... Speaker, 321 ... CPU, 322 ... Flash ROM, 323 ... DRAM, 324 ... Internal bus, 325 ... Remote control receiver, 326 ... Remote control transmitter, 400 ... HDMI cable
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Human Computer Interaction (AREA)
- Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
- Controls And Circuits For Display Device (AREA)
- Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
Abstract
Description
左眼画像データおよび右眼画像データを含む立体画像データを出力する立体画像データ出力部と、
上記左眼画像データおよび上記右眼画像データによる画像に重畳する重畳情報のデータを出力する重畳情報データ出力部と、
上記左眼画像データおよび上記右眼画像データによる画像に重畳する重畳情報をシフトさせて視差を付与するための視差情報を出力する視差情報出力部と、
ペイロード部に上記立体画像データ出力部から出力される立体画像データを含み、ヘッダ部のユーザデータ領域に上記重畳情報データ出力部から出力される重畳情報のデータと上記視差情報出力部から出力される視差情報とを含むビデオエレメンタリーストリームを送信する送信部とを備え、
上記ビデオエレメンタリーストリームのヘッダ部のユーザデータ領域に含まれる上記重畳情報のデータおよび上記視差情報は、ユーザデータ識別情報により識別可能とされている
立体画像データ送信装置にある。
ペイロード部に左眼画像データおよび右眼画像データを含む立体画像データを含み、ヘッダ部のユーザデータ領域に、上記左眼画像データおよび上記右眼画像データによる画像に重畳する重畳情報のデータと、上記左眼画像データおよび上記右眼画像データによる画像に重畳する重畳情報をシフトさせて視差を付与するための視差情報を含み、上記ヘッダ部のユーザデータ領域に含まれている上記重畳情報のデータおよび上記視差情報がユーザデータ識別情報により識別可能とされているビデオエレメンタリーストリームを受信する受信部と、
上記受信部で受信された上記ビデオエレメンタリーストリームのペイロード部から上記立体画像データを取得する立体画像データ取得部と、
上記受信部で受信された上記ビデオエレメンタリーストリームのヘッダ部のユーザデータ領域から上記重畳情報のデータを取得する重畳情報データ取得部と、
上記受信部で受信された上記ビデオエレメンタリーストリームのヘッダ部のユーザデータ領域から上記視差情報を取得する視差情報取得部と、
上記立体画像データ取得部で取得された上記立体画像データに含まれる上記左眼画像データおよび上記右眼画像データと、上記重畳情報データ取得部で取得された上記重畳情報のデータと、上記視差情報取得部で取得された上記視差情報とを用い、左眼画像および右眼画像に重畳する同一の上記重畳情報に視差を付与し、上記重畳情報が重畳された左眼画像のデータおよび上記重畳情報が重畳された右眼画像のデータを得る画像データ処理部と
を備える立体画像データ受信装置にある。
左眼画像データおよび右眼画像データを含む立体画像データを出力する立体画像データ出力部と、
上記左眼画像データおよび上記右眼画像データによる画像に重畳する重畳情報をシフトさせて視差を付与するための視差情報を出力する視差情報出力部と、
上記立体画像データ出力部から出力される立体画像データおよび上記視差情報出力部から出力される視差情報を送信する送信部とを備え、
上記視差情報には、上記左眼画像データによる画像に重畳する重畳情報および上記右眼画像データによる画像に重畳する重畳情報のうち該視差情報に基づいてシフトさせる重畳情報を指定するシフト対象指定情報が付加されている
立体画像データ送信装置にある。
左眼画像データおよび右眼画像データを含む立体画像データと、上記左眼画像データおよび上記右眼画像データによる画像に重畳する重畳情報をシフトさせて視差を付与するための視差情報とを受信する受信部を備え、
上記視差情報には、上記左眼画像データによる画像に重畳する重畳情報および上記右眼画像データによる画像に重畳する重畳情報のうち該視差情報に基づいてシフトさせる重畳情報を指定するシフト対象指定情報が付加されており、
上記受信部で受信された上記立体画像データに含まれる上記左眼画像データおよび上記右眼画像データと、上記受信部で受信された上記視差情報と、該視差情報に付加されている上記シフト対象指定情報を用い、左眼画像および右眼画像に重畳する同一の上記重畳情報に視差を付与し、上記重畳情報が重畳された左眼画像のデータおよび上記重畳情報が重畳された右眼画像のデータを得る画像データ処理部をさらに備える
立体画像データ受信装置にある。
左眼画像データおよび右眼画像データを含む立体画像データを出力する立体画像データ出力部と、
上記左眼画像データおよび上記右眼画像データによる画像に重畳する重畳情報をシフトさせて視差を付与するための視差情報を出力する視差情報出力部と、
上記立体画像データ出力部から出力される立体画像データおよび上記視差情報出力部から出力される視差情報を送信する送信部とを備え、
上記視差情報は、上記重畳情報が表示される所定数のフレーム期間の各フレームで順次更新される視差情報である
立体画像データ送信装置にある。
左眼画像データおよび右眼画像データを含む立体画像データと、上記左眼画像データおよび上記右眼画像データによる画像に重畳する重畳情報をシフトさせて視差を付与するための視差情報とを受信する受信部を備え、
上記視差情報は、上記重畳情報が表示される所定数のフレーム期間の各フレームで順次更新される視差情報であり、
上記受信部で受信された上記立体画像データに含まれる上記左眼画像データおよび上記右眼画像データと、上記受信部で受信された上記視差情報を用い、左眼画像および右眼画像に重畳する同一の上記重畳情報に視差を付与し、上記重畳情報が重畳された左眼画像のデータおよび上記重畳情報が重畳された右眼画像のデータを得る画像データ処理部をさらに備える
立体画像データ受信装置にある。
左眼画像データおよび右眼画像データを含む立体画像データを出力する立体画像データ出力部と、
上記左眼画像データおよび上記右眼画像データによる画像に重畳する重畳情報をシフトさせて視差を付与するための視差情報を出力する視差情報出力部と、
上記立体画像データ出力部から出力される立体画像データおよび上記視差情報出力部から出力される視差情報を送信する送信部とを備え、
上記視差情報は、上記重畳情報が表示される所定数のフレーム期間の各フレームで共通に使用される視差情報であり、
上記視差情報には、上記各フレームで共通に使用される視差情報であるか、上記各フレームで順次更新される視差情報であるかを示す情報が付加されている
立体画像データ送信装置にある。
左眼画像データおよび右眼画像データを含む立体画像データと、上記左眼画像データおよび上記右眼画像データによる画像に重畳する重畳情報をシフトさせて視差を付与するための視差情報とを受信する受信部を備え、
上記視差情報は、上記重畳情報が表示される所定数のフレーム期間の各フレームで共通に使用される視差情報であり、
上記受信部で受信された上記立体画像データに含まれる上記左眼画像データおよび上記右眼画像データと、上記受信部で受信された上記視差情報を用い、左眼画像および右眼画像に重畳する同一の上記重畳情報に視差を付与し、上記重畳情報が重畳された左眼画像のデータおよび上記重畳情報が重畳された右眼画像のデータを得る画像データ処理部をさらに備える
立体画像データ受信装置にある。
左眼画像データおよび右眼画像データを含む立体画像データを出力する立体画像データ出力部と、
上記左眼画像データおよび上記右眼画像データによる画像に重畳する重畳情報をシフトさせて視差を付与するための視差情報を出力する視差情報出力部と、
上記立体画像データ出力部から出力される立体画像データおよび上記視差情報出力部から出力される視差情報を送信する送信部とを備え、
上記視差情報は、上記重畳情報が表示される所定数のフレーム期間の各フレームで共通に使用される視差情報であるか、または上記記重畳情報が表示される所定数のフレーム期間の各フレームで順次更新される視差情報であり、
上記視差情報には、上記各フレームで共通に使用される視差情報であるか、上記各フレームで順次更新される視差情報であるかを示すフラグ情報が付加されている
立体画像データ送信装置にある。
左眼画像データおよび右眼画像データを含む立体画像データと、上記左眼画像データおよび上記右眼画像データによる画像に重畳する重畳情報をシフトさせて視差を付与するための視差情報とを受信する受信部を備え、
上記視差情報は、上記重畳情報が表示される所定数のフレーム期間の各フレームで共通に使用される視差情報であるか、または上記記重畳情報が表示される所定数のフレーム期間の各フレームで順次更新される視差情報であり、
上記視差情報には、上記各フレームで共通に使用される視差情報であるか、上記各フレームで順次更新される視差情報であるかを示すフラグ情報が付加されており、
上記受信部で受信された上記立体画像データに含まれる上記左眼画像データおよび上記右眼画像データと、上記受信部で受信された上記視差情報と、該視差情報に付加されている上記フラグ情報とを用い、左眼画像および右眼画像に重畳する同一の上記重畳情報に視差を付与し、上記重畳情報が重畳された左眼画像のデータおよび上記重畳情報が重畳された右眼画像のデータを得る画像データ処理部をさらに備える
立体画像データ受信装置にある。
1.実施の形態
2.変形例
[立体画像表示システムの構成例]
図1は、実施の形態としての立体画像表示システム10の構成例を示している。この立体画像表示システム10は、放送局100と、セットトップボックス(STB:Set Top Box)200と、テレビ受信機300を有している。
Multimedia Interface)ケーブル400を介して接続されている。セットトップボックス200には、HDMI端子202が設けられている。テレビ受信機300には、HDMI端子302が設けられている。HDMIケーブル400の一端はセットトップボックス200のHDMI端子202に接続され、このHDMIケーブル400の他端はテレビ受信機300のHDMI端子302に接続されている。
放送局100は、ビットストリームデータを、放送波にのせて送信する。このビットストリームデータには、左眼画像データおよび右眼画像データを含む立体画像データ、音声データ、重畳情報データ、さらには視差情報(視差ベクトル)などが含まれる。ここで、重畳情報データは、クローズド・キャプションデータ、サブタイトルデータ、グラフィクスデータ、テキストデータなどである。
図2は、放送局100において、上述のビットストリームデータを生成する送信データ生成部110の構成例を示している。この構成例は、視差ベクトルを数値情報として送信する例である。この送信データ生成部110は、カメラ111L,111Rと、ビデオフレーミング部112と、ビデオエンコーダ113と、視差ベクトル検出部114と、視差ベクトルエンコーダ115を有している。
ここでは、立体画像データ(3D画像データ)の伝送方式として、以下の第1~第3の方式を挙げるが、これら以外の伝送方式であってもよい。ここでは、図3に示すように、左眼(L)および右眼(R)の画像データが、それぞれ、決められた解像度、例えば、1920×1080pのピクセルフォーマットの画像データである場合を例にとって説明する。
視差ベクトルの検出例について説明する。ここでは、左眼画像に対する右眼画像の視差ベクトルを検出する例について説明する。図5に示すように、左眼画像を検出画像とし、右眼画像を参照画像とする。この例では、(xi,yi)および(xj,yj)の位置における視差ベクトルが検出される。
AAC等の符号化を施し、オーディオのエレメンタリーストリームを生成する。
AAC等の符号化が施され、符号化オーディオデータを含むオーディオエレメンタリーストリームが生成される。このオーディオエレメンタリーストリームはマルチプレクサ122に供給される。
なお、上述の図2に示す送信データ生成部110は、視差ベクトルの伝送内容(図8参照)を独立したエレメンタリーストリームとして受信側に伝送する構成となっている。しかし、視差ベクトルの伝送内容を他のストリームの中に埋め込んで伝送することも考えられる。例えば、視差ベクトルの伝送内容は、ビデオのストリームにユーザデータとして埋め込まれて伝送される。また、例えば、視差ベクトルの伝送内容は、サブタイトル、グラフィクス、あるいはテキストのストリームに埋め込まれて伝送される。
図15は、送信データ生成部110Bの構成例を示している。この例も、視差ベクトルを数値情報として送信する例である。この送信データ生成部110Bは、視差ベクトルの伝送内容を、サブタイトルあるいはグラフィクスのデータストリームに埋め込んで伝送する構成となっている。この図15において、図2と対応する部分には同一符号を付し、その詳細説明は省略する。
また、上述の図2、図13、図15に示す送信データ生成部110,110A,110Bは、視差ベクトルを数値情報として送信する(図8参照)。しかし、視差ベクトルを数値情報として伝送する代わりに、画像に重畳するための重畳情報(例えば、サブタイトル情報、グラフィクス情報、テキスト情報等)のデータに送信側で視差情報を予め反映させて送信することも考えられる。
図2、図13、図15に示す送信データ生成部110,110A,110Bにおいて、視差ベクトル検出部114では、左眼画像データおよび右眼画像データに基づき、画像内の所定位置における視差ベクトル(視差情報)が検出される。送信データ生成部110,110A,110Bでは、視差ベクトル検出部114で検出された画像内の所定位置における視差情報をそのまま受信側に送信する構成となっている。
(1)汎用目的の視差情報セットのみを送る(特定サービスとの関連を持たない場合)
(2)汎用目的の視差情報セットのみを送る(特定サービスとの関連を持つ場合)
(3)汎用目的の視差情報セットと特定サービス用途向けの視差情報セットの双方を送る
(4)特定サービス用途向けの視差情報セットのみを送る
Of Block)毎の視差ベクトルを求める。グループは、ブロックの上位層に当たり、複数個の近接するブロックをまとめてグループ化することで得られる。図25(b)の例では、各グループは、破線枠で括られる4個のブロックにより構成されている。そして、各グループの視差ベクトルは、例えば、そのグループ内の全ブロックの視差ベクトルから、最も値の大きな視差ベクトルが選択されることで得られる。
視差情報セット作成部131で作成される視差情報セットについて、上述したケース(1)~(4)のそれぞれについてさらに説明する。
ケース(1)は、上述したように、汎用目的の視差情報セットのみを送るケースである。このケース(1)では、汎用目的の視差情報セットのみが作成される。なお、このケース(1)では、視差情報セットに含まれる視差情報は、特定サービスと関連を持たない。
ケース(2)は、上述したように、汎用目的の視差情報セットのみを送るケースである。このケース(2)では、汎用目的の視差情報セットのみが作成される。なお、このケース(2)では、視差情報セットに含まれる視差情報は特定サービスと関連を持つ。
ケース(3)は、上述したように、汎用目的の視差情報セットと、特定サービス用途向けの視差情報セットの双方を送るケースである。このケース(3)では、汎用目的の視差情報セットと特定サービス用途向けの視差情報セットの双方が作成される。この場合、汎用目的の視差情報セットに付加される識別情報(Disparity_Sequence_id)は「0」とされる。また、特定サービス用途向けの視差情報セットに付加される識別情報(Disparity_Sequence_id)は、「0」以外の値とされる。識別情報(Disparity_Sequence_id)のサービスとの関連付けは、例えば、上述したように、ユーザデータにその情報が挿入されることで行われる。あるいは、この関連付けは、例えば、上述したように、ディスパリティ・リンケージ・デスクリプタ(Disparity_Linkage_Descriptor)で行われる。
ケース(4)は、上述したように、特定サービス用途向けの視差情報セットのみを送るケースである。このケース(4)では、特定サービス用途向けの視差情報セットのみが作成される。この場合、特定サービス用途向けの視差情報セットに付加される識別情報(Disparity_Sequence_id)は、「0」以外の値とされる。例えば、識別情報(Disparity_Sequence_id)のサービスとの関連付けは、上述したように、ユーザデータにその情報が挿入されることで行われる。あるいは、この関連付けは、例えば、上述したように、ディスパリティ・リンケージ・デスクリプタ(Disparity_Linkage_Descriptor)で行われる。
AAC等の符号化が施され、符号化オーディオデータを含むオーディオエレメンタリーストリームが生成される。このオーディオエレメンタリーストリームはマルチプレクサ122に供給される。
関連付けされている各ストリームにおけるサービス・コンポーネントと視差情報との関連付けは、ディスパリティ・リンケージ・ディスクリプタ(Disparity_Linkage_Descriptor)を用いて行われる。また、このディスパリティ・リンケージ・デスクリプタにより、各ストリームにおける視差情報セット識別情報(Disparity_Sequence_id)とサービスとの関連付けも行われる。このディスパリティ・リンケージ・デスクリプタは、多重化ストリームのヘッダ領域、あるいは、多重化ストリーム内のPMT等のデスクリプタ・テーブルに配置される。
=0x47413934”は「ATSC」を示し、そのうち、“Data_Type
= 0x03” は「クローズド・キャプション」を示し、“Data_Type
= 0x06” は、「Bar_data (letterbox
or pillarbox areas)」を示す。また、“Service_id = 0x44544731”は、「AFD」を示す。
図51は、視差情報(Disparity_Information)を含むユーザデータの詳細構成の一例を示している。この例は、符号化方式がMPEG2である場合の例である。開始コード(user_data_start_code)の後のデータ識別子が、視差情報識別子(Disparity_Information_identifier)であるとき、その後のデータ本体として、視差情報(Disparity_Information)が配置される。
view」、例えば右眼画像のみに適用するか、それとも「both views」、つまり左眼画像および右眼画像のそれぞれに適用するのかを示す1ビットのデータである。“1”である場合には、「both
views」に適用することを示す。“0”である場合には、「secondview」に適用することを示す。「Disparity_precision」は、視差ベクトルに対して重畳情報をどの程度ずらすかを示す2ビットのデータである。“0”である場合には、視差ベクトルの1/2に相当する分だけずらすことを示す。“1”である場合には、視差ベクトルに相当する分だけずらすことを示す。
of service)分だけ、コンポーネント・リンケージ情報(Component_Linkage_Info)が配置される。このコンポーネント・リンケージ情報には、図55に示すように、コンポーネントのエレメント数だけ、そのコンポーネント・エレメントと視差ベクトルとの対応付け情報が配置される。
「送信データ生成部の他の構成例」
図22に示す送信データ生成部110Dでは、視差情報セットとして、汎用目的の視差情報セットおよび特定のサービス用途向けの視差情報セットのいずれか、あるいは双方を送信可能とされている。しかし、クローズド・キャプション情報あるいはサブタイトル情報などの限定用途の視差情報を送信することとし、既存の放送規格(ATSC、CEAなど)に容易に連携できるシンプルなデータ構造とすることも考えられる。
Caption)エンコーダ133と、ビデオエンコーダ113と、オーディオエンコーダ117と、マルチプレクサ122を有している。
AAC等の符号化が施され、符号化オーディオデータを含むオーディオエレメンタリーストリームが生成される。このオーディオエレメンタリーストリームはマルチプレクサ122に供給される。このマルチプレクサ122では、各エンコーダから供給されるエレメンタリーストリームのパケットが多重化され、伝送データとしてのビットストリームデータBSDが得られる。
次に、視差情報(caption_disparity_data)を含むユーザデータの詳細構成を説明する。図64は、その一例を示している。図65は、図64に示す構成内の各情報の内容を示している。この例は、符号化方式がMPEG2である場合の例である。32ビットの開始コード(user_data_start_code)の後に放送規格「ATSC」を識別する32ビットのコード“0x47413934”が配置されている。
図1に戻って、セットトップボックス200は、放送局100から放送波にのせて送信されてくるビットストリームデータ(トランスポートストリーム)を受信する。このビットストリームデータには、左眼画像データおよび右眼画像データを含む立体画像データ、音声データ、重畳情報データ、さらには視差情報(視差ベクトル)が含まれる。ここで、重畳情報データは、例えば、サブタイトルデータ、グラフィクスデータ、テキストデータ(クローズド・キャプションデータを含む)等である。
セットトップボックス200の構成例を説明する。図79は、セットトップボックス200の構成例を示している。このセットトップボックス200は、ビットストリーム処理部201と、HDMI端子202と、アンテナ端子203と、デジタルチューナ204と、映像信号処理回路205と、HDMI送信部206と、音声信号処理回路207を有している。また、このセットトップボックス200は、CPU211と、フラッシュROM212と、DRAM213と、内部バス214と、リモコン受信部215と、リモコン送信機216を有している。
図80は、ビットストリーム処理部201の構成例を示している。このビットストリーム処理部201は、上述の図2、図62に示す送信データ生成部110,110Eに対応させた構成となっている。このビットストリーム処理部201は、デマルチプレクサ220と、ビデオデコーダ221と、サブタイトル・グラフィクスデコーダ222と、テキストデコーダ223と、オーディオデコーダ224と、視差ベクトルデコーダ225を有している。また、このビットストリーム処理部201は、立体画像用サブタイトル・グラフィクス発生部226と、立体画像用テキスト発生部227と、ビデオ重畳部228と、マルチチャネルスピーカコントロール部229を有している。
ここで、図80に示すビットストリーム処理部201の立体画像用サブタイトル・グラフィクス発生部226および立体画像用テキスト発生部227における重畳情報への視差の付与について、さらに説明する。
図85に示すビットストリーム処理部201Aは、上述の図13、図22に示す送信データ生成部110A,110Dに対応させた構成となっている。この図85において、図80と対応する部分には同一符号を付し、その詳細説明は省略する。
ここで、図85に示すビットストリーム処理部201Aの立体画像用サブタイトル・グラフィクス発生部226および立体画像用テキスト発生部227における重畳情報への視差の付与について、さらに説明する。
図90に示すビットストリーム処理部201Bは、上述の図15に示す送信データ生成部110Bに対応させた構成となっている。この図90において、図80と対応する部分には同一符号を付し、その詳細説明は省略する。
図91に示すビットストリーム処理部201Cは、上述の図21に示す送信データ生成部110Cに対応させた構成となっている。この図91において、図80と対応する部分には同一符号を付し、その詳細説明は省略する。
図92に示すビットストリーム処理部201Dは、上述の図63に示す送信データ生成部110Fに対応させた構成となっている。この図92において、図80と対応する部分には同一符号を付し、適宜、その詳細説明を省略する。
図94は、セットトップボックス200Aの他の構成例を示している。この図94において、図79、図85と対応する部分には同一符号を付し、適宜、その詳細説明は省略する。このセットトップボックス200Aは、ビットストリーム処理部201Dと、HDMI端子202と、アンテナ端子203と、デジタルチューナ204と、映像・グラフィック処理回路205Aと、HDMI送信部206と、音声信号処理回路207を有している。
1」は、左眼画像への重畳位置に対して、対応する視差ベクトルに基づいて、オフセット1(Offset 1)だけずれた位置に重畳される。同様に、このビットストリーム処理部201Dにおいて、サブタイトル情報「Subtitle 2」は、左眼画像への重畳位置に対して、対応する視差ベクトルに基づいて、オフセット2(Offset 2)だけずれた位置に重畳される。また、映像・グラフィック処理回路205Aにおいて、グラフィクス情報「Graphics 1」は、左眼画像への重畳位置に対して、オフセット3(Offset 3)だけずれた位置に重畳される。
図1に戻って、テレビ受信機300は、セットトップボックス200からHDMIケーブル400を介して送られてくる立体画像データを受信する。このテレビ受信機300は、3D信号処理部301を有している。この3D信号処理部301は、立体画像データに対して、伝送方式に対応した処理(デコード処理)を行って、左眼画像データおよび右眼画像データを生成する。すなわち、この3D信号処理部301は、図2、図13、図15、図21に示す送信データ生成部110,110A,110B,110Cにおけるビデオフレーミング部112とは逆の処理を行う。そして、この3D信号処理部301は、立体画像データを構成する左眼画像データおよび右眼画像データを取得する。
テレビ受信機300の構成例を説明する。図96は、テレビ受信機300の構成例を示している。このテレビ受信機300は、3D信号処理部301と、HDMI端子302と、HDMI受信部303と、アンテナ端子304と、デジタルチューナ305と、ビットストリーム処理部306を有している。
DisplayPanel)等で構成されている。
図97は、図1の立体画像表示システム10における、セットトップボックス200のHDMI送信部(HDMIソース)206と、テレビ受信機300のHDMI受信部(HDMIシンク)303の構成例を示している。
Field)には、伝送データの種類に応じて、3種類の区間が存在する。この3種類の区間は、ビデオデータ区間(Video
Data period)、データアイランド区間(Data Islandperiod)、およびコントロール区間(Control period)である。
上述したように、HDMI送信部206は、例えば、CPU211(図79参照)からの要求に応じて、HDMIケーブル400を介して接続されているHDMI受信部303から、E-EDIDを、DDC83を介して読み出す。そして、CPU211は、このE-EDIDに基づき、HDMI受信部303の性能の設定、例えば立体画像データの取り扱いが可能か否か等を認識する。
NAME”で表される表示装置の名前を示す情報が配置されている。基本ブロックには、続いて、“Monitor
Range Limits”で表される、アスペクト比が4:3および16:9である場合についての表示可能な画素数を示す情報が配置されている。
Vendor Specific DataBlock)のデータ構造例を示している。このVender
Specific領域には、1バイトのブロックである第0ブロック乃至第Nブロックが設けられている。
VisualInterface)に対応しているかを示すフラグが配置されている。
図103は、立体画像データのTMDS伝送データ構造の一つであるフレームパッキング(Frame packing)方式の3Dビデオフォーマット(3D Video Format)を示している。この3Dビデオフォーマットは、立体画像データとして、プログレッシブ方式の左眼(L)および右眼(R)の画像データを伝送するためのフォーマットである。
alternative)方式の3Dビデオフォーマット(3D VideoFormat)を示している。この3Dビデオフォーマットは、立体画像データとして、プログレッシブ方式の左眼(L)および右眼(R)の画像データを伝送するためのフォーマットである。この3Dビデオフォーマットでは、左眼(L)および右眼(R)の画像データとして、1920×1080pのピクセルフォーマットの画像データの伝送が行われる。
なお、上述実施の形態おいては、1つのコンポーネント・エレメントIDに複数の視差ベクトルを関連付けるために、DHI(図54、図55)内に「number_of_mutiple_link」の情報を配置している。この「number_of_mutiple_link」により、コンポーネント・エレメントIDと関連付けされる視差ベクトルの個数が予め指定される。しかし、例えば、パーティション・ポジションID(Partition_position_id)に、コネクト・フラグ(Connect_flag)を付加して、1つのコンポーネント・エレメントIDに複数の視差ベクトルを関連付けする方法も考えられる。
Claims (16)
- 左眼画像データおよび右眼画像データを含む立体画像データを出力する立体画像データ出力部と、
上記左眼画像データおよび上記右眼画像データによる画像に重畳する重畳情報のデータを出力する重畳情報データ出力部と、
上記左眼画像データおよび上記右眼画像データによる画像に重畳する重畳情報をシフトさせて視差を付与するための視差情報を出力する視差情報出力部と、
ペイロード部に上記立体画像データ出力部から出力される立体画像データを含み、ヘッダ部のユーザデータ領域に上記重畳情報データ出力部から出力される重畳情報のデータと上記視差情報出力部から出力される視差情報とを含むビデオエレメンタリーストリームを送信する送信部とを備え、
上記ビデオエレメンタリーストリームのヘッダ部のユーザデータ領域に含まれる上記重畳情報のデータおよび上記視差情報は、ユーザデータ識別情報により識別可能とされている
立体画像データ送信装置。 - 上記視差情報には、上記左眼画像データによる画像に重畳する重畳情報および上記右眼画像データによる画像に重畳する重畳情報のうち該視差情報に基づいてシフトさせる重畳情報を指定するシフト対象指定情報が付加されている
請求項1に記載の立体画像データ送信装置。 - 上記視差情報は、上記重畳情報が表示される所定数のフレーム期間の各フレームで順次更新される視差情報である
請求項1に記載の立体画像データ送信装置。 - 上記視差情報は、上記所定数のフレーム期間の最初のフレームの視差情報と、2番目以降の、前のフレームの視差情報に対するオフセット情報とからなる
請求項3に記載の立体画像データ送信装置。 - 上記視差情報は、上記重畳情報が表示される所定数のフレーム期間の各フレームで共通に使用される視差情報である
請求項1に記載の立体画像データ送信装置。 - 上記視差情報は、上記重畳情報が表示される所定数のフレーム期間の各フレームで共通に使用される視差情報であるか、または上記記重畳情報が表示される所定数のフレーム期間の各フレームで順次更新される視差情報であり、
上記視差情報には、上記各フレームで共通に使用される視差情報であるか、上記各フレームで順次更新される視差情報であるかを示すフラグ情報が付加されている
請求項1に記載の立体画像データ送信装置。 - ペイロード部に左眼画像データおよび右眼画像データを含む立体画像データを含み、ヘッダ部のユーザデータ領域に、上記左眼画像データおよび上記右眼画像データによる画像に重畳する重畳情報のデータと、上記左眼画像データおよび上記右眼画像データによる画像に重畳する重畳情報をシフトさせて視差を付与するための視差情報を含み、上記ヘッダ部のユーザデータ領域に含まれている上記重畳情報のデータおよび上記視差情報がユーザデータ識別情報により識別可能とされているビデオエレメンタリーストリームを受信する受信部と、
上記受信部で受信された上記ビデオエレメンタリーストリームのペイロード部から上記立体画像データを取得する立体画像データ取得部と、
上記受信部で受信された上記ビデオエレメンタリーストリームのヘッダ部のユーザデータ領域から上記重畳情報のデータを取得する重畳情報データ取得部と、
上記受信部で受信された上記ビデオエレメンタリーストリームのヘッダ部のユーザデータ領域から上記視差情報を取得する視差情報取得部と、
上記立体画像データ取得部で取得された上記立体画像データに含まれる上記左眼画像データおよび上記右眼画像データと、上記重畳情報データ取得部で取得された上記重畳情報のデータと、上記視差情報取得部で取得された上記視差情報とを用い、左眼画像および右眼画像に重畳する同一の上記重畳情報に視差を付与し、上記重畳情報が重畳された左眼画像のデータおよび上記重畳情報が重畳された右眼画像のデータを得る画像データ処理部と
を備える立体画像データ受信装置。 - 左眼画像データおよび右眼画像データを含む立体画像データを出力する立体画像データ出力部と、
上記左眼画像データおよび上記右眼画像データによる画像に重畳する重畳情報をシフトさせて視差を付与するための視差情報を出力する視差情報出力部と、
上記立体画像データ出力部から出力される立体画像データおよび上記視差情報出力部から出力される視差情報を送信する送信部とを備え、
上記視差情報には、上記左眼画像データによる画像に重畳する重畳情報および上記右眼画像データによる画像に重畳する重畳情報のうち該視差情報に基づいてシフトさせる重畳情報を指定するシフト対象指定情報が付加されている
立体画像データ送信装置。 - 左眼画像データおよび右眼画像データを含む立体画像データと、上記左眼画像データおよび上記右眼画像データによる画像に重畳する重畳情報をシフトさせて視差を付与するための視差情報とを受信する受信部を備え、
上記視差情報には、上記左眼画像データによる画像に重畳する重畳情報および上記右眼画像データによる画像に重畳する重畳情報のうち該視差情報に基づいてシフトさせる重畳情報を指定するシフト対象指定情報が付加されており、
上記受信部で受信された上記立体画像データに含まれる上記左眼画像データおよび上記右眼画像データと、上記受信部で受信された上記視差情報と、該視差情報に付加されている上記シフト対象指定情報を用い、左眼画像および右眼画像に重畳する同一の上記重畳情報に視差を付与し、上記重畳情報が重畳された左眼画像のデータおよび上記重畳情報が重畳された右眼画像のデータを得る画像データ処理部をさらに備える
立体画像データ受信装置。 - 左眼画像データおよび右眼画像データを含む立体画像データを出力する立体画像データ出力部と、
上記左眼画像データおよび上記右眼画像データによる画像に重畳する重畳情報をシフトさせて視差を付与するための視差情報を出力する視差情報出力部と、
上記立体画像データ出力部から出力される立体画像データおよび上記視差情報出力部から出力される視差情報を送信する送信部とを備え、
上記視差情報は、上記重畳情報が表示される所定数のフレーム期間の各フレームで順次更新される視差情報である
立体画像データ送信装置。 - 上記視差情報は、上記所定数のフレーム期間の最初のフレームの視差情報と、2番目以降の、前のフレームの視差情報に対するオフセット情報とからなる
請求項10に記載の立体画像データ送信装置。 - 左眼画像データおよび右眼画像データを含む立体画像データと、上記左眼画像データおよび上記右眼画像データによる画像に重畳する重畳情報をシフトさせて視差を付与するための視差情報とを受信する受信部を備え、
上記視差情報は、上記重畳情報が表示される所定数のフレーム期間の各フレームで順次更新される視差情報であり、
上記受信部で受信された上記立体画像データに含まれる上記左眼画像データおよび上記右眼画像データと、上記受信部で受信された上記視差情報を用い、左眼画像および右眼画像に重畳する同一の上記重畳情報に視差を付与し、上記重畳情報が重畳された左眼画像のデータおよび上記重畳情報が重畳された右眼画像のデータを得る画像データ処理部をさらに備える
立体画像データ受信装置。 - 左眼画像データおよび右眼画像データを含む立体画像データを出力する立体画像データ出力部と、
上記左眼画像データおよび上記右眼画像データによる画像に重畳する重畳情報をシフトさせて視差を付与するための視差情報を出力する視差情報出力部と、
上記立体画像データ出力部から出力される立体画像データおよび上記視差情報出力部から出力される視差情報を送信する送信部とを備え、
上記視差情報は、上記重畳情報が表示される所定数のフレーム期間の各フレームで共通に使用される視差情報であり、
上記視差情報には、上記各フレームで共通に使用される視差情報であるか、上記各フレームで順次更新される視差情報であるかを示す情報が付加されている
立体画像データ送信装置。 - 左眼画像データおよび右眼画像データを含む立体画像データと、上記左眼画像データおよび上記右眼画像データによる画像に重畳する重畳情報をシフトさせて視差を付与するための視差情報とを受信する受信部を備え、
上記視差情報は、上記重畳情報が表示される所定数のフレーム期間の各フレームで共通に使用される視差情報であり、
上記受信部で受信された上記立体画像データに含まれる上記左眼画像データおよび上記右眼画像データと、上記受信部で受信された上記視差情報を用い、左眼画像および右眼画像に重畳する同一の上記重畳情報に視差を付与し、上記重畳情報が重畳された左眼画像のデータおよび上記重畳情報が重畳された右眼画像のデータを得る画像データ処理部をさらに備える
立体画像データ受信装置。 - 左眼画像データおよび右眼画像データを含む立体画像データを出力する立体画像データ出力部と、
上記左眼画像データおよび上記右眼画像データによる画像に重畳する重畳情報をシフトさせて視差を付与するための視差情報を出力する視差情報出力部と、
上記立体画像データ出力部から出力される立体画像データおよび上記視差情報出力部から出力される視差情報を送信する送信部とを備え、
上記視差情報は、上記重畳情報が表示される所定数のフレーム期間の各フレームで共通に使用される視差情報であるか、または上記記重畳情報が表示される所定数のフレーム期間の各フレームで順次更新される視差情報であり、
上記視差情報には、上記各フレームで共通に使用される視差情報であるか、上記各フレームで順次更新される視差情報であるかを示すフラグ情報が付加されている
立体画像データ送信装置。 - 左眼画像データおよび右眼画像データを含む立体画像データと、上記左眼画像データおよび上記右眼画像データによる画像に重畳する重畳情報をシフトさせて視差を付与するための視差情報とを受信する受信部を備え、
上記視差情報は、上記重畳情報が表示される所定数のフレーム期間の各フレームで共通に使用される視差情報であるか、または上記記重畳情報が表示される所定数のフレーム期間の各フレームで順次更新される視差情報であり、
上記視差情報には、上記各フレームで共通に使用される視差情報であるか、上記各フレームで順次更新される視差情報であるかを示すフラグ情報が付加されており、
上記受信部で受信された上記立体画像データに含まれる上記左眼画像データおよび上記右眼画像データと、上記受信部で受信された上記視差情報と、該視差情報に付加されている上記フラグ情報とを用い、左眼画像および右眼画像に重畳する同一の上記重畳情報に視差を付与し、上記重畳情報が重畳された左眼画像のデータおよび上記重畳情報が重畳された右眼画像のデータを得る画像データ処理部をさらに備える
立体画像データ受信装置。
Priority Applications (6)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN2010800031387A CN102210154A (zh) | 2009-06-29 | 2010-06-22 | 立体图像数据发送装置和立体图像数据接收装置 |
EP10794029.8A EP2451172A4 (en) | 2009-06-29 | 2010-06-22 | TRANSMITTER FOR STEREOSCOPIC IMAGE DATA AND RECEIVER FOR STEREOSCOPIC IMAGE DATA |
US13/059,045 US8937642B2 (en) | 2009-06-29 | 2010-06-22 | Stereo image data transmitting apparatus and stereo image data receiving apparatus |
BRPI1004294A BRPI1004294A2 (pt) | 2009-06-29 | 2010-06-22 | aparelhos de transmissão e de recepção de dados de imagem estéreo |
US13/479,618 US20120257014A1 (en) | 2009-06-29 | 2012-05-24 | Stereo image data transmitting apparatus and stereo image data receiving apparatus |
US13/479,603 US8860786B2 (en) | 2009-06-29 | 2012-05-24 | Stereo image data transmitting apparatus and stereo image data receiving apparatus |
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2009-153686 | 2009-06-29 | ||
JP2009153686 | 2009-06-29 | ||
JP2010-060021 | 2010-03-16 | ||
JP2010060021A JP5407957B2 (ja) | 2009-06-29 | 2010-03-16 | 立体画像データ送信装置および立体画像データ受信装置 |
Related Child Applications (3)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/059,045 A-371-Of-International US8937642B2 (en) | 2009-06-29 | 2010-06-22 | Stereo image data transmitting apparatus and stereo image data receiving apparatus |
US13/479,603 Continuation US8860786B2 (en) | 2009-06-29 | 2012-05-24 | Stereo image data transmitting apparatus and stereo image data receiving apparatus |
US13/479,618 Continuation US20120257014A1 (en) | 2009-06-29 | 2012-05-24 | Stereo image data transmitting apparatus and stereo image data receiving apparatus |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2011001859A1 true WO2011001859A1 (ja) | 2011-01-06 |
Family
ID=43410936
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2010/060587 WO2011001859A1 (ja) | 2009-06-29 | 2010-06-22 | 立体画像データ送信装置および立体画像データ受信装置 |
Country Status (8)
Country | Link |
---|---|
US (3) | US8937642B2 (ja) |
EP (1) | EP2451172A4 (ja) |
JP (1) | JP5407957B2 (ja) |
KR (1) | KR20120036789A (ja) |
CN (1) | CN102210154A (ja) |
BR (1) | BRPI1004294A2 (ja) |
TW (1) | TWI466532B (ja) |
WO (1) | WO2011001859A1 (ja) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102202224A (zh) * | 2011-06-22 | 2011-09-28 | 清华大学 | 用于平面视频立体转换的字幕去抖方法及字幕去抖装置 |
WO2011155418A1 (ja) * | 2010-06-10 | 2011-12-15 | ソニー株式会社 | 立体画像データ送信装置、立体画像データ送信方法、立体画像データ受信装置および立体画像データ受信方法 |
WO2012057048A1 (ja) * | 2010-10-29 | 2012-05-03 | ソニー株式会社 | 立体画像データ送信装置、立体画像データ送信方法、立体画像データ受信装置および立体画像データ受信方法 |
WO2012060198A1 (ja) * | 2010-11-02 | 2012-05-10 | ソニー株式会社 | 立体画像データ送信装置、立体画像データ送信方法、立体画像データ受信装置および立体画像データ受信方法 |
CN102625121A (zh) * | 2011-01-25 | 2012-08-01 | Jvc建伍株式会社 | 3d图像处理装置及3d图像处理方法 |
JP2012205301A (ja) * | 2011-03-25 | 2012-10-22 | Samsung Electronics Co Ltd | シャッタメガネ制御方法及びそれを遂行する表示システム |
Families Citing this family (25)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP5407968B2 (ja) * | 2009-06-29 | 2014-02-05 | ソニー株式会社 | 立体画像データ送信装置および立体画像データ受信装置 |
JP2011172216A (ja) * | 2010-01-25 | 2011-09-01 | Panasonic Corp | 再生装置 |
EP2547101A4 (en) | 2010-04-14 | 2013-12-18 | Samsung Electronics Co Ltd | METHOD AND DEVICE FOR PRODUCING A BROADCASTING BITSTREAM FOR DIGITAL BROADCASTING WITH SUBTITLES AND METHOD AND DEVICE FOR RECEIVING A BROADBAND RADIO BITSTREAM FOR DIGITAL BROADCASTING WITH SUBTITLES |
KR20120004203A (ko) * | 2010-07-06 | 2012-01-12 | 삼성전자주식회사 | 디스플레이 방법 및 장치 |
JP5811602B2 (ja) * | 2010-12-16 | 2015-11-11 | ソニー株式会社 | 画像生成装置、プログラム、画像表示システム、および画像表示装置 |
CA2839256C (en) * | 2011-06-21 | 2017-07-11 | Lg Electronics Inc. | Method and apparatus for processing broadcast signal for 3-dimensional broadcast service |
JP2013051660A (ja) * | 2011-08-04 | 2013-03-14 | Sony Corp | 送信装置、送信方法および受信装置 |
KR20130024504A (ko) * | 2011-08-31 | 2013-03-08 | 삼성전기주식회사 | 삼차원 카메라 시스템 및 주시각 조정 방법 |
JP2013066075A (ja) * | 2011-09-01 | 2013-04-11 | Sony Corp | 送信装置、送信方法および受信装置 |
US9407897B2 (en) | 2011-09-30 | 2016-08-02 | Panasonic Intellectual Property Management Co., Ltd. | Video processing apparatus and video processing method |
KR20130088636A (ko) * | 2012-01-31 | 2013-08-08 | 삼성전자주식회사 | 영상 전송 장치 및 방법, 그리고 영상 재생 장치 및 방법 |
WO2013114887A1 (en) * | 2012-02-02 | 2013-08-08 | Panasonic Corporation | Methods and apparatuses for 3d media data generation, encoding, decoding and display using disparity information |
EP2683168B1 (en) | 2012-02-16 | 2019-05-01 | Sony Corporation | Transmission device, transmission method and receiver device |
JP2014027448A (ja) * | 2012-07-26 | 2014-02-06 | Sony Corp | 情報処理装置、情報処理方法、及びプログラム |
TWI520577B (zh) * | 2012-08-10 | 2016-02-01 | 晨星半導體股份有限公司 | 立體影像輸出裝置與相關的立體影像輸出方法 |
US10193566B2 (en) * | 2012-09-21 | 2019-01-29 | Nippon Telegraph And Telephone Corporation | Encoding/decoding system for parallel data |
US9654762B2 (en) * | 2012-10-01 | 2017-05-16 | Samsung Electronics Co., Ltd. | Apparatus and method for stereoscopic video with motion sensors |
CN102970556B (zh) * | 2012-12-03 | 2014-12-03 | 深圳市维尚视界立体显示技术有限公司 | 3dv 视频封装、播放方法及其系统 |
US9998800B2 (en) * | 2013-03-12 | 2018-06-12 | Electronics And Telecommunications Research Institute | 3D broadcast service providing method and apparatus, and 3D broadcast service reproduction method and apparatus for using image of asymmetric aspect ratio |
US9832479B2 (en) * | 2013-10-04 | 2017-11-28 | Ati Technologies Ulc | Motion estimation apparatus and method for multiview video |
GB2546720B (en) * | 2015-11-02 | 2019-11-13 | Advanced Risc Mach Ltd | Method of and apparatus for graphics processing |
WO2018163374A1 (ja) * | 2017-03-09 | 2018-09-13 | 三菱電機株式会社 | 誤り訂正装置および誤り訂正方法 |
CN114040141A (zh) * | 2017-10-20 | 2022-02-11 | 杭州海康威视数字技术股份有限公司 | 数据传输方法及摄像机、电子设备 |
CN109698933B (zh) * | 2017-10-20 | 2020-07-24 | 杭州海康威视数字技术股份有限公司 | 数据传输方法及摄像机、电子设备、计算机可读存储介质 |
TWI784330B (zh) * | 2020-10-21 | 2022-11-21 | 財團法人工業技術研究院 | 物件追蹤方法及其處理裝置與系統 |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH11289555A (ja) * | 1998-04-02 | 1999-10-19 | Toshiba Corp | 立体映像表示装置 |
JP2004274125A (ja) * | 2003-03-05 | 2004-09-30 | Sony Corp | 画像処理装置および方法 |
JP2005006114A (ja) | 2003-06-12 | 2005-01-06 | Sharp Corp | 放送データ送信装置、放送データ送信方法および放送データ受信装置 |
JP2009135686A (ja) * | 2007-11-29 | 2009-06-18 | Mitsubishi Electric Corp | 立体映像記録方法、立体映像記録媒体、立体映像再生方法、立体映像記録装置、立体映像再生装置 |
JP2009153686A (ja) | 2007-12-26 | 2009-07-16 | Lion Corp | 床用清掃シート |
Family Cites Families (26)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE19918046B4 (de) * | 1998-04-23 | 2007-02-15 | Lg Electronics Inc. | Speicherstruktur für Bild-in-Bild-Anzeige bei einer digitalen Videoanzeigeeinheit sowie Verfahren hierfür |
EP1418766A3 (en) * | 1998-08-28 | 2010-03-24 | Imax Corporation | Method and apparatus for processing images |
AU2003252388A1 (en) | 2002-08-27 | 2004-03-19 | Sharp Kabushiki Kaisha | Content reproduction device capable of reproducing a content in optimal reproduction mode |
US20060015919A1 (en) * | 2004-07-13 | 2006-01-19 | Nokia Corporation | System and method for transferring video information |
JP4535954B2 (ja) | 2005-04-18 | 2010-09-01 | 日本電信電話株式会社 | 2眼式立体表示装置およびプログラム |
EP1875440B1 (en) * | 2005-04-19 | 2008-12-03 | Koninklijke Philips Electronics N.V. | Depth perception |
JP2006325165A (ja) | 2005-05-20 | 2006-11-30 | Excellead Technology:Kk | テロップ発生装置、テロップ発生プログラム、及びテロップ発生方法 |
US7679641B2 (en) * | 2006-04-07 | 2010-03-16 | Real D | Vertical surround parallax correction |
JP4693900B2 (ja) | 2006-04-07 | 2011-06-01 | シャープ株式会社 | 画像処理装置 |
WO2008010929A2 (en) * | 2006-07-18 | 2008-01-24 | Thomson Licensing | Methods and apparatus for adaptive reference filtering |
JP2010510558A (ja) | 2006-10-11 | 2010-04-02 | コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ | 三次元グラフィックスのデータの作成 |
KR101299683B1 (ko) * | 2006-10-20 | 2013-08-26 | 삼성전자주식회사 | 디스플레이장치, 디스플레이 시스템 및 그 제어방법 |
KR101362941B1 (ko) * | 2006-11-01 | 2014-02-17 | 한국전자통신연구원 | 스테레오스코픽 콘텐츠 재생에 이용되는 메타 데이터의복호화 방법 및 장치 |
JP2008164367A (ja) | 2006-12-27 | 2008-07-17 | Matsushita Electric Ind Co Ltd | 固体撮像装置、カメラ、車両及び監視装置 |
JP4942106B2 (ja) | 2007-06-27 | 2012-05-30 | 独立行政法人情報通信研究機構 | 奥行データ出力装置及び奥行データ受信装置 |
JP2009017198A (ja) | 2007-07-04 | 2009-01-22 | Panasonic Corp | データ処理装置およびデータ処理方法 |
SE531899C2 (sv) | 2007-07-10 | 2009-09-01 | Agency 9 Ab | System för grafikhantering |
JP2009101822A (ja) | 2007-10-23 | 2009-05-14 | Mazda Motor Corp | 車両の追従走行制御装置 |
KR100955578B1 (ko) * | 2007-12-18 | 2010-04-30 | 한국전자통신연구원 | 스테레오스코픽 콘텐츠 장면 재생 방법 및 그 장치 |
CN102077246A (zh) * | 2008-06-24 | 2011-05-25 | 汤姆森特许公司 | 利用运动补偿进行图像的深度提取的系统和方法 |
MX2011002553A (es) * | 2008-09-18 | 2011-04-04 | Panasonic Corp | Dispositivo de reproduccion de video estereoscopico y dispositivo de presentacion visual de video estereoscopico. |
CN102232294B (zh) | 2008-12-01 | 2014-12-10 | 图象公司 | 用于呈现具有内容自适应信息的三维动态影像的方法和系统 |
EP2356820B1 (en) * | 2008-12-02 | 2017-07-19 | LG Electronics Inc. | 3d caption display method and 3d display apparatus for implementing the same |
AU2010215135B2 (en) | 2009-02-17 | 2016-05-12 | Koninklijke Philips Electronics N.V. | Combining 3D image and graphical data |
US8290338B2 (en) * | 2009-05-27 | 2012-10-16 | Panasonic Corporation | Recording medium, playback device, encoding device, integrated circuit, and playback output device |
JP5402715B2 (ja) * | 2009-06-29 | 2014-01-29 | ソニー株式会社 | 立体画像データ送信装置、立体画像データ送信方法、立体画像データ受信装置および立体画像データ受信方法 |
-
2010
- 2010-03-16 JP JP2010060021A patent/JP5407957B2/ja not_active Expired - Fee Related
- 2010-06-15 TW TW99119494A patent/TWI466532B/zh not_active IP Right Cessation
- 2010-06-22 BR BRPI1004294A patent/BRPI1004294A2/pt not_active IP Right Cessation
- 2010-06-22 KR KR20117003349A patent/KR20120036789A/ko not_active Application Discontinuation
- 2010-06-22 US US13/059,045 patent/US8937642B2/en not_active Expired - Fee Related
- 2010-06-22 EP EP10794029.8A patent/EP2451172A4/en not_active Withdrawn
- 2010-06-22 CN CN2010800031387A patent/CN102210154A/zh active Pending
- 2010-06-22 WO PCT/JP2010/060587 patent/WO2011001859A1/ja active Application Filing
-
2012
- 2012-05-24 US US13/479,618 patent/US20120257014A1/en not_active Abandoned
- 2012-05-24 US US13/479,603 patent/US8860786B2/en not_active Expired - Fee Related
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH11289555A (ja) * | 1998-04-02 | 1999-10-19 | Toshiba Corp | 立体映像表示装置 |
JP2004274125A (ja) * | 2003-03-05 | 2004-09-30 | Sony Corp | 画像処理装置および方法 |
JP2005006114A (ja) | 2003-06-12 | 2005-01-06 | Sharp Corp | 放送データ送信装置、放送データ送信方法および放送データ受信装置 |
JP2009135686A (ja) * | 2007-11-29 | 2009-06-18 | Mitsubishi Electric Corp | 立体映像記録方法、立体映像記録媒体、立体映像再生方法、立体映像記録装置、立体映像再生装置 |
JP2009153686A (ja) | 2007-12-26 | 2009-07-16 | Lion Corp | 床用清掃シート |
Cited By (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2011155418A1 (ja) * | 2010-06-10 | 2011-12-15 | ソニー株式会社 | 立体画像データ送信装置、立体画像データ送信方法、立体画像データ受信装置および立体画像データ受信方法 |
WO2012057048A1 (ja) * | 2010-10-29 | 2012-05-03 | ソニー株式会社 | 立体画像データ送信装置、立体画像データ送信方法、立体画像データ受信装置および立体画像データ受信方法 |
WO2012060198A1 (ja) * | 2010-11-02 | 2012-05-10 | ソニー株式会社 | 立体画像データ送信装置、立体画像データ送信方法、立体画像データ受信装置および立体画像データ受信方法 |
JP2012100102A (ja) * | 2010-11-02 | 2012-05-24 | Sony Corp | 立体画像データ送信装置、立体画像データ送信方法、立体画像データ受信装置および立体画像データ受信方法 |
EP2506584A1 (en) * | 2010-11-02 | 2012-10-03 | Sony Corporation | Three-dimensional image data transmitting device, three-dimensional image data transmitting method, three-dimensional image data receiving device, and three-dimensional image data receiving method |
EP2506584A4 (en) * | 2010-11-02 | 2015-03-04 | Sony Corp | DEVICE FOR TRANSMITTING THREE-DIMENSIONAL IMAGE DATA, METHOD FOR TRANSMITTING THREE-DIMENSIONAL IMAGE DATA, DEVICE FOR RECEIVING THREE-DIMENSIONAL IMAGE DATA AND METHOD FOR RECEIVING THREE-DIMENSIONAL IMAGE DATA |
CN102625121A (zh) * | 2011-01-25 | 2012-08-01 | Jvc建伍株式会社 | 3d图像处理装置及3d图像处理方法 |
JP2012205301A (ja) * | 2011-03-25 | 2012-10-22 | Samsung Electronics Co Ltd | シャッタメガネ制御方法及びそれを遂行する表示システム |
CN102202224A (zh) * | 2011-06-22 | 2011-09-28 | 清华大学 | 用于平面视频立体转换的字幕去抖方法及字幕去抖装置 |
CN102202224B (zh) * | 2011-06-22 | 2013-03-27 | 清华大学 | 用于平面视频立体转换的字幕去抖方法及字幕去抖装置 |
Also Published As
Publication number | Publication date |
---|---|
EP2451172A1 (en) | 2012-05-09 |
BRPI1004294A2 (pt) | 2016-03-15 |
US20120257014A1 (en) | 2012-10-11 |
JP2011030192A (ja) | 2011-02-10 |
US8937642B2 (en) | 2015-01-20 |
KR20120036789A (ko) | 2012-04-18 |
TW201116044A (en) | 2011-05-01 |
US20110141235A1 (en) | 2011-06-16 |
JP5407957B2 (ja) | 2014-02-05 |
US8860786B2 (en) | 2014-10-14 |
US20120249735A1 (en) | 2012-10-04 |
CN102210154A (zh) | 2011-10-05 |
EP2451172A4 (en) | 2014-04-02 |
TWI466532B (zh) | 2014-12-21 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP5407957B2 (ja) | 立体画像データ送信装置および立体画像データ受信装置 | |
JP5407968B2 (ja) | 立体画像データ送信装置および立体画像データ受信装置 | |
JP5402715B2 (ja) | 立体画像データ送信装置、立体画像データ送信方法、立体画像データ受信装置および立体画像データ受信方法 | |
JP5446913B2 (ja) | 立体画像データ送信装置および立体画像データ送信方法 | |
JP5531972B2 (ja) | 立体画像データ送信装置、立体画像データ送信方法、立体画像データ受信装置および立体画像データ受信方法 | |
TWI437873B (zh) | Three-dimensional image data transmission device, three-dimensional image data transmission method, three-dimensional image data receiving device and three-dimensional image data receiving method | |
JP5521486B2 (ja) | 立体画像データ送信装置および立体画像データ送信方法 | |
TWI452892B (zh) | Image data transmission device, control method and program | |
WO2011001854A1 (ja) | 立体画像データ送信装置、立体画像データ送信方法、立体画像データ受信装置および立体画像データ受信方法 | |
JP2011166757A (ja) | 送信装置、送信方法および受信装置 | |
JP2011010255A (ja) | 立体画像データ送信方法、立体画像データ受信装置および立体画像データ受信方法 | |
JP2013176141A (ja) | 立体画像データ受信装置および立体画像データ受信方法 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 201080003138.7 Country of ref document: CN |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2010794029 Country of ref document: EP |
|
ENP | Entry into the national phase |
Ref document number: 20117003349 Country of ref document: KR Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 13059045 Country of ref document: US |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 10794029 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
ENP | Entry into the national phase |
Ref document number: PI1004294 Country of ref document: BR Kind code of ref document: A2 Effective date: 20110214 |