WO2012057048A1 - 立体画像データ送信装置、立体画像データ送信方法、立体画像データ受信装置および立体画像データ受信方法 - Google Patents
立体画像データ送信装置、立体画像データ送信方法、立体画像データ受信装置および立体画像データ受信方法 Download PDFInfo
- Publication number
- WO2012057048A1 WO2012057048A1 PCT/JP2011/074368 JP2011074368W WO2012057048A1 WO 2012057048 A1 WO2012057048 A1 WO 2012057048A1 JP 2011074368 W JP2011074368 W JP 2011074368W WO 2012057048 A1 WO2012057048 A1 WO 2012057048A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- data
- information
- image data
- unit
- disparity
- Prior art date
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/025—Systems for the transmission of digital non-picture data, e.g. of text during the active part of a television frame
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/128—Adjusting depth or disparity
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/172—Processing image signals image signals comprising non-image signal components, e.g. headers or format information
- H04N13/183—On-screen display [OSD] information, e.g. subtitles or menus
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/194—Transmission of image signals
Definitions
- the present invention relates to a stereoscopic image data transmission device, a stereoscopic image data transmission method, a stereoscopic image data reception device, and a stereoscopic image data reception method, and more particularly to a stereoscopic image data transmission device that can satisfactorily display superimposition information such as captions. .
- Patent Document 1 proposes a transmission method using a television broadcast radio wave of stereoscopic image data.
- stereoscopic image data including left-eye image data and right-eye image data is transmitted, and stereoscopic image display using binocular parallax is performed in the television receiver.
- FIG. 38 shows the relationship between the display position of the left and right images of an object (object) on the screen and the playback position of the stereoscopic image in stereoscopic image display using binocular parallax.
- the right and left line of sight intersects in front of the screen surface.
- the position is in front of the screen surface.
- DPa represents a horizontal disparity vector related to the object A.
- the right and left lines of sight intersect on the screen surface. It becomes on the surface.
- the left image Lc is shifted to the left side and the right image Rc is shifted to the right side, the right and left lines of sight intersect at the back of the screen surface.
- the playback position is behind the screen.
- DPc represents a horizontal disparity vector related to the object C.
- subtitles when subtitles are superimposed on an image (overlay display), the viewer may feel inconsistency in perspective unless it is displayed in front of the closest object (object) in the perspective. is there.
- other graphics information or text information when superimposed on an image, it is expected that parallax adjustment is performed according to the perspective of each object in the image to maintain the consistency of perspective.
- An object of the present invention is to maintain perspective consistency with each object in an image when displaying superimposition information such as subtitles.
- Another object of the present invention is to enable stable transmission of disparity information without affecting the code system of an existing 8-unit code, for example, in the ARIB system.
- the concept of this invention is An image data output unit for outputting left-eye image data and right-eye image data constituting a stereoscopic image;
- a superimposition information data output unit for outputting superimposition information data to be superimposed on an image based on the left eye image data and the right eye image data;
- a parallax information output unit for outputting parallax information for shifting the superimposition information to be superimposed on the image based on the left-eye image data and the right-eye image data and providing parallax;
- a data transmission unit that transmits a multiplexed data stream including the first data stream including the image data and the second data stream including the superimposition information data and the disparity information;
- the second data stream has a data unit for sending display control information as management information of the superimposition information,
- the parallax information is a stereoscopic image data transmission device inserted in the data unit.
- the image data output unit outputs left-eye image data and right-eye image data constituting a stereoscopic image. Further, the superimposition information data output unit outputs superimposition information data to be superimposed on the image based on the left eye image data and the right eye image data.
- the superimposition information means information such as captions superimposed on the image.
- the parallax information output unit outputs parallax information for adding parallax by shifting the superimposition information to be superimposed on the image based on the left eye image data and the right eye image data.
- the data transmission unit transmits a multiplexed data stream having the first data stream and the second data stream.
- the first data stream includes left eye image data and right eye image data.
- the second data stream includes superimposition information data and disparity information.
- the second data stream has a data unit that transmits display control information as superimposition information management information.
- the disparity information is inserted in this data unit.
- the superimposition information data is ARIB subtitle text data
- the disparity information is inserted into a display control information data unit included in the subtitle management data.
- the disparity information inserted in this data unit is distinguished from other display control information by the type information.
- the disparity information inserted in the data unit is disparity information corresponding to specific superimposition information displayed on the same screen, or disparity information corresponding to a plurality of superimposition information displayed on the same screen.
- disparity information is inserted into the second data stream as superimposition information management information, and the superimposition information and the disparity information are associated with each other.
- appropriate parallax can be given to the superimposition information superimposed on the left eye image and the right eye image using the corresponding parallax information. For this reason, in the display of superimposition information such as captions, the consistency of perspective with each object in the image can be maintained in an optimum state.
- the second data stream has a data unit for sending display control information as superimposition information management information, and disparity information is inserted into this data unit. Therefore, for example, in the ARIB system, it is possible to stably transmit disparity information without affecting the existing 8-unit code system.
- the disparity information is the first disparity information used in common within a predetermined number of frame periods in which the superimposition information is displayed, or within the first disparity information and the predetermined number of frame periods.
- the flag information indicating the presence of the second parallax information may be inserted in the data unit. In this case, it is possible to select whether to transmit only the first parallax information or to transmit the second parallax information. By transmitting the second parallax information, the parallax to be added to the superimposition information can be dynamically changed on the receiving side in conjunction with the change in the image content.
- the second disparity information may be composed of disparity information of the first frame in a predetermined number of frame periods and disparity information of frames at each subsequent update frame interval.
- the update frame interval is arbitrarily set. For example, when the change of the disparity information in the time direction (frame direction) is severe, the update frame interval can be shortened so that the change of the disparity information in the time direction can be more correctly transmitted to the reception side. For example, when the change in the time direction of the disparity information is moderate, it is possible to suppress the data amount of the disparity information by increasing the update frame interval.
- flag information indicating whether or not the disparity information is updated may be added to the second disparity information for each frame at each update frame interval.
- the transmission of the disparity information within the period can be omitted using this flag information, and the data amount of the disparity information can be suppressed. It becomes.
- information for adjusting an update frame interval may be added to the second disparity information for each frame for each update frame period.
- this adjustment information it is possible to arbitrarily adjust the update frame interval in the direction of shortening or lengthening the update frame interval, and the change in the time direction of the disparity information can be more accurately transmitted to the receiving side. It becomes.
- information specifying a frame period may be inserted into a data unit that transmits display control information into which disparity information is inserted.
- the disparity information update frame interval intended on the transmission side can be correctly transmitted to the reception side.
- the video frame period is referred to on the receiving side.
- the present invention for example, in a data unit that transmits display control information into which disparity information is inserted, information indicating a correspondence level with respect to disparity information that is essential when displaying superimposition information may be inserted. . In this case, it is possible to control the correspondence to the parallax information on the reception side with this information.
- the second data stream may include flag information indicating whether or not the management information layer supports extended display of superimposition information.
- the receiving side can easily grasp whether or not the display is compatible with extended display of superimposition information, for example, three-dimensional display.
- a data receiving unit for receiving a multiplexed data stream including the first data stream and the second data stream;
- the first data stream includes left eye image data and right eye image data constituting a stereoscopic image,
- the second data stream shifts superimposition information data to be superimposed on an image based on the left eye image data and the right eye image data, and superimposition information to be superimposed on an image based on the left eye image data and the right eye image data.
- the second data stream has a data unit for sending display control information as management information of the superimposition information, and the disparity information is inserted in the data unit,
- An image data acquisition unit for acquiring the left eye image data and the right eye image data from the first data stream included in the multiplexed data stream;
- a superimposition information data acquisition unit for acquiring the superimposition information data from the second data stream included in the multiplexed data stream;
- a disparity information acquisition unit that acquires the disparity information from the second data stream included in the multiplexed data stream;
- the left eye image data, the right eye image data, the parallax information, and the superimposition information data are used to give parallax to the same superimposition information superimposed on the left eye image and the right eye image, and the superimposition information is
- the stereoscopic image data receiving apparatus further includes an image data processing unit that obtains data of the superimposed left eye image and data of the right eye image on which the superimposition information is superimposed.
- the data reception unit receives a multiplexed data stream including the first data stream and the second data stream.
- the first data stream includes left eye image data and right eye image data constituting a stereoscopic image.
- the superimposition information data superimposed on the image based on the left eye image data and the right eye image data and the superimposition information superimposed on the image based on the left eye image data and the right eye image data are shifted.
- parallax information for adding parallax.
- the second data stream has a data unit for sending display control information as superimposition information management information, and the disparity information is inserted into this data unit.
- the superimposition information data is ARIB subtitle text data
- the disparity information is inserted into a display control information data unit included in the subtitle management data.
- the disparity information inserted in this data unit is distinguished from other display control information by the type information.
- the image data acquisition unit acquires left-eye image data and right-eye image data from the first data stream included in the multiplexed data stream. Further, the superimposition information data acquisition unit acquires the superimposition information data from the second data stream included in the multiplexed data stream. Also, the disparity information acquisition unit acquires disparity information from the second data stream included in the multiplexed data stream.
- the image data processing unit uses the left eye image data and the right eye image data, the superimposition information data, and the parallax information, and gives the parallax to the same superimposition information to be superimposed on the left eye image and the right eye image.
- left-eye image data on which superimposition information is superimposed and right-eye image data on which superimposition information is superimposed are obtained.
- disparity information is inserted into the second data stream as superimposition information management information, and the superimposition information and the disparity information are associated with each other. Therefore, the image data processing unit can give appropriate parallax to the superimposition information superimposed on the left eye image and the right eye image using the corresponding parallax information. Therefore, in the display of superimposition information such as subtitles, it is possible to maintain the perspective consistency with each object in the image in an optimum state.
- the disparity information inserted in the data unit is disparity information that is sequentially updated within a predetermined number of frame periods in which superimposition information is displayed, and is the first frame of the predetermined number of frame periods. And the disparity information of the frame at each subsequent update frame interval.
- the parallax to be added to the superimposition information can be dynamically changed in conjunction with the change in the image content.
- the image data processing unit performs an interpolation process on disparity information of a plurality of frames constituting disparity information that is sequentially updated within a predetermined number of frame periods. You may make it produce
- the interpolation process may be a linear interpolation process, but may be accompanied by a low-pass filter process in the time direction (frame direction), for example.
- the disparity information is inserted as the management information of the superimposition information into the second data stream including the superimposition information, and the superimposition information and the disparity information are associated with each other. Therefore, on the reception side, appropriate parallax can be given to the superimposition information superimposed on the left eye image and the right eye image using the corresponding parallax information. Therefore, in the display of superimposition information such as subtitles, it is possible to maintain the perspective consistency with each object in the image in an optimum state.
- the second data stream has a data unit for sending display control information as superimposition information management information, and disparity information is inserted into this data unit. Therefore, for example, in the ARIB system, it is possible to stably transmit disparity information without affecting the existing 8-unit code system.
- FIG. 1 It is a block diagram which shows the structural example of the set top box which comprises a stereo image display system. It is a block diagram which shows the structural example of the bit stream process part which comprises a set top box.
- FIG. It is a block diagram which shows the structural example of the television receiver which comprises a stereo image display system. It is a block diagram which shows the other structural example of a stereo image display system.
- stereoscopic image display using binocular parallax it is a figure for demonstrating the relationship between the display position of the left-right image of the object on a screen, and the reproduction
- FIG. 1 shows a configuration example of a stereoscopic image display system 10 as an embodiment.
- the stereoscopic image display system 10 includes a broadcasting station 100, a set top box (STB) 200, and a television receiver (TV: Television) 300.
- STB set top box
- TV television receiver
- the set top box 200 and the television receiver 300 are connected via an HDMI (High Definition Multimedia Interface) cable 400.
- the set top box 200 is provided with an HDMI terminal 202.
- the television receiver 300 is provided with an HDMI terminal 302.
- One end of the HDMI cable 400 is connected to the HDMI terminal 202 of the set top box 200, and the other end of the HDMI cable 400 is connected to the HDMI terminal 302 of the television receiver 300.
- the broadcasting station 100 transmits the bit stream data BSD on a broadcast wave.
- the broadcast station 100 includes a transmission data generation unit 110 that generates bit stream data BSD.
- the bit stream data BSD includes image data, audio data, superimposition information data, disparity information (disparity vector), and the like.
- the image data (hereinafter referred to as “stereoscopic image data” as appropriate) includes left-eye image data and right-eye image data constituting a stereoscopic image.
- the superimposition information is graphics information, text information, or the like, but is a caption in this embodiment.
- FIG. 2 shows a configuration example of the transmission data generation unit 110 in the broadcast station 100.
- the transmission data generation unit 110 transmits disparity information (disparity vector) with a data structure that can be easily linked to ARIB (Association of Radio Industries and Businesses), which is one of existing broadcasting standards.
- the transmission data generation unit 110 includes a data extraction unit (archive unit) 130, a parallax information generation unit 131, a video encoder 113, an audio encoder 117, a caption generation unit 132, a caption encoder 133, and a multiplexer 122. is doing.
- a data recording medium 130a is detachably attached to the data extraction unit 130, for example.
- audio data and parallax information are recorded in association with stereoscopic image data including left eye image data and right eye image data.
- the data extraction unit 130 extracts and outputs stereoscopic image data, audio data, parallax information, and the like from the data recording medium 130a.
- the data recording medium 130a is a disk-shaped recording medium, a semiconductor memory, or the like.
- the stereoscopic image data recorded on the data recording medium 130a is stereoscopic image data of a predetermined transmission method.
- An example of a transmission method of stereoscopic image data (3D image data) will be described.
- the following first to third transmission methods are listed, but other transmission methods may be used.
- the case where the image data of the left eye (L) and the right eye (R) is image data of a predetermined resolution, for example, a 1920 ⁇ 1080 pixel format. Let's take an example.
- the first transmission method is a top-and-bottom method. As shown in FIG. 4A, in the first half of the vertical direction, the data of each line of the left eye image data is transmitted, and the vertical direction In the latter half of the method, the data of each line of the left eye image data is transmitted. In this case, since the lines of the left eye image data and the right eye image data are thinned out to 1 ⁇ 2, the vertical resolution is halved with respect to the original signal.
- the second transmission method is a side-by-side (Side By Side) method.
- pixel data of the left eye image data is transmitted, and in the second half in the horizontal direction.
- the pixel data of the right eye image data is transmitted.
- the pixel data in the horizontal direction is thinned out to 1/2.
- the horizontal resolution is halved with respect to the original signal.
- the third transmission method is a frame sequential method, in which left-eye image data and right-eye image data are sequentially switched for each frame and transmitted as shown in FIG. 4 (c).
- This frame sequential method may be referred to as a full frame method or a backward compatible method.
- the disparity information recorded in the data recording medium 130a is, for example, a disparity vector for each pixel (pixel) constituting the image.
- a detection example of a disparity vector will be described.
- the parallax vector of the right eye image with respect to the left eye image will be described.
- the left eye image is a detected image
- the right eye image is a reference image.
- the disparity vectors at the positions (xi, yi) and (xj, yj) are detected.
- a case where a disparity vector at the position of (xi, yi) is detected will be described as an example.
- a 4 ⁇ 4, 8 ⁇ 8, or 16 ⁇ 16 pixel block (parallax detection block) Bi is set in the left eye image with the pixel at the position (xi, yi) at the upper left. Then, a pixel block matching the pixel block Bi is searched in the right eye image.
- a search range centered on the position of (xi, yi) is set in the right eye image, and each pixel in the search range is sequentially set as a pixel of interest, for example, 4 ⁇ 4 similar to the above-described pixel block Bi. 8 ⁇ 8 or 16 ⁇ 16 comparison blocks are sequentially set.
- the sum of the absolute differences for each corresponding pixel is obtained.
- the pixel block Bi when the pixel value of the pixel block Bi is L (x, y) and the pixel value of the comparison block is R (x, y), the pixel block Bi, a certain comparison block, The sum of absolute differences between the two is represented by ⁇
- n pixels are included in the search range set in the right eye image, n total sums S1 to Sn are finally obtained, and the minimum sum Smin is selected. Then, the position of the upper left pixel (xi ′, yi ′) is obtained from the comparison block from which the sum Smin is obtained. Thus, the disparity vector at the position (xi, yi) is detected as (xi′ ⁇ xi, yi′ ⁇ yi).
- the left eye image has the pixel at the position (xj, yj) at the upper left, for example, 4 ⁇ 4, 8 ⁇ 8, or 16 ⁇ Sixteen pixel blocks Bj are set and detected in the same process.
- the caption generation unit 132 generates caption data (ARIB method caption text data).
- the caption encoder 133 generates a caption data stream (caption elementary stream) including the caption data generated by the caption generator 132.
- FIG. 7A shows a configuration example of a caption data stream.
- three caption units (captions) “1st Caption Unit”, “2nd Caption Unit”, and “3rd Caption Unit” are displayed on the same screen. Is shown.
- the caption data of each caption unit is inserted into the caption data stream as the caption text data (caption code) of the caption text data group.
- setting data such as the display area of each caption unit is inserted into the caption data stream as data of a caption management data group.
- the caption unit display areas of "1st Caption Unit", “2nd Caption Unit” and “3rd Caption Unit” are indicated by (x1, y1), (x2, y2), and (x3, y3), respectively. .
- the parallax information creation unit 131 has a viewer function.
- the disparity information creating unit 131 performs a downsizing process on the disparity information output from the data extracting unit 130, that is, the disparity vector for each pixel (pixel), and generates a disparity vector belonging to a predetermined region.
- FIG. 8 shows an example of data in the relative depth direction given as the luminance value of each pixel (pixel).
- the data in the relative depth direction can be handled as a disparity vector for each pixel by a predetermined conversion.
- the luminance value of the person portion is high. This means that the value of the parallax vector of the person portion is large, and therefore, in stereoscopic image display, this means that the person portion is perceived as being raised.
- the luminance value of the background portion is low. This means that the value of the parallax vector in the background portion is small, and therefore, in stereoscopic image display, this means that the background portion is perceived as a sunken state.
- FIG. 9 shows an example of a disparity vector for each block.
- the block corresponds to an upper layer of pixels (picture elements) located at the lowermost layer.
- This block is configured by dividing an image (picture) region into a predetermined size in the horizontal direction and the vertical direction.
- the disparity vector of each block is obtained, for example, by selecting the disparity vector having the largest value from the disparity vectors of all pixels (pixels) existing in the block.
- the disparity vector of each block is indicated by an arrow, and the length of the arrow corresponds to the magnitude of the disparity vector.
- FIG. 10 shows an example of the downsizing process performed by the parallax information creating unit 131.
- the disparity information creating unit 131 obtains a disparity vector for each block using a disparity vector for each pixel (pixel) as illustrated in FIG.
- a block corresponds to an upper layer of pixels located at the lowest layer, and is configured by dividing an image (picture) region into a predetermined size in the horizontal direction and the vertical direction.
- the disparity vector of each block is obtained, for example, by selecting the disparity vector having the largest value from the disparity vectors of all the pixels (pixels) existing in the block.
- the disparity information creating unit 131 obtains a disparity vector for each group (Group Of ⁇ ⁇ ⁇ ⁇ Block) using the disparity vector for each block, as illustrated in FIG.
- a group is an upper layer of a block, and is obtained by grouping a plurality of adjacent blocks together.
- each group is composed of four blocks bounded by a broken line frame.
- the disparity vector of each group is obtained, for example, by selecting the disparity vector having the largest value from the disparity vectors of all blocks in the group.
- the disparity information creating unit 131 obtains a disparity vector for each partition (Partition) using the disparity vector for each group as shown in FIG.
- the partition is an upper layer of the group and is obtained by grouping a plurality of adjacent groups together.
- each partition is composed of two groups bounded by a broken line frame.
- the disparity vector of each partition is obtained, for example, by selecting the disparity vector having the largest value from the disparity vectors of all groups in the partition.
- the disparity information creating unit 131 obtains the disparity vector of the entire picture (entire image) located in the highest layer using the disparity vector for each partition.
- the whole picture includes four partitions that are bounded by a broken line frame.
- the disparity vector for the entire picture is obtained, for example, by selecting the disparity vector having the largest value from the disparity vectors for all partitions included in the entire picture.
- the disparity information creating unit 131 performs the downsizing process on the disparity vector for each pixel (pixel) located in the lowest layer, and the disparity vectors of the respective regions in each layer of the block, the group, the partition, and the entire picture Can be requested.
- the disparity vectors of four layers of blocks, groups, partitions, and entire pictures are obtained.
- the number of hierarchies, how to cut areas in each hierarchy, and the number of areas are not limited to this.
- the disparity information creating unit 131 creates disparity vectors corresponding to a predetermined number of caption units (captions) displayed on the same screen by the above-described downsizing process. In this case, the disparity information creating unit 131 creates a disparity vector (individual disparity vector) for each caption unit, or creates a disparity vector common to each caption unit (common disparity vector). This selection depends on, for example, user settings.
- the disparity information creating unit 131 When creating the individual disparity vector, the disparity information creating unit 131 obtains a disparity vector belonging to the display area by the above-described downsizing process based on the display area of each caption unit. Further, when creating the common disparity vector, the disparity information creating unit 131 obtains the disparity vector of the entire picture (entire image) by the above-described downsizing process (see FIG. 10D). Note that when creating the common parallax vector, the parallax information creation unit 131 may obtain a parallax vector belonging to the display area of each caption unit and select the parallax vector having the largest value.
- the caption encoder 133 includes the disparity vector (disparity information) created by the disparity information creating unit 131 as described above in the caption data stream.
- the caption data of each caption unit displayed on the same screen is inserted into the caption data stream as the caption text data (caption code) of the caption text data group.
- a disparity vector is inserted into the caption data stream as caption management data (control code) of the caption management data group.
- the disparity information creating unit 131 creates individual disparity vectors corresponding to each caption unit, as shown in FIG. “Disparity 1” is an individual disparity vector corresponding to “1st Caption Unit”. “Disparity 2” is a disparity vector corresponding to “2nd Caption Unit”. “Disparity 3” is an individual disparity vector corresponding to “3rd Caption Unit”.
- FIG. 11A shows a configuration example of a caption data stream (PES stream) generated by the caption encoder 133.
- the caption text information of each caption unit and the extended display control information (data unit ID) associated with each caption text information are inserted into the PES stream of the caption text data group. Further, extended display control information (disparity information) corresponding to the caption text information of each caption unit is inserted into the PES stream of the caption management data group.
- the extended display control information (data unit ID) of the caption text data group is required to associate each extended display control information (disparity information) of the caption management data group with each caption text information of the caption text data group.
- the disparity information as each extended display control information of the caption management data group is an individual disparity vector of the corresponding caption unit.
- the setting data such as the display area of each caption unit is inserted as subtitle management data (control code) into the PES stream of the subtitle management data group, although not shown.
- the caption unit display areas of "1st Caption Unit”, “2nd Caption Unit” and “3rd Caption Unit” are indicated by (x1, y1), (x2, y2), and (x3, y3), respectively. .
- FIG. 11 (c) shows a first view (1st view) on which each caption unit (caption) is superimposed, for example, a right eye image.
- FIG. 11D shows a second view (1st View) in which each caption unit is superimposed, for example, a left eye image.
- the individual disparity vector corresponding to each caption unit is used to give disparity between each caption unit superimposed on the right eye image and each caption unit superimposed on the left eye image, for example. Used for.
- disparity information creating unit 131 creates a common disparity vector.
- three caption units (captions) “1st ⁇ Caption Unit ”,“ 2nd Caption Unit ”, and“ 3rd Caption Unit ” are displayed on the same screen.
- the disparity information creating unit 131 creates a common disparity vector “Disparity” common to the caption units, as shown in FIG.
- FIG. 12A shows a configuration example of a caption data stream (PES stream) generated by the caption encoder 133.
- the caption sentence information of each caption unit is inserted into the PES stream of the caption sentence data group.
- extended display control information (disparity information) corresponding to the caption text information of each caption unit is inserted into the PES stream of the caption management data group.
- the disparity information as the extended display control information of the caption management data group is a common disparity vector of each caption unit.
- the setting data such as the display area of each caption unit is inserted as subtitle management data (control code) into the PES stream of the subtitle management data group, although not shown.
- the caption unit display areas of "1st Caption Unit”, “2nd Caption Unit” and “3rd Caption Unit” are indicated by (x1, y1), (x2, y2), and (x3, y3), respectively. .
- FIG. 12 (c) shows a first view (1st view) on which each caption unit (caption) is superimposed, for example, a right eye image.
- FIG. 12D shows a second view (1st View) in which each caption unit is superimposed, for example, a left eye image.
- the common disparity vector common to each caption unit is used to give disparity between each caption unit superimposed on the right eye image and each caption unit superimposed on the left eye image, for example. Used for.
- FIGS. 13A and 13B show a case where the positions of both the caption unit superimposed on the first view and the second view are shifted.
- the shift value (offset value) D [i] of each caption unit in the first view and the second view from the value “disparity [i]” of the disparity vector “Disparity” corresponding to each caption unit.
- disparity [i] is an even number
- the position of each caption unit to be superimposed on the first view is shifted to the left by “disparity [i] / 2”.
- the position of each caption unit to be superimposed on the second view is shifted to the right by (disparity [i] / 2).
- disparity (i) is an odd number
- “D [i] is obtained in the second view.
- i] (disparity [i] -1) / 2 ".
- the position of each caption unit to be superimposed on the first view is shifted to the left by “(disparity [i] +1) / 2”.
- the position of each caption unit to be superimposed on the second view is shifted to the right by “(disparity [i] ⁇ 1) / 2”.
- FIG. 14 shows a packet structure of a caption code.
- Data_group_id indicates data group identification, and here indicates a caption sentence data group.
- Data_group_size indicates the number of bytes of the subsequent data group data.
- this data group data is caption text data (caption_data).
- One or more data units are arranged in the caption text data. Each data unit is separated by a data unit separation code (unit_separator).
- a caption code is arranged as data unit data (data_unit_data) in each data unit.
- FIG. 15 shows a packet structure of control codes included in the PES stream of the caption management data group.
- Data_group_size indicates the number of bytes of subsequent data group data. In the case of a caption management data group, this data group data is caption management data (caption_management_data).
- This subtitle management data has one or more data units. Each data unit is separated by a data unit separation code (unit_separator). A control code is arranged as data unit data (data_unit_data) in each data unit. In this embodiment, the value of the disparity vector is given as an 8-unit code.
- FIG. 16 shows the structure of a data group in a caption data stream (PES stream).
- a 6-bit field of “data_group_id” indicates data group identification, and identifies the types of caption management data and caption text data.
- a 16-bit field of “data_group_size” indicates the number of bytes of subsequent data group data in this data group field.
- Data group data is stored in “data_group_data_byte”.
- “CRC — 16” is a 16-bit cyclic redundancy check code. The encoding section of the CRC code is from the beginning of “data_group_id” to the end of “data_group_data_byte”.
- caption management data group “data_group_data_byte” in the data group structure of FIG. 16 is caption management data (caption_management_data).
- caption data “data_group_data_byte” in the data group structure of FIG. 16 is caption data (caption_data).
- FIG. 17 schematically shows the structure of caption management data.
- “Advanced_rendering_version” is 1-bit flag information newly defined in this embodiment and indicating whether or not the extended display of subtitles is supported. On the receiving side, based on the flag information arranged in the management information layer as described above, it is possible to easily grasp whether or not the extended display of subtitles is supported.
- a 24-bit field of “data_unit_loop_length” indicates the number of bytes of the subsequent data unit in the caption management data field. In “data_unit”, a data unit to be transmitted in this caption management data field is stored.
- FIG. 18 schematically shows the structure of caption data. The structure is the same as the caption management data described above except that there is no “advanced_rendering_version”.
- FIG. 19 shows the structure (Syntax) of the data unit (data_unit) included in the caption data stream.
- An 8-bit field of “unit_separator” indicates a data unit separation code and is “0x1F”.
- the 8-bit field of “data_unit_parameter” is a data unit parameter that identifies the type of data unit.
- FIG. 20 shows data unit types, data unit parameters, and functions.
- the data unit parameter indicating the data unit of the main text is “0x20”.
- the data unit parameter indicating the geometric data unit is “0x28”.
- a data unit parameter indicating a bitmap data unit is set to “0x35”.
- a data unit for extended display control for storing display control information is newly defined, and a data unit parameter indicating this data unit is set to, for example, “0x4F”.
- the 24-bit field “data_unit_size” indicates the number of bytes of the subsequent data unit data in this data unit field.
- Data unit data is stored in “data_unit_data_byte”.
- FIG. 21 shows the structure (Syntax) of the data unit (data_unit) for extended display control.
- the data unit parameter is “0x4F”
- display control information is stored in “Advanced_Rendering_Control” as “data_unit_data_byte”.
- FIG. 22 shows the structure (Syntax) of “Advanced_Rendering_Control”.
- FIG. 22 shows a structure in the case of inserting stereo video parallax information as display control information. That is, FIG. 22 shows the structure of “Advanced_Rendering_Control” in the data unit (data_unit) of extended display control included in the caption management data group.
- the 8-bit field of “start_code” indicates the beginning of “Advanced_Rendering_Control”.
- a 16-bit field of “data_unit_id” indicates a data unit ID.
- the 16-bit field of “data_length” indicates the number of subsequent data bytes in this advanced rendering control field.
- the 8-bit field of “Advanced_rendering_type” is an advanced rendering type that specifies the type of display control information.
- the data unit parameter is, for example, “0x01”, indicating that the display control information is “stereo video parallax information”. Disparity information is stored in “disparity_information”.
- FIG. 23 also shows the structure (Syntax) of “Advanced_Rendering_Control”.
- FIG. 23 shows a structure when a data unit ID is inserted as display control information. That is, FIG. 23 illustrates the structure of “Advanced_Rendering_Control” in the data unit (data_unit) of extended display control included in the caption data group.
- the 8-bit field of “start_code” indicates the beginning of “Advanced_Rendering_Control”.
- a 16-bit field of “data_unit_id” indicates a data unit ID.
- the 16-bit field of “data_length” indicates the number of subsequent data bytes in this advanced rendering control field.
- the 8-bit field of “Advanced_rendering_type” is an advanced rendering type that specifies the type of display control information.
- the data unit parameter is “0x00”, for example, and the display control information is “data unit ID”.
- 26 and 27 show the main data definition contents in the structure of “Advanced_Rendering_Control” described above, and in the structure of “disparity_information” shown in FIGS. 24 and 25 described later.
- Interval_PTS [32..0] specifies the frame period (interval of one frame) in the update frame interval of disparity information (disparity) in units of 90 KHz. That is, “interval_PTS [32..0]” represents a value obtained by measuring the frame period with a clock of 90 KHz in 33-bit length.
- disparity information by specifying the frame period with “interval_PTS [32..0]”, it is possible to correctly convey the disparity information update frame interval intended on the transmission side to the reception side.
- this information is not added, for example, the video frame period is referred to on the receiving side.
- “Rendering_level” indicates a level corresponding to disparity information (disparity) that is essential on the reception side (decoder side) when displaying captions. “00” indicates that subtitle three-dimensional display using parallax information is optional. “01” indicates that subtitle three-dimensional display using disparity information (default_disparity) commonly used within the subtitle display period is essential. “10” indicates that subtitle three-dimensional display using disparity information (disparity_update) sequentially updated within the subtitle display period is essential.
- Temporal_extension_flag is 1-bit flag information indicating the presence / absence of disparity information (disparity_update) that is sequentially updated within the caption display period. In this case, “1” indicates that it exists, and “0” indicates that it does not exist.
- the 8-bit field “default_disparity” indicates default disparity information. This disparity information is disparity information when updating is not performed, that is, disparity information that is commonly used within a caption display period.
- the disparity information includes “disparity_temporal_extension ()”.
- BSP Base : Segment Period
- FIG. 28 illustrates an example of updating disparity information for each base segment period (BSP).
- the base segment period means an update frame interval.
- the disparity information sequentially updated within the caption display period includes the disparity information of the first frame in the caption display period, and the disparity information of the frame for each subsequent base segment period (updated frame interval). It is made up of.
- the 2-bit field of“ temporal_division_size ” indicates the number of frames included in the base segment period (update frame interval). “00” indicates 16 frames. “01” indicates 25 frames. “10” indicates 30 frames. Further, “11” indicates 32 frames.
- “Temporal_division_count” indicates the number of base segments included in the caption display period.
- “Disparity_curve_no_update_flag” is 1-bit flag information indicating whether or not disparity information is updated. “1” indicates that the disparity information is not updated at the edge of the corresponding base segment, that is, skipping, and “0” indicates that the disparity information is updated at the edge of the corresponding base segment.
- FIG. 29 shows an example of updating disparity information for each base segment period (BSP).
- BSP base segment period
- the disparity information includes “shifting_interval_counts” of the corresponding base segment. Further, when the disparity information is updated when “disparity_curve_no_update_flag” is “0”, the disparity information includes “disparity_update”.
- a 6-bit field of “shifting_interval_counts” indicates a draw factor for adjusting the base segment period (update frame interval), that is, the number of subtracted frames.
- the base segment period is adjusted by the draw factor with respect to the update timing of the disparity information at time points C to F.
- the presence of this adjustment information makes it possible to adjust the base segment period (update frame interval), and more accurately convey changes in the time direction (frame direction) of disparity information to the receiving side. .
- the adjustment in the direction of increasing by the number of added frames may be considered.
- bidirectional adjustment is possible by setting a 6-bit field of “shifting_interval_counts” as a signed integer.
- interval_PTS [32..0] is added to the structure (Syntax) of “disparity_information” shown in FIG.
- a structure (Syntax) of “disparity_information” without “interval_PTS [32..0]” may be considered.
- the structure of “disparity_information” is as shown in FIG.
- the video encoder 113 performs encoding such as MPEG4-AVC, MPEG2, VC-1, etc. on the stereoscopic image data supplied from the data extraction unit 130 to generate a video elementary stream.
- the audio encoder 117 performs encoding such as MPEG-2Audio AAC on the audio data supplied from the data extraction unit 130 to generate an audio elementary stream.
- the multiplexer 122 multiplexes the elementary streams output from the video encoder 113, the audio encoder 117, and the caption encoder 133.
- the multiplexer 122 outputs bit stream data (transport stream) BSD as transmission data (multiplexed data stream).
- the operation of the transmission data generation unit 110 shown in FIG. 2 will be briefly described.
- the stereoscopic image data output from the data extraction unit 130 is supplied to the video encoder 113.
- the stereoscopic image data is encoded by MPEG4-AVC, MPEG2, VC-1, or the like, and a video elementary stream including the encoded video data is generated. This video elementary stream is supplied to the multiplexer 122.
- the caption generation unit 132 generates ARIB format caption data. This caption data is supplied to the caption encoder 133.
- the caption encoder 133 generates a caption elementary stream (caption data stream) including the caption data generated by the caption generator 132. This subtitle elementary stream is supplied to the multiplexer 122.
- the disparity vector for each pixel (pixel) output from the data extracting unit 130 is supplied to the disparity information creating unit 131.
- disparity vectors horizontal disparity vectors
- corresponding to a predetermined number of caption units (captions) displayed on the same screen are created by downsizing processing.
- the disparity information creating unit 131 creates a disparity vector (individual disparity vector) for each caption unit, or a disparity vector common to all caption units (common disparity vector).
- the disparity vector created by the disparity information creating unit 131 is supplied to the caption encoder 133.
- the disparity vector is included in the caption data stream (see FIGS. 11 to 12).
- the caption data of each caption unit displayed on the same screen is inserted into the caption data stream as the caption text data (caption code) of the caption text data group.
- disparity vectors are inserted into the caption data stream as caption management data (control codes) of the caption management data group.
- the disparity vector is inserted into an extended display control data unit that transmits newly defined display control information (see FIGS. 21, 23 to 25, and 30).
- the audio data output from the data extraction unit 130 is supplied to the audio encoder 117.
- the audio encoder 117 performs encoding such as MPEG-2Audio AAC on the audio data, and generates an audio elementary stream including the encoded audio data. This audio elementary stream is supplied to the multiplexer 122.
- the elementary stream from the video encoder 113, the audio encoder 117, and the caption encoder 133 is supplied to the multiplexer 122.
- the multiplexer 122 the elementary streams supplied from the encoders are packetized and multiplexed to obtain bit stream data (transport stream) BSD as transmission data.
- the bit stream data BSD output from the multiplexer 122 is a multiplexed data stream including a video data stream and a caption data stream.
- the video data stream includes stereoscopic image data.
- the caption data stream includes ARIB caption data (caption unit) and disparity vectors (disparity information).
- disparity vectors are inserted into the caption data stream as the caption data management information, and the caption data and the disparity vectors are associated with each other. Therefore, on the reception side (set top box 200), appropriate parallax can be given to caption units (captions) superimposed on the left-eye image and right-eye image using the corresponding disparity vector (disparity information). Therefore, in the display of caption units (captions), perspective consistency with each object in the image can be maintained in an optimum state.
- the above-described caption data stream has a newly defined extended display control data unit for sending display control information.
- the disparity vector (disparity information) is inserted into the data unit for the extended display control. Therefore, in the ARIB system, disparity information can be stably transmitted without affecting the existing 8-unit code system. That is, backward compatibility can be guaranteed.
- disparity information can be extended and defined in the code table of 8-unit code.
- the configuration in the 8-unit code table becomes complicated, and the legacy problem of the mounted equipment adversely affects the stable operation of the existing code system. There is a risk.
- disparity information can also be sent as a metadata stream independent of the caption data stream.
- the mechanism for managing the synchronization relationship may be complicated.
- Data positioned between the above 8 unit code and the external data stream is located inside the caption data stream and exists in the caption management data, but is positioned outside the 8 unit code. Is a unit. Therefore, according to the present invention, the parameter type of the caption management data is newly extended, a new data unit for extended display control for sending display control information is newly defined, and disparity information is transmitted by this data unit.
- disparity information (see “default_disparity” in FIG. 24) that is commonly used in the caption display period is inserted into the newly defined extended display control data unit. Is done. Also, disparity information (see “disparity_update” in FIG. 25) that is sequentially updated within the caption display period can be inserted into this data unit. Then, flag information indicating the presence of disparity information that is sequentially updated within the caption display period is inserted into the data unit for extended display control (see “temporal_extension_flag” in FIG. 24).
- the disparity information sequentially updated within the caption display period inserted into the extended display control data unit is basically based on the base segment period (BSP).
- the parallax information is used (see FIG. 28). That is, this disparity information includes disparity information of the first frame in the caption display period and disparity information of the frame for each subsequent base segment period.
- the base segment period update frame interval
- can be arbitrarily set see “temporal_division_size” in FIG. 25).
- the update frame interval is shortened so that the change of the disparity information in the time direction is more correctly performed on the receiving side (set top box 200). It becomes possible to convey. For example, when the change in the time direction of the disparity information is moderate, it is possible to suppress the data amount of the disparity information by increasing the update frame interval.
- This flag information is flag information indicating whether or not disparity information is updated (see “disparity_curve_no_update_flag” in FIG. 25). This flag information is added for each frame for each base segment period (update frame interval). In this case, when a period in which the change in the time direction of the disparity information is similar continues, transmission of the disparity information within the period can be omitted using this flag information (a base marked with “skip” in FIG. 29). It is possible to suppress the data amount of the disparity information) (see the edge of the segment).
- This adjustment information is adjustment information for adjusting the base segment period (update frame interval) (see “shifting_interval_counts” in FIG. 25).
- this adjustment information it is possible to arbitrarily adjust the base segment period in a direction to shorten or lengthen the base segment period. As a result, it is possible to more accurately notify the receiving side (set top box 200) the change in the time direction (frame direction) of the disparity information (see adjustment by “Draw factor” in FIG. 29).
- the transmission data generation unit 110 shown in FIG. 2 information specifying the frame period is inserted into the data unit of the extended display control (see “interval_PTS [32..0]” in FIG. 24).
- the base segment period (update frame interval) of the disparity information intended on the transmission side (broadcast station 100) can be correctly transmitted to the reception side (set top box 200).
- the video frame period is referred to on the receiving side.
- the extended display control data unit includes information indicating a disparity information (disparity) correspondence level that is essential on the reception side (decoder side) during caption display. (See “rendering_level” in FIG. 24). In this case, it is possible to control the correspondence to the parallax information on the reception side with this information.
- 1-bit flag information indicating whether or not caption extended display is supported is inserted in the caption management data layer (“advanced_rendering_version” in FIG. 17). reference).
- the flag information arranged in the management information layer as described above it is possible to easily grasp whether or not the extended display of subtitles is supported.
- the set-top box 200 receives bit stream data (transport stream) BSD transmitted from the broadcasting station 100 on broadcast waves.
- the bit stream data BSD includes stereoscopic image data and audio data including left eye image data and right eye image data.
- the bit stream data BSD includes caption unit caption data, and further, a disparity vector (disparity information) for giving disparity to the caption unit.
- the set top box 200 has a bit stream processing unit 201.
- the bit stream processing unit 201 extracts stereoscopic image data, audio data, caption unit caption data, disparity vectors, and the like from the bit stream data BSD.
- the bit stream processing unit 201 generates left eye image data and right eye image data on which captions are superimposed using stereoscopic image data, caption unit caption data, and the like.
- left-eye caption data and right-eye caption data to be superimposed on the left-eye image and right-eye image are generated based on the disparity vector and caption unit caption data.
- the left-eye caption and the right-eye caption are the same caption.
- the superimposed position in the image is shifted in the horizontal direction by, for example, the parallax vector between the left-eye caption and the right-eye caption.
- parallax is given between the left-eye caption and the right-eye caption, and the recognition position of the caption is set in front of the image.
- FIG. 31 (a) shows a display example of caption units (captions) on an image.
- captions are superimposed on an image composed of a background and a foreground object.
- FIG. 31B shows the perspective of the background, the foreground object, and the caption, and indicates that the caption is recognized at the forefront.
- FIG. 32A shows a display example of caption units (captions) on the same image as FIG. 31A.
- FIG. 32B shows a left-eye caption LGI superimposed on the left-eye image and a right-eye caption RGI superimposed on the right-eye image.
- FIG. 32C shows that parallax is given between the left-eye caption LGI and the right-eye caption RGI because the caption is recognized at the forefront.
- FIG. 33 shows a configuration example of the set top box 200.
- the set top box 200 includes a bit stream processing unit 201, an HDMI terminal 202, an antenna terminal 203, a digital tuner 204, a video signal processing circuit 205, an HDMI transmission unit 206, and an audio signal processing circuit 207. ing.
- the set top box 200 includes a CPU 211, a flash ROM 212, a DRAM 213, an internal bus 214, a remote control receiving unit 215, and a remote control transmitter 216.
- the antenna terminal 203 is a terminal for inputting a television broadcast signal received by a receiving antenna (not shown).
- the digital tuner 204 processes the television broadcast signal input to the antenna terminal 203 and outputs predetermined bit stream data (transport stream) BSD corresponding to the user's selected channel.
- the bitstream processing unit 201 extracts stereoscopic image data, audio data, caption unit caption data, disparity vectors, and the like from the bitstream data BSD.
- the bit stream processing unit 201 synthesizes left-eye caption data and right-eye caption data with stereoscopic image data, and generates and outputs display stereoscopic image data.
- the bit stream processing unit 201 outputs audio data.
- the detailed configuration of the bit stream processing unit 201 will be described later.
- the video signal processing circuit 205 performs image quality adjustment processing on the stereoscopic image data output from the bit stream processing unit 201 as necessary, and supplies the processed stereoscopic image data to the HDMI transmission unit 206.
- the audio signal processing circuit 207 performs sound quality adjustment processing or the like on the audio data output from the bit stream processing unit 201 as necessary, and supplies the processed audio data to the HDMI transmission unit 206.
- the HDMI transmitting unit 206 transmits baseband image (video) and audio data from the HDMI terminal 202 by communication conforming to HDMI. In this case, since transmission is performed using the HDMI TMDS channel, image and audio data are packed and output from the HDMI transmission unit 206 to the HDMI terminal 202.
- the CPU 211 controls the operation of each part of the set top box 200.
- the flash ROM 212 stores control software and data.
- the DRAM 213 constitutes a work area for the CPU 211.
- the CPU 211 develops software and data read from the flash ROM 212 on the DRAM 213 to activate the software, and controls each part of the set top box 200.
- the remote control receiving unit 215 receives the remote control signal (remote control code) transmitted from the remote control transmitter 216 and supplies it to the CPU 211.
- the CPU 211 controls each part of the set top box 200 based on the remote control code.
- the CPU 211, flash ROM 212 and DRAM 213 are connected to the internal bus 214.
- a television broadcast signal input to the antenna terminal 203 is supplied to the digital tuner 204.
- the digital tuner 204 processes the television broadcast signal and outputs predetermined bit stream data (transport stream) BSD corresponding to the user's selected channel.
- the bit stream data BSD output from the digital tuner 204 is supplied to the bit stream processing unit 201.
- the bit stream processing unit 201 extracts stereoscopic image data, audio data, caption unit caption data, disparity vectors, and the like from the bit stream data BSD.
- the left-eye caption data and the right-eye caption data are combined with the stereoscopic image data to generate display stereoscopic image data.
- the bit stream processing unit 201 assigns parallax between the left eye caption and the right eye caption based on the disparity vector.
- the display stereoscopic image data generated by the bit stream processing unit 201 is supplied to the video signal processing circuit 205.
- image quality adjustment processing or the like is performed on the display stereoscopic image data as necessary.
- the processed display stereoscopic image data output from the video signal processing circuit 205 is supplied to the HDMI transmission unit 206.
- the audio data obtained by the bit stream processing unit 201 is supplied to the audio signal processing circuit 207.
- the audio signal processing circuit 207 performs processing such as sound quality adjustment processing on the audio data as necessary.
- the processed audio data output from the audio signal processing circuit 207 is supplied to the HDMI transmission unit 206.
- the stereoscopic image data and audio data supplied to the HDMI transmission unit 206 are transmitted from the HDMI terminal 202 to the HDMI cable 400 via the HDMI TMDS channel.
- FIG. 34 shows a configuration example of the bit stream processing unit 201.
- the bit stream processing unit 201 has a configuration corresponding to the transmission data generation unit 110 shown in FIG.
- the bit stream processing unit 201 includes a demultiplexer 221, a video decoder 222, and a caption decoder 223. Furthermore, the bit stream processing unit 201 includes a stereoscopic image caption generation unit 224, a parallax information extraction unit 225, a video superimposition unit 226, an audio decoder 227, and a parallax information processing unit 228.
- the demultiplexer 221 extracts video, audio, and subtitle packets from the bit stream data BSD and sends them to each decoder.
- the video decoder 222 performs processing opposite to that of the video encoder 113 of the transmission data generation unit 110 described above. That is, a video elementary stream is reconstructed from the video packet extracted by the demultiplexer 221, and decoding processing is performed to obtain stereoscopic image data including left-eye image data and right-eye image data.
- the transmission method of the stereoscopic image data is, for example, the above-described first transmission method (“Top & Bottom” method), the second transmission method (“Side By Side” method), and the third transmission method (“Frame Sequential ”method) (see FIG. 4).
- the caption decoder 223 performs the reverse process of the caption encoder 133 of the transmission data generation unit 110 described above. That is, the caption decoder 223 reconstructs a caption elementary stream (caption data stream) from the caption packet extracted by the demultiplexer 221, performs a decoding process, and performs caption data (ARIB method) of each caption unit. Subtitle data).
- the disparity information extracting unit 225 extracts disparity vectors (disparity information) corresponding to each caption unit from the caption stream obtained through the caption decoder 223. In this case, a disparity vector (individual disparity vector) for each caption unit, or a disparity vector common to each caption unit (common disparity vector) is obtained (see FIGS. 11 to 12).
- the caption data stream includes ARIB-style caption (caption unit) data and disparity information (disparity vector).
- the disparity information is inserted as management information for caption data. Therefore, the disparity information extracting unit 225 can extract the disparity vector in association with the caption data of each caption unit.
- the disparity information extracting unit 225 acquires disparity information (see “default_disparity” in FIG. 24) that is commonly used within the caption display period. In addition, the disparity information extracting unit 225 may further acquire disparity information (see “disparity_update” in FIG. 25) that is sequentially updated within the caption display period.
- the disparity information (disparity vector) extracted by the disparity information extracting unit 225 is sent to the stereoscopic image caption generating unit 224 through the disparity information processing unit 228.
- the disparity information sequentially updated within the caption display period is composed of the disparity information of the first frame in the caption display period and the disparity information of the frame for each subsequent base segment period (update frame interval). ing.
- the parallax information processing unit 228 sends the parallax information used in common during the caption display period to the stereoscopic image caption generation unit 224 as it is.
- the disparity information processing unit 228 performs interpolation processing on the disparity information sequentially updated within the caption display period, and generates disparity information at an arbitrary frame interval, for example, one frame interval within the caption display period.
- the stereoscopic image caption generation unit 224 sends the parallax information used in common during the caption display period to the stereoscopic image caption generation unit 224 as it is.
- the disparity information processing unit 228 performs interpolation processing on the disparity information sequentially updated within the caption display period, and generates disparity information at an arbitrary frame interval, for example, one frame interval within the caption display period.
- the disparity information processing unit 228 performs not the linear interpolation process but the interpolation process with the low-pass filter (LPF) process in the time direction (frame direction), and the disparity information at a predetermined frame interval after the interpolation process.
- the time direction (frame direction) is smoothed.
- FIG. 35 illustrates an example of an interpolation process involving the above-described LPF process in the parallax information processing unit 228. This example corresponds to the parallax information update example of FIG. 29 described above.
- the stereoscopic image caption generation unit 224 generates left-eye caption data and right-eye caption data to be superimposed on the left-eye image and the right-eye image, respectively. This generation process is performed based on the caption data of each caption unit obtained by the caption decoder 223 and the disparity information (disparity vector) supplied through the disparity information processing unit 228.
- the stereoscopic image caption generation unit 224 outputs left-eye caption data and right-eye caption data (bitmap data).
- the left eye and right eye captions are the same information.
- the superimposed position in the image is shifted in the horizontal direction by, for example, the parallax vector between the left-eye caption and the right-eye caption.
- the same subtitle superimposed on the left eye image and the right eye image can be used with parallax adjusted according to the perspective of each object in the image. The perspective consistency between each object within is maintained.
- the stereoscopic image caption generation unit 224 uses the disparity information.
- the stereoscopic image caption generation unit 224 uses any one of the parallax information that is sequentially updated within the caption display period from the parallax information processing unit 228. Which one is used is, for example, as described above, information indicating the disparity information (disparity) correspondence level that is included in the extended display control data unit and is essential on the reception side (decoder side) when displaying captions. (See “rendering_level” in FIG. 24). In this case, for example, when “00”, it depends on the user setting.
- the disparity information that is sequentially updated within the caption display period it is possible to dynamically change the disparity to be given to the left eye and the right eye in conjunction with the change in the image content.
- the video superimposing unit 226 uses the left-eye and right-eye caption data generated by the stereoscopic image caption generation unit 224 for the stereoscopic image data (left-eye image data and right-eye image data) obtained by the video decoder 222. (Bitmap data) is superimposed to obtain display stereoscopic image data Vout.
- the video superimposing unit 226 outputs the display stereoscopic image data Vout to the outside of the bit stream processing unit 201.
- the audio decoder 227 performs a process reverse to that of the audio encoder 117 of the transmission data generation unit 110 described above. That is, the audio decoder 227 reconstructs an audio elementary stream from the audio packet extracted by the demultiplexer 221 and performs a decoding process to obtain audio data Aout. The audio decoder 227 outputs the audio data Aout to the outside of the bit stream processing unit 201.
- bit stream processing unit 201 shown in FIG. 34 will be briefly described.
- the bit stream data BSD output from the digital tuner 204 (see FIG. 33) is supplied to the demultiplexer 221.
- the demultiplexer 221 extracts video, audio, and subtitle packets from the bit stream data BSD and supplies them to each decoder.
- the video decoder 222 reconstructs a video elementary stream from the video packet extracted by the demultiplexer 221, further performs a decoding process, and generates stereoscopic image data including left-eye image data and right-eye image data. can get.
- the stereoscopic image data is supplied to the video superimposing unit 226.
- the subtitle decoder 223 reconstructs a subtitle elementary stream from the subtitle packet extracted by the demultiplexer 221, further performs decoding processing, and subtitle data of each caption unit (ARIB method subtitle data). Is obtained.
- the caption data of each caption unit is supplied to the stereoscopic image caption generation unit 224.
- the disparity information extracting unit 225 extracts disparity vectors (disparity information) corresponding to each caption unit from the caption stream obtained through the caption decoder 223. In this case, a disparity vector (individual disparity vector) for each caption unit, or a disparity vector common to each caption unit (common disparity vector) is obtained.
- the disparity information extracting unit 225 acquires disparity information that is commonly used within the caption display period, or disparity information that is sequentially updated within the caption display period.
- the disparity information (disparity vector) extracted by the disparity information extracting unit 225 is sent to the stereoscopic image caption generating unit 224 through the disparity information processing unit 228.
- the disparity information processing unit 2228 the following processing is performed on the disparity information sequentially updated within the caption display period. That is, the disparity information processing unit 228 performs interpolation processing with LPF processing in the time direction (frame direction), and generates disparity information at an arbitrary frame interval, for example, one frame interval within the caption display period. And sent to the stereoscopic image caption generation unit 224.
- the left-eye caption and the right-eye caption to be superimposed on the left-eye image and the right-eye image, respectively, based on the caption data of each caption unit and the parallax vector corresponding to each caption unit.
- Data bitmap data
- the superimposed position in the image is shifted in the horizontal direction by the amount of the parallax vector, for example, with respect to the left-eye caption.
- the left eye caption and right eye caption data are supplied to the video superimposing unit 226.
- the video superimposing unit 226 superimposes the left-eye caption data and the right-eye caption data (bitmap data) generated by the stereoscopic image caption generation unit 224 on the stereoscopic image data obtained by the video decoder 222 for display.
- Stereoscopic image data Vout is obtained.
- the display stereoscopic image data Vout is output to the outside of the bit stream processing unit 201.
- the audio decoder 227 reconstructs an audio elementary stream from the audio packet extracted by the demultiplexer 221, further performs decoding processing, and audio data Aout corresponding to the display stereoscopic image data Vout described above. Is obtained.
- the audio data Aout is output to the outside of the bit stream processing unit 201.
- disparity vectors are inserted into the received caption data stream as caption data management information, and caption data and the disparity vectors are associated with each other. Yes. Therefore, the bit stream processing unit 201 can give appropriate disparity to the caption unit (caption) superimposed on the left eye image and the right eye image using the corresponding disparity vector (disparity information). Therefore, in the display of caption units (captions), perspective consistency with each object in the image can be maintained in an optimum state.
- the disparity information extracting unit 225 of the bitstream processing unit 201 updates the disparity information that is commonly used in the caption display period or sequentially updated in the caption display period together with the disparity information. Parallax information is acquired.
- the stereoscopic image caption generation unit 224 uses the disparity information that is sequentially updated within the caption display period, so that the disparity to be given to the left-eye and right-eye captions is dynamically linked with changes in the image content. It can be changed.
- the disparity information processing unit 228 of the bitstream processing unit 201 performs interpolation processing on disparity information that is sequentially updated within the caption display period, and within the caption display period. Disparity information at arbitrary frame intervals is generated. In this case, even when the disparity information is transmitted from the transmission side (broadcast station 100) every base segment period (update frame interval) such as 16 frames, the disparity given to the left-eye and right-eye captions is reduced. It is possible to control at fine intervals, for example, for each frame.
- the television receiver 300 receives stereoscopic image data sent from the set top box 200 via the HDMI cable 400.
- the television receiver 300 includes a 3D signal processing unit 301.
- the 3D signal processing unit 301 performs processing (decoding processing) corresponding to the transmission method on the stereoscopic image data to generate left-eye image data and right-eye image data.
- FIG. 36 illustrates a configuration example of the television receiver 300.
- the television receiver 300 includes a 3D signal processing unit 301, an HDMI terminal 302, an HDMI receiving unit 303, an antenna terminal 304, a digital tuner 305, and a bit stream processing unit 306.
- the television receiver 300 includes a video / graphic processing circuit 307, a panel drive circuit 308, a display panel 309, an audio signal processing circuit 310, an audio amplification circuit 311, and a speaker 312.
- the television receiver 300 includes a CPU 321, a flash ROM 322, a DRAM 323, an internal bus 324, a remote control receiving unit 325, and a remote control transmitter 326.
- the antenna terminal 304 is a terminal for inputting a television broadcast signal received by a receiving antenna (not shown).
- the digital tuner 305 processes the television broadcast signal input to the antenna terminal 304 and outputs predetermined bit stream data (transport stream) BSD corresponding to the user's selected channel.
- the bit stream processing unit 306 has the same configuration as the bit stream processing unit 201 in the set top box 200 shown in FIG.
- the bit stream processing unit 306 extracts stereoscopic image data, audio data, caption unit caption data, disparity vectors, and the like from the bit stream data BSD.
- the bit stream processing unit 306 combines the left-eye caption data and the right-eye caption data with the stereoscopic image data to generate and output stereoscopic image data for display.
- the bit stream processing unit 306 outputs audio data.
- the HDMI receiving unit 303 receives uncompressed image data and audio data supplied to the HDMI terminal 302 via the HDMI cable 400 by communication conforming to HDMI.
- the HDMI receiving unit 303 has a version of, for example, HDMI 1.4a, and can handle stereoscopic image data.
- the 3D signal processing unit 301 performs decoding processing on the stereoscopic image data received by the HDMI receiving unit 303 or obtained by the bit stream processing unit 306 to generate left eye image data and right eye image data. To do. In this case, the 3D signal processing unit 301 performs decoding processing corresponding to the transmission method (see FIG. 4) on the stereoscopic image data obtained by the bit stream processing unit 306. The 3D signal processing unit 301 performs a decoding process corresponding to the TMDS transmission data structure on the stereoscopic image data received by the HDMI receiving unit 303.
- the video / graphic processing circuit 307 generates image data for displaying a stereoscopic image based on the left eye image data and right eye image data generated by the 3D signal processing unit 301.
- the video / graphic processing circuit 307 performs image quality adjustment processing on the image data as necessary. Further, the video / graphic processing circuit 307 synthesizes superimposition information data such as a menu and a program guide with the image data as necessary.
- the panel drive circuit 308 drives the display panel 309 based on the image data output from the video / graphic processing circuit 307.
- the display panel 309 includes, for example, an LCD (Liquid Crystal Display), a PDP (Plasma Display Panel), or the like.
- the audio signal processing circuit 310 performs necessary processing such as D / A conversion on the audio data received by the HDMI receiving unit 303 or obtained by the bit stream processing unit 306.
- the audio amplification circuit 311 amplifies the audio signal output from the audio signal processing circuit 310 and supplies the amplified audio signal to the speaker 312.
- the CPU 321 controls the operation of each unit of the television receiver 300.
- the flash ROM 322 stores control software and data.
- the DRAM 323 constitutes a work area for the CPU 321.
- the CPU 321 develops software and data read from the flash ROM 322 on the DRAM 323 to activate the software, and controls each unit of the television receiver 300.
- the remote control receiving unit 325 receives the remote control signal (remote control code) transmitted from the remote control transmitter 326 and supplies it to the CPU 321.
- the CPU 321 controls each part of the television receiver 300 based on the remote control code.
- the CPU 321, flash ROM 322, and DRAM 323 are connected to the internal bus 324.
- the HDMI receiving unit 303 receives stereoscopic image data and audio data transmitted from the set top box 200 connected to the HDMI terminal 302 via the HDMI cable 400.
- the stereoscopic image data received by the HDMI receiving unit 303 is supplied to the 3D signal processing unit 301.
- the audio data received by the HDMI receiving unit 303 is supplied to the audio signal processing circuit 310.
- the TV broadcast signal input to the antenna terminal 304 is supplied to the digital tuner 305.
- the digital tuner 305 processes the television broadcast signal and outputs predetermined bit stream data (transport stream) BSD corresponding to the user's selected channel.
- the bit stream data BSD output from the digital tuner 305 is supplied to the bit stream processing unit 306.
- the bit stream processing unit 306 extracts stereoscopic image data, audio data, caption unit caption data, disparity vectors, and the like from the bit stream data BSD. Further, in the bit stream processing unit 306, the left-eye caption data and the right-eye caption data are combined with the stereoscopic image data, and display stereoscopic image data is generated.
- the display stereoscopic image data generated by the bit stream processing unit 306 is supplied to the 3D signal processing unit 301. Also, the audio data obtained by the bit stream processing unit 306 is supplied to the audio signal processing circuit 310.
- the 3D signal processing unit 301 performs decoding processing on the stereoscopic image data received by the HDMI receiving unit 303 or obtained by the bit stream processing unit 306 to generate left eye image data and right eye image data. Is done.
- the left eye image data and right eye image data are supplied to the video / graphic processing circuit 307.
- image data for displaying a stereoscopic image is generated based on the left eye image data and the right eye image data, and image quality adjustment processing and superimposition information data synthesis processing are performed as necessary. Is also done.
- the image data obtained by the video / graphic processing circuit 307 is supplied to the panel drive circuit 308. Therefore, a stereoscopic image is displayed on the display panel 309.
- the left eye image based on the left eye image data and the right eye image based on the right eye image data are alternately displayed on the display panel 309 in a time division manner.
- the viewer can see only the left eye image with the left eye and the right eye with the shutter glasses by alternately opening the left eye shutter and the right eye shutter in synchronization with the display on the display panel 309. Only the right eye image can be seen, and a stereoscopic image can be perceived.
- the audio signal processing circuit 310 necessary processing such as D / A conversion is performed on the audio data received by the HDMI receiving unit 303 or obtained by the bit stream processing unit 306.
- the audio data is amplified by the audio amplification circuit 311 and then supplied to the speaker 312. Therefore, sound corresponding to the display image on the display panel 309 is output from the speaker 312.
- the stereoscopic image display system 10 includes the broadcasting station 100, the set top box 200, and the television receiver 300.
- the television receiver 300 includes a bit stream processing unit 306 that functions in the same manner as the bit stream processing unit 201 in the set-top box 200, as shown in FIG. Therefore, as shown in FIG. 37, a stereoscopic image display system 10A including a broadcasting station 100 and a television receiver 300 is also conceivable.
- the set-top box 200 and the television receiver 300 are connected by an HDMI digital interface.
- the present invention can be similarly applied even when these are connected by a digital interface (including wireless as well as wired) similar to the HDMI digital interface.
- the caption unit (caption) is handled as the superimposition information.
- the present invention can be similarly applied to other information that handles superimposition information such as graphics information and text information.
- the present invention can be applied to a stereoscopic image system that displays superimposition information such as captions superimposed on an image.
- CPU 215 Remote control reception unit 216: Remote control transmitter 221 ... Demultiplexer 222 ... Video decoder 223 ... Subtitle decoder 224 ... Stereo image caption generation unit 225 ... Parallax information extraction unit 226: Video superimposing unit 227: Audio decoder 228: Parallax information processing unit 300: Television receiver (TV) DESCRIPTION OF SYMBOLS 301 ... 3D signal processing part 302 ... HDMI terminal 303 ... HDMI receiving part 304 ... Antenna terminal 305 ... Digital tuner 306 ... Bit stream processing part 307 ... Video / graphic processing circuit 308 ... Panel drive circuit 309 ... Display panel 310 ... Audio signal processing circuit 311 ... Audio amplification circuit 312 ... Speaker 321 ... CPU 325 ... Remote control receiver 326 ... Remote control transmitter 400 ... HDMI cable
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Human Computer Interaction (AREA)
- Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
Abstract
Description
立体画像を構成する左眼画像データおよび右眼画像データを出力する画像データ出力部と、
上記左眼画像データおよび上記右眼画像データによる画像に重畳する重畳情報のデータを出力する重畳情報データ出力部と、
上記左眼画像データおよび上記右眼画像データによる画像に重畳する重畳情報をシフトさせて視差を付与するための視差情報を出力する視差情報出力部と、
上記画像データ含む第1のデータストリームと、上記重畳情報のデータおよび上記視差情報を含む第2のデータストリームとを有する多重化データストリームを送信するデータ送信部とを備え、
上記第2のデータストリームは、上記重畳情報の管理情報としての表示制御情報を送出するデータユニットを有し、
上記視差情報は、上記データユニットに挿入されている
立体画像データ送信装置。
第1のデータストリームと第2のデータストリームとが含まれる多重化データストリームを受信するデータ受信部を備え、
上記第1のデータストリームは、立体画像を構成する左眼画像データおよび右眼画像データを含み、
上記第2のデータストリームは、上記左眼画像データおよび上記右眼画像データによる画像に重畳する重畳情報のデータと、上記左眼画像データおよび上記右眼画像データによる画像に重畳する重畳情報をシフトさせて視差を付与するための視差情報を含み、
上記第2のデータストリームは、上記重畳情報の管理情報としての表示制御情報を送出するデータユニットを有し、上記視差情報は該データユニットに挿入されており、
上記多重化データストリームに含まれる上記第1のデータストリームから上記左眼画像データおよび上記右眼画像データを取得する画像データ取得部と、
上記多重化データストリームに含まれる上記第2のデータストリームから上記重畳情報のデータを取得する重畳情報データ取得部と、
上記多重化データストリームに含まれる上記第2のデータストリームから上記視差情報を取得する視差情報取得部と、
上記左眼画像データおよび上記右眼画像データと、上記視差情報と、上記重畳情報のデータを用い、左眼画像および右眼画像に重畳する同一の重畳情報に視差を付与し、上記重畳情報が重畳された左眼画像のデータおよび上記重畳情報が重畳された右眼画像のデータを得る画像データ処理部とをさらに備える
立体画像データ受信装置にある。
1.実施の形態
2.変形例
[立体画像表示システムの構成例]
図1は、実施の形態としての立体画像表示システム10の構成例を示している。この立体画像表示システム10は、放送局100と、セットトップボックス(STB:Set Top Box)200と、テレビ受信機(TV:Television)300を有している。
放送局100は、ビットストリームデータBSDを、放送波に載せて送信する。放送局100は、ビットストリームデータBSDを生成する送信データ生成部110を備えている。このビットストリームデータBSDには、画像データ、音声データ、重畳情報のデータ、さらには視差情報(視差ベクトル)などが含まれる。ここで、画像データ(以下、適宜、「立体画像データ」という)は、立体画像を構成する左眼画像データおよび右眼画像データを含む。重畳情報は、グラフィクス情報、テキスト情報などであるが、この実施の形態においては字幕である。
図2は、放送局100において、送信データ生成部110の構成例を示している。この送信データ生成部110は、既存の放送規格の一つであるARIB(Association of Radio Industries and Businesses)に容易に連携できるデータ構造で視差情報(視差ベクトル)を送信する。この送信データ生成部110は、データ取り出し部(アーカイブ部)130と、視差情報作成部131と、ビデオエンコーダ113と、オーディオエンコーダ117と、字幕発生部132と、字幕エンコーダ133と、マルチプレクサ122を有している。
図1に戻って、セットトップボックス200は、放送局100から放送波に載せて送信されてくるビットストリームデータ(トランスポートストリーム)BSDを受信する。このビットストリームデータBSDには、左眼画像データおよび右眼画像データを含む立体画像データ、音声データが含まれている。また、ビットストリームデータBSDには、キャプション・ユニットの字幕データ、さらには、このキャプション・ユニットに視差を付与するための視差ベクトル(視差情報)が含まれている。
セットトップボックス200の構成例を説明する。図33は、セットトップボックス200の構成例を示している。このセットトップボックス200は、ビットストリーム処理部201と、HDMI端子202と、アンテナ端子203と、デジタルチューナ204と、映像信号処理回路205と、HDMI送信部206と、音声信号処理回路207を有している。また、このセットトップボックス200は、CPU211と、フラッシュROM212と、DRAM213と、内部バス214と、リモコン受信部215と、リモコン送信機216を有している。
図34は、ビットストリーム処理部201の構成例を示している。このビットストリーム処理部201は、上述の図2に示す送信データ生成部110に対応した構成となっている。このビットストリーム処理部201は、デマルチプレクサ221と、ビデオデコーダ222と、字幕デコーダ223を有している。さらに、このビットストリーム処理部201は、立体画像用字幕発生部224と、視差情報取り出し部225と、ビデオ重畳部226と、オーディオデコーダ227と、視差情報処理部228を有している。
図1に戻って、テレビ受信機300は、セットトップボックス200からHDMIケーブル400を介して送られてくる立体画像データを受信する。このテレビ受信機300は、3D信号処理部301を有している。この3D信号処理部301は、立体画像データに対して、伝送方式に対応した処理(デコード処理)を行って、左眼画像データおよび右眼画像データを生成する。
テレビ受信機300の構成例を説明する。図36は、テレビ受信機300の構成例を示している。このテレビ受信機300は、3D信号処理部301と、HDMI端子302と、HDMI受信部303と、アンテナ端子304と、デジタルチューナ305と、ビットストリーム処理部306を有している。
なお、上述実施の形態においては、立体画像表示システム10が、放送局100、セットトップボックス200およびテレビ受信機300で構成されているものを示した。しかし、テレビ受信機300は、図36に示すように、セットトップボックス200内のビットストリーム処理部201と同等に機能するビットストリーム処理部306を備えている。したがって、図37に示すように、放送局100およびテレビ受信機300で構成される立体画像表示システム10Aも考えられる。
100・・・放送局
110・・・送信データ生成部
113・・・ビデオエンコーダ
117・・・オーディオエンコーダ
122・・・マルチプレクサ
130・・・データ取り出し部
130a・・・データ記録媒体
131・・・視差情報作成部
132・・・字幕発生部
133・・・字幕エンコーダ
200・・・セットトップボックス(STB)
201・・・ビットストリーム処理部
202・・・HDMI端子
203・・・アンテナ端子
204・・・デジタルチューナ
205・・・映像信号処理回路
206・・・HDMI送信部
207・・・音声信号処理回路
211・・・CPU
215・・・リモコン受信部
216・・・リモコン送信機
221・・・デマルチプレクサ
222・・・ビデオデコーダ
223・・・字幕デコーダ
224・・・立体画像用字幕発生部
225・・・視差情報取り出し部
226・・・ビデオ重畳部
227・・・オーディオデコーダ
228・・・視差情報処理部
300・・・テレビ受信機(TV)
301・・・3D信号処理部
302・・・HDMI端子
303・・・HDMI受信部
304・・・アンテナ端子
305・・・デジタルチューナ
306・・・ビットストリーム処理部
307・・・映像・グラフィック処理回路
308・・・パネル駆動回路
309・・・表示パネル
310・・・音声信号処理回路
311・・・音声増幅回路
312・・・スピーカ
321・・・CPU
325・・・リモコン受信部
326・・・リモコン送信機
400・・・HDMIケーブル
Claims (16)
- 立体画像を構成する左眼画像データおよび右眼画像データを出力する画像データ出力部と、
上記左眼画像データおよび上記右眼画像データによる画像に重畳する重畳情報のデータを出力する重畳情報データ出力部と、
上記左眼画像データおよび上記右眼画像データによる画像に重畳する重畳情報をシフトさせて視差を付与するための視差情報を出力する視差情報出力部と、
上記画像データ含む第1のデータストリームと、上記重畳情報のデータおよび上記視差情報を含む第2のデータストリームとを有する多重化データストリームを送信するデータ送信部とを備え、
上記第2のデータストリームは、上記重畳情報の管理情報としての表示制御情報を送出するデータユニットを有し、
上記視差情報は、上記データユニットに挿入されている
立体画像データ送信装置。 - 上記視差情報は、上記重畳情報が表示される所定数のフレーム期間内で共通に使用される第1の視差情報、または該第1の視差情報および上記所定数のフレーム期間内で順次更新される第2の視差情報であり、
上記データユニットには、上記第2の視差情報の存在を示すフラグ情報が挿入されている
請求項1に記載の立体画像データ送信装置。 - 上記第2の視差情報は、上記所定数のフレーム期間の最初のフレームの視差情報と、その後の更新フレーム間隔毎のフレームの視差情報とからなる
請求項2に記載の立体画像データ送信装置。 - 上記第2の視差情報には、上記更新フレーム間隔毎のフレームのそれぞれについて、上記視差情報の更新の有無を示すフラグ情報が付加されている
請求項3に記載の立体画像データ送信装置。 - 上記第2の視差情報には、上記更新フレーム期間毎のフレームのそれぞれについて、上記更新フレーム間隔を調整する情報が付加されている
請求項3に記載の立体画像データ送信装置。 - 上記データユニットには、フレーム周期を指定する情報が挿入されている
請求項3に記載の立体画像データ送信装置。 - 上記データユニットに挿入されている視差情報は、同一画面に表示される特定の重畳情報に対応した視差情報、または同一画面に表示される複数の重畳情報に対応した視差情報である
請求項1に記載の立体画像データ送信装置。 - 上記データユニットには、上記重畳情報の表示の際に必須の、上記視差情報に対する対応レベルを示す情報が挿入されている
請求項2に記載の立体画像データ送信装置。 - 上記第2のデータストリームには、上記管理情報のレイヤに、上記第2のデータストリームが上記重畳情報の拡張表示に対応しているか否かを示すフラグ情報が含まれている
請求項1に記載の立体画像データ送信装置。 - 上記重畳情報のデータは、ARIB方式の字幕分データであり、
上記第2のデータストリームにおいて、上記視差情報は、字幕管理データに含まれる上記表示制御情報のデータユニットに挿入される
請求項1に記載の立体画像データ送信装置。 - 立体画像を構成する左眼画像データおよび右眼画像データを出力する画像データ出力ステップと、
上記左眼画像データおよび上記右眼画像データによる画像に重畳する重畳情報のデータを出力する重畳情報データ出力ステップと、
上記左眼画像データおよび上記右眼画像データによる画像に重畳する重畳情報をシフトさせて視差を付与するための視差情報を出力する視差情報出力ステップと、
上記画像データ含む第1のデータストリームと、上記重畳情報のデータおよび上記視差情報を含む第2のデータストリームとを有する多重化データストリームを送信するデータ送信ステップとを備え、
上記第2のデータストリームは、上記重畳情報の管理情報としての表示制御情報を送出するデータユニットを有し、
上記視差情報は、上記データユニットに挿入されている
立体画像データ送信方法。 - 第1のデータストリームと第2のデータストリームとが含まれる多重化データストリームを受信するデータ受信部を備え、
上記第1のデータストリームは、立体画像を構成する左眼画像データおよび右眼画像データを含み、
上記第2のデータストリームは、上記左眼画像データおよび上記右眼画像データによる画像に重畳する重畳情報のデータと、上記左眼画像データおよび上記右眼画像データによる画像に重畳する重畳情報をシフトさせて視差を付与するための視差情報を含み、
上記第2のデータストリームは、上記重畳情報の管理情報としての表示制御情報を送出するデータユニットを有し、上記視差情報は該データユニットに挿入されており、
上記多重化データストリームに含まれる上記第1のデータストリームから上記左眼画像データおよび上記右眼画像データを取得する画像データ取得部と、
上記多重化データストリームに含まれる上記第2のデータストリームから上記重畳情報のデータを取得する重畳情報データ取得部と、
上記多重化データストリームに含まれる上記第2のデータストリームから上記視差情報を取得する視差情報取得部と、
上記上記左眼画像データおよび上記右眼画像データと、上記視差情報と、上記重畳情報のデータを用い、左眼画像および右眼画像に重畳する同一の重畳情報に視差を付与し、上記重畳情報が重畳された左眼画像のデータおよび上記重畳情報が重畳された右眼画像のデータを得る画像データ処理部とをさらに備える
立体画像データ受信装置。 - 上記データユニットに挿入されている視差情報は、上記重畳情報が表示される所定数のフレーム期間内で順次更新される視差情報であり、上記所定数のフレーム期間の最初のフレームの視差情報と、その後の更新フレーム間隔毎のフレームの視差情報とからなる
請求項12に記載の立体画像データ受信装置。 - 上記画像データ処理部は、上記所定数のフレーム期間内で順次更新される視差情報を構成する複数フレームの視差情報に対して補間処理を施し、上記所定数のフレーム期間内における任意のフレーム間隔の視差情報を生成して使用する
請求項13に記載の立体画像データ受信装置。 - 上記補間処理は、時間方向のローパスフィルタ処理を伴う
請求項14に記載の立体画像データ受信装置。 - 第1のデータストリームと第2のデータストリームとが含まれる多重化データストリームを受信するデータ受信ステップを備え、
上記第1のデータストリームは、立体画像を構成する左眼画像データおよび右眼画像データを含み、
上記第2のデータストリームは、上記左眼画像データおよび上記右眼画像データによる画像に重畳する重畳情報のデータと、上記左眼画像データおよび上記右眼画像データによる画像に重畳する重畳情報をシフトさせて視差を付与するための視差情報を含み、
上記第2のデータストリームは、上記重畳情報の管理情報としての表示制御情報を送出するデータユニットを有し、上記視差情報は該データユニットに挿入されており、
上記多重化データストリームに含まれる上記第1のデータストリームから上記左眼画像データおよび上記右眼画像データを取得する画像データ取得ステップと、
上記多重化データストリームに含まれる上記第2のデータストリームから上記重畳情報のデータを取得する重畳情報データ取得ステップと、
上記多重化データストリームに含まれる上記第2のデータストリームから上記視差情報を取得する視差情報取得ステップと、
上記左眼画像データおよび上記右眼画像データと、上記視差情報と、上記重畳情報のデータを用い、左眼画像および右眼画像に重畳する同一の重畳情報に視差を付与し、上記重畳情報が重畳された左眼画像のデータおよび上記重畳情報が重畳された右眼画像のデータを得る画像データ処理ステップとをさらに備える
立体画像データ受信方法。
Priority Applications (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN2011800102175A CN102771131A (zh) | 2010-10-29 | 2011-10-21 | 立体图像数据发送装置、立体图像数据发送方法、立体图像数据接收装置和立体图像数据接收方法 |
BR112012015128A BR112012015128A2 (pt) | 2010-10-29 | 2011-10-21 | dispositivos de transmissão de dados de imagem estereoscópica e de recepção de dados de imagem estereoscópica, e, métodos de transmissão de dados de imagem estereoscópica e de recepção de dados de imagem estereoscópica |
US13/518,775 US20120262454A1 (en) | 2010-10-29 | 2011-10-21 | Stereoscopic image data transmission device, stereoscopic image data transmission method, stereoscopic image data reception device, and stereoscopic image data reception method |
EP11836182.3A EP2503785A4 (en) | 2010-10-29 | 2011-10-21 | STEREOSCOPIC IMAGE DATA TRANSMISSION DEVICE, STEREOSCOPIC IMAGE DATA TRANSMISSION METHOD, STEREOSCOPIC IMAGE DATA RECEIVING DEVICE, AND STEREOSCOPIC IMAGE DATA RECEIVING METHOD |
KR1020127016903A KR20130132239A (ko) | 2010-10-29 | 2011-10-21 | 입체화상 데이터 송신 장치, 입체화상 데이터 송신 방법, 입체화상 데이터 수신 장치 및 입체화상 데이터 수신 방법 |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2010244837A JP2012099986A (ja) | 2010-10-29 | 2010-10-29 | 立体画像データ送信装置、立体画像データ送信方法、立体画像データ受信装置および立体画像データ受信方法 |
JP2010-244837 | 2010-10-29 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2012057048A1 true WO2012057048A1 (ja) | 2012-05-03 |
Family
ID=45993758
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2011/074368 WO2012057048A1 (ja) | 2010-10-29 | 2011-10-21 | 立体画像データ送信装置、立体画像データ送信方法、立体画像データ受信装置および立体画像データ受信方法 |
Country Status (8)
Country | Link |
---|---|
US (1) | US20120262454A1 (ja) |
EP (1) | EP2503785A4 (ja) |
JP (1) | JP2012099986A (ja) |
KR (1) | KR20130132239A (ja) |
CN (1) | CN102771131A (ja) |
AR (1) | AR083529A1 (ja) |
BR (1) | BR112012015128A2 (ja) |
WO (1) | WO2012057048A1 (ja) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102685558A (zh) * | 2012-05-17 | 2012-09-19 | 安徽天虹数码技术有限公司 | Ts流字幕插入方法 |
WO2014109321A1 (ja) * | 2013-01-09 | 2014-07-17 | ソニー株式会社 | 送信装置、送信方法、受信装置および受信方法 |
Families Citing this family (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8566694B2 (en) * | 2011-04-04 | 2013-10-22 | Xerox Corporation | Multi-dimensional documents for parallel content display on a single screen for multiple viewers |
JP2013066075A (ja) * | 2011-09-01 | 2013-04-11 | Sony Corp | 送信装置、送信方法および受信装置 |
CN110111339B (zh) * | 2019-04-28 | 2023-08-15 | 航天智造(上海)科技有限责任公司 | 条纹图像目标区域提取方法 |
CN110211053B (zh) * | 2019-04-28 | 2023-09-15 | 航天智造(上海)科技有限责任公司 | 用于三维测量的快速精确相位匹配方法 |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2005006114A (ja) | 2003-06-12 | 2005-01-06 | Sharp Corp | 放送データ送信装置、放送データ送信方法および放送データ受信装置 |
WO2010095411A1 (ja) * | 2009-02-19 | 2010-08-26 | パナソニック株式会社 | 記録媒体、再生装置、集積回路 |
WO2010119815A1 (ja) * | 2009-04-15 | 2010-10-21 | ソニー株式会社 | データ構造、記録媒体、再生装置および再生方法、並びにプログラム |
WO2011001859A1 (ja) * | 2009-06-29 | 2011-01-06 | ソニー株式会社 | 立体画像データ送信装置および立体画像データ受信装置 |
Family Cites Families (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5621470A (en) * | 1992-12-18 | 1997-04-15 | Sid-Ahmed; Maher A. | Interpixel and interframe interpolation of television pictures with conversion from interlaced to progressive scanning |
US8228994B2 (en) * | 2005-05-20 | 2012-07-24 | Microsoft Corporation | Multi-view video coding based on temporal and view decomposition |
US8358331B2 (en) * | 2008-12-02 | 2013-01-22 | Lg Electronics Inc. | 3D caption display method and 3D display apparatus for implementing the same |
WO2010095410A1 (ja) * | 2009-02-20 | 2010-08-26 | パナソニック株式会社 | 記録媒体、再生装置、集積回路 |
JP2010250562A (ja) * | 2009-04-15 | 2010-11-04 | Sony Corp | データ構造、記録媒体、再生装置および再生方法、並びにプログラム |
JP5627860B2 (ja) * | 2009-04-27 | 2014-11-19 | 三菱電機株式会社 | 立体映像配信システム、立体映像配信方法、立体映像配信装置、立体映像視聴システム、立体映像視聴方法、立体映像視聴装置 |
JP5429034B2 (ja) * | 2009-06-29 | 2014-02-26 | ソニー株式会社 | 立体画像データ送信装置、立体画像データ送信方法、立体画像データ受信装置および立体画像データ受信方法 |
-
2010
- 2010-10-29 JP JP2010244837A patent/JP2012099986A/ja active Pending
-
2011
- 2011-10-21 CN CN2011800102175A patent/CN102771131A/zh active Pending
- 2011-10-21 EP EP11836182.3A patent/EP2503785A4/en not_active Withdrawn
- 2011-10-21 US US13/518,775 patent/US20120262454A1/en not_active Abandoned
- 2011-10-21 KR KR1020127016903A patent/KR20130132239A/ko not_active Application Discontinuation
- 2011-10-21 WO PCT/JP2011/074368 patent/WO2012057048A1/ja active Application Filing
- 2011-10-21 AR ARP110103912A patent/AR083529A1/es not_active Application Discontinuation
- 2011-10-21 BR BR112012015128A patent/BR112012015128A2/pt not_active IP Right Cessation
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2005006114A (ja) | 2003-06-12 | 2005-01-06 | Sharp Corp | 放送データ送信装置、放送データ送信方法および放送データ受信装置 |
WO2010095411A1 (ja) * | 2009-02-19 | 2010-08-26 | パナソニック株式会社 | 記録媒体、再生装置、集積回路 |
WO2010119815A1 (ja) * | 2009-04-15 | 2010-10-21 | ソニー株式会社 | データ構造、記録媒体、再生装置および再生方法、並びにプログラム |
WO2011001859A1 (ja) * | 2009-06-29 | 2011-01-06 | ソニー株式会社 | 立体画像データ送信装置および立体画像データ受信装置 |
Non-Patent Citations (1)
Title |
---|
See also references of EP2503785A4 * |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102685558A (zh) * | 2012-05-17 | 2012-09-19 | 安徽天虹数码技术有限公司 | Ts流字幕插入方法 |
CN102685558B (zh) * | 2012-05-17 | 2015-02-11 | 安徽天虹数码技术有限公司 | Ts流字幕插入方法 |
WO2014109321A1 (ja) * | 2013-01-09 | 2014-07-17 | ソニー株式会社 | 送信装置、送信方法、受信装置および受信方法 |
Also Published As
Publication number | Publication date |
---|---|
AR083529A1 (es) | 2013-03-06 |
KR20130132239A (ko) | 2013-12-04 |
JP2012099986A (ja) | 2012-05-24 |
US20120262454A1 (en) | 2012-10-18 |
EP2503785A4 (en) | 2013-10-23 |
CN102771131A (zh) | 2012-11-07 |
EP2503785A1 (en) | 2012-09-26 |
BR112012015128A2 (pt) | 2016-05-31 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8860786B2 (en) | Stereo image data transmitting apparatus and stereo image data receiving apparatus | |
JP5429034B2 (ja) | 立体画像データ送信装置、立体画像データ送信方法、立体画像データ受信装置および立体画像データ受信方法 | |
JP5531972B2 (ja) | 立体画像データ送信装置、立体画像データ送信方法、立体画像データ受信装置および立体画像データ受信方法 | |
JP5454444B2 (ja) | 立体画像データ送信装置、立体画像データ送信方法、立体画像データ受信装置および立体画像データ受信方法 | |
JP5633259B2 (ja) | 立体画像データ送信装置、立体画像データ送信方法および立体画像データ受信装置 | |
WO2012060198A1 (ja) | 立体画像データ送信装置、立体画像データ送信方法、立体画像データ受信装置および立体画像データ受信方法 | |
JP2011030183A (ja) | 立体画像データ送信装置および立体画像データ送信方法 | |
JP2011030182A (ja) | 立体画像データ送信装置、立体画像データ送信方法、立体画像データ受信装置および立体画像データ受信方法 | |
WO2013031549A1 (ja) | 送信装置、送信方法および受信装置 | |
JP2011249945A (ja) | 立体画像データ送信装置、立体画像データ送信方法、立体画像データ受信装置および立体画像データ受信方法 | |
WO2013005571A1 (ja) | 送信装置、送信方法および受信装置 | |
JP5682149B2 (ja) | 立体画像データ送信装置、立体画像データ送信方法、立体画像データ受信装置および立体画像データ受信方法 | |
WO2012057048A1 (ja) | 立体画像データ送信装置、立体画像データ送信方法、立体画像データ受信装置および立体画像データ受信方法 | |
JP2011166757A (ja) | 送信装置、送信方法および受信装置 | |
JP2012120143A (ja) | 立体画像データ送信装置、立体画像データ送信方法、立体画像データ受信装置および立体画像データ受信方法 | |
WO2012026342A1 (ja) | 立体画像データ送信装置、立体画像データ送信方法、立体画像データ受信装置および立体画像データ受信方法 | |
WO2013011834A1 (ja) | 送信装置、送信方法および受信装置 | |
WO2012063675A1 (ja) | 立体画像データ送信装置、立体画像データ送信方法および立体画像データ受信装置 | |
WO2013018489A1 (ja) | 送信装置、送信方法および受信装置 | |
WO2013172142A1 (ja) | 送信装置、送信方法、受信装置および受信方法 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 201180010217.5 Country of ref document: CN |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2011836182 Country of ref document: EP |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 11836182 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 13518775 Country of ref document: US |
|
ENP | Entry into the national phase |
Ref document number: 20127016903 Country of ref document: KR Kind code of ref document: A |
|
REG | Reference to national code |
Ref country code: BR Ref legal event code: B01A Ref document number: 112012015128 Country of ref document: BR |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
ENP | Entry into the national phase |
Ref document number: 112012015128 Country of ref document: BR Kind code of ref document: A2 Effective date: 20120619 |