WO2011016240A1 - 映像再生装置 - Google Patents
映像再生装置 Download PDFInfo
- Publication number
- WO2011016240A1 WO2011016240A1 PCT/JP2010/004934 JP2010004934W WO2011016240A1 WO 2011016240 A1 WO2011016240 A1 WO 2011016240A1 JP 2010004934 W JP2010004934 W JP 2010004934W WO 2011016240 A1 WO2011016240 A1 WO 2011016240A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- video
- graphics
- depth
- synthesized
- frame
- Prior art date
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/79—Processing of colour television signals in connection with recording
- H04N9/87—Regeneration of colour television signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/128—Adjusting depth or disparity
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/189—Recording image signals; Reproducing recorded image signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/41—Structure of client; Structure of client peripherals
- H04N21/426—Internal components of the client ; Characteristics thereof
- H04N21/42646—Internal components of the client ; Characteristics thereof for reading from or writing on a non-volatile solid state storage medium, e.g. DVD, CD-ROM
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/432—Content retrieval operation from a local storage medium, e.g. hard-disk
- H04N21/4325—Content retrieval operation from a local storage medium, e.g. hard-disk by playing back content from the storage medium
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/79—Processing of colour television signals in connection with recording
- H04N9/80—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
- H04N9/82—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only
- H04N9/8205—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only involving the multiplexing of an additional signal and the colour video signal
- H04N9/8227—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only involving the multiplexing of an additional signal and the colour video signal the additional signal being at least another television signal
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/76—Television signal recording
- H04N5/84—Television signal recording using optical recording
- H04N5/85—Television signal recording using optical recording on discs or drums
Definitions
- This technical field relates to video playback technology, and in particular, to a video playback device that outputs a stereoscopically visible video signal.
- the video playback device When displaying a stereoscopic video, the video playback device generates a left-eye video and a right-eye video from content data recorded in a video source such as an optical disk, and sends the generated video to the video display device.
- the video display device alternately presents a left-eye video and a right-eye video.
- Patent Document 1 discloses a video playback device that enables stereoscopic video display by presenting a right-eye video and a left-eye video having parallax to a user. The user synthesizes an image viewed through the right eye (right eye image) and an image viewed through the left eye (left eye image) in the brain, and recognizes it as a stereoscopic image.
- ⁇ Subtitles and graphics video synthesized with the main video are generated in the video playback device based on the subtitles and graphics data included in the video source.
- the video playback device uses a memory and a processor to generate subtitles and graphics video from the data of the video source so as to meet the timing of synthesis with the main video, and generate the generated video for each main video.
- a video signal of the synthesized video is generated by synthesizing with the video of the frame, and the synthesized video signal is output at the playback timing of the synthesized video.
- the user can only recognize the graphics video for the left eye with respect to the graphics video, and the graphics video can be recognized as a stereoscopic video. Not only does this occur, but the localization in the depth direction (depth direction) of the graphics image is not correctly determined, and the user perceives that the image is abnormal (unnatural).
- the lack of at least one of the left-eye video and the right-eye video of the stereoscopic video to be synthesized with the main video is not preferable even if it is temporary.
- the present embodiment has been made to solve the above-mentioned problems, and is used for synthesis when the second video (for example, graphics or subtitles) to be combined with the first video (for example, main video) is missing.
- This is a video playback device that can output a natural stereoscopic video that does not give the user a sense of incongruity even when it cannot.
- One aspect of the present invention is a video playback apparatus that synthesizes and outputs a first video that is a stereoscopic video and a second video, and the first video and the second video that are to be synthesized into one frame.
- the acquisition unit cannot acquire the second video to be synthesized into one frame at the time of generating the video, the second to be synthesized with the first video in a frame different from the one frame
- a video playback device for synthesizing a video with a first video to be synthesized into one frame.
- Another aspect of the semiconductor device is a semiconductor device that synthesizes and outputs a first video that is a stereoscopic video and a second video, and the first video and the second video that are to be synthesized into one frame.
- An acquisition means for acquiring the image a synthesis means for generating a synthesized video by synthesizing the first video and the second video obtained by the acquisition means, and a signal output means for outputting the synthesized video
- the synthesizing unit is synthesized with the first video in a frame different from the one frame when the obtaining unit cannot acquire the second video to be synthesized into one frame when the synthesized video is generated.
- a semiconductor device that combines a second video to be combined with a first video to be combined into one frame.
- Still another aspect thereof is a video playback method for synthesizing and outputting a first video that is a stereoscopic video and a second video in the video playback device, and the acquisition means is combined into one frame.
- the first video and the second video to be generated, and the synthesizing unit generates the synthesized video by synthesizing the first video and the second video acquired in the obtaining step in synchronism with each other.
- And outputting a video, and the synthesizing unit is different from the one frame when the acquisition unit cannot acquire the second video to be synthesized into one frame at the time of generating the synthesized video.
- This is a video playback method in which a second video to be synthesized with a first video in a frame is synthesized with a first video to be synthesized into one frame.
- the video playback apparatus can be used even when the second video (subtitle video, graphics video, etc.) of the stereoscopic video synthesized with the first video (main video, etc.) cannot be used for synthesis due to omission or the like. Can output natural 3D images without feeling uncomfortable.
- FIG. 1 is a configuration diagram of a 3D video playback display system including a video playback device according to an embodiment.
- Configuration diagram of video display device of 3D video playback display system Configuration diagram of video playback apparatus according to embodiments Detailed block diagram of second AV processing unit of video reproduction apparatus according to Embodiment 1
- Diagram explaining the plane synthesis model Schematic diagram of stereoscopic video signal sent out by video playback apparatus according to the embodiment Diagram explaining the concept of depth in stereoscopic video
- the figure explaining the offset given to a picture Detailed block diagram of the synthesis unit of the second AV processing unit of the video reproduction apparatus Flowchart of processing for outputting graphics video Diagram explaining alternative video output when graphics video is missing
- Diagram explaining alternative video output when graphics video is missing The figure explaining the process which a graphics decoder and a synchronization part perform along time
- stereoscopic video (“3D video”) refers to video that can be viewed stereoscopically by a user (viewer).
- a video signal including such a stereoscopic video is referred to as a stereoscopic video signal.
- the video playback device is a video playback device capable of outputting a stereoscopic video signal.
- the stereoscopic video signal is generated by a video playback device based on content data (for example, a video stream signal) recorded on an optical disc (video source) such as a BD-ROM.
- the video playback device generates a video of the main video (first video) and a graphics video or a subtitle video (second video) based on the content data acquired from the video source, and the first video is the first video.
- This video playback device is used for synthesizing at another synthesizing timing when the generation of the second video (caption video, graphics video, etc.) to be synthesized with the first video at a certain synthesizing timing is not in time for the synthesizing timing.
- the second video to be output is output as the second video at the synthesis timing.
- the video reproduction apparatus appropriately adjusts (depth correction) the depth (depth) of the second video. By doing so, the discomfort that the user feels with respect to the stereoscopic video due to the temporary loss of the second video is reduced.
- FIG. 1 is an overall configuration diagram of a stereoscopic video playback display system including a video playback device 12 according to the first embodiment.
- a video display device 11 is a display device such as a display for displaying a stereoscopic video.
- the video playback device 12 receives a data signal such as video from a storage medium such as an optical disc such as a BD-ROM, a network such as the Internet, and a broadcast wave (video source), and plays back and outputs a stereoscopic video signal.
- the stereoscopic glasses 13 are polarizing glasses that include liquid crystal shutters that appropriately block stereoscopic images displayed by the video display device 11.
- FIG. 2 is a diagram showing details of the configuration of the video display device 11 in each embodiment described in this specification.
- the video display device 11 includes a first input / output unit 101, a first AV processing unit 102, a display unit 103, a first remote control signal reception unit 104, a transmission unit 105, and the like, and includes a first remote control signal reception unit.
- the operation starts in response to an operation from the user.
- the stereoscopic video signal output from the video playback device 12 is input to the first input / output unit 101.
- the first input / output unit 101 may be configured by a signal input unit compatible with HDMI (High Definition Multimedia Interface) Ver1.4.
- the stereoscopic video signal is displayed on the display unit 103 after the resolution of the video and the timing of the audio are adjusted in the first AV processing unit 102.
- the transmission unit 105 sends a switching signal for controlling the opening / closing of the shutters of the left and right lenses to the stereoscopic glasses 13 worn by the user.
- FIG. 3 is a diagram illustrating a configuration of the video reproduction device 12 according to the present embodiment.
- the disk drive unit 201 is a disk drive device into which a disk 200 such as a BD-ROM can be inserted, and outputs data recorded on the disk 200 to the second AV processing unit 202 in response to an instruction from the CPU 205.
- the second AV processing unit 202 receives a playback instruction from the user through the second remote control signal receiving unit 204, acquires content data recorded on the disc 200 from the disc drive unit 201, and generates a stereoscopic video signal. And output to the second input / output unit 203.
- the CPU 205 receives a user instruction through the second remote control signal receiving unit 204 and outputs an instruction to the second AV processing unit 202 based on the instruction. In addition, the CPU 205 controls data transfer between the second AV processing unit 202 and the flash memory 206 and between the second AV processing unit 202 and the memory 207 as necessary. Second remote control signal receiving unit 204 receives a user instruction output from a remote control (not shown).
- the flash memory 206 stores software for operating the CPU 205 and the second AV processing unit 202, and data necessary for combining graphics video such as depth information (depth information) described later with the main video.
- a memory 207 is a memory for temporarily storing data such as video when decoding content data to generate a stereoscopic video signal.
- video data data of the main video
- graphics video second video
- graphics data data of graphics video
- graphics data data of graphics video
- subtitle data data such as subtitle data
- the second AV processing unit 202 decodes these data and develops them into a plurality of videos (planes), generates a synthesized video from these videos, and inputs a video signal (“synthesized video signal”) to the second input.
- output unit 203 output from the output unit 203.
- the video reproduction device 12 may include a recording medium such as a hard disk drive (HDD) in addition to the disk drive unit 201.
- the video playback device 12 further includes an antenna for receiving broadcast radio waves, a tuner that converts the received broadcast radio waves into digital stream signals (video, audio, graphics, subtitles, and the like). Also good.
- the video playback device 12 reads the content data recorded on the disc 200 to generate a main video and a graphics video, and synthesizes both videos to generate a stereoscopic video.
- a composite video signal indicating video is output.
- the composite video signal is a video signal including a left-eye video and a right-eye video.
- the video display device 11 displays the left-eye video and the right-eye video included in the input composite video signal on a display in a time-sharing manner.
- the time-sharing switching timing is set as an infrared synchronization signal (switching signal) in FIG. Tell the 3D glasses 13.
- the stereoscopic glasses 13 switch the transmittance of the liquid crystal shutter according to the synchronization signal received from the video display device 11, and alternately block the field of view of the left eye side or the right eye side of the stereoscopic glasses 13, thereby Present only to the left eye, and present the right eye image only to the right eye. In this way, stereoscopic image display is realized.
- FIG. 4 is a block diagram showing details of the second AV processing unit 202 in the first embodiment.
- the second AV processing unit 202 includes a stream control unit 501, a depth determination unit 502, a video decoder 503, a graphics decoder 504, a synthesis unit 505, and the like.
- the second remote control signal receiving unit 204 receives an instruction from the user via a remote control (not shown). In accordance with the instruction, the CPU 205 controls the second AV processing unit 202, the flash memory 206, the memory 207, and the like.
- the stream control unit 501 acquires video data (main image data), graphics data (graphics image data), and depth information recorded on the disc 200 from the disk drive unit 201. Then, the stream control unit 501 outputs video data to the video decoder 503, graphics data to the graphics decoder 504, and depth information to the depth determination unit 502.
- the video decoder 503 receives video data from the stream control unit 501.
- the video data is compressed, and the compressed video data is expanded to generate a video image (an image of the main image only).
- the graphics decoder 504 receives the graphics data recorded on the disc 200 from the stream control unit 501 and decompresses it to generate a graphics video (video of only graphics video).
- Most graphics data is composed of several parts such as an image of a button for prompting a user operation.
- the graphics decoder 504 generates a graphics video by decoding them and drawing (writing) on the memory 207.
- the graphics decoder 504 may create a graphics video using device-specific graphics data recorded in the flash memory 206.
- the disc 200 stores left-eye video and right-eye video of video data (main video data) and graphics data (graphics video data), respectively.
- main video data video data
- graphics data graphics data
- the main video is also described as a stereoscopic video, but the main video may be a non-stereo video. Even in that case, the composite video signal can be configured as a stereoscopic video signal.
- information on the depth position (depth) regarding the video data and the graphics data is also recorded on the disc 200 as depth information.
- the depth position information (depth information) may be recorded as average depth information of objects in the video data and graphics data.
- the depth information is recorded on the disc 200 separately from the video data and graphics data. To be recorded.
- the depth determination unit 502 calculates the depth of the graphics data from the depth information recorded on the disc 200, and outputs it to the synthesis unit 505 as depth data.
- the left-eye video and right-eye video are shifted in the horizontal direction on the display screen (offset adjustment) so that the apparent distance in the depth direction of the video is farther than the display position (display image position). Show or show closer to the display position.
- Data indicating the apparent depth position (depth) determined by the depth determination unit 502 is referred to as depth data.
- the synthesizing unit 505 synthesizes a plurality of planes such as a main video and graphics video to generate a synthesized video signal. That is, the synthesis unit 505 synthesizes the video video (main video) from the video decoder 503 and the graphics video from the graphics decoder 504 based on the depth data sent from the depth determination unit 502.
- Plane Compositing Model With reference to FIG. 5, a process from superimposing (combining) a graphics video on a main video and outputting a synthesized video will be described.
- the model shown in FIG. 5 is generally called a plane synthesis model.
- the left-eye video image (L video image) 305L and the right-eye video image (R video image) 305R (the main image image) from the video data 305S (# 1 # VIDEO STREAM) read from the disc 200 or the like.
- Left-eye video and right-eye video are generated and output to the combining unit 505.
- the graphics decoder 504 from the graphics data 307S (# 2 VIDEO STREAM) read from the disk 200 or the like, the left-eye graphics video (L graphics video) 307L and the right-eye graphics video (R graphics video) 307R (graphics) Left-eye video and right-eye video) are generated and output to the combining unit 505.
- the synthesizing unit 505 superimposes (synthesizes) the main video 305L and R and the graphics video 307L and R and outputs it as a synthesized video 313.
- the synthesis unit 505 can further superimpose the menu graphics 301 and the caption data 303 on the synthesized video.
- FIG. 6 is an example of a signal format used when the video reproduction device 12 outputs a video signal (synthesized video signal) of the synthesized video 313 as a stereoscopic video signal.
- the composite video signal 313SGL includes, for example, a video signal (L) for the left eye video and a video signal (R) for the right eye video alternately to form a video signal for one frame of the composite video.
- the video playback device 12 transmits the video signal (L) of the left eye video in 1/120 seconds, and transmits the video signal of the right eye video in the next 1/120 seconds.
- FIG. 8 is a diagram for explaining offset processing for a video.
- By performing an offset process on the video it becomes possible to adjust the depth (depth) of the video.
- depth depth
- a right-eye video and a left-eye video having parallax are used, and both videos are shifted in the opposite direction by a predetermined distance in the horizontal direction of the screen (by performing offset processing).
- Depth feeling distance in the depth direction
- the second AV processing unit 202 adjusts the depth of the video by changing the degree of offset (shift distance) based on the depth data described above.
- FIG. 8A is a diagram showing the left-eye video 401L and the right-eye video 401R when the degree of offset (offset amount) is zero.
- the apparent distance that the user feels with respect to an object (for example, a bus) in the video matches the video display surface.
- FIG. 8B shows a case where the offset 405 is given to the reference frame 403 (the position of the image when the offset amount is zero) so that the left-eye image and the right-eye images 401L and R are gathered in the center direction of the drawing.
- the user feels that the apparent distance of the object in the video is closer to the video display surface.
- FIG. 8C is a diagram showing a case where an offset 407 is given to the reference frame 403 so that the left eye video and the right eye video 401L and R are moved away from each other. At this time, the user feels that the apparent distance of the object in the video is farther than the video display surface.
- the depth of the video in the stereoscopic video can be adjusted by adjusting the magnitude of the offset given to the video and the direction of the offset.
- the left-eye image 401L and the right-eye image 401R may be different images in consideration of parallax or the same image.
- FIG. 9 is a block diagram showing details of the synthesis unit 505.
- the synthesizing unit 505 includes a graphics video acquiring unit 700 that acquires graphics video directly from the graphics decoder 504 or indirectly via the memory 207, and a graphics synthesizing unit 703 that combines the main video and the graphics video.
- the graphics video acquisition unit 700 includes a synchronization unit 701 that outputs graphics video in synchronization with the main video, and a depth provision unit 702 that adjusts the depth (depth) of the graphics video.
- the synchronization unit 701 outputs the graphics video at a predetermined output timing so as to be in time for the synthesis processing performed in the graphics synthesis unit 703 while synchronizing both videos, and the depth assigning unit 702 applies to the input graphics video. Then, the depth (depth) of the graphics video is appropriately adjusted based on the depth data output from the depth determination unit 502 described above, and the depth-corrected graphics video is output to the graphics synthesis unit 703.
- the graphics video decoded by the graphics decoder 504 is temporarily written (drawn) in the memory 207, and the synchronization unit 701 confirms that the left-eye video and the right-eye video of the graphics video are written in the memory 207.
- the graphics video written in the memory 207 is read out so as to be in time (the timing for performing the synthesis processing that is set so that the generation of the synthesized video is in time by the timing at which the synthesized video is output from the graphics synthesis unit 703).
- the graphics video is output to the depth assigning unit 702.
- the synchronization unit 701 of the graphics video acquisition unit 700 outputs the graphics video in time for the synthesis timing of the main video and the graphics video in the graphics synthesis unit 703. For example, normally, in 60 Hz NTSC, the left-eye video and the right-eye video of the graphics video are output to the depth assigning unit 702 every 1/120 seconds, that is, about every 8 mSec.
- the depth assigning unit 702 performs an offset process on the graphics video based on the depth data output from the depth determination unit 502, adjusts the depth of the graphics video, and outputs the depth-corrected graphics video to the graphics synthesis unit 703. .
- the graphics synthesis unit 703 generates a (stereo) synthesized video by synthesizing the main video and the graphics video input via the depth adding unit 702 in synchronization with the main video, and generates the (three-dimensional) synthesized video signal. 2 to the input / output unit 203.
- the depth assigning unit 702 when an offset is already added to the graphics video output by the graphics decoder 504, that is, when the graphics video is generated based on the graphics data recorded on the disc 200 by the 2-plane method or the like, usually, It is not necessary for the depth assigning unit 702 to add an offset to the graphics video.
- an offset is not added to the graphics video in advance, that is, the graphics video is generated and output by the graphics decoder 504 based on the graphics data recorded on the disc 200 by the 2-plane plus 1 offset method or the 1-plane plus 1 offset method. In such a case, the depth assigning unit 702 always performs an offset process on the graphics video based on the depth data.
- the depth-corrected graphics video output from the depth assigning unit 702 in time for the synthesis processing timing (combination timing) in the graphics synthesis unit 703 is synthesized with the main video in the graphics synthesis unit 703, and the playback timing of the synthesized video Is output as a composite video signal.
- the graphics decoder 504 decodes the graphics data and outputs the graphics video for the left eye within a limited time, Furthermore, it is necessary to decode the graphics data and output a graphics image for the right eye. Also, it takes time to confirm the synchronization between the main video and the graphics video. Therefore, due to the decoding performance and rendering performance of the graphics decoder 504, the left-eye video and the right-eye video to be synthesized in synchronism with the main video are displayed at a predetermined timing (for example, the graphics video output timing from the synchronization unit 701).
- the synchronization unit 701 cannot output the graphics video to be combined with the main video at the playback timing of the main video.
- the video reproduction device 12 is configured so that at least one of the left-eye video and the right-eye video of the graphics video is output at the timing at which the graphics video is to be output from the synchronization unit 701.
- the left-eye video and the right-eye video of the graphics video for another output timing are output as the graphics video to be output at the output timing.
- the video playback device 12 avoids output of video that the user feels uncomfortable even when the original graphics video is missing.
- the graphics video output from the graphics decoder 504 is a left-eye graphics video and a right-eye graphics video that have parallax and no offset. . Therefore, even during normal times (when the graphics video is normally sent from the graphics decoder 504), the depth assigning unit 702 performs an offset process on the graphics video based on the depth data of the graphics video of each frame (depth). The depth-corrected graphics video is output to the graphics composition unit 703.
- the depth assigning unit 702 is as follows.
- the graphics video at another output timing (graphic video of another frame) is offset (depth corrected) and output. To do.
- FIG. 10 is a flowchart of the graphics video output process performed by the graphics video acquisition unit 700 in the vicinity of the output timing in the time domain t (each time domain corresponds to one frame of the composite video signal).
- the synchronization unit 701 determines whether both the left-eye video (Lt) and the right-eye video (Rt) of the graphics video to be output at the graphics video output timing of the time domain t have already been input (whether they can be used). (S101).
- the synchronization unit 701 When both the left-eye video Lt and the right-eye video Rt are input (“YES” in S101), the synchronization unit 701 outputs the left-eye video Lt and the right-eye video Rt to the depth assigning unit 702 at appropriate timing (S102).
- the depth assigning unit 702 corrects the depth of the left-eye video Lt and the right-eye video Rt using the depth data Dt for the graphics video of the time domain t (offset processing), and depth-corrected left-eye video Lt ′, and
- the depth-corrected right-eye video Rt ′ is output to the graphics synthesis unit 703 (S103).
- the graphics synthesis unit 703 synthesizes the graphics video and the main video and outputs the synthesized video signal.
- the synchronization unit 701 displays the graphics video in the previous time domain t ⁇ 1. It is determined whether or not both the left-eye image Lt-1 and the right-eye image Rt-1 are input (can be used) (S104).
- the synchronization unit 701 When both the left-eye video Lt-1 and the right-eye video Rt-1 are input to the synchronization unit 701 (when available) (“YES” in S104), the synchronization unit 701 performs the left-eye video Lt at an appropriate timing. ⁇ 1 and the right eye image Rt ⁇ 1 are output to the depth assigning unit 702 (S105).
- the depth providing unit 702 uses the depth data Dt of the original graphics video in the time domain t for the input left-eye video Lt-1 and right-eye video Rt-1, and outputs the left-eye video Lt-1 and the right-eye video Rt-1.
- the depth-corrected left-eye video Lt-1 ′′ and the depth-added right-eye video Rt-1 ′′ are output to the graphics synthesis unit 703 (S106).
- the graphics synthesis unit 703 synthesizes the graphics video and the main video and outputs the synthesized video signal.
- the video playback device 12 can use at least one of the left-eye video and the right-eye video of the graphics video in the time domain in each time domain when it cannot be used for the synthesis process with the main video.
- the depth-corrected graphics video is generated by correcting the depth so that the graphics video of another time domain (different frame) is originally given the depth of the graphics video synthesized with the main video in the time domain,
- the synthesized video is output by synthesizing the depth-corrected graphics video and the main video.
- FIG. 11A shows a graphics decoder 504, a synchronization unit 701, and a depth in each time region (frame) when graphics data is recorded in a video source (content data of the disc 200) in a 2-plane plus 1 offset format. It is the figure which put together the output from the provision part 702, and the depth data D (output from the depth determination part 502). Here, it is assumed that the depth data is recorded on the disc 200 for each frame (one hour area) of graphics video (or main video).
- a graphics video indicated by a solid rectangle as an output from the graphics decoder means a graphics video that can be used by the synchronization unit 701 at a graphics video output timing of the synchronization unit 701 in each time domain, and a broken line A graphics image indicated by a rectangle indicates a graphics image that cannot be used by the synchronization unit 701 at the same output timing in each time domain.
- the graphics decoder 504 includes the left-eye video (L1, L2,%) And the right-eye video (R1, R2) of the graphics video so that the synchronization unit 701 can output the graphics video in time domain (TD1 to TD5). , ...) is output.
- L1, L2, etc. the graphics decoder 504 has not been able to output the right-eye video R4 (701) of the graphics video by a predetermined timing with respect to the time domain TD4.
- the synchronization unit 701 determines that there is no input of the right-eye video R4 (701), and in the time domain TD4, the left-eye video L3 (703) of the graphics video (of the previous frame) in the previous time domain TD3. ) And the right eye image R3 (705).
- the depth is corrected (offset processing) using the corresponding depth data D1 to D3 for the L3 and right eye images R1 to R3, and the left and right eye images (L1 ′ to L3 ′ and R1 ′) of the graphics image to which the depth is given. To R3 ′).
- the graphics video L3 and R3 are corrected (offset processing) using the depth data D4 of the time domain TD4, and the left and right eye videos L3 '' of the graphics video to which the depth is given. And R3 ′′ are output.
- the video reproduction device 12 when a graphics video is missing (cannot be used) in a certain time domain (frame), the left-eye video and the right-eye of the graphics video in another time domain (frame).
- a graphics video is missing (cannot be used) in a certain time domain (frame)
- the left-eye video and the right-eye of the graphics video in another time domain (frame).
- the depth of the graphics video changes smoothly with the passage of time, it is possible to output a 3D video that changes smoothly.
- depth data does not require decoding or the like unlike graphics data and video data, the possibility of missing depth data at a predetermined timing in each time domain is extremely low. If the depth data cannot be used, the depth data of the current time domain graphics video may be estimated using the depth data of the previous time domain graphics video.
- the depth information of the graphics video is not recorded on the disc 200 as the data for each frame (for each time region)
- other information for example, the depth data of the main video
- the depth data may be determined based on input from the user. For example, if the user wants to add effects such as a fade-out effect that pushes the displayed image far away and erases it to the graphics image, the instruction is received from the user and the fade-out recorded in the flash memory 206 or the like is received. Depth data may be obtained from the sequence and used.
- the depth information may be recorded as a sequence without being recorded every frame (every time region).
- the depth information may be sequence information such as moving away at 1 meter per second. Even in such a case, the depth data of the graphics video in each time region (frame) can be obtained by a simple calculation, and therefore the depth data of the graphics video in each frame may be calculated by the calculation.
- the graphics video in the previous time domain is output again, but it is re-outputted.
- the graphics image may not be the previous time domain graphics image. Two or more previous time-domain graphics images may be output again. Further, if available, a later time domain graphics image may be used.
- the synchronization unit 701 uses the previous time-domain graphics video and adds the current time-domain graphics to the graphics video. Depth correction (offset processing) is performed using the depth data of the video, and the left-eye video and right-eye video that have been depth-corrected are output. By doing so, the output of an abnormal graphics image is prevented, and the depth of the graphics image changes smoothly with the passage of time, so that a smoothly changing stereoscopic image can be output.
- the graphics video output from the graphics decoder 504 is a left-eye graphics video and a right-eye graphics video having a parallax and having an offset. Therefore, in normal times (when the graphics video is normally sent from the graphics decoder 504), the depth assigning unit 702 generally outputs the input graphics video without performing offset processing on the graphics video.
- the depth assigning unit 702 In a non-normal time (when the graphics video at the output timing cannot be used at the graphics video output timing of the synchronization unit 701 and a graphics video of another time domain is output from the synchronization unit 701), the depth assigning unit 702 As shown below, using the depth data of the original graphics video to be synthesized with the main video in the current time domain and the depth data of the graphics video of another time domain, offset processing is performed on the graphics video of the other time domain. Go (correct depth) and output as depth corrected graphics video.
- FIG. 11B shows a graphics decoder 504, a synchronization unit 701, and a depth assigning unit 702 in each time domain (frame) when graphics data is recorded in the video source (content data of the disc 200) in a 2-plane format.
- FIG. 6 is a diagram summarizing the output from, and depth data D (output from the depth determination unit 502).
- the graphics video indicated by the solid line rectangle as the output from the graphics decoder means a graphics video that can be used by the synchronization unit 701 at the graphics video output timing of the synchronization unit 701 in each time domain.
- the graphics video indicated by is a graphics video that cannot be used by the synchronization unit 701 at the same output timing in each time domain.
- the graphics decoder 504 cannot output the right-eye video R4 (701) of the graphics video by a predetermined timing in the time domain TD4.
- the synchronization unit 701 determines that there is no input of the right-eye video R4 (701), and in the time domain TD4, the left-eye video L3 (703) and the right-eye video R3 (705) of the graphics video in the previous time domain TD3. Is output at a predetermined output timing.
- the depth assigning unit 702 uses the graphics video depth data D3 of the time domain TD3 and the original graphics video depth data D4 of the time domain TD4 for the graphics video L3 and R3 (D4-D3). Only the depth is corrected (offset processing), and the left-eye image L3 ′ and the right-eye image R3 ′ of the graphics image to which the depth is added are output. Here, the depth correction is performed by the difference in depth (D4-D3) for the following reason.
- the graphics video L3 and R3 in the time domain TD3 has already been subjected to depth correction (offset addition) by only the depth D3.
- the depth assigning unit 702 reduces the depth of the graphics video L3 and R3 by the depth (D4-D3). Further correct and output.
- FIG. 12 is a diagram illustrating, along the time axis, an example of graphics video writing and reading processing when the memory 207 includes a memory area for three videos for each of the left-eye video and the right-eye video of the graphics video.
- the areas 1 to 3 (areas 4 to 6) of the graphics L memory (R memory) of the memory 207 are storage areas for temporarily storing the left-eye video (right-eye video) of the graphics video.
- the left and right eye videos L2 and R2 of the graphics video are drawn (written) in the graphics domain 1 and the graphics domain 3, and the graphics video L1 from the graphics domain 2 and the graphics domain 4 in the meantime.
- R1 are read out, input to the synchronization unit 701, and output from the synchronization unit 701.
- the left and right eye videos L3 and R3 of the graphics video are drawn in the graphics area 2 and the graphics area 4 by switching the graphics area every time the time domain advances in synchronization with the playback timing of the synthesized video.
- the left and right eye images L2 and R2 of the graphics video are read from the graphics area 1 and the graphics area 3, input to the synchronization unit 701, and output from the synchronization unit 701.
- the graphics video is not output in the drawing state, and the graphics video is synthesized in accordance with the timing of the main video, Composite video can be output.
- the synchronization unit 701 determines that there is no input of both the left-eye and right-eye images of the graphics image (“NO” in step S101 in FIG. 10), and the graphics images L3 and R3 in the previous time domain TD3 are determined.
- the data is output again (step S105 in FIG. 10).
- the left eye and right eye videos L5 and R5 of the graphics video in the time domain TD5 are written in the graphics areas 3 and 6, respectively.
- FIG. 13 is a diagram illustrating, along the time axis, an example of graphics video writing and reading processing when the memory 207 includes a memory area for two videos for each of the left-eye video and the right-eye video of the graphics video.
- the synchronization unit 701 determines that there is no input of both the left and right eye images of the graphics image (“NO” in step S101 in FIG. 10), and the graphics image in the previous time domain TD3. L3 and R3 are output again (step S105 in FIG. 10).
- the left eye and right eye videos L5 and R5 of the graphics video in the time domain TD5 are written in the graphics areas 1 and 3, respectively.
- FIG. 14 is a diagram illustrating another example of graphics video writing and reading processing along the time axis when the memory 207 includes a memory area for two videos for each of the left-eye video and the right-eye video of the graphics video. .
- the synchronization unit 701 determines that there is no input of both the left-eye and right-eye images of the graphics image (“NO” in step S101 in FIG. 10), and the graphics image L3 in the previous time domain TD3. And R3 are output again (step S105 in FIG. 10).
- the left-eye and right-eye videos L5 and R5 of the graphics video in the time domain TD5 are not written to the graphics areas 1 and 3 as in the previous example.
- the time domain TD4 only the re-output of the left-eye and right-eye videos L3 and R3 of the graphics video in the time domain TD3 is performed.
- the left and right eye videos L4 and R4 of the graphics video in the time domain TD4 are output, and the left and right eye videos L6 and R6 of the graphics video for the time domain TD6 are written.
- the images L6 and R6 are output.
- the graphics video of the previous time domain TDx-1 is re-outputted in the time domain TDx, and the output of the TDx graphics video is output in the time domain TDx + 1.
- the synchronization unit 701 may be configured to perform the above. Needless to say, even in the case of this example, the depth assigning unit 702 may correct the depth of the graphics video output by the synchronization unit 701 using the depth data for the current time domain.
- the video playback device 12 is a video playback device that synthesizes and outputs a first video (main video) and a second video (graphics video), which are three-dimensional video images, and outputs them in one frame.
- a synthesizing unit 505 that generates a synthesized video by synthesizing the video and outputs the synthesized video, and the synthesizing unit 505 includes a single video and graphics decoder 503 and 504 when the synthesized video is generated.
- the first video is The second image to be made to synthesize the first image to be synthesized into one frame.
- the video playback apparatus has a main video (first video) and a graphics video (second video) synthesized with the main video from content data recorded on an optical disc (video source) such as a BD-ROM. (Stereoscopic image)), and the two images are combined to output the combined image as a stereoscopic image. If at least one of the left-eye video and the right-eye video of the graphics video used for the synthesis cannot be used at the timing of synthesis of the main video and the graphics video, the video playback device uses another time domain (another synthesized video frame). The left-eye image and right-eye image of the graphics image used for composition are used for composition.
- the video playback device corrects the depth of the graphics video so that the depth of the graphics video matches the depth of the graphics video that should originally be used in the synthesis, and then combines the synthesized video with the main video. Output.
- the present video reproduction apparatus prevents the loss of graphics video and prevents unnatural fluctuations in the depth of the graphics video.
- the video playback device 12 according to the present embodiment even when the graphics video cannot be prepared at the playback timing of the main video to be synchronized, by using the previously output graphics video, Prevent abnormal graphics video output.
- the left-eye graphics video and the graphics video for the left eye are selected from the previous time domain (composite video frame) graphics videos.
- a graphics image in which both right-eye graphics images are available (available) is used. This prevents the user from feeling uncomfortable with the video due to the difference between the left-eye graphics video and the right-eye graphics video other than the parallax and the offset.
- the second AV processing unit 202 appropriately corrects and outputs the depth of the graphics video using depth information recorded on the disc 200 and depth data based on the depth information recorded on the flash memory 206. .
- the missing (unusable) graphics video is interpolated using the left-eye graphics video and the right-eye graphics video at the previous playback timing, for example, a graphics video whose depth continuously changes can be obtained.
- Even in the case of reproduction it is possible to output a 3D image with a smoothly changing depth without displaying a 3D image that feels uncomfortable.
- Embodiment 2 Another embodiment of the video reproduction apparatus will be described.
- the same processing as the processing for graphics video in the first embodiment is performed.
- the depth correction for the caption video is performed according to the depth information recorded on the disc 200 and an instruction input by the user.
- the same components as those of the first embodiment are denoted by the same reference numerals, and the description thereof is omitted as appropriate.
- FIG. 15 is a block diagram of the second AV processing unit 202 of the video reproduction device 12 according to the second embodiment.
- a subtitle decoder 506 is a decoder that decodes subtitle data input from the disc 200 and generates a subtitle video to be superimposed on the main video (video video).
- Subtitle video obtained by decoding by the subtitle decoder 506 can be switched on / off for superimposition on the main video by the CPU 205 sending an instruction to the synthesis unit 505 based on the received user operation.
- the second AV processing unit 202 appropriately corrects the depth of the subtitle video using the depth information recorded on the disc 200 and the depth data based on the depth information recorded on the flash memory 206 and outputs the corrected subtitle video. .
- FIG. 16 is a diagram illustrating details of the configuration of the second AV processing unit 202 of the video reproduction device 12 according to the first and second embodiments.
- the second AV processing unit 202 may be configured as one or more semiconductor devices.
- the second AV processing unit 202 inputs stream data including 3D content from a video source such as the disc 200, outputs the first video stream to the first decoder 553, and outputs the second video stream to the second video stream.
- a stream control unit 501 that outputs to the decoder 554 and outputs the depth information included in the stream data to the depth determination unit 502, and a depth determination unit 502 that determines the video depth of the 3D content based on the depth information and outputs it as depth data.
- a first decoder 553 that decodes the video stream of the first video and outputs the first video (main video), and outputs a second video (graphics video or subtitle video) by decoding the video stream of the second video
- a second decoder 554 for synthesizing and synthesizing the second video in synchronism with the first video and outputting a synthesized video signal Comprising the, the signal output section 555 for outputting a composite video signal.
- the synthesis unit 505 has a function equivalent to that of the synthesis unit 505 of the first to third embodiments.
- the signal output unit 555 may be an output terminal.
- the second AV processing unit 202 in the first and second embodiments can be configured as a semiconductor device.
- FIG. 17 is a block diagram of the second AV processing unit 202 of the video reproduction device 12 according to the third embodiment.
- a depth determination decoder 507 is a decoder for determining the depth of the video decoded by the graphics decoder 504.
- the depth determination decoder 507 obtains the average depth from the contents of the graphics video output from the graphics decoder 504, and outputs it as depth data.
- the graphics video output from the graphics decoder 504 is a stereoscopic video
- the distance (depth) in the depth direction of the stereoscopic video is proportional to the amount of left-right shift (offset amount) of the left-eye video of the graphics video.
- the decoder for depth determination 507 compares the left-eye video and the right-eye video of the graphics video, detects the shift amount between the two videos, and acquires the average shift amount as depth data.
- the depth determination decoder 507 can estimate (extrapolate) the depth data. For example, as shown in FIG. 18, when the depth data D1 to D3 can be calculated for the time domain TD1 to the time domain TD3, the depth determination decoder 507 sets the depth value of the time domain TD4 to the time domain TD1 to the time domain. it can be acquired as an external ⁇ degree data ExD4 by recursively calculated from the depth data D1 ⁇ d3 of TD3.
- the video playback device 12 is a video playback device that synthesizes and outputs a first video (main video) and a second video (graphics video) that are three-dimensional video, and outputs them in one frame.
- a synthesizing unit 505 that generates a synthesized video by synthesizing the video and outputs the synthesized video, and the synthesizing unit 505 includes a single video and graphics decoder 503 and 504 when the synthesized video is generated.
- the first video is The second image to be made to synthesize the first image to be synthesized into one frame.
- the depth when the second video is displayed as a stereoscopic video is determined by the depth information, and the synthesis unit 505 performs video when synthesizing the first video and the second video to be synthesized into one frame.
- the graphics decoders 503 and 504 fail to obtain the second video to be synthesized into one frame, the graphics decoders 503 and 504 are configured based on the content of the second video to be synthesized with the first video in another frame. Depth information for the second video to be combined with the first frame is obtained, and the second video to be combined with the first video in another frame is combined with the first video using the obtained depth information. .
- the depth determination decoder 507 detects the depth from the graphics video and outputs it as depth data. Therefore, if the generation of graphics video is not in time by the output timing of the graphics video from the synchronization unit 701 in a certain time domain, it is estimated (extrapolated) from the depth data of the previous graphics video in the time domain. The depth of the graphics video in another time domain (frame) is corrected using the depth data thus obtained, so that the graphics video is synthesized with the main video at the synthesis timing of the time domain.
- the video playback apparatus 12 is a stereoscopic video that changes smoothly with time without outputting an abnormal video even if depth information is not recorded separately from the video on a video source such as the disc 200. Can be realized.
- FIG. 19 is a diagram illustrating details of the configuration of the second AV processing unit 202 of the video reproduction device 12 according to the third embodiment.
- the second AV processing unit 202 may be configured as one or more semiconductor devices.
- the second AV processing unit 202 inputs stream data including 3D content from a video source such as the disc 200, outputs the first video stream to the first decoder 553, and outputs the second video stream to the second video stream.
- a stream control unit 501 that outputs to the decoder 554, a first decoder 553 that decodes the video stream of the first video and outputs the first video (main video), and a second video that decodes the video stream of the second video
- a second decoder 554 that outputs (graphics video or subtitle video)
- a depth determination decoder 507 that acquires depth data of the second video based on the content of the second video output by the second decoder 554, and a first A synthesizing unit 505 for synthesizing and synthesizing the second video with the video and outputting a synthesized video signal
- a signal output unit 555 for outputting the synthesized video signal; Equipped with a.
- the synthesis unit 505 has a function equivalent to that of the synthesis unit 505 of the first to third embodiments.
- the signal output unit 555 may be an output terminal.
- the second AV processing unit 202 of the third embodiment can also be configured as a semiconductor device.
- the video playback device 12 of the first to third embodiments can be variously modified. Hereinafter, an example of a modification of the video reproduction device 12 will be described.
- the synthesized video is not limited to these.
- a configuration is also possible in which another video image (sub-image) is output in synchronism with the video image (main image) and the two images are combined. Even in such a case, the main video and the sub-video are recognized as abnormal video by the user if the left-eye video and the right-eye video are not compatible.
- smooth stereoscopic video representation is possible by correcting and outputting the depth of the video while maintaining correspondence between the left-eye and right-eye videos.
- the graphics image may be an image in which the left-eye and right-eye images are the same. In that case, the graphics image looks planar unless the depth is adjusted, but by adjusting the depth, the graphics image can be localized in the depth direction.
- the sequence recorded in the flash memory 206 is shown.
- it is not limited to this and may be a memory card or the like memory 207. It doesn't matter.
- the depth data used for the depth correction of the graphics video may use information other than the depth information of the graphics video, for example, depth information of the main video. Even in that case, it is possible to achieve an equivalent effect.
- the depth correction may be performed by regularly performing offset processing.
- the graphics video in another time domain used when the graphics video is not ready in time may not be the graphics video in the previous time domain.
- content recorded on a BD-ROM is taken as an example as a video source.
- the video source is not limited to a BD-ROM, and a video source medium such as a broadcast wave, other optical disks, and a hard disk is used. Including all.
- the main video (first video) and the graphics video or subtitle video (second video) are generated from video data included in the same video source (content).
- the source and the video source (content) of the second video may be different from each other.
- the first video is a main video of a certain content, but the first video is not limited to this and may be an arbitrary video.
- the second video that is a stereoscopic video is not limited to the caption video or the graphics video, and may be an arbitrary video.
- the first video may be a stereoscopic video or a non-stereoscopic video.
- the video reproduction apparatus according to the embodiment can synthesize a second video that is a stereoscopic video with the first video and output the synthesized video as a stereoscopic video.
- the second video which is a stereoscopic video, may be recorded in an arbitrary format such as a 2-plane system, a 2-plane plus 1 offset system, or a 1-plane plus 1 offset system in the video source.
- the video playback device 12 of the present embodiment is not limited to an optical disc such as a BD-ROM, but can be applied to a stereoscopic video playback device for displaying a stereoscopic video by overlaying video and graphics, such as a television or a game machine.
- the video reproduction apparatus transmits a stereoscopic video signal by a frame sequential method
- the stereoscopic video signal method is not limited to this, and any method may be used.
- the stereoscopic video signal may be a video signal including a left-eye video (left-eye video) and a right-eye video (right-eye video) as a video signal for one frame.
- the video playback device of the present embodiment there is no particular limitation on the format of the stereoscopic video signal.
- the video playback device 12 of each embodiment can synthesize (superimpose) the second video ((three-dimensional) graphics video) on the first video.
- the video reproduction device 12 performs the synthesis between the first video and the second video.
- the video playback device 12 uses the second video in another time domain (frame) as the depth data of the second video in the current time domain (frame). Output with depth correction.
- the graphics video of Embodiments 1 and 3 and the caption video of Embodiment 2 are examples of stereoscopic graphics video.
- the video playback device 12 has the left eye of the stereoscopic graphics video in the time domain (frame) before that time domain (frame). Using the graphics video and the right-eye graphics video, the missing period is interpolated and output.
- the video playback device is useful as a video playback device that plays back a 3D video composed of a plurality of videos.
- Video display device 12 Video playback device 13 Stereoscopic glasses 101 First input / output unit 102 First AV processing unit 103 Display unit 104 First remote control signal receiving unit 105 Transmitting unit 200 Disc 201 Disc drive unit 202 Second AV Processing unit 203 Second input / output unit 204 Second remote control signal receiving unit 205 CPU 206 Flash memory 207 Memory 501 Stream control unit 502 Depth determination unit 503 Video decoder 504 Graphics decoder 505 Synthesis unit 506 Subtitle decoder 507 Depth determination decoder 700 Graphics signal acquisition unit 701 Synchronization unit 702 Depth addition unit 703 Graphics synthesis unit
Abstract
Description
1.立体映像再生表示システムの構成
実施の形態1による立体映像再生表示システムについて説明する。図1は、実施の形態1による映像再生装置12を備えた立体映像再生表示システムの全体構成図である。図1において、映像表示装置11は、立体映像を表示するディスプレイなどの表示装置である。映像再生装置12は、BD-ROMといった光ディスク等の記憶媒体、インターネット等のネットワ-ク、および、放送波など(映像ソース)から映像等のデータ信号を受け取り、立体映像信号を再生して出力する装置である。立体メガネ13は、映像表示装置11が表示する立体映像を適切に遮断する液晶シャッタを備えた偏光メガネである。
2.1.全体動作
以上のように構成された立体映像再生表示システムにおいて、映像再生装置12は、ディスク200に記録されたコンテンツのデータを読み出して本編映像およびグラフィクス映像を生成し、両映像を合成して立体映像を示す合成映像信号を出力する。合成映像信号は、左目映像および右目映像を含んでなる映像信号である。
次に、第2のAV処理部202の構成の詳細および第2のAV処理部202における本編映像とグラフィクス映像との合成処理について説明する。
図5を参照し、本編映像にグラフィクス映像を重畳(合成)して、合成映像を出力するまでのプロセスを説明する。図5に示されるモデルは、一般に、プレーン合成モデルと称される。
図6は、映像再生装置12が合成映像313の映像信号(合成映像信号)を立体映像信号として出力する場合に用いる信号フォーマットの例図である。合成映像信号313SGLには、例えば、左目映像の映像信号(L)と右目映像の映像信号(R)とが交互に含まれ、合成映像1フレーム分の映像信号が構成される。映像再生装置12は、60フレーム/秒で合成映像を出力する場合、1/120秒で左目映像の映像信号(L)を伝送し、次の1/120秒で右目映像の映像信号を伝送する。
図7を参照し、立体映像である合成映像313におけるデプスについて説明する。図7では、ユーザが立体映像を鑑賞した場合に認識する立体感(奥行き方向の拡がり)を図面左右方向(デプス方向)で表現している。本図が例示する立体映像(合成映像313)では、本編映像305として背景とその手前(ユーザ側)に立つ人が含まれる。合成映像313では、本編映像305の手前(ユーザ側)にグラフィクス映像307が重畳されている。同図で、ユーザが感じるグラフィクス映像307の奥行き方向(図面左右方向)見かけ上の距離を矢印315の方向に変化させることが、グラフィクス映像307のデプスを調整することに対応する。
これより、合成部505における、グラフィクス映像に対するデプス調整処理について説明する。
グラフィクス映像取得部700の同期部701は、グラフィクス合成部703における本編映像とグラフィクス映像との合成のタイミングに間に合うようにグラフィクス映像を出力する。例えば、通常、60HzNTSCでは、グラフィクス映像の左目映像および右目映像は、1/120秒ごとに、つまり約8mSecごとに、深度付与部702へ出力される。深度付与部702は、深度決定部502から出力された深度データにもとづき、グラフィクス映像に対してオフセット処理を行い、グラフィクス映像のデプスを調整して深度補正済グラフィクス映像をグラフィクス合成部703へ出力する。グラフィクス合成部703は、本編映像と、本編映像に同期して深度付与部702を介して入力されるグラフィクス映像とを合成して(立体)合成映像を生成し、(立体)合成映像信号を第2の入出力部203へ出力する。
グラフィクス映像を立体映像として生成する場合、グラフィクスデコーダ504は、限られた時間内で、グラフィクスデータをデコードして左目用のグラフィクス映像を出力し、さらに、グラフィクスデータをデコードして右目用のグラフィクス映像を出力する必要がある。また、本編映像とグラフィクス映像の同期を確認するための時間も必要である。そのため、グラフィクスデコーダ504のデコード性能、描画性能に起因して、本編映像に同期して合成すべきグラフィクス映像の左目映像および右目映像が、所定のタイミング(例えば、同期部701からのグラフィクス映像出力タイミング)までに同期部701またはメモリ207に入力されない事態が起こり得る。また、グラフィクス映像の左目映像および右目映像の一方しか、所定のタイミングまでに同期部701またはメモリ207に入力されない事態も起こり得る。後者の場合であっても、その入力された片目分のグラフィクス映像を使用することはできない。なぜなら、例えば、コンテンツ内に2プレーン方式で記録されグラフィクスデコーダ504で生成されたグラフィクス映像では、左目映像と右目映像とに視差をつけて、立体視を実現している。そのため、左目用のグラフィクス映像と、右目用のグラフィクス映像との両方がそろっていないと、グラフィクス映像に関して視差以外の情報差がユーザに与えられてしまい、ユーザはグラフィクスを十分な立体像として認識できないだけでなく、突然、深さ方向の位置の認識間違いが起こり、異常な映像として認識する。以上のような事態が生じると、同期部701は、本編映像の再生タイミングにおいて、本編映像に合成されるべきグラフィクス映像を出力することができない。
以下、グラフィクスデータが、コンテンツのデータ内に2プレーンプラス1オフセット方式で記録されている場合の処理について説明する。
次に、グラフィクスデータが、コンテンツのデータ内に1プレーンプラス1オフセット方式で記録されている場合の処理について説明する。
最後に、グラフィクスデータが、コンテンツのデータ内に2プレーン方式で記録されている場合の処理について説明する。
図12は、メモリ207がグラフィクス映像の左目映像および右目映像それぞれについて映像3枚分のメモリ領域を備えている場合におけるグラフィクス映像の書き込みおよび読み出し処理の例を時間軸に沿って示す図である。メモリ207のグラフィクスLメモリ(Rメモリ)の領域1~3(領域4~6)はそれぞれ、グラフィクス映像の左目映像(右目映像)を一時的に記憶するための記憶領域である。
図13は、メモリ207がグラフィクス映像の左目映像および右目映像それぞれについて映像2枚分のメモリ領域を備えている場合におけるグラフィクス映像の書き込みおよび読み出し処理の例を時間軸に沿って示す図である。
図14は、メモリ207がグラフィクス映像の左目映像および右目映像それぞれについて映像2枚分のメモリ領域を備えている場合におけるグラフィクス映像の書き込みおよび読み出し処理の別例を時間軸に沿って示す図である。
映像再生装置の別の実施の形態について説明する。本実施の形態においては、字幕データのデコードが同期すべき本編映像と合成処理されるタイミングまでに用意できなかった場合に、実施の形態1におけるグラフィクス映像に対する処理と同様の処理を行い、以前に表示した字幕映像を再度表示する。字幕映像に対する深度補正は、ディスク200に記録されている深度情報や、ユーザが入力する指示に従って行われる。なお、実施の形態2の説明においては、実施の形態1と同様の構成については同一の符号をつけ説明を適宜省略する。
図16は、実施の形態1および2の映像再生装置12の第2のAV処理部202の構成の詳細を示す図である。第2のAV処理部202は、例えば、1つまたは2つ以上の半導体装置として構成されてよい。第2のAV処理部202は、ディスク200等映像ソースから3Dコンテンツを含んだストリームデータを入力し、第1映像の映像ストリームを第1デコーダ553へ出力し、第2映像の映像ストリームを第2デコーダ554へ出力し、ストリームデータに含まれる深度情報を深度決定部502へ出力するストリーム制御部501と、深度情報に基づいて3Dコンテンツの映像の深度を決定し深度データとして出力する深度決定部502と、第1映像の映像ストリームをデコードして第1映像(本編映像)を出力する第1デコーダ553と、第2映像の映像ストリームをデコードして第2映像(グラフィクス映像または字幕映像)を出力する第2デコーダ554と、第1映像に第2映像を同期させて合成し合成映像信号を出力する合成部505と、合成映像信号を出力する信号出力部555と、を備える。ここで、合成部505は、実施の形態1~3の合成部505と同等の機能を有する。なお、信号出力部555は、出力端子でよい。
映像再生装置のさらに別の実施の形態について説明する。本実施の形態では、ディスク200等の映像ソース中に深度情報が映像と別のデータとして記録されていなくても、グラフィクス映像の深度データを取得することができる映像再生装置の構成を説明する。なお、実施の形態3の説明においては、実施の形態1および2と同様の構成については同一の符号をつけ説明を適宜省略する。
本実施の形態1~3の映像再生装置12は、様々に変形可能である。以下、映像再生装置12の変形の一例について説明する。
以上、各実施の形態の映像再生装置12は、第1映像に対し、第2映像((立体)グラフィクス映像)を合成(重畳)可能である。当該合成に際し、映像再生装置12は、第1映像と第2映像との間の同期をとって合成する。第2映像((立体)グラフィクス映像)に欠落があった場合、映像再生装置12は、別の時間領域(フレーム)の第2映像を現在の時間領域(フレーム)の第2映像の深度データで深度補正して出力する。
12 映像再生装置
13 立体メガネ
101 第1の入出力部
102 第1のAV処理部
103 表示部
104 第1のリモコン信号受信部
105 送信部
200 ディスク
201 ディスクドライブ部
202 第2のAV処理部
203 第2の入出力部
204 第2のリモコン信号受信部
205 CPU
206 フラッシュメモリ
207 メモリ
501 ストリーム制御部
502 深度決定部
503 ビデオデコーダ
504 グラフィクスデコーダ
505 合成部
506 字幕デコーダ
507 深度決定用デコーダ
700 グラフィクス信号取得部
701 同期部
702 深度付加部
703 グラフィクス合成部
Claims (6)
- 立体映像である第1の映像と第2の映像とを合成して出力する映像再生装置であって、
一のフレームに合成されるべき第1の映像と第2の映像を取得する取得手段と、
前記取得手段により取得した第1の映像と第2の映像を同期して合成することで合成映像を生成し、前記合成映像を出力する合成手段とを備え、
前記合成手段は、前記合成映像の生成時において、前記取得手段が前記一のフレームに合成されるべき第2の映像を取得できなかったときに、前記一のフレームとは別のフレームにおいて第1の映像と合成されるべき第2の映像を前記一のフレームに合成されるべき第1の映像と合成する、
映像再生装置。 - 前記第2の映像が立体映像表示されたときの奥行きが深度情報により決定され、
前記合成手段は、前記一のフレームに合成されるべき第1の映像と第2の映像の合成時において、前記取得手段が前記一のフレームに合成されるべき第2の映像を取得できなかったときに、前記一のフレームに合成されるべき第2の映像に対する深度情報を用いて前記別のフレームにおいて第1の映像と合成されるべき第2の映像を第1の映像と合成する、
請求項1記載の映像再生装置。 - 前記第2の映像が立体映像表示されたときの奥行きが深度情報により決定され、
前記合成手段は、前記一のフレームに合成されるべき第1の映像と第2の映像の合成時において、前記取得手段が前記一のフレームに合成されるべき第2の映像を取得できなかったときに、前記別のフレームにおいて第1の映像と合成されるべき第2の映像の内容に基づいて前記一のフレームに合成されるべき第2の映像に対する深度情報を求め、その求めた深度情報を用いて前記別のフレームにおいて第1の映像と合成されるべき第2の映像を第1の映像と合成する、
請求項1記載の映像再生装置。 - 前記第2の映像はグラフィックス映像または字幕映像である、請求項1記載の映像再生装置。
- 立体映像である第1の映像と第2の映像とを合成して出力する半導体装置であって、
一のフレームに合成されるべき第1の映像と第2の映像を取得する取得手段と、
前記取得手段により取得した第1の映像と第2の映像を同期して合成することで合成映像を生成する合成手段と、
前記合成映像を出力する信号出力手段とを備え、
前記合成手段は、前記合成映像の生成時において、前記取得手段が前記一のフレームに合成されるべき第2の映像を取得できなかったときに、前記一のフレームとは別のフレームにおいて第1の映像と合成されるべき第2の映像を前記一のフレームに合成されるべき第1の映像と合成する、
半導体装置。 - 映像再生装置において立体映像である第1の映像と第2の映像とを合成して出力する映像再生方法であって、
取得手段が、一のフレームに合成されるべき第1の映像と第2の映像を取得するステップと、
合成手段が、前記取得するステップにより取得した第1の映像と第2の映像を同期して合成することで合成映像を生成し、前記合成映像を出力するステップとを備え、
前記合成手段は、前記合成映像の生成時において、前記取得手段が前記一のフレームに合成されるべき第2の映像を取得できなかったときに、前記一のフレームとは別のフレームにおいて第1の映像と合成されるべき第2の映像を前記一のフレームに合成されるべき第1の映像と合成する、
映像再生方法。
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/388,708 US20120128327A1 (en) | 2009-08-05 | 2010-08-05 | Video reproducing device |
EP10806240A EP2464129A4 (en) | 2009-08-05 | 2010-08-05 | VIDEO PLAYBACK DEVICE |
CN2010800339078A CN102474663A (zh) | 2009-08-05 | 2010-08-05 | 影像再现装置 |
JP2011525793A JPWO2011016240A1 (ja) | 2009-08-05 | 2010-08-05 | 映像再生装置 |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2009182151 | 2009-08-05 | ||
JP2009-182151 | 2009-08-05 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2011016240A1 true WO2011016240A1 (ja) | 2011-02-10 |
Family
ID=43544149
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2010/004934 WO2011016240A1 (ja) | 2009-08-05 | 2010-08-05 | 映像再生装置 |
Country Status (6)
Country | Link |
---|---|
US (1) | US20120128327A1 (ja) |
EP (1) | EP2464129A4 (ja) |
JP (1) | JPWO2011016240A1 (ja) |
CN (1) | CN102474663A (ja) |
TW (1) | TW201123844A (ja) |
WO (1) | WO2011016240A1 (ja) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2012118231A1 (en) * | 2011-03-02 | 2012-09-07 | Sharp Kabushiki Kaisha | Method and device for displaying a pair of stereoscopic images |
CN103517062A (zh) * | 2012-06-15 | 2014-01-15 | 晨星软件研发(深圳)有限公司 | 影像的同步方法及其装置 |
WO2018079081A1 (ja) * | 2016-10-25 | 2018-05-03 | ソニーセミコンダクタソリューションズ株式会社 | 表示制御装置、電子機器、表示制御装置の制御方法、および、プログラム |
Families Citing this family (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9341464B2 (en) | 2011-10-17 | 2016-05-17 | Atlas5D, Inc. | Method and apparatus for sizing and fitting an individual for apparel, accessories, or prosthetics |
WO2013058985A1 (en) | 2011-10-17 | 2013-04-25 | Kimmel Zebadiah M | Method and apparatus for detecting deterioration of health status |
US9817017B2 (en) * | 2011-10-17 | 2017-11-14 | Atlas5D, Inc. | Method and apparatus for monitoring individuals while protecting their privacy |
TWI520577B (zh) * | 2012-08-10 | 2016-02-01 | 晨星半導體股份有限公司 | 立體影像輸出裝置與相關的立體影像輸出方法 |
CN103686116B (zh) * | 2012-09-04 | 2016-04-20 | 晨星软件研发(深圳)有限公司 | 立体影像输出装置与相关的立体影像输出方法 |
US9600993B2 (en) | 2014-01-27 | 2017-03-21 | Atlas5D, Inc. | Method and system for behavior detection |
KR102237479B1 (ko) * | 2014-06-03 | 2021-04-07 | (주)아이리스아이디 | 홍채 인식 단말기 및 방법 |
US10013756B2 (en) | 2015-03-13 | 2018-07-03 | Atlas5D, Inc. | Methods and systems for measuring use of an assistive device for ambulation |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH0750782A (ja) * | 1993-08-06 | 1995-02-21 | Mitsubishi Electric Corp | 字幕信号復調回路 |
JPH11164329A (ja) * | 1997-11-27 | 1999-06-18 | Toshiba Corp | 立体映像表示装置 |
JP2002074393A (ja) * | 2000-05-31 | 2002-03-15 | Sony Computer Entertainment Inc | 画像描画方法、画像描画装置、記録媒体及びプログラム |
JP2006084964A (ja) | 2004-09-17 | 2006-03-30 | Seiko Epson Corp | 立体画像表示装置 |
JP2008016946A (ja) * | 2006-07-03 | 2008-01-24 | Matsushita Electric Ind Co Ltd | 映像再生装置、映像再生方法および映像再生プログラム |
Family Cites Families (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1558664A (zh) * | 2004-01-17 | 2004-12-29 | 上海交通大学 | 一种高清晰数字电视节目字幕与背景图像的合成方法 |
JP3746506B2 (ja) * | 2004-03-08 | 2006-02-15 | 一成 江良 | 立体視化パラメータ埋込装置及び立体視画像再生装置 |
JP4730183B2 (ja) * | 2006-04-17 | 2011-07-20 | 株式会社日立製作所 | 映像表示装置 |
US20080226128A1 (en) * | 2007-03-12 | 2008-09-18 | Conversion Works, Inc. | System and method for using feature tracking techniques for the generation of masks in the conversion of two-dimensional images to three-dimensional images |
JP2009135686A (ja) * | 2007-11-29 | 2009-06-18 | Mitsubishi Electric Corp | 立体映像記録方法、立体映像記録媒体、立体映像再生方法、立体映像記録装置、立体映像再生装置 |
US8471898B2 (en) * | 2008-02-05 | 2013-06-25 | Disney Enterprises, Inc. | Medial axis decomposition of 2D objects to synthesize binocular depth |
-
2010
- 2010-08-05 JP JP2011525793A patent/JPWO2011016240A1/ja not_active Withdrawn
- 2010-08-05 US US13/388,708 patent/US20120128327A1/en not_active Abandoned
- 2010-08-05 EP EP10806240A patent/EP2464129A4/en not_active Withdrawn
- 2010-08-05 TW TW099126077A patent/TW201123844A/zh unknown
- 2010-08-05 CN CN2010800339078A patent/CN102474663A/zh active Pending
- 2010-08-05 WO PCT/JP2010/004934 patent/WO2011016240A1/ja active Application Filing
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH0750782A (ja) * | 1993-08-06 | 1995-02-21 | Mitsubishi Electric Corp | 字幕信号復調回路 |
JPH11164329A (ja) * | 1997-11-27 | 1999-06-18 | Toshiba Corp | 立体映像表示装置 |
JP2002074393A (ja) * | 2000-05-31 | 2002-03-15 | Sony Computer Entertainment Inc | 画像描画方法、画像描画装置、記録媒体及びプログラム |
JP2006084964A (ja) | 2004-09-17 | 2006-03-30 | Seiko Epson Corp | 立体画像表示装置 |
JP2008016946A (ja) * | 2006-07-03 | 2008-01-24 | Matsushita Electric Ind Co Ltd | 映像再生装置、映像再生方法および映像再生プログラム |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2012118231A1 (en) * | 2011-03-02 | 2012-09-07 | Sharp Kabushiki Kaisha | Method and device for displaying a pair of stereoscopic images |
CN103517062A (zh) * | 2012-06-15 | 2014-01-15 | 晨星软件研发(深圳)有限公司 | 影像的同步方法及其装置 |
WO2018079081A1 (ja) * | 2016-10-25 | 2018-05-03 | ソニーセミコンダクタソリューションズ株式会社 | 表示制御装置、電子機器、表示制御装置の制御方法、および、プログラム |
US10867587B2 (en) | 2016-10-25 | 2020-12-15 | Sony Semiconductor Solutions Corporation | Display control apparatus, electronic equipment, and control method of display control apparatus |
Also Published As
Publication number | Publication date |
---|---|
TW201123844A (en) | 2011-07-01 |
US20120128327A1 (en) | 2012-05-24 |
JPWO2011016240A1 (ja) | 2013-01-10 |
EP2464129A1 (en) | 2012-06-13 |
EP2464129A4 (en) | 2012-08-01 |
CN102474663A (zh) | 2012-05-23 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2011016240A1 (ja) | 映像再生装置 | |
EP2380357B1 (en) | Method and device for overlaying 3d graphics over 3d video | |
JP4959830B2 (ja) | 映像再生装置 | |
JP5357317B2 (ja) | 立体視映像再生装置 | |
CN103024408B (zh) | 立体图像转换装置和方法、立体图像输出装置 | |
WO2011135857A1 (ja) | 画像変換装置 | |
JP4875127B2 (ja) | 三次元画像処理装置 | |
JP2011172216A (ja) | 再生装置 | |
US8704876B2 (en) | 3D video processor and 3D video processing method | |
JP5066244B2 (ja) | 映像再生装置及び映像再生方法 | |
EP2408214A2 (en) | Playback apparatus, playback method, and program | |
JP5400467B2 (ja) | 映像出力装置、その制御方法、及びプログラム | |
WO2012014489A1 (ja) | 映像信号処理装置及び映像信号処理方法 | |
JP2011035858A (ja) | 映像処理装置 | |
JP2011114378A (ja) | 映像表示システム及びその制御方法 | |
WO2012017687A1 (ja) | 映像再生装置 | |
JP5422597B2 (ja) | 三次元映像処理装置 | |
WO2011074070A1 (ja) | 画像出力装置 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 201080033907.8 Country of ref document: CN |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 10806240 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 13388708 Country of ref document: US Ref document number: 2011525793 Country of ref document: JP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2010806240 Country of ref document: EP |