WO2012156940A1 - Method for generating, transmitting and receiving stereoscopic images, and related devices - Google Patents
Method for generating, transmitting and receiving stereoscopic images, and related devices Download PDFInfo
- Publication number
- WO2012156940A1 WO2012156940A1 PCT/IB2012/052486 IB2012052486W WO2012156940A1 WO 2012156940 A1 WO2012156940 A1 WO 2012156940A1 IB 2012052486 W IB2012052486 W IB 2012052486W WO 2012156940 A1 WO2012156940 A1 WO 2012156940A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- images
- pixels
- composite
- regions
- Prior art date
Links
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/44—Receiver circuitry for the reception of television signals according to analogue transmission standards
- H04N5/445—Receiver circuitry for the reception of television signals according to analogue transmission standards for displaying additional information
- H04N5/45—Picture in picture, e.g. displaying simultaneously another television channel in a region of the screen
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/161—Encoding, multiplexing or demultiplexing different image signal components
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N2213/00—Details of stereoscopic systems
- H04N2213/005—Aspects relating to the "3D+depth" image format
Definitions
- the present invention concerns the generation, storage, transmission, reception and reproduction of stereoscopic video streams, i.e. video streams which, when appropriately processed in a visualization device, produce sequences of images which are perceived as being three-dimensional by a viewer.
- stereoscopic video streams i.e. video streams which, when appropriately processed in a visualization device, produce sequences of images which are perceived as being three-dimensional by a viewer.
- the perception of three-dimensionality can be obtained by reproducing two images, one for the viewer's right eye and the other for the viewer's left eye.
- a stereoscopic video stream therefore transports information about two sequences of images, corresponding to the right and left perspectives of an object or a scene.
- the invention relates in particular to a method and a device for multiplexing the two images of the right and left perspectives (hereafter referred to as right image and left image) within a composite image which represents a frame of the stereoscopic video stream, hereafter also referred to as container frame.
- the invention also relates to a method and a device for de-multiplexing said composite image, i.e. for extracting therefrom the right and left images entered by the multiplexing device.
- a first example is the so-called side-by-side multiplexing, wherein the right image and the left image are sub-sampled horizontally and are arranged side by side in the same frame of a stereoscopic video stream.
- This type of multiplexing has the drawback that the horizontal resolution is halved while the vertical resolution is left unchanged.
- top-bottom multiplexing Another example is the so-called top-bottom multiplexing, wherein the right image and the left image are sub-sampled vertically and are arranged one on top of the other in the same frame of a stereoscopic video stream.
- This type of multiplexing has the drawback that the vertical resolution is halved while the horizontal resolution is left unchanged.
- This method allows the ratio between horizontal and vertical resolution to be kept constant, but it reduces the diagonal resolution and also alters the correlation among the pixels of the image by introducing high-frequency spatial spectral components which would otherwise be absent. This may reduce the efficiency of the subsequent compression step (e.g. MPEG2 or MPEG4 or H.264 compression) while also increasing the bit-rate of the compressed video stream.
- the subsequent compression step e.g. MPEG2 or MPEG4 or H.264 compression
- One of these methods provides for executing a 70% scaling of the right and left images; the scaled images are then broken up into blocks of 8x8 pixels.
- the blocks of each scaled image can be compacted into an area equal to approximately half the composite image.
- This method has the drawback that the redistribution of the blocks modifies the spatial correlation among the blocks that compose the image by introducing high-frequency spatial spectral components, thereby reducing compression efficiency.
- Another of these methods applies diagonal scaling to each right and left image, so that the original image is deformed into a parallelogram.
- the two parallelograms are then broken up into triangular regions, and a rectangular composite image is composed wherein the triangular regions obtained by breaking up the two parallelograms are reorganized and rearranged.
- the triangular regions of the right and left images are organized in a manner such that they are separated by a diagonal of the composite image.
- this solution also suffers from the drawback of altering the ratio (balance) between horizontal and vertical resolution.
- the subdivision into a large number of triangular regions rearranged within the stereoscopic frame causes the subsequent compression step (e.g. MPEG2, MPEG4 or H.264), prior to transmission on the communication channel, to generate artifacts in the boundary areas between the triangular regions.
- Said artifacts may, for example, be produced by a motion estimation procedure carried out by a compression process according to the H.264 standard.
- a further drawback of this solution concerns the computational complexity required by the operations for scaling the right and left images, and by the following operations for segmenting and rototranslating the triangular regions.
- Said method is related to the subdivision of the other image into three rectangular regions, and on how to arrange said three regions in the composite image.
- the general idea at the basis of the present invention is to enter two images into a composite image whose number of pixels is greater than or equal to the sum of the pixels of the two images to be multiplexed, e.g. the right image and the left image.
- the pixels of the first image (e.g. the left image) are entered into the composite image without undergoing any changes, whereas the second image is subdivided into two regions whose pixels are arranged in free areas of the composite image.
- This solution offers the advantage that one of the two images is left unchanged, which results in better quality of the reconstructed image.
- the second image is broken up into two regions, so as to maximize the spatial correlation among the pixels and to reduce the generation of artifacts during the compression phase.
- Subdividing one of the two stereoscopic images into three regions prevents most of the existing decoders from reconstructing the image without the addition of ad hoc functions, due to the lack of appropriate resources; reducing the subdivision into two regions may allow existing decoders with Picture in Picture (PIP) functionality to use it for reassembling the image thus reducing the amount of software changes needed to implement the invention in current decoders.
- PIP Picture in Picture
- pixels of said right image (R) and pixels of said left image are selected, and said selected pixels are entered into a composite image of said stereoscopic video stream, the method being characterized in that all the pixels of said right image and all the pixels of said left image are entered into different positions in said composite image, by leaving one of said two images unchanged and breaking up the other one into two regions (Rl, R2) comprising a plurality of pixels and entering said regions into said composite image.
- Further objects of the present invention are a method for reconstructing a pair of images by starting from a composite image, a device for generating composite images, a device for reconstructing a pair of images starting from a composite image, and a stereoscopic video stream.
- Fig. 1 shows a block diagram of a device for multiplexing the right image and the left image into a composite image
- Fig. 2 is a flow chart of a method executed by the device of Fig. 1 ;
- Fig. 3 shows a first phase of constructing a composite image according to one embodiment of the present invention
- Fig. 4 shows a first form of disassembly of an image to be entered into a composite image
- Fig. 5a and 5b show a first and a second form of a composite image that includes the image of Fig.4.
- Fig. 6 shows a second form of disassembly of an image to be entered into a composite image.
- Fig. 7a and 7b show a first and a second form of a composite image that includes the image of Fig. 6.
- Fig. 8 shows a third form of disassembly of an image to be entered into a composite image.
- Fig. 9a and 9b show a first and a second form of a composite image that includes the image of Fig. 8.
- Fig. 10 shows a fourth form of disassembly of an image to be entered into a composite image.
- Fig. 1 la and l ib show a first and a second form of a composite image that includes the image of Fig. 10.
- Fig. 12 shows a boundary region of the disassembled image to be replied in the composite image.
- Fig. 13 shows a possible way to place the boundary region of Fig 12 in the composite image.
- Fig. 14 shows what sub-region of the boundary region of the figures 12 and 13 can be extracted from the composite image.
- Fig. 15 shows how the sub-region of Fig. 14 can be overwritten in the reassembled image for eliminating the artifacts in the reconstructed image after reassembling.
- Fig. 16 shows a block diagram of a receiver for receiving a composite image generated according to the method of the present invention.
- Fig. 17 shows some phases of reconstructing the left and right images contained in a composite image according to any form shown in the previous figures.
- Fig. 1 shows the block diagram of a device 100 for generating a stereoscopic video stream 101.
- the device 100 receives two sequences of images 102 and 103, e.g. two video streams, intended for the left eye (L) and for the right eye (R), respectively.
- two sequences of images 102 and 103 e.g. two video streams, intended for the left eye (L) and for the right eye (R), respectively.
- the device 100 allows to implement a method for multiplexing two images of the two sequences 102 and 103.
- the device 100 comprises a disassembler module 104 for breaking up an input image (the right image in the example of Fig. 1) into two sub-images, each corresponding to one region of the received image, and an assembler module 105 capable of entering the pixels of received images into a single composite image to be provided at its output.
- a disassembler module 104 for breaking up an input image (the right image in the example of Fig. 1) into two sub-images, each corresponding to one region of the received image, and an assembler module 105 capable of entering the pixels of received images into a single composite image to be provided at its output.
- step 200 The method starts in step 200. Subsequently (step 201), one of the two input images (right or left) is broken up into two regions, as shown in Fig. 3.
- the disassembled image is a frame R of a video stream 720p, i.e. a progressive format with a resolution of 1280 x 720 pixels.
- the frame R of Fig. 3 comes from the video stream 103 which carries the images intended for the right eye, and is disassembled into two regions Rl and R2.
- the disassembly of the image R is obtained by dividing it into two parts.
- the rectangular region Rl has a size of 640x360 pixels and is obtained by taking the first 640 pixels of the first 360 rows.
- the region R2 is L-shaped, and is obtained by taking the pixels from 641 to 1280 of the first 360 rows and all the pixels of the last 360 rows.
- the operation of disassembling the image R is carried out by the module 104, which receives an input image R (in this case the frame R) and outputs two sub-images (i.e. two groups of pixels) corresponding to the two regions Rl, and R2. Subsequently (steps 202 and 203) the composite image C is constructed, which comprises the information pertaining to both the right and the left input images; in the example described herein, said composite image C is a frame of the output stereoscopic video stream, and therefore it is also referred to as container frame.
- step 202 the input image received by the device 100 and not disassembled by the device 104 (the left image L in the example of Fig.
- a container frame suitable for containing both will be a frame of 1920x1080 pixels, e.g. a frame of a video stream of the 1080p type (progressive format with 1920 x 1080 pixels.
- the left image L is entered into the container frame C and positioned in the upper left corner. This is obtained by copying the 1280x720 pixels of the image L into an area CI consisting of the first 1280 pixels of the first 720 rows of the container frame C.
- the image disassembled in step 201 by the module 104 is entered into the container frame.
- the pixels of the sub-images outputted by the module 104 are copied by preserving the respective spatial relations.
- the regions Rl, and R2 are copied into respective areas of the frame C without undergoing any deformation.
- FIG. 5a An example of the container frame C outputted by the module 105 is shown in Fig. 5a.
- the rectangular region Rl is copied into the last 640 pixels of the first 360 rows of the composite frame C (area C2), i.e. next to the previously copied image L.
- the L-shaped region R2 is copied under the area C2, i.e. in the area C3, which comprises the last 640 pixels of the rows from 361 to 720 plus the last 1280 pixels of the last 360 rows.
- region C2' there remains a rectangular region in the frame C composed by the first 640 pixels of the last 360 rows (region C2') which can be used for other purposes, e.g. for any ancillary data or signalling: it is represented lightly darkened in Fig. 5a and in the other figures as well.
- the same RGB values are assigned to the remaining pixels of the frame C; for example, said remaining pixels may be all black.
- the video stream outputted by the device 100 can be compressed to a considerable extent while preserving good possibilities that the image will be reconstructed very faithfully to the transmitted one without creating significant artifacts.
- the division of the frame R into two regions Rl, and R2 corresponds to the division of the frame into the smallest possible number of regions, taking into account the space available in the composite image and the space occupied by the left image entered unchanged into the container frame.
- Said smallest number is, in other words, the minimum number of regions necessary to occupy the space left available in the container frame C by the left image.
- the minimum number of regions into which the image must be disassembled is defined as a function of the format of the source images (right and left images) and of the target composite image (container frame C).
- the image R can be split in only two regions Rl and R2, in the way shown in Fig. 4.
- the two images L and R are positioned at two opposite corners of the composite image C, in particular at the top left corner and at the bottom right corner respectively.
- the part Rl of the image R that is superimposed to the image L can be shifted either in the top right corner, as it is shown in the figure, or in the bottom left corner.
- the part R2 of the image R not superimposed to the image L, placed at the bottom right corner has the form of an irregular polygon with six sides. This way the second image is broken up into the minimum number of regions (two).
- Fig. 5a represents just a first way to dispose the two images in the composite frame C according to the present invention:
- Fig. 5b shows a layout alternative to that of Fig. 5a, in which the region Rl has been placed in the first 640 pixels of the last 360 rows of C (area C2'), while the area C2 remains free of video information.
- Fig. 5a and 5b can be considered as alternative to each other ("dual arrangements"), since they simply differ in the allocation of Rl, which is placed in the upper right corner of C in the former case and in the lower left corner of C in the latter case.
- FIG. 6 A second way to break up the image R in order to be placed in the composite frame C is shown in Fig. 6; Rl is obtained by extracting the last 640 pixels of the last 360 rows of R.
- the L-shaped sub-image R2 is composed by the remaining pixel of R, namely the first 360 rows plus the first 640 pixels of the last 360 rows.
- Fig. 7a and 7b show the dual arrangements in which the regions Rl and R2 as obtained in Fig. 6 can be placed in the composite frame C after having placed the image L in its bottom right corner (area CI "), composed by the last 1280 pixels of the last 720 rows of C.
- the L-shaped R2 region is placed in upper left corner of C.
- the only difference between the two figures is the area of C occupied by the Rl sub-image, which is placed in the lower left (area C2') and upper right (area C2) corner, respectively.
- the rectangular spare region occupies the upper right corner (area C2) and lower left corner (area C2'), respectively.
- FIG. 8 A third way to disassemble the image R in order to be placed in the composite frame C is shown in Fig. 8;
- Rl is obtained by extracting the first 640 pixels of the last 360 rows of R.
- the L-shaped sub-image R2 is composed by the remaining pixel of R, namely the first 360 rows plus the last 640 pixels of the last 360 rows.
- Fig. 9a and 9b show the dual arrangements in which the regions Rl and R2 as obtained in Fig. 6 can be positioned in the composite frame C after having placed the image L in its bottom left corner (region CI "), composed by the first 1280 pixels of the last 720 rows of C.
- the L-shaped R2 region is placed in the upper right corner of C.
- the two figures differ in the position of the rectangular region Rl, which is placed in the lower right (area C6) and upper left (area C4) corner, respectively.
- the rectangular spare region occupies the upper left (area C2) and lower right corner (area C2'), respectively.
- a fourth way to disassemble the image R is depicted in Fig. 10.
- the last 640 pixels of the first 360 rows are extracted to form the sub-image Rl.
- the L-shaped region R2 is composed by the remaining pixel of R, namely the first 640 pixels of the first 360 rows plus the last 360 rows.
- Fig. 11a and l ib show the dual arrangements in which the regions Rl and R2 as obtained in Fig. 6 can be positioned in the composite frame C after having placed the image L in its upper right corner (region CI'"), composed by the last 1280 pixels of the first 720 rows of C.
- the L-shaped R2 region is placed in the lower left corner of C.
- the two figures differ in the position of the rectangular region Rl, which is placed in the top left (area C6) and bottom right (area C4) corner, respectively.
- the rectangular spare region occupies the upper left (area C2) and lower right corner (area C2'), respectively.
- an additional L-shaped region R3 comprising the boundary region between Rl and R2 as shown in Figure 12, can be replicated and inserted in the spare area C2' as shown in Figure 13.
- Such R3 region can have a constant width or two different widths, h and k, for the horizontal and vertical arms, respectively.
- the parameters h and k are integers greater than zero.
- the R3 region can eventually be placed symmetrically with respect to the internal boundary of R.
- the artifacts appear prevailingly close to the internal boundaries within the reconstructed image Rout.
- the pixels of Rl ' (corresponding to Rl after compression and decompression) and R2' (corresponding to R2 after compression and decompression) placed near the internal boundaries of Rout can be discarded in the replication and can be replaced by the internal pixels of the region R3' obtained after the compression and decompression operations of R3. Pixels at the edges of R3' should be discarded, since they are close to another internal boundary and therefore may be affected by artifacts.
- the L shaped region R3 is put in the spare area C2' adjacent to its bottom right corner, so to maximize the length of the R3 arms that can be placed in the available region.
- the frame C thus obtained in any of the ways described so far is subsequently compressed and transmitted or saved to a storage medium (e.g. a DVD).
- compression means are provided which are adapted to compress an image or a video signal, along with means for recording and/or transmitting the compressed image or video signal.
- Fig. 16 shows a block diagram of a receiver 1100 which decompresses the received container frame (if compressed), reconstructs the two right and left images, and makes them available to a visualization device (e.g. a television set) allowing fruition of 3D contents.
- the receiver 1100 may be a set-top-box or a receiver built in a television set.
- the same remarks made for the receiver 1100 are also applicable to a reader (e.g. a DVD reader) which reads a container frame (possibly compressed) and processes it in order to obtain one pair of frames corresponding to the right and left images entered into the container frame (possibly compressed) read by the reader.
- a reader e.g. a DVD reader
- the receiver receives (via cable or antenna) a compressed stereoscopic video stream 1101 and decompresses it by means of a decompression module 1102, thereby obtaining a video stream comprising a sequence of frames C corresponding to the frames C.
- a decompression module 1102 thereby obtaining a video stream comprising a sequence of frames C corresponding to the frames C.
- the frames C correspond to the container frames C carrying the information about the right and left images, except for any artifacts introduced by the compression process.
- These frames C are then supplied to a reconstruction module 1103, which executes an image reconstruction method as described below.
- the decompression module 1102 may be omitted and the video signal may be supplied directly to the reconstruction module 1103.
- the reconstruction process starts in step 1300, when the decompressed container frame C is received.
- the reconstruction process depends on the particular arrangements decided during the assembling process. Let us consider for example the composite frame shown in Fig. 5a.
- the reconstruction module 1103 extracts (step 1301) the left image L' (corresponding to the source image L) by copying the first 720x1280 pixels of the decompressed frame into a new frame Lout which is smaller than the container frame, e.g. a frame of a 720p stream.
- the image Lout thus reconstructed is outputted to the receiver 1100 (step 1302).
- the method provides for extracting the right source image R from the container frame C.
- the phase of extracting the right image begins by copying (step 1303) the area C2 included in the frame C. More in detail, the last 640 pixels of the first 360 rows of C are copied into the corresponding first 640 columns of the first 360 rows of the new 720x1280 frame representing the reconstructed image Rout.
- the area C3 containing the decompressed region R2' (which was R2 before compression and decompression operations) is extracted (step 1305).
- the pixels of the area C3 are copied in the L shaped remaining part of Rout, namely in the last 360 columns of the first 360 rows plus in the last 360 rows of Rout, thus obtaining the reconstructed image corresponding to the image R as assembled in Fig. 3.
- the receiver 1100 first performs the same operations already described for reconstructing Lout and Rout and then, as an additional step (1305 in Fig.17) extracts the internal region of R3' (called Ri3) and overwrites the corresponding pixels around the internal boundaries of Rout, using at least some of the pixels of R3'.
- a strip of m vertical and n horizontal pixels staying in the inner part of R3 forming a region called Ri3' is copied in the corresponding internal boundary region of Rout.
- m and n can be integers greater than zero that can assume low values typically in a range between 3 and 16; they can be equal to each other or not, giving to Ri3 a constant or non constant width.
- the same technique can be used, mutatis mutandis, in case a rectangular shape of R3 has been used for covering only one of its arms, either horizontal or vertical.
- region R3 ' and Ri3 are optional.
- a possibility would be to transmit region R3 and leave the freedom, at the decoder side, to use it or not: this would lead to two types of decoders, a simplified one and a more complex one with a better performance.
- the R3' region can be mixed on top of the reconstructed image Rout with the so called "soft edge” technique which consists in cross fading the pixel values of the internal boundary region of Rout with the corresponding pixel values of R3' so that R3' contribution is maximized at the boundary between Rl ' and R2' and minimized at the R3' boundaries.
- step 1307 The process for reconstructing the right and left images contained in the container frame C is thus completed (step 1307). Said process is repeated for each frame of the video stream received by the receiver 1100, so that the output will consist of two video streams 1104 and 1105 for the right image and for the left image, respectively.
- the electronic modules that provide the above described devices may be variously subdivided and distributed; furthermore, they may be provided in the form of hardware modules or as software algorithms implemented by a processor, in particular a video processor equipped with suitable memory areas for temporarily storing the input frames received.
- These modules may therefore execute in parallel or in series one or more of the video processing steps of the image multiplexing and de-multiplexing methods according to the present invention.
- the invention relates to any de-multiplexing method which allows a right image and a left image to be extracted from a composite image by reversing one of the above-described multiplexing processes falling within the protection scope of the present invention.
- the invention therefore also relates to a method for generating a pair of images starting from a composite image, which comprises the steps of:
- a first one e.g. the left image
- said right and left images by copying one single group of contiguous pixels from a region of said composite image
- a second image (e.g. the right image) by copying other groups of contiguous pixels from two different regions of said composite image.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
- Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
Abstract
Description
Claims
Priority Applications (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201280024020.1A CN103703761A (en) | 2011-05-17 | 2012-05-17 | Method for generating, transmitting and receiving stereoscopic images, and related devices |
EP12731690.9A EP2710799A1 (en) | 2011-05-17 | 2012-05-17 | Method for generating, transmitting and receiving stereoscopic images, and related devices |
KR1020137033537A KR20140044332A (en) | 2011-05-17 | 2012-05-17 | Method for generating, transmitting and receiving stereoscopic images, and related devices |
JP2014510935A JP2014517606A (en) | 2011-05-17 | 2012-05-17 | Method for generating, transmitting and receiving stereoscopic images, and related apparatus |
US14/118,032 US20140168365A1 (en) | 2011-05-17 | 2012-05-17 | Method for generating, transmitting and receiving stereoscopic images, and related devices |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
IT000439A ITTO20110439A1 (en) | 2011-05-17 | 2011-05-17 | METHOD FOR GENERATING, TRANSMITTING AND RECEIVING STEREOSCOPIC IMAGES, AND RELATED DEVICES |
ITTO2011A000439 | 2011-05-17 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2012156940A1 true WO2012156940A1 (en) | 2012-11-22 |
Family
ID=44555000
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/IB2012/052486 WO2012156940A1 (en) | 2011-05-17 | 2012-05-17 | Method for generating, transmitting and receiving stereoscopic images, and related devices |
Country Status (7)
Country | Link |
---|---|
US (1) | US20140168365A1 (en) |
EP (1) | EP2710799A1 (en) |
JP (1) | JP2014517606A (en) |
KR (1) | KR20140044332A (en) |
CN (1) | CN103703761A (en) |
IT (1) | ITTO20110439A1 (en) |
WO (1) | WO2012156940A1 (en) |
Families Citing this family (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR102346747B1 (en) * | 2015-05-07 | 2022-01-04 | 에스케이플래닛 주식회사 | System for cloud streaming service, method of cloud streaming service of providing multi-view screen based on resize and apparatus for the same |
JP6389540B2 (en) * | 2017-02-06 | 2018-09-12 | ソフトバンク株式会社 | Movie data generation device, display system, display control device, and program |
CN108765289B (en) * | 2018-05-25 | 2022-02-18 | 李锐 | Digital image extracting, splicing, restoring and filling method |
CN109714585B (en) * | 2019-01-24 | 2021-01-22 | 京东方科技集团股份有限公司 | Image transmission method and device, display method and device, and storage medium |
CN118570398B (en) * | 2024-07-31 | 2024-11-01 | 浙江荷湖科技有限公司 | Self-adaptive motion artifact detection method and three-dimensional reconstruction method using same |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2003088682A1 (en) | 2002-04-09 | 2003-10-23 | Teg Sensorial Technologies Inc. | Stereoscopic video sequences coding system and method |
US20050041736A1 (en) * | 2003-05-07 | 2005-02-24 | Bernie Butler-Smith | Stereoscopic television signal processing method, transmission system and viewer enhancements |
WO2008153863A2 (en) | 2007-06-07 | 2008-12-18 | Real D | Stereoplexing for video and film applications |
WO2009081335A1 (en) * | 2007-12-20 | 2009-07-02 | Koninklijke Philips Electronics N.V. | Image encoding method for stereoscopic rendering |
ITTO20110035A1 (en) * | 2011-01-19 | 2011-04-20 | Sisvel S P A | VIDEO FLOW CONSISTING OF COMBINED FRAME VIDEO, PROCEDURE AND DEVICES FOR ITS GENERATION, TRANSMISSION, RECEPTION AND REPRODUCTION |
Family Cites Families (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5691768A (en) * | 1995-07-07 | 1997-11-25 | Lucent Technologies, Inc. | Multiple resolution, multi-stream video system using a single standard decoder |
US6057884A (en) * | 1997-06-05 | 2000-05-02 | General Instrument Corporation | Temporal and spatial scaleable coding for video object planes |
US8749615B2 (en) * | 2007-06-07 | 2014-06-10 | Reald Inc. | Demultiplexing for stereoplexed film and video applications |
CN101720047B (en) * | 2009-11-03 | 2011-12-21 | 上海大学 | Method for acquiring range image by stereo matching of multi-aperture photographing based on color segmentation |
-
2011
- 2011-05-17 IT IT000439A patent/ITTO20110439A1/en unknown
-
2012
- 2012-05-17 WO PCT/IB2012/052486 patent/WO2012156940A1/en active Application Filing
- 2012-05-17 US US14/118,032 patent/US20140168365A1/en not_active Abandoned
- 2012-05-17 EP EP12731690.9A patent/EP2710799A1/en not_active Withdrawn
- 2012-05-17 CN CN201280024020.1A patent/CN103703761A/en active Pending
- 2012-05-17 KR KR1020137033537A patent/KR20140044332A/en not_active Application Discontinuation
- 2012-05-17 JP JP2014510935A patent/JP2014517606A/en active Pending
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2003088682A1 (en) | 2002-04-09 | 2003-10-23 | Teg Sensorial Technologies Inc. | Stereoscopic video sequences coding system and method |
US20050041736A1 (en) * | 2003-05-07 | 2005-02-24 | Bernie Butler-Smith | Stereoscopic television signal processing method, transmission system and viewer enhancements |
WO2008153863A2 (en) | 2007-06-07 | 2008-12-18 | Real D | Stereoplexing for video and film applications |
WO2009081335A1 (en) * | 2007-12-20 | 2009-07-02 | Koninklijke Philips Electronics N.V. | Image encoding method for stereoscopic rendering |
ITTO20110035A1 (en) * | 2011-01-19 | 2011-04-20 | Sisvel S P A | VIDEO FLOW CONSISTING OF COMBINED FRAME VIDEO, PROCEDURE AND DEVICES FOR ITS GENERATION, TRANSMISSION, RECEPTION AND REPRODUCTION |
Also Published As
Publication number | Publication date |
---|---|
ITTO20110439A1 (en) | 2012-11-18 |
EP2710799A1 (en) | 2014-03-26 |
US20140168365A1 (en) | 2014-06-19 |
JP2014517606A (en) | 2014-07-17 |
KR20140044332A (en) | 2014-04-14 |
CN103703761A (en) | 2014-04-02 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
AU2010334367B2 (en) | Method for generating, transmitting and receiving stereoscopic images, and related devices | |
EP2599319B1 (en) | Method for combining images relating to a three-dimensional content | |
KR101676504B1 (en) | Demultiplexing for stereoplexed film and video applications | |
KR101549274B1 (en) | Stereoplexing for video and film applications | |
KR101939971B1 (en) | Frame compatible depth map delivery formats for stereoscopic and auto-stereoscopic displays | |
KR20100031126A (en) | Stereoplexing for video and film applications | |
US20140168365A1 (en) | Method for generating, transmitting and receiving stereoscopic images, and related devices | |
JP6019520B2 (en) | Method and associated device for generating, transmitting and receiving stereoscopic images | |
US9571811B2 (en) | Method and device for multiplexing and demultiplexing composite images relating to a three-dimensional content |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 12731690 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2014510935 Country of ref document: JP Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2012731690 Country of ref document: EP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWE | Wipo information: entry into national phase |
Ref document number: 14118032 Country of ref document: US |
|
ENP | Entry into the national phase |
Ref document number: 20137033537 Country of ref document: KR Kind code of ref document: A |