WO2014010820A1 - Method and apparatus for estimating image motion using disparity information of a multi-view image - Google Patents
Method and apparatus for estimating image motion using disparity information of a multi-view image Download PDFInfo
- Publication number
- WO2014010820A1 WO2014010820A1 PCT/KR2013/003833 KR2013003833W WO2014010820A1 WO 2014010820 A1 WO2014010820 A1 WO 2014010820A1 KR 2013003833 W KR2013003833 W KR 2013003833W WO 2014010820 A1 WO2014010820 A1 WO 2014010820A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- view
- point
- motion
- image
- information
- Prior art date
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/60—Type of objects
- G06V20/64—Three-dimensional objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
- G06T7/285—Analysis of motion using a sequence of stereo image pairs
Definitions
- the method and apparatus relates to a method and apparatus for estimating and processing a motion of a multi-view image by using disparity information of the multi-view image.
- Motion estimation of multi-view images to generate an image between two images may be performed by a method of estimating a motion between two sequential images or by a method of estimating a motion between two images having different points of view.
- the latter may be performed by acquiring disparity information, and the former may be performed by acquiring a motion vector.
- motion estimation values of the points of view may lack consistency.
- the method and apparatus providing a method and apparatus for estimating and processing a motion of a multi-view image by using disparity information of the multi-view image.
- the motion estimation information between the two sequential images may be acquired by considering a disparity between two points of view in order to strengthen the consistency of temporal motion estimation information between two points of view and to improve the performance of motion estimation.
- FIG. 1 is a diagram which describes a method of processing a multi-view image, according to an embodiment of the present invention
- FIG. 2 is a flowchart illustrating a method of processing a multi-view image, according to an exemplary embodiment of the inventive concept
- FIGS. 3A and 3B are block diagrams which describe a method of processing a multi-view image, according to an exemplary embodiment of the inventive concept
- FIG. 4 is a flowchart illustrating the method of FIGS. 3A and 3B according to an exemplary embodiment of the inventive concept
- FIG. 5 is a block diagram for describing a method of processing a multi-view image, according to another exemplary embodiment of the inventive concept
- FIG. 6 is a flowchart illustrating the method of FIG. 5 according to another exemplary embodiment of the inventive concept
- FIGS. 7A to 8C are multi-view images according to an exemplary embodiment of the inventive concept
- FIG. 9 is a flowchart which illustrates a method of processing a multi-view image by acquiring a corresponding area, according to another exemplary embodiment of the inventive concept.
- FIG. 10 is a flowchart illustrating a method of processing a multi-view image by acquiring a corresponding area, according to another exemplary embodiment of the inventive concept.
- FIG. 11 is a block diagram of an apparatus for processing a multi-view image, according to an exemplary embodiment of the inventive concept.
- a method of processing a multi-view image includes: extracting disparity information between an image of a first point of view and an image of a second point of view; and using the extracted disparity information to estimate a motion between two sequential images of the first point of view or the second point of view.
- the estimating may include: acquiring a corresponding area in images of the first point of view and the second point of view using the extracted disparity information; estimating a motion of the corresponding area between two sequential images of each of the first point of view and the second point of view; and compensating for motion information estimated at each of the first point of view and the second point of view using motion information estimated at the other point of view.
- the estimating may include: estimating a motion between two sequential images of the first point of view or the second point of view; and using the extracted disparity information to compensate for estimated motion information between the two sequential images of the first point of view or the second point of view.
- the estimating may include: estimating a motion between two sequential images of the first point of view; and estimating a motion between two sequential images of the second point of view using the extracted disparity information and motion information estimated at the first point of view.
- the extracting may include extracting, at the same time, disparity information between two images of the first point of view and the second point of view.
- a time of extracting at least one of the two sequential images may be equal to or adjacent to a time that the disparity information is extracted.
- the estimating may include acquiring a motion vector between the two sequential images of the first point of view or the second point of view.
- the method may further include generating an image between the two sequential images of the first point of view or the second point of view based on the estimated motion information.
- An exemplary embodiment may provide an apparatus for processing a multi-view image, the apparatus including: a processor which is configured to extract disparity information between an image of a first point of view and an image of a second point of view in the multi-view image and is further configured to estimate a motion using the extracted disparity information between two sequential images of the first point of view or the second point of view.
- the apparatus may further include a memory which stores the multi-view image.
- Another exemplary embodiment may provide an apparatus for processing a multi-view image, the apparatus including: at least two cameras each photographing two objects which are spaced apart from each other and spaced apart from the cameras; and a processor which is configured to extract disparity information between an image of a first point of view taken by a first camera and an image of a second point of view taken by a second camera in the multi-view image and is further configured to estimate a motion using the extracted disparity information between two sequential images of the first point of view or the second point of view.
- Another exemplary embodiment may provide an apparatus for processing a multi-view image, the apparatus including: a first motion estimation module which estimates a motion between two sequential images of a first point of view; a second motion estimation module which estimates a motion between two sequential images of a second point of view; and a disparity extraction module which extracts disparity information between an image of the first point of view and an image of the second point of view; wherein the two motion estimation modules estimate a motion between two sequential images based on the extracted disparity information.
- FIG. 1 is a diagram which describes a method of processing a multi-view image, according to an exemplary embodiment of the inventive concept.
- image data of multi-view images may exist at times T0, T1, T2, T3, and T4 according to each point of view.
- a display device capable of displaying a multi-view image displays multi-view images, of which image data along time exists according to each point of view as shown in FIG. 1, it is necessary to acquire motion estimation information between the images to additionally generate an image between every two sequential images, so that images are naturally displayed.
- the motion estimation information between the two sequential images may be acquired by considering a disparity between two points of view in order to strengthen the consistency of temporal motion estimation information between two points of view and to improve the performance of motion estimation.
- FIG. 2 is a flowchart which illustrates a method of processing a multi-view image, according to an exemplary embodiment of the inventive concept.
- a multi-view image processing apparatus extracts disparity information between images of a first point of view and a second point of view.
- the images of the first point of view and the second point of view from which the disparity information is extracted may be images taken at the same time, and the disparity information may be extracted from metadata in multi-view image data.
- a method of extracting disparity information is not limited thereto, and various methods of extracting disparity information may be used.
- the multi-view image processing apparatus estimates a motion between two sequential images of each of the first point of view and the second point of view by considering the extracted disparity information.
- a time of at least one of the two sequential images may be equal to or adjacent to a time that the disparity information is extracted.
- the multi-view image processing apparatus may estimate the motion between the two sequential images so that images of the first point of view and the second point of view, which may be generated by motion estimation information, have an estimated disparity.
- FIGS. 3A and 3B are block diagrams which describe a method of processing a multi-view image, according to an exemplary embodiment of the inventive concept.
- motion estimation modules V n and V n-k 300 and 320, a disparity extraction module 310, and a motion estimation compensation module 330 may be programs stored in a memory of a multi-view image processing apparatus according to an exemplary embodiment of the inventive concept.
- the motion estimation module Vn 300 may estimate a motion between two sequential images V n T m and V n T m+1 of a first point of view.
- estimable motion information may be represented by a motion vector .
- the motion estimation module V n-k 320 may estimate a motion between two sequential images V n-k T m and V n-k T m+1 of a second point of view, and motion information estimated by the motion estimation module V n-k 320 may be represented by a motion vector .
- the disparity extraction module 310 may extract disparity information between the image VnTm of the first point of view and the image V n-k T m of the second point of view.
- the motion estimation module Vn 300 and the motion estimation module V n-k 320 may estimate a motion between two sequential images by considering the disparity information .
- the motion estimation compensation module 330 for compensating for the motion vectors and estimated by the motion estimation modules V n and V n-k 300 and 320 may be further included.
- the motion estimation compensation module 330 may compensate for the motion vectors and estimated by the motion estimation modules Vn and Vn-k 300 and 320 by considering the disparity information , or motion estimation information or of the other point of view, respectively.
- the motion estimation compensation module 330 may compensate for the motion estimation information of the second point of view using the disparity information or the motion estimation information of the first point of view.
- the compensated motion estimation information of the second point of view may be represented by a motion vector .
- the motion estimation compensation module 330 may compensate for the motion estimation information of the first point of view using the disparity information or the motion estimation information of the second point of view.
- the compensated motion estimation information of the first point of view may be represented by a motion vector .
- FIG. 4 is a flowchart illustrating the method of FIGS. 3A and 3B according to an exemplary embodiment of the inventive concept.
- a multi-view image processing apparatus extracts disparity information between an image of the first point of view and an image of the second point of view.
- the multi-view image processing apparatus estimates a motion between two sequential images of each of the first point of view and the second point of view.
- the multi-view image processing apparatus compensates for the motion information estimated in operation S403 using the extracted disparity information. For example, the motion information may be compensated for so that an image of the first point of view and an image of the second point of view, which may be generated by motion information, have a disparity which corresponds to the extracted disparity information.
- FIG. 5 is a block diagram which describes a method of processing a multi-view image, according to another exemplary embodiment of the inventive concept.
- motion estimation modules V n and V n-k 500 and 520 and a disparity extraction module 510 may be programs stored in a memory of a multi-view image processing apparatus according to an exemplary embodiment of the inventive concept.
- the motion estimation module V n 500 may estimate a motion between two sequential images V n T m and V n T m+1 of a first point of view.
- estimable motion information may be represented by a motion vector .
- the motion estimation module V n-k 520 may estimate a motion between two sequential images V n-k T m and V n-k T m+1 of a second point of view, and motion information estimable by the motion estimation module V n-k 520 may be represented by a motion vector .
- the disparity extraction module 510 may extract disparity information between the image V n T m of the first point of view and the image V n-k T m of the second point of view.
- the motion estimation module V n-k 520 may estimate motion information between two sequential images V n-k T m and V n-k T m+1 of the second point of view using the disparity information and motion information acquirable by the motion vector , and the motion information may be acquired as the motion vector .
- the multi-view image processing apparatus may estimate motion information of one point of view using disparity information between the two points of view and may estimate motion information of the other point of view.
- FIG. 6 is a flowchart illustrating the method of FIG. 5 according to another exemplary embodiment of the inventive concept.
- the multi-view image processing apparatus extracts disparity information between an image of a first point of view and an image of a second point of view.
- the multi-view image processing apparatus estimates a motion between two sequential images of the first point of view.
- the multi-view image processing apparatus estimates a motion between two sequential images of the second point of view using the disparity information and the estimated motion information of the first point of view. For example, the motion information of the second point of view may be estimated so that an image of the second point of view has the extracted disparity from a corresponding image of the first point of view, which may be generated based on the motion information estimated at the first point of view.
- FIGS. 7A to 8C are multi-view images according to an exemplary embodiment of the inventive concept.
- FIGS. 7A and 7B show two images of different points of view, which are captured by two cameras at the same time.
- images generated with respect to different points of view have a global difference due to a position difference between the two cameras.
- a left area 710 of the image shown in FIG. 7A is not shown in FIG. 7B.
- a right area 740 of the image shown in FIG. 7B is not shown in FIG. 7A. If the image shown in FIG. 7A is moved to the right, the moved image may be similar to the image shown in FIG. 7B.
- a multi-view image processing apparatus may acquire a corresponding area 730 in FIG. 7B, which corresponds to a current area 720 in FIG. 7A, using disparity information.
- FIG. 8A shows cameras 810 and 820, each photographing two objects 830 and 840, positions of which are apart from the cameras 810 and 820 and are different from each other.
- FIG. 8B shows an image captured by the camera 810 located on the left side of FIG. 8A
- FIG. 8C shows an image captured by the camera 820 located on the right side of FIG. 8A.
- a disparity of the object 830 apart farther from the cameras 810 and 820 is less than a disparity of the object 840 apart nearer from the cameras 810 and 820. That is, a disparity of each area in a multi-view image may vary according to a distance between an object and cameras.
- the multi-view image processing apparatus may extract disparity information for each area in a multi-view image and may estimate a motion between images by acquiring a corresponding area of each area using the disparity information, thereby improving the performance of the motion estimation performance.
- FIG. 9 is a flowchart which illustrates a method of processing a multi-view image by acquiring a corresponding area, according to another exemplary embodiment of the inventive concept.
- a multi-view image processing apparatus extracts disparity information between an image of a first point of view and an image of a second point of view.
- the multi-view image processing apparatus acquires a corresponding area in an image of each of the first point of view and the second point of view.
- the multi-view image processing apparatus estimates a motion between two sequential images of each of the first point of view and the second point of view based on the corresponding area.
- the multi-view image processing apparatus compensates for the motion information estimated in operation S905 using the extracted disparity information based on the corresponding area.
- FIG. 10 is a flowchart which illustrates a method of processing a multi-view image by acquiring a corresponding area, according to another exemplary embodiment of the inventive concept.
- a multi-view image processing apparatus extracts disparity information between an image of a first point of view and an image of a second point of view.
- the multi-view image processing apparatus acquires a corresponding area in an image of each of the first point of view and the second point of view.
- the multi-view image processing apparatus estimates a motion between two sequential images of the first point of view based on the corresponding area.
- the multi-view image processing apparatus estimates a motion between two sequential images of the second point of view based on the corresponding area using the disparity information and the motion information estimated at the first point of view.
- the motion information of the second point of view may be estimated based on the corresponding area so that an image of the second point of view has the extracted disparity from a corresponding image of the first point of view, which may be generated based on the motion information estimated at the first point of view.
- FIG. 11 is a block diagram of an apparatus 1100 for processing a multi-view image, according to an exemplary embodiment of the inventive concept.
- the apparatus 1100 may include a memory 1140 which stores a multi-view image 1141 and programs for processing the multi-view image 1141, and a processor 1130 which processes the multi-view image 1141.
- the apparatus 1100 may further include a display 1120 which displays a processed multi-view image and an input unit 1110 which receives an input from a user.
- the processor 1130 may process the multi-view image 1141 by executing the programs stored in the memory 1140.
- the programs stored in the memory 1140 may be classified into a plurality of modules according to their functions, e.g., a disparity extraction module 1142, a motion estimation module 1143, a motion estimation compensation module 1144, a multi-view image decoding module 1145, and a multi-view image reproduction module 1146.
- the disparity extraction module 1142 may extract disparity information between an image of a first point of view and an image of a second point of view.
- the extracted disparity information may relate to an entire image or a partial image.
- the images of the first point of view and the second point of view, from which the disparity information is extracted, may be generated at the same time.
- the motion estimation module 1143 may estimate a motion between two sequential images of the first point of view or the second point of view based on the extracted disparity information.
- the motion estimation module 1143 may acquire a corresponding area of each of the first point of view and the second point of view using the disparity information and perform motion estimation on each corresponding area.
- the motion estimation module 1143 may estimate a motion between two sequential images of each of the first point of view and the second point of view and compensate for motion information estimated at each of the first point of view and the second point of view, using disparity information.
- the motion estimation module 1143 may estimate a motion between two sequential images of the first point of view, and may estimate a motion between two sequential images of the second point of view using disparity information and the motion information estimated at the first point of view.
- the first point of view and the second point of view are not limited to specific points of view.
- a time of at least one of the two sequential images may be equal or adjacent to a time the disparity information is extracted. That is, the motion estimation module 1143 may estimate temporally changed motion information of an image at the time the disparity information is extracted.
- the motion estimation module 1143 may estimate a motion between sequential images by acquiring a motion vector.
- the motion estimation compensation module 1144 may compensate for an estimated motion vector by considering disparity information or motion estimation information of another point of view.
- the motion estimation compensation module 1144 may compensate for motion information estimated at the second point of view using disparity information or motion information estimated at the first point of view. In addition, the motion estimation compensation module 1144 may compensate for motion information estimated at the first point of view using disparity information or motion information estimated at the second point of view.
- the multi-view image decoding module 1145 may generate an image between two sequential images of each point of view using the motion information estimated by the motion estimation module 1143 or the motion information compensated by the motion estimation compensation module 1144.
- the multi-view image reproduction module 1146 may control the display 1120 to display a multi-view image using the image generated by the multi-view image decoding module 1145. According to an exemplary embodiment of the inventive concept, the multi-view image reproduction module 1146 may control the display 1120 to display images using the image generated by the multi-view image decoding module 1145.
- the inventive concept can also be embodied as computer-readable codes on a non-transitory computer-readable recording medium.
- the non-transitory computer-readable recording medium is any data storage device that can store data which can be thereafter read by a computer system. Examples of the computer-readable recording medium include read-only memory (ROM), random-access memory (RAM), CD-ROMs, magnetic tapes, floppy disks, optical data storage devices, and so forth.
Abstract
Description
Claims (15)
- A method of processing a multi-view image, the method comprising:extracting disparity information between an image of a first point of view and an image of a second point of view; andestimating a motion between two sequential images of the first point of view or the second point of view using the extracted disparity information.
- The method of claim 1, wherein the estimating comprises:acquiring a corresponding area in images of the first point of view and the second point of view using the extracted disparity information;estimating a motion between two sequential images of each of the first point of view and the second point of view based on the corresponding area; andcompensating for motion information estimated at each of the first point of view and the second point of view using motion information estimated at the other point of view.
- The method of claim 1, wherein the estimating comprises:estimating a motion between two sequential images of the first point of view or the second point of view; andcompensating for estimated motion information between the two sequential images of the first point of view or the second point of view using the extracted disparity information.
- The method of claim 1, wherein the estimating comprises:estimating a motion between two sequential images of the first point of view; andestimating a motion between two sequential images of the second point of view using the extracted disparity information and motion information estimated at the first point of view.
- The method of claim 1, further comprising generating an image between the two sequential images of the first point of view or the second point of view based on the estimated motion information.
- The method of claim 1, wherein the estimating comprises:estimating a motion between two sequential images of each of the first point of view and the second point of view; andcompensating for motion information estimated at the second point of view using the extracted disparity information or motion information estimated at the first point of view.
- The method of claim 1, wherein the estimating comprises:estimating a motion between two sequential images of each of the first point of view and the second point of view; andcompensating for motion information estimated at the first point of view using the extracted disparity information or motion information estimated at the second point of view.
- An apparatus for processing a multi-view image, the apparatus comprising:a memory for storing a multi-view image; anda processor which extracts disparity information between an image of a first point of view and an image of a second point of view in the multi-view image and estimates a motion using the extracted disparity information between two sequential images of the first point of view or the second point of view.
- The apparatus of claim 8, wherein the processor acquires a corresponding area in images of the first point of view and the second point of view using the extracted disparity information, estimates a motion between two sequential images of each of the first point of view and the second point of view based on the corresponding area, and compensates for motion information estimated at each of the first point of view and the second point of view using motion information estimated at the other point of view.
- The apparatus of claim 8, wherein the processor estimates a motion between two sequential images of the first point of view or the second point of view and compensates for estimated motion information between the two sequential images of the first point of view or the second point of view using the extracted disparity information.
- The apparatus of claim 8, wherein the processor estimates a motion between two sequential images of the first point of view and estimates a motion between two sequential images of the second point of view using the extracted disparity information and the motion information estimated at the first point of view.
- The apparatus of claim 8, wherein the processor generates an image between the two sequential images of the first point of view or the second point of view, based on the estimated motion information.
- The apparatus of claim 8, wherein the processor estimates a motion between two sequential images of each of the first point of view and the second point of view and compensates for motion information estimated at the second point of view using the extracted disparity information or motion information estimated at the first point of view.
- The apparatus of claim 8, wherein the processor estimates a motion between two sequential images of each of the first point of view and the second point of view and compensates for motion information estimated at the first point of view using the extracted disparity information or motion information estimated at the second point of view.
- A non-transitory computer-readable recording medium storing a computer-readable program which executes a method of processing a video signal, wherein the program, when executed by a processor of a computer causes the computer to execute a method comprising:extracting disparity information between an image of a first point of view and an image of a second point of view; andestimating a motion between two sequential images of the first point of view or the second point of view, using the extracted disparity information.
Priority Applications (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
BR112015000303A BR112015000303A2 (en) | 2012-07-10 | 2013-05-03 | method for processing a multi-view image, apparatus for processing a multi-view image, and computer readable permanent recording media |
JP2015521531A JP2015527804A (en) | 2012-07-10 | 2013-05-03 | Method and apparatus for estimating motion of video using parallax information of multi-view video |
CN201380039625.2A CN104488264B (en) | 2012-07-10 | 2013-05-03 | The method and apparatus of image motion are estimated using the different information of multi-view image |
EP13816970.1A EP2859727B1 (en) | 2012-07-10 | 2013-05-03 | Method and apparatus for estimating image motion using disparity information of a multi-view image |
AU2013287585A AU2013287585B2 (en) | 2012-07-10 | 2013-05-03 | Method and apparatus for estimating image motion using disparity information of a multi-view image |
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201261669757P | 2012-07-10 | 2012-07-10 | |
US61/669,757 | 2012-07-10 | ||
KR10-2012-0100654 | 2012-09-11 | ||
KR1020120100654A KR101966920B1 (en) | 2012-07-10 | 2012-09-11 | Method and apparatus for estimating motion of image using disparity information of multi view image |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2014010820A1 true WO2014010820A1 (en) | 2014-01-16 |
Family
ID=50142202
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/KR2013/003833 WO2014010820A1 (en) | 2012-07-10 | 2013-05-03 | Method and apparatus for estimating image motion using disparity information of a multi-view image |
Country Status (8)
Country | Link |
---|---|
US (1) | US9798919B2 (en) |
EP (1) | EP2859727B1 (en) |
JP (1) | JP2015527804A (en) |
KR (1) | KR101966920B1 (en) |
CN (1) | CN104488264B (en) |
AU (1) | AU2013287585B2 (en) |
BR (1) | BR112015000303A2 (en) |
WO (1) | WO2014010820A1 (en) |
Families Citing this family (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20160119532A1 (en) * | 2015-01-22 | 2016-04-28 | Mediatek Inc. | Method And Apparatus Of Utilizing Image/Video Data From Multiple Sources |
KR101665513B1 (en) | 2015-01-30 | 2016-10-12 | 삼성전자 주식회사 | Computer tomography apparatus and method for reconstructing a computer tomography image thereof |
KR101783964B1 (en) * | 2015-04-09 | 2017-10-10 | 삼성전자주식회사 | Tomography apparatus and method for reconstructing a tomography image thereof |
US10389128B2 (en) * | 2016-01-25 | 2019-08-20 | Rolls-Royce Corporation | Power control system |
US20180091797A1 (en) * | 2016-09-27 | 2018-03-29 | The Boeing Company | Apparatus and method of compensating for relative motion of at least two aircraft-mounted cameras |
CN108693970B (en) * | 2017-04-11 | 2022-02-18 | 杜比实验室特许公司 | Method and apparatus for adapting video images of a wearable device |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR20070011147A (en) * | 2005-07-18 | 2007-01-24 | 한국전자통신연구원 | Apparatus of predictive coding/decoding using view-temporal reference picture buffers and method using the same |
KR20070022568A (en) * | 2005-08-22 | 2007-02-27 | 삼성전자주식회사 | Method and apparatus for encoding multiview video |
US20110254930A1 (en) | 2010-04-14 | 2011-10-20 | Canon Kabushiki Kaisha | Image processing apparatus and image processing method |
KR20110139882A (en) * | 2010-06-24 | 2011-12-30 | 삼성전자주식회사 | Method and apparatus for multiview depth image coding and decoding |
KR20120002112A (en) * | 2010-06-30 | 2012-01-05 | 엘지이노텍 주식회사 | Apparatus and method for coding multi-view video |
JP2012023662A (en) * | 2010-07-16 | 2012-02-02 | Sony Corp | Image processor and image processing method |
Family Cites Families (19)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP3519594B2 (en) * | 1998-03-03 | 2004-04-19 | Kddi株式会社 | Encoding device for stereo video |
KR101038452B1 (en) * | 2003-08-05 | 2011-06-01 | 코닌클리케 필립스 일렉트로닉스 엔.브이. | Multi-view image generation |
KR100779875B1 (en) * | 2005-01-14 | 2007-11-27 | 주식회사 휴맥스 | Method for setting reference frame order for multi-view coding and computer readable medium storing thereof |
MX2008002391A (en) * | 2005-08-22 | 2008-03-18 | Samsung Electronics Co Ltd | Method and apparatus for encoding multiview video. |
KR100667830B1 (en) | 2005-11-05 | 2007-01-11 | 삼성전자주식회사 | Method and apparatus for encoding multiview video |
KR20070076356A (en) * | 2006-01-18 | 2007-07-24 | 엘지전자 주식회사 | Method and apparatus for coding and decoding of video sequence |
EP2052546A4 (en) * | 2006-07-12 | 2010-03-03 | Lg Electronics Inc | A method and apparatus for processing a signal |
KR101336204B1 (en) * | 2006-08-18 | 2013-12-03 | 주식회사 케이티 | A method and apparatus for encoding or decoding frames of different views in multiview video using global disparity |
BRPI0721077A2 (en) * | 2006-12-28 | 2014-07-01 | Nippon Telegraph & Telephone | CODING METHOD AND VIDEO DECODING METHOD, SAME APPARELS, SAME PROGRAMS, AND STORAGE Means WHICH STORE THE PROGRAMS |
CN101647279A (en) | 2007-01-24 | 2010-02-10 | Lg电子株式会社 | A method and an apparatus for processing a video signal |
JP2009049979A (en) | 2007-07-20 | 2009-03-05 | Fujifilm Corp | Image processing device, image processing method, image processing system, and program |
ES2727567T3 (en) * | 2007-12-27 | 2019-10-17 | Psholix Ag | Procedure and device for real-time multivist image generation |
CN102177721B (en) * | 2008-10-10 | 2015-09-16 | 皇家飞利浦电子股份有限公司 | The method of the parallax information that processing signals comprises |
KR101502362B1 (en) | 2008-10-10 | 2015-03-13 | 삼성전자주식회사 | Apparatus and Method for Image Processing |
JP2011019202A (en) * | 2009-07-10 | 2011-01-27 | Sony Corp | Image signal processing apparatus and image display |
JP2011239169A (en) * | 2010-05-10 | 2011-11-24 | Sony Corp | Stereo-image-data transmitting apparatus, stereo-image-data transmitting method, stereo-image-data receiving apparatus, and stereo-image-data receiving method |
JP2013165306A (en) | 2010-06-01 | 2013-08-22 | Panasonic Corp | Stereoscopic video display device |
JP5633259B2 (en) | 2010-09-06 | 2014-12-03 | ソニー株式会社 | Stereo image data transmitting device, stereo image data transmitting method, and stereo image data receiving device |
JP5977964B2 (en) * | 2012-03-08 | 2016-08-24 | 日本放送協会 | Multi-view video frame interpolation apparatus, method and program |
-
2012
- 2012-09-11 KR KR1020120100654A patent/KR101966920B1/en active IP Right Grant
-
2013
- 2013-05-03 CN CN201380039625.2A patent/CN104488264B/en not_active Expired - Fee Related
- 2013-05-03 BR BR112015000303A patent/BR112015000303A2/en active Search and Examination
- 2013-05-03 WO PCT/KR2013/003833 patent/WO2014010820A1/en active Application Filing
- 2013-05-03 EP EP13816970.1A patent/EP2859727B1/en active Active
- 2013-05-03 AU AU2013287585A patent/AU2013287585B2/en not_active Ceased
- 2013-05-03 JP JP2015521531A patent/JP2015527804A/en active Pending
- 2013-05-09 US US13/890,791 patent/US9798919B2/en active Active
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR20070011147A (en) * | 2005-07-18 | 2007-01-24 | 한국전자통신연구원 | Apparatus of predictive coding/decoding using view-temporal reference picture buffers and method using the same |
KR20070022568A (en) * | 2005-08-22 | 2007-02-27 | 삼성전자주식회사 | Method and apparatus for encoding multiview video |
US20110254930A1 (en) | 2010-04-14 | 2011-10-20 | Canon Kabushiki Kaisha | Image processing apparatus and image processing method |
KR20110139882A (en) * | 2010-06-24 | 2011-12-30 | 삼성전자주식회사 | Method and apparatus for multiview depth image coding and decoding |
KR20120002112A (en) * | 2010-06-30 | 2012-01-05 | 엘지이노텍 주식회사 | Apparatus and method for coding multi-view video |
JP2012023662A (en) * | 2010-07-16 | 2012-02-02 | Sony Corp | Image processor and image processing method |
Also Published As
Publication number | Publication date |
---|---|
US9798919B2 (en) | 2017-10-24 |
EP2859727B1 (en) | 2020-02-19 |
EP2859727A4 (en) | 2016-07-06 |
CN104488264B (en) | 2017-09-01 |
KR101966920B1 (en) | 2019-04-08 |
EP2859727A1 (en) | 2015-04-15 |
AU2013287585B2 (en) | 2016-11-10 |
JP2015527804A (en) | 2015-09-17 |
CN104488264A (en) | 2015-04-01 |
KR20140008226A (en) | 2014-01-21 |
US20140015936A1 (en) | 2014-01-16 |
AU2013287585A1 (en) | 2015-02-05 |
BR112015000303A2 (en) | 2017-06-27 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2014010820A1 (en) | Method and apparatus for estimating image motion using disparity information of a multi-view image | |
WO2016027930A1 (en) | Portable device and method for controlling the same | |
WO2015005577A1 (en) | Camera pose estimation apparatus and method | |
WO2019164379A1 (en) | Method and system for facial recognition | |
WO2013015549A2 (en) | Plane-characteristic-based markerless augmented reality system and method for operating same | |
WO2010027137A2 (en) | Apparatus and method for frame interpolation based on accurate motion estimation | |
WO2012108721A2 (en) | Device and method for providing augmented reality using image information | |
WO2012093811A1 (en) | Method for support in such a way as to allow collection of objects comprised in an input image, and a recording medium able to be read by terminal devices and computers | |
WO2020027607A1 (en) | Object detection device and control method | |
WO2016013885A1 (en) | Method for retrieving image and electronic device thereof | |
US10154247B2 (en) | Virtual reality communication systems and methods thereof | |
WO2012115435A2 (en) | Method and apparatus for encoding and decoding multi view video | |
WO2010076988A2 (en) | Image data obtaining method and apparatus therefor | |
WO2018084536A1 (en) | Time slice image provision server, method and user terminal | |
WO2020017825A1 (en) | Method of combining content from multiple frames and electronic device therefor | |
WO2019164288A1 (en) | Method for providing text translation managing data related to application, and electronic device thereof | |
WO2017099510A1 (en) | Method for segmenting static scene on basis of image statistical information and method therefor | |
WO2016186236A1 (en) | Color processing system and method for three-dimensional object | |
WO2022025565A1 (en) | System and method for generating bokeh image for dslr quality depth-of-field rendering and refinement and training method for the same | |
WO2019098421A1 (en) | Object reconstruction device using motion information and object reconstruction method using same | |
CN108520547B (en) | Image processing method, device, equipment and storage medium | |
CN102917234A (en) | Image processing device and method, and program | |
WO2019009579A1 (en) | Stereo matching method and apparatus using support point interpolation | |
WO2020171257A1 (en) | Image processing method and device thereof | |
WO2011007970A1 (en) | Method and apparatus for processing image |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 13816970 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2013816970 Country of ref document: EP |
|
ENP | Entry into the national phase |
Ref document number: 2015521531 Country of ref document: JP Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
ENP | Entry into the national phase |
Ref document number: 2013287585 Country of ref document: AU Date of ref document: 20130503 Kind code of ref document: A |
|
REG | Reference to national code |
Ref country code: BR Ref legal event code: B01A Ref document number: 112015000303 Country of ref document: BR |
|
ENP | Entry into the national phase |
Ref document number: 112015000303 Country of ref document: BR Kind code of ref document: A2 Effective date: 20150107 |