WO2012113797A1 - Dispositif et procédé de gestion de la position du plan focal dans une scène stéréoscopique - Google Patents

Dispositif et procédé de gestion de la position du plan focal dans une scène stéréoscopique Download PDF

Info

Publication number
WO2012113797A1
WO2012113797A1 PCT/EP2012/052943 EP2012052943W WO2012113797A1 WO 2012113797 A1 WO2012113797 A1 WO 2012113797A1 EP 2012052943 W EP2012052943 W EP 2012052943W WO 2012113797 A1 WO2012113797 A1 WO 2012113797A1
Authority
WO
WIPO (PCT)
Prior art keywords
translation
focal plane
value
images
scene
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Ceased
Application number
PCT/EP2012/052943
Other languages
English (en)
French (fr)
Inventor
Alexandre Delattre
Jérôme LARRIEU
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nintendo European Research and Development SAS
Original Assignee
Mobiclip SAS
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Mobiclip SAS filed Critical Mobiclip SAS
Priority to JP2013554871A priority Critical patent/JP5981460B2/ja
Priority to EP12704435.2A priority patent/EP2678823B1/fr
Priority to US14/000,462 priority patent/US9800865B2/en
Publication of WO2012113797A1 publication Critical patent/WO2012113797A1/fr
Anticipated expiration legal-status Critical
Ceased legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/30Image reproducers
    • H04N13/398Synchronisation thereof; Control thereof
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/50Depth or shape recovery
    • G06T7/55Depth or shape recovery from multiple images
    • G06T7/593Depth or shape recovery from multiple images from stereo images
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/97Determining parameters from multiple pictures
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • H04N13/128Adjusting depth or disparity
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10004Still image; Photographic image
    • G06T2207/10012Stereo images
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20068Projection on vertical or horizontal image axis

Definitions

  • the present invention relates to the field of scopic stereo vision systems and more particularly to the management of the position of the focal plane in a scopic stereo scene.
  • Stereoscopic images are typically composed of a left image and a right image of the same scene. These two images are taken by cameras from a slightly shifted point of view. Generally, the two cameras are parallel and spaced a few centimeters apart.
  • a scopic stereo video is composed of two video streams, a left video stream and a right video stream.
  • Stereoscopy is due to the parallax effect. This effect results in the same object appearing in the two images, left and right, slightly translated. The value of the translation depends on the depth, that is, the distance to the cameras, from the object.
  • an object is viewed on a stereo scopic screen 1.3.
  • the eyes 1.1 and 1.2 of the user are represented.
  • the object is located on the left image at point 1.6 and on the right image at point 1.5. These points are distant from a translation value 1.7, this value expressed in translation in the right image relative to the left image is a negative translation.
  • this value expressed in translation in the right image relative to the left image is a negative translation.
  • an object is viewed on a stereoscopic screen 2.3.
  • the eyes 2.1 and 2.2 of the user are represented.
  • the object is located on the left image at point 2.6 and on the right image at point 2.5. These points are distant from a translation value 2.7, this value expressed in translation in the right image relative to the left image is a positive translation.
  • this point When the eyes of the user will accommodate to visualize this point, it will appear in point 2.4 as being behind the screen.
  • an object is viewed on a stereoscopic screen 3.3.
  • the eyes 3. 1 and 3.2 of the user are represented.
  • the object is on the left image at point 3.6 and on the right image at point 3.5. These points merge, the translation for this point is zero between the two images.
  • the eyes of the user will accommodate to visualize this point, it will appear in point 3.4 as being at the level of the screen.
  • the focal plane of a stereoscopic scene is the plane perpendicular to the optical axes of the cameras located at a given depth such that all the points representing objects in this plane are represented in the left and right images with zero parallax.
  • the optical axes of the cameras performing the stereoscopic shooting are typically parallel, the objects at infinity are superimposed perfectly in the two images while the closer objects undergo a translation all the stronger as the object is close to the cameras. .
  • the images taken by such a system therefore have a focal plane to infinity.
  • the entire scene is located in front of this screen. It is found that the visualization of a stereoscopic video where the focal plane is thus located at infinity tends to be uncomfortable to the eye.
  • a focal plane located globally in front of the stage with rare objects that can be located in front of the focal plane is much more comfortable to visualize.
  • the aim of the invention is to propose a method for managing the focal plane within an image or sequence of stereo scopic images.
  • a translation of the right and left images is performed to place the focal plane in front of the visualized scene.
  • the depth of the major objects of the scene is estimated by dividing the images into a plurality of zones and estimating the depth of each zone. The depth of the focal plane is then calculated so that most of the scene is behind the focal plane.
  • the invention relates to a method for managing the position of the focal plane in a stereoscopic scene represented by a left image and a right image which comprises a step of determining at least one zone in the two images; a step of calculating the dominant translation value for each of the zones; a step of determining a relative translation value between the left and right images according to the dominant translation values of each of the zones and a relative translation step of the two left and right images of said previously determined value.
  • the step of calculating the dominant translation value for each of the zones comprises a step of projecting the left zone and the right zone on a left line and a straight line; a step of calculating the correlation between the left line and the straight line for a plurality of translation values and a step of determining the translation value for which the correlation is the best.
  • the step of determining a relative translation value between the left and right images comprises a step of classifying the dominant translation values of each zone by group of close values; a step of calculating the average value of the dominant translation values within each group and a step of choosing the highest average value that corresponds to the nearest group, as the relative translation value between the left and right images.
  • the method further comprises a step of eliminating areas for which the dominant translation value obtained is aberrant.
  • the method further comprises a step of eliminating the groups of zones not containing a number of zones at least equal to a given threshold.
  • the invention also relates to a device for shooting or stereoscopic viewing, characterized in that it comprises means for managing the position of the focal plane in a stereoscopic scene represented by a left image and a right image comprising means for determining at least one area in both images; means for calculating the dominant translation value for each of the zones; means for determining a relative translation value between the left and right images based on the dominant translation values of each of the zones and means for performing a relative translation of the two left and right images of said previously determined value.
  • Fig. 1 illustrates the positions of a point appearing behind the focal plane.
  • Fig. 2 illustrates the positions of a point appearing in front of the focal plane.
  • Fig. 3 illustrates the positions of a point appearing in the focal plane.
  • Fig. 4 illustrates the method of displacement of the focal plane in the embodiment of the invention.
  • Fig. 5 illustrates an exemplary embodiment of the focal plane management method.
  • Fig. 6 illustrates a zone division of an image.
  • Fig. 7 illustrates the projection on a line of a zone.
  • FIG. 4 illustrates the method of displacement of the focal plane.
  • a scene includes objects 4.1, 4.2 and 4.3 each located at a given depth as viewed by an observer 4.4.
  • the scene is displayed on a 4.5 screen.
  • the focal plane of the scene merges with the screen.
  • Objects appear in front of the screen or behind the screen if they are in front of the focal plane or behind the focal plane.
  • the scene is created from a left image 4.7 and a right image 4.8 which each represent the scene and its objects.
  • the focal plane corresponds to the objes whose location in the left and right images is superimposed. Objects out of the focal plane undergo a parallax, negative for objects in front of the focal plane and positive for objects behind the focal plane.
  • the cameras When shooting, the cameras have a parallel optical axis. As a result, the left and right images are superimposed for objects at infinity. When such scenes are viewed on the screen, the entire scene appears in front of the focal plane and thus as "out" of the screen. The visualization then requires constant adjustments of accommodation in an area close to the eyes and is uncomfortable.
  • the relative translation value that must be applied to the left and right images to move the focal plane and position it to a given depth value is calculated based on a measurement of the depths of the objects present in the scene. When these depths have been measured, a depth is determined so that most of the scene is behind the focal plane. This is considered to be the case, for example, when more than 50% of the surface of the images appears as located behind the focal plane.
  • the depth is not necessary to actually calculate the depth values. Since the depth is related to the parallax value and therefore to the translation value, it is sufficient to calculate the translation values for the different objects of the scene and to apply a translation value relative to the images calculated as a function of these values of translation for objects.
  • Several focal plane management policies can be adopted here. According to a first exemplary embodiment, it is sought to place the focal plane so that most of the scene is behind the focal plane.
  • FIG. 5 An exemplary embodiment of the focal plane management method according to this first embodiment is illustrated in FIG. 5.
  • a first step 5.1 at least one zone is determined in the left and right images. Indeed, according to some embodiments, the entire image is analyzed or focuses on an area of interest in the image, typically a central area. According to other embodiments, the images are divided into a plurality of zones. The values, called Z, of dominant translation of each of the zones of the image are then calculated. The cutting can be done, for example, according to the method illustrated in FIG. 6. In the exemplary embodiment, an image 6.1 is divided into a set of zones 6.2 here determined by a 9-zone fair grid.
  • Each zone is projected on a line according to the diagram of FIG. 7.
  • An area 7.2 or each column is summed according to the operation 7.3 to give a single line 7.4. In this way we obtain a line that is representative of the area along the horizontal axis.
  • the correlation can be done by any other means, for example by sum of the squares of the differences.
  • the translation can also be determined by calculating the Fourier transform of the left and right lines, which is then divided into members. The inverse Fourier transform is then calculated. The index of the maximum standard value in the sense of the complex numbers then gives the optimal translation value.
  • the translation can also be determined by direct calculation of parallaxes in two dimensions, for example by comparison of sub-blocks of the image, ie without prior projection of the images on lines.
  • This method is naturally more precise but slower. Correlation on the projection of each zone on a line brings a gain in terms of computation time compared to a correlation made over the entire zone. Since the translation we are trying to measure is only a horizontal translation, the result is reliable. This is particularly important when the process is applied in real time when rendering the scene on a device with a limited computing resource.
  • a step 5.2 it is advantageous to eliminate the areas for which the dominant translation value obtained is aberrant if there is one. That is, values that are too small or too large or too different from the other values obtained. This step is optional.
  • the translation values obtained are grouped into groups of similar values.
  • the idea here is to group areas of the image representing close objects in terms of depth.
  • a threshold is fixed on the minimum size of a group in number of zones during a step 5.4. Only those groups with a minimum number of members are retained. For example, groups consisting of a single element are eliminated. We try to divide the image into significant groups of depths. We focus on objects of a certain size on the screen.
  • step 5.5 the averages of the different dominant translation values obtained for each zone are calculated. We then choose the highest average value that corresponds to the closest significant group. In this way, the focal plane is placed in front of the stage. It is always possible that certain objects, for example eliminated during the thresholding operation 5.4, or too small to significantly influence the translation of a given zone, are in front of the focal plane. However, most of the scene is thus positioned behind the focal plane.
  • step 5.6 the relative translation between the left and right images is then performed to effectively place the focal plane at the determined depth.
  • This translation can be performed on the right image relative to the left image, or on the left image relative to the right image. It is also possible to distribute the translation on both images.
  • it is sought to place the main object of the scene at the focal plane. To do this, for example, we define a single zone in the image, typically a central zone. The depth of this zone, or simply the corresponding translation, is then calculated as before and it is applied to the images to place the focal plane at this zone. Alternatively, one can also choose to place the focal plane at the depth level of the group with the most members in the algorithm described above.
  • This method can be used on a still image consisting of a single left image and a single right image. It can also be used on a stereoscopic image sequence. In the latter case, it can be performed as processing on the sequence outside the viewing process or be performed in real time during viewing or when shooting.
  • This last embodiment is particularly interesting when it is integrated in a camera for stereoscopic shooting and rendering. It can, for example, integrate in a device such as a mobile phone incorporating a stereoscopic view system, a portable game console, or a stereoscopic camera.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
  • Processing Or Creating Images (AREA)
  • Image Analysis (AREA)
PCT/EP2012/052943 2011-02-23 2012-02-21 Dispositif et procédé de gestion de la position du plan focal dans une scène stéréoscopique Ceased WO2012113797A1 (fr)

Priority Applications (3)

Application Number Priority Date Filing Date Title
JP2013554871A JP5981460B2 (ja) 2011-02-23 2012-02-21 立体シーンにおける焦点面の位置を管理するためのデバイスおよび方法
EP12704435.2A EP2678823B1 (fr) 2011-02-23 2012-02-21 Dispositif et procédé de gestion de la position du plan focal dans une scène stéréoscopique
US14/000,462 US9800865B2 (en) 2011-02-23 2012-02-21 Device and method for managing the position of the focal plane in a stereoscopic scene

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
FR1151467A FR2971875B1 (fr) 2011-02-23 2011-02-23 Dispositif et procede de gestion de la position du plan focal dans une scene stereoscopique
FR11/51467 2011-02-23

Publications (1)

Publication Number Publication Date
WO2012113797A1 true WO2012113797A1 (fr) 2012-08-30

Family

ID=44508464

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/EP2012/052943 Ceased WO2012113797A1 (fr) 2011-02-23 2012-02-21 Dispositif et procédé de gestion de la position du plan focal dans une scène stéréoscopique

Country Status (5)

Country Link
US (1) US9800865B2 (enExample)
EP (1) EP2678823B1 (enExample)
JP (1) JP5981460B2 (enExample)
FR (1) FR2971875B1 (enExample)
WO (1) WO2012113797A1 (enExample)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109087352B (zh) * 2018-08-16 2021-07-13 数坤(北京)网络科技股份有限公司 一种心脏冠脉优势型自动判别方法

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6163337A (en) * 1996-04-05 2000-12-19 Matsushita Electric Industrial Co., Ltd. Multi-view point image transmission method and multi-view point image display method
US20040114218A1 (en) * 2001-04-12 2004-06-17 Adam Karlsson Method in microscopy and a microscope, where subimages are recorded and puzzled in the same coordinate system to enable a precise positioning of the microscope stage
US20070047040A1 (en) * 2005-08-31 2007-03-01 Samsung Electronics Co., Ltd. Apparatus and method for controlling depth of three-dimensional image
EP1884893A1 (en) * 2006-08-03 2008-02-06 Mitsubishi Electric Information Technology Centre Europe B.V. Sparse integral image description with application to motion analysis
EP2280555A2 (en) * 2009-07-24 2011-02-02 FUJIFILM Corporation Imaging Device And Imaging Method

Family Cites Families (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0228055B1 (en) * 1985-12-27 1992-03-04 Mitsubishi Denki Kabushiki Kaisha Driving equipment for objective lens
JPH0984014A (ja) * 1995-09-20 1997-03-28 Toppan Printing Co Ltd 画像動き予測方法および予測装置
JPH11243546A (ja) * 1998-02-25 1999-09-07 Sony Corp 画像符号化装置及び方法
JP2003028635A (ja) * 2001-07-16 2003-01-29 Honda Motor Co Ltd 画像測距装置
EP1335162A3 (en) 2002-02-06 2006-03-15 Essef Corporation d.b.a. Pentair Water Treatment Pressure vessel
JP4398216B2 (ja) * 2003-10-17 2010-01-13 富士重工業株式会社 情報表示装置および情報表示方法
KR100524077B1 (ko) * 2003-11-13 2005-10-26 삼성전자주식회사 중간 영상 합성을 위한 시간적 평활화 장치 및 방법
JP2010278878A (ja) * 2009-05-29 2010-12-09 Fujifilm Corp 立体画像表示装置及びその表示画像切替方法
US10134150B2 (en) * 2010-08-10 2018-11-20 Monotype Imaging Inc. Displaying graphics in multi-view scenes

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6163337A (en) * 1996-04-05 2000-12-19 Matsushita Electric Industrial Co., Ltd. Multi-view point image transmission method and multi-view point image display method
US20040114218A1 (en) * 2001-04-12 2004-06-17 Adam Karlsson Method in microscopy and a microscope, where subimages are recorded and puzzled in the same coordinate system to enable a precise positioning of the microscope stage
US20070047040A1 (en) * 2005-08-31 2007-03-01 Samsung Electronics Co., Ltd. Apparatus and method for controlling depth of three-dimensional image
EP1884893A1 (en) * 2006-08-03 2008-02-06 Mitsubishi Electric Information Technology Centre Europe B.V. Sparse integral image description with application to motion analysis
EP2280555A2 (en) * 2009-07-24 2011-02-02 FUJIFILM Corporation Imaging Device And Imaging Method

Also Published As

Publication number Publication date
JP2014507906A (ja) 2014-03-27
EP2678823B1 (fr) 2020-09-30
US9800865B2 (en) 2017-10-24
FR2971875A1 (fr) 2012-08-24
FR2971875B1 (fr) 2017-11-03
US20140055581A1 (en) 2014-02-27
JP5981460B2 (ja) 2016-08-31
EP2678823A1 (fr) 2014-01-01

Similar Documents

Publication Publication Date Title
EP3101624B1 (en) Image processing method and image processing device
TWI497980B (zh) 處理3d立體影像的系統與方法
US20110228051A1 (en) Stereoscopic Viewing Comfort Through Gaze Estimation
EP3221841B1 (fr) Procede et dispositif de filtrage adaptatif temps reel d'images de disparite ou de profondeur bruitees
EP2828834A2 (fr) Modèle et procédé de production de modèle 3d photo-réalistes
EP2070338A2 (fr) Dispositif d'affichage pour la visualisation stéréoscopique
FR3011368A1 (fr) Procede et dispositif pour le renforcement de la forme des bords pour l'amelioration visuelle du rendu base sur des images de profondeur d'un flux video en trois dimensions
FR2801123A1 (fr) Procede de creation automatique de maquette numerique a partir de couples d'images stereoscopiques
KR20180048082A (ko) 집적영상 디스플레이의 화질 평가 장치 및 방법
EP3759430B1 (fr) Systeme de modelisation tridimensionnelle d'une scene par photogrammetrie multi-vue
US20130286010A1 (en) Method, Apparatus and Computer Program Product for Three-Dimensional Stereo Display
EP2678823B1 (fr) Dispositif et procédé de gestion de la position du plan focal dans une scène stéréoscopique
FR2969791A1 (fr) Procede de correction de l'hyperstereoscopie et systeme de visualisation de casque associe
FR3052565B1 (fr) Dispositif individuel d'immersion visuelle pour personne en mouvement
EP2678825B1 (fr) Procédé de calibrage d'un dispositif de prise de vue stéréoscopique
CA3131716C (fr) Procede de generation d'un point de vue stereoscopique a valeur d'entraxe modifiable
EP3072110B1 (fr) Procédé d'estimation du mouvement d'un objet
WO2025141251A1 (fr) Procede de traitement d'images d'une scene pour restituer un changement de point de vue de la scene
Lavrushkin et al. Stereoscopic quality assessment of 1,000 VR180 videos using 8 metrics
WO2025141252A1 (fr) Procede de rendu 3d pour visiocasques et dispositifs d'affichages 3d a partir d'une image d'une scene
Iatsun et al. Spatio-temporal modeling of visual attention for stereoscopic 3D video
Navarro et al. Extended depth-of-field in Integral Imaging by depth-dependent deconvolution
WO2025012524A1 (fr) Procede et dispositif d'affichage personnalise d'une scene a partir d'image(s) de ladite scene, sur un support d'affichage
WO2000004503A1 (fr) Procede de modelisation d'objets ou de scenes 3d
FR3030971A1 (fr) Procede d'estimation de flou independant du contenu de la scene et utilisation dans un systeme de mise au point automatique

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 12704435

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2013554871

Country of ref document: JP

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE

WWE Wipo information: entry into national phase

Ref document number: 2012704435

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 14000462

Country of ref document: US