WO2017141139A1 - A method for image transformation - Google Patents
A method for image transformation Download PDFInfo
- Publication number
- WO2017141139A1 WO2017141139A1 PCT/IB2017/050688 IB2017050688W WO2017141139A1 WO 2017141139 A1 WO2017141139 A1 WO 2017141139A1 IB 2017050688 W IB2017050688 W IB 2017050688W WO 2017141139 A1 WO2017141139 A1 WO 2017141139A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- viewing direction
- main viewing
- display unit
- user
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 25
- 230000009466 transformation Effects 0.000 title description 11
- 230000004048 modification Effects 0.000 claims abstract description 48
- 238000012986 modification Methods 0.000 claims abstract description 48
- 238000004590 computer program Methods 0.000 claims description 25
- 230000006870 function Effects 0.000 claims description 21
- 230000004044 response Effects 0.000 claims description 9
- 230000004886 head movement Effects 0.000 claims description 4
- 210000003128 head Anatomy 0.000 description 19
- 230000000694 effects Effects 0.000 description 11
- 230000008859 change Effects 0.000 description 8
- 239000013598 vector Substances 0.000 description 7
- 238000004458 analytical method Methods 0.000 description 6
- 238000010586 diagram Methods 0.000 description 5
- 230000003287 optical effect Effects 0.000 description 5
- 235000019646 color tone Nutrition 0.000 description 3
- 238000000844 transformation Methods 0.000 description 3
- 244000089409 Erythrina poeppigiana Species 0.000 description 2
- 235000009776 Rathbunia alamosensis Nutrition 0.000 description 2
- 239000003086 colorant Substances 0.000 description 2
- 238000004891 communication Methods 0.000 description 2
- 238000012937 correction Methods 0.000 description 2
- 239000011521 glass Substances 0.000 description 2
- 230000003993 interaction Effects 0.000 description 2
- 230000008569 process Effects 0.000 description 2
- 238000012545 processing Methods 0.000 description 2
- 241001442234 Cosa Species 0.000 description 1
- 230000001133 acceleration Effects 0.000 description 1
- 238000004422 calculation algorithm Methods 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 230000002708 enhancing effect Effects 0.000 description 1
- 238000013213 extrapolation Methods 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 239000007787 solid Substances 0.000 description 1
- 230000005236 sound signal Effects 0.000 description 1
- 238000011426 transformation method Methods 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/06—Topological mapping of higher dimensional structures onto lower dimensional surfaces
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B30/00—Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
- G06F3/012—Head tracking input arrangements
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/14—Digital output to display device ; Cooperation and interconnection of the display device with other functional units
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T15/00—3D [Three Dimensional] image rendering
- G06T15/10—Geometric effects
- G06T15/20—Perspective computation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/80—Analysis of captured images to determine intrinsic or extrinsic camera parameters, i.e. camera calibration
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/111—Transformation of image signals corresponding to virtual viewpoints, e.g. spatial image interpolation
- H04N13/117—Transformation of image signals corresponding to virtual viewpoints, e.g. spatial image interpolation the virtual viewpoint locations being selected by the viewers or determined by viewer tracking
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/30—Image reproducers
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/30—Image reproducers
- H04N13/332—Displays for viewing with the aid of special glasses or head-mounted displays [HMD]
- H04N13/344—Displays for viewing with the aid of special glasses or head-mounted displays [HMD] with head-mounted left-right displays
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/30—Image reproducers
- H04N13/366—Image reproducers using viewer tracking
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/30—Image reproducers
- H04N13/398—Synchronisation thereof; Control thereof
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/698—Control of cameras or camera modules for achieving an enlarged field of view, e.g. panoramic image capture
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G2320/00—Control of display operating conditions
- G09G2320/02—Improving the quality of display appearance
- G09G2320/0261—Improving the quality of display appearance in the context of movement of objects on the screen or movement of the observer relative to the screen
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G2340/00—Aspects of display data processing
- G09G2340/04—Changes in size, position or resolution of an image
- G09G2340/045—Zooming at least part of an image, i.e. enlarging it or shrinking it
Definitions
- Various embodiments relate to video effects in immersive displays, and more particularly to a method for image transformation.
- a viewer In an immersive display, a viewer is surrounded by one or more images displayed either as a subsection of a sphere or as the whole sphere around the viewer. The viewer may see the display image on a physical screen surface around him or use a head-mounted display (HMD) display.
- HMD head-mounted display
- a head mounted display is a display device, worn on the head that has a small display optic in front of the eyes.
- One use case for the head mounted display is the ability to watch live or pre-recorded videos.
- Another use case is the ability to watch computer created content, such as three- dimensional (3D) games. When watching a 360 degree, i.e.
- a user may be able to feel much more immersed inside the world of the video compared to watching a conventional two- dimensional (2D) display.
- 3D stereo image effect in the video may enhance the immersive feeling even further.
- 360 degree stereo panorama videos are a known way to distribute 3D videos meant to be viewed in head mounted displays. While the computer created virtual reality content may be rendered to provide the viewer with an immersive experience having a deep feeling of "presence", like interacting with and moving inside the 3D content stereo panorama videos based on filmed (captured e.g. by a stereo/multi-camera device) virtual reality content is only provided with a limited set of 3D effects and viewer interaction. The lack of many 3D effects in the video playback may reduce the feeling of presence and may give the video artificial look.
- a method comprising: obtaining, by an immersive display unit, at least one image to be displayed on the immersive display unit; obtaining an input indicating a main viewing direction of a user of the immersive display unit; determining a first value for at least one image modification parameter for an image element residing in the main viewing direction; determining a second value for the at least one image modification parameter for image elements deviating from said main viewing direction, the second value being a function of an angle between the main viewing direction and a direction from the user towards an image location deviating from the main viewing direction; and displaying the at least one image according to the determined values of the at least one image modification parameter on the immersive display unit.
- the input is determined on the basis of a movement of the head of the user. According to an embodiment, the input is obtained via a user interface of the immersive display unit.
- the input is obtained automatically in response to the user positioning within the immersive display unit.
- the input further indicates strength of the input in said main viewing direction.
- the at least one image modification parameter is a zoom ratio of the image element.
- the zoom ratio of the n th equally spaced image element around a 360 degree display circle is calculated as:
- m is the total number of image elements of equal width around a 360 degree display circle
- a(n) is the angle deviating from the main viewing direction, corresponding to the n th image element before zooming
- a '(n) is the angle of a modified n th image element after zooming
- a '(n) being a function of a(n) and the user input value for image magnification in the main viewing direction.
- the at least one image modification parameter is at least one of:
- the immersive display unit is a head mounted device (HMD) with head tracking system or a curved display at least partly surrounding the head of the user.
- HMD head mounted device
- the at least one image comprises video image data.
- an apparatus comprising an immersive display unit comprising at least one processor, memory including computer program code, the memory and the computer program code configured to, with the at least one processor, cause the apparatus to at least: obtain at least one image to be displayed on the immersive display unit; obtain an input indicating a main viewing direction of a user of the immersive display unit; determine a first value for at least one image modification parameter for an image element residing in the main viewing direction; determine a second value for the at least one image modification parameter for image elements deviating from said main viewing direction, the second value being a function of an angle between the main viewing direction and a direction from the user towards an image location deviating from the main viewing direction; and display the at least one image according to the determined values of the at least one image modification parameter on the immersive display unit.
- a computer readable storage medium stored with code thereon for use by an apparatus, which when executed by a processor, causes the apparatus to perform the method according to any of the above embodiments.
- Fig. la shows an example of a multi -camera system as a simplified block diagram, in accordance with an embodiment
- Fig. lb shows a perspective view of a multi-camera system, in accordance with an embodiment
- Fig. 2 shows an example of a video playback apparatus as a simplified block diagram, in accordance with an embodiment
- Fig. 3 shows a flow chart of an image transformation process according to an embodiment of the invention
- Figs. 4a, 4b illustrate an example of moving the display virtually closer to a viewer according to an embodiment of the invention
- Fig. 5 illustrates an example of geometric analysis for determining a modification ratio according to an embodiment of the invention
- Fig. 6 is a graph illustrating an example of image element movements according to an embodiment of the invention.
- Fig. 7 is a graph illustrating an example of the magnification as a function of the original location of the image element according to an embodiment of the invention.
- Fig. 8 shows a qualitative explanation of the results of the image transformation process
- Figs. 9a, 9b show the basic principle of changing the image properties as a function of an angle in a cylindrical display and in a spherical display, respectively.
- a video player When 360 degree stereo panorama video is viewed in an immersive multimedia display unit, for example, a head mounted display with a video player software, a video player may be able to create the similar effect of the viewer moving in the immersed space which is present in the real world.
- the forward and/or backward motion of a head of the user may make the video more realistic looking as the objects in the foreground appear to move slightly in relation to background objects when the head is moved.
- stereo panorama videos based on filmed virtual reality content that is captured e.g. by a stereo- or multi-camera device, the effect of the movement of foreground and background objects relative to each other is difficult to achieve, at least in a computationally lightweight manner.
- Figure la illustrates an example of a multi -camera system 100, which may be able to capture and produce 360 degree stereo panorama video.
- the multi-camera system 100 comprises two or more camera units 102.
- the number of camera units 102 is eight, but may also be less than eight or more than eight.
- Each camera unit 102 is located at a different location in the multi- camera system, and may have a different orientation with respect to other camera units 102, so that they may capture a part of the 360 degree scene from different viewpoints substantially simultaneously.
- a pair of camera units 102 of the multi-camera system 100 may correspond with left and right eye viewpoints at a time.
- the camera units 102 may have an omnidirectional constellation, so that it has a 360° viewing angle in a 3D-space.
- such multi-camera system 100 may be able to see each direction of a scene so that each spot of the scene around the multi-camera system 100 can be viewed by at least one camera unit 102 or a pair of camera units 102.
- any two camera units 102 of the multi-camera system 100 may be regarded as a pair of camera units 102.
- a multi-camera system of two cameras may have only one pair of camera units
- a multi-camera system of three cameras may have three pairs of camera units
- a multi-camera system of four cameras may have six pairs of camera units, etc.
- a multi-camera system 100 comprising N camera units 102, where N is an integer greater than one, may have N(N-l)/2 pairs of camera units 102. Accordingly, images captured by the camera units 102 at a certain time may be considered as N(N-l)/2 pairs of captured images.
- the multi-camera system 100 of Figure la may also comprise a processor 104 for controlling operations of the multi-camera system 100.
- a memory 106 for storing data and computer code to be executed by the processor 104, and a transceiver 108 for communicating with, for example, a communication network and/or other devices in a wireless and/or wired manner.
- the multi-camera systemlOO may further comprise a user interface (UI) 110 for displaying information to the user, for generating audio signals, and/or for receiving user inputs.
- UI user interface
- the multi-camera system 100 need not comprise each feature mentioned above, or may comprise other features as well.
- Figure la also illustrates some operational elements which may be implemented, for example, as a computer code which can be executed in the processor 104, in hardware, or both to perform a desired function.
- An optical flow estimation 114 may perform optical flow estimation for pair of images of different camera units 102. Transform vectors or other information indicative of an amount interpolation/extrapolation to be applied to different parts of a viewport may have been stored into the memory 106 or they may be calculated e.g. as a function of the location of a pixel in question. The operation of the elements will be described later in more detail. It should be noted that there may also be other operational elements in the multi-camera system 100 than those depicted in Figure la.
- Figure lb shows a perspective view of the multi-camera system 100, in accordance with an embodiment.
- seven camera units 102a— 102g can be seen, but the multi-camera system 100 may comprise even more camera units which are not visible from this perspective view.
- Figure lb also shows two microphones 112a, 112b, but the apparatus may also comprise one microphone or more than two microphones.
- the multi-camera system 100 may be controlled by another device, wherein the multi-camera system 100 and the other device may communicate with each other and a user may use a user interface of the other device for entering commands, parameters, etc. and the user may be provided with information from the multi-camera system 100 via the user interface of the other device.
- a viewport is a part of the scene which is displayed by a head mounted display at a time. Both left and right eye images may have overlapping, but slightly different viewports.
- a camera space, or camera coordinates stands for a coordinate system of an individual camera unit 102 whereas a world space, or world coordinates, stands for a coordinate system of the multi-camera system 100 as a whole.
- An optical flow may be used to describe how objects, surfaces, and edges in a visual scene move or transform, when an observing point moves from a location of one camera to a location of another camera. In some embodiments, there need not be any actual movement but it may virtually be determined how the view of the scene might change when a viewing point is moved from one camera unit to another camera unit.
- FIG. 2 shows an example of a video playback apparatus 200 as a simplified block diagram, in accordance with an embodiment.
- a non-limiting example of video playback apparatus 200 includes, an immersive display unit.
- An example of the immersive display unit includes, but is not limited to, a head mounted display.
- the video playback apparatus 200 may comprise, for example, one or two displays 202 for video playback. When two displays are used a first display 202a may display images for a left eye and a second display 202b may display images for a right eye, in accordance with an embodiment. In case of only one display 202, that display 202 may be used to display images for the left eye on the left side of the display 202 and to display images for the right eye on the right side of the display 202.
- the one or two displays are configured to create one or more images surrounding the viewer partly or completely, such as a head-mounted display with head tracking system or a cylindrical or a spherical display curving in 2D or in 3D at least partly, but possibly 360° around the viewer.
- the video playback apparatus 200 may be provided with encoded data streams via a communication interface 204 and a processor 206 may perform control operations for the video playback apparatus 200 and may also perform operations to reconstruct video streams for displaying on the basis of received encoded data streams.
- the video playback apparatus 200 may further comprise a processor 206 for reconstructing video streams for displaying on the basis of received encoded data streams.
- the decoding 208 is implemented, for example, as a software code, which can be executed by the processor 206 to perform the desired function, in hardware, or in both.
- the video playback apparatus 200 may further comprise a user input 212 for receiving e.g.
- the video playback apparatus 200 may comprise an image modification unit 214 which may perform image modification on the basis of modification information provided by a user input and at least one image modification function and transform image elements as will be described later in this specification.
- the image modification unit 214 can be implemented, for example, as a software code, which can be executed by the processor to perform the desired function, in hardware, or in both.
- the video playback device 200 need not comprise each of the above elements or may also comprise other elements.
- the decoding element 208 may be a separate device wherein that device may perform decoding operations and provide decoded data stream to the video playback device 200 for further processing and displaying decoded video streams.
- Video information to be processed may have been captured and processed by two or more camera units 102 to obtain a panorama video, for example a 360 degree panorama video. From this panorama video, a first stream of images representing e.g. left eye views and a second stream of images representing e.g. right eye views of the scene may be encoded.
- an immersive display unit for example a head mounted display, arranged to create an image surrounding a user obtains (300) at least one image to be displayed on the immersive display, and an input indicating a main viewing direction of the immersive display unit is obtained (302).
- the input indicating the main viewing direction of the immersive display unit is given by a user. In another embodiment, the input indicating the main viewing direction is received automatically.
- a first value for at least one image modification parameter for an image element residing in the main viewing direction is determined (304).
- a second value is determined (306) for at least one image modification parameter for image elements deviating from the main viewing direction as a function of an angle between the main viewing direction and a direction from the user towards the image location deviating from the main viewing direction.
- the at least one image is displayed (308) according to the determined values of the at least one image modification parameter on the display.
- the immersive display unit may be a head mounted device (HMD) with a head tracking system or a curved display at least partly surrounding the head of the user.
- HMD head mounted device
- the user of the HMD may give a simple user input for indicating the main viewing direction.
- the input for indicating the main viewing direction may be obtained automatically in response to the user positioning him/herself within the immersive display unit.
- a value of at least one image modification parameter in the main viewing direction is used as a reference value, and the values of the image modification parameter in other directions of the 3D image are calculated as a function of the angle deviating from the main viewing direction.
- the user input is determined on the basis of a movement of the head of the user.
- a movement of the head such as a nod of the head forward
- the user input is obtained via a user interface of the HMD.
- the main viewing direction may be submitted via the user interface before the user starts wearing the HMD.
- there is an external user interface device connected to the HMD and the user may give the user input via the external user interface device even if wearing the HMD.
- the user input further indicates strength of the user input in the main viewing direction.
- the strength value of the user input may be used as the basis for adjusting the value of the image modification parameter in the main viewing direction.
- the strength may be obtained e.g. by a movement sensor connected to the HMD, which determines the acceleration and/or distance of the head movement. If an external user interface device connected to the HMD is used, the strength may be given as a numerical value or otherwise proportional to a predefined scale.
- the at least one image modification parameter is a zoom ratio of the image element.
- the element of the display screen residing in the main viewing direction may be enlarged, while the image parts on the left and right side of the viewer do not change in size, but just move a little backwards.
- Figures 4a and 4b show top views of cylindrical displays and a viewer in the center of the display.
- the viewer wears the HMD and sees a cylindrical display around him/her.
- the part of the display screen residing in the main viewing direction i.e. in front of the viewer
- the viewer has the feeling that the display around him has moved to the dashed line position of Fig. 4b, although the physical display stays all the time in the original location (solid line).
- the arc 500 represents the physical cylindrical display around the viewer 502 at a radius r.
- the image element 504 is represented by an arc, which in its initial state starts at angle a from the main viewing direction (i.e. to the left from the upright direction on the image). The viewer sees the width of the arc of the image element 504 in angle ⁇ .
- the angle ⁇ is here shown as much larger than 1 degree. In practical implementations, ⁇ would be « 1 degree.
- the viewer 502 stays in the center of the cylindrical display.
- the dimensions of the image around the viewer change so that viewer feels that the circle of the display has moved distance k closer to the viewer.
- the virtual position of the "moved" display circle is marked in dashed line 500'. The viewer would see the original arc of the image element 504 now in a new location, in angle ⁇ , as the image element 504' .
- the actual physical cylindrical display 500 around the viewer remains at the same location.
- the physical cylindrical display 500 should show the original image element 504 as the black arc 504' ' on the display circle.
- Equations (1) and (2) can be combined as:
- Equation (3) now defines how the original angle a (corresponding to the start edge of the chosen image element) changes to a+ ⁇ when the circle display of radius r is virtually moved by distance k.
- a magnification factor M corresponding to ⁇ / ⁇ , can be calculated as follows:
- the 360 degree display circle may be divided into m image elements of equal width, whereby the angle of one element from the viewer's perspective corresponds to 360°/m.
- its starting angle i.e. the angle of the first side of the image element from the main viewing direction
- a(n) (n-l)*360°/m.
- the starting angle of each image element changes from a(n) to a '(n ), which corresponds to a+ ⁇ , as shown above.
- the angle difference in the modified image between a '(n) and a '(n+l) is a '(n+1 ) - a '(n).
- M magnification factor
- M [a '(n+l) - a '(n)] / (360 o /m) (4)
- m is the number of image elements of equal width around a 360 degree display circle with the viewer at the circle center
- a(n) is the angle deviating from the main viewing direction, corresponding to the n image element before zooming
- a '(n) is the angle of a modified n image element after zooming
- a '(n) being a function of a(n) and the user input value for image magnification in the main viewing direction.
- FIG. 6 shows an example of image element movements according to Eq. (3).
- the parameter m is selected as 100, which is large enough to obtain smooth output curves.
- the dashed line shows the original location of the image elements, whereas the solid line shows how the location of the image elements after the image modification. It can be seen that, for example, at 0 and 180 degrees there is no movement, whereas at around 90 and 270 degrees (i.e. directly left and right from the main viewing direction) the movements of the image elements are at maximum.
- the dashed line shows the original magnification of the image elements, which equals to 1.
- the solid line shows the magnification ratio M according to Eq. (4); i.e. image element width in the new location vs. the element width in the original situation.
- the magnification is maximum (1.25) at 0 degrees (i.e. in the main viewing direction straight ahead from the viewer) and minimum (0.83) in the opposite direction. At around 90 and 270 degrees (left/right sides of the viewer), there is no magnification.
- FIG 8 shows a qualitative explanation of the above embodiments.
- the viewer sees the image elements around the sphere such that each image element has the same angular slot marked with evenly spaced black lines.
- the size of image elements is modified such that the new element borders are marked with grey lines and black knobs.
- the image elements just on the sides of the viewer have not changed their size but a moved slightly backwards, and the image element just behind the viewer have squeezed.
- the at least one image modification parameter is at least one of: a blurring strength of the image element
- the above embodiments may be applied to other image modification parameters, as well.
- the blurring strength can be applied such that the image elements close to the main viewing direction are shown sharp in focus, whereas the image elements deviating from the main viewing direction are shown blurred (out of focus), and the blurring strength depends on the deviating angle according to a pre-defined formula.
- the tone of the color (brighter/darker) may be adjusted in the image elements deviating from the main viewing direction.
- the color of the image elements may change according to a predefined scheme (e.g. red-purple-blue) in the image elements deviating from the main viewing direction.
- the embodiments may be applied to still images, even in GIF format, presented on displays showing images which surround the viewer.
- the embodiments may be applied to directional audio.
- the audio level may be raised in the direction where the image is enlarged.
- two or more image modification parameters may be adjusted simultaneously.
- the above embodiments have been mainly described in relation to cylindrical 2D 360 degree displays.
- Figure 9a shows the main direction vector r and the deviated direction vector s and the angle a between the vectors in a cylindrical display.
- Figure 9b shows the main direction vector r and the deviated direction vector s and the angle a between the vectors inside a spherical display.
- the viewer inside a spherical display or wearing a 360 degree head- mounted display may be watching a virtual reality video where s/he is moving, e.g. flying, in direction r.
- the surrounding image elements may change: the view in front of the viewer may remain in focus and in the original color, possibly provided with magnification at the same time, while the views by the sides of the viewer, i.e. on left, right, top and below, may become blurry and change colors or color tones. This would provide the viewer with the feeling that s/he is moving through more and less blurry rings of different colors.
- the various embodiments may be applied when a user views the immersive video content and interacts with it in real-time by the applying the various image transformations (zooming, blurring etc.).
- the various embodiments may be applied in a post-production (e.g. editing) phase of the content.
- the editor e.g. a person or a computer program
- the footage with the transformations implemented may then be recorded for further use.
- the image transformation procedure needs to be performed for the two display views, for the two eyes, separately.
- parallax correction procedures for enhancing the 3D effect by further adjusting the relative movement of foreground and background objects, thereby making the scene more realistic looking.
- Various parallax corrections are known, as such, in the zooming of 3D images.
- parallax motion effect may be achieved by embedding optical flow data from stereo images into the video stream and at the playback time slightly warping left and right eye images according to the viewport and optical flow data.
- the warping may be done so that the video quality or stereo effect does not degrade at all in the middle of the viewport, which is the area that may be critical for the perceived video quality. Instead the video quality and stereo effect may get degraded gradually towards the edges of the viewport, but human stereo vision ability is naturally already reduced in that direction.
- the image transformation may relate to creating the feeling of image rotation around one of the axis in the x,y,z, space. For example, the viewer may look straight ahead at one point in the horizon and starts to see the horizon of the image to shake around the point.
- the above embodiments may provide various advantages.
- the feeling of immersiveness and interactivity is increased to the filmed virtual reality (VR) content that usually have image content wherein the viewer cannot move around and has no interaction possibility.
- the image modification algorithm is simple with a short processing time, thereby enabling the transformation to be performed even on live streaming VR content.
- the various embodiments may be implemented in hardware or special purpose circuits or any combination thereof. While various embodiments may be illustrated and described as block diagrams or using some other pictorial representation, it is well understood that these blocks, apparatus, systems, techniques or methods described herein may be implemented in, as non- limiting examples, hardware, software, firmware, special purpose circuits or logic, general purpose hardware or controller or other computing devices, or some combination thereof.
- Embodiments of the inventions may be practiced in various components such as integrated circuit modules.
- a skilled man appreciates that any of the embodiments described above may be implemented as a combination with one or more of the other embodiments, unless there is explicitly or implicitly stated that certain embodiments are only alternatives to each other.
- the implementation may include a computer readable storage medium stored with code thereon for use by an apparatus, which when executed by a processor, causes the apparatus to perform the various embodiments or a subset of them.
- the implementation may include a computer program embodied on a non-transitory computer readable medium, the computer program comprising instructions causing, when executed on at least one processor, at least one apparatus to apparatus to perform the various embodiments or a subset of them.
- an apparatus may comprise circuitry and electronics for handling, receiving and transmitting data, computer program code in a memory, and a processor that, when running the computer program code, causes the terminal device to carry out the features of an embodiment.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Optics & Photonics (AREA)
- General Engineering & Computer Science (AREA)
- Human Computer Interaction (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Computing Systems (AREA)
- Geometry (AREA)
- Computer Graphics (AREA)
- Controls And Circuits For Display Device (AREA)
Abstract
A method comprising: obtaining, by an immersive display unit, at least one image to be displayed on the display; obtaining an input indicating a main viewing direction of a user of the immersive display unit; determining a first value for at least one image modification parameter for an image element residing in the main viewing direction; determining a second value for the at least one image modification parameter for image elements deviating from said main viewing direction, the second value being a function of an angle between the main viewing direction and a direction from the user towards an image location deviating from the main viewing direction; and displaying the at least one image according to the determined values of the at least one image modification parameter on the immersive display unit.
Description
A METHOD FOR IMAGE TRANSFORMATION
Field
Various embodiments relate to video effects in immersive displays, and more particularly to a method for image transformation.
Background of the invention
In an immersive display, a viewer is surrounded by one or more images displayed either as a subsection of a sphere or as the whole sphere around the viewer. The viewer may see the display image on a physical screen surface around him or use a head-mounted display (HMD) display. A head mounted display is a display device, worn on the head that has a small display optic in front of the eyes. One use case for the head mounted display is the ability to watch live or pre-recorded videos. Another use case is the ability to watch computer created content, such as three- dimensional (3D) games. When watching a 360 degree, i.e. the whole sphere, panoramic video with a head mounted display equipped with head tracking technology, a user may be able to feel much more immersed inside the world of the video compared to watching a conventional two- dimensional (2D) display. 3D stereo image effect in the video may enhance the immersive feeling even further.
360 degree stereo panorama videos are a known way to distribute 3D videos meant to be viewed in head mounted displays. While the computer created virtual reality content may be rendered to provide the viewer with an immersive experience having a deep feeling of "presence", like interacting with and moving inside the 3D content stereo panorama videos based on filmed (captured e.g. by a stereo/multi-camera device) virtual reality content is only provided with a limited set of 3D effects and viewer interaction. The lack of many 3D effects in the video playback may reduce the feeling of presence and may give the video artificial look.
One such issue in the 360 degree stereo panorama video format, particularly when using head mounted displays, is that the experienced distance to an object in front of the viewer remains the same even if the viewer moves his/her head closer to the object. This is due to the fact the viewer is bound to the same position from where the camera has captured the view. In real life, when the head is moved forward/backward, the viewpoint moves slightly along the head and the foreground objects appear to move in relation to the background objects.
Summary
Various embodiments include a method, an apparatus and a computer program, which are characterized by what is stated in the independent claims. Various embodiments of the invention are disclosed in the dependent claims.
According to a first embodiment, there is disclosed a method comprising: obtaining, by an immersive display unit, at least one image to be displayed on the immersive display unit; obtaining an input indicating a main viewing direction of a user of the immersive display unit; determining a first value for at least one image modification parameter for an image element residing in the main viewing direction; determining a second value for the at least one image modification parameter for image elements deviating from said main viewing direction, the second value being a function of an angle between the main viewing direction and a direction from the user towards an image location deviating from the main viewing direction; and displaying the at least one image according to the determined values of the at least one image modification parameter on the immersive display unit.
According to an embodiment, the input is determined on the basis of a movement of the head of the user. According to an embodiment, the input is obtained via a user interface of the immersive display unit.
According to an embodiment, the input is obtained automatically in response to the user positioning within the immersive display unit.
According to an embodiment, the input further indicates strength of the input in said main viewing direction.
According to an embodiment, the at least one image modification parameter is a zoom ratio of the image element.
According to an embodiment, the zoom ratio of the nth equally spaced image element around a 360 degree display circle is calculated as:
M = [α ' (η+1) - α ' (n)] / (360°/m),
where m is the total number of image elements of equal width around a 360 degree display circle, a(n) is the angle deviating from the main viewing direction, corresponding to the nth image element before zooming, and a '(n) is the angle of a modified nth image element after zooming,
and a '(n) being a function of a(n) and the user input value for image magnification in the main viewing direction.
According to an embodiment, the at least one image modification parameter is at least one of:
- a blurring strength of the image element;
a color tone strength of the image element;
a color selection of the image element.
According to an embodiment, the immersive display unit is a head mounted device (HMD) with head tracking system or a curved display at least partly surrounding the head of the user.
According to an embodiment, the at least one image comprises video image data.
According to a second embodiment, there is provided an apparatus comprising an immersive display unit comprising at least one processor, memory including computer program code, the memory and the computer program code configured to, with the at least one processor, cause the apparatus to at least: obtain at least one image to be displayed on the immersive display unit; obtain an input indicating a main viewing direction of a user of the immersive display unit; determine a first value for at least one image modification parameter for an image element residing in the main viewing direction; determine a second value for the at least one image modification parameter for image elements deviating from said main viewing direction, the second value being a function of an angle between the main viewing direction and a direction from the user towards an image location deviating from the main viewing direction; and display the at least one image according to the determined values of the at least one image modification parameter on the immersive display unit.
According to a third embodiment, there is provided a computer readable storage medium stored with code thereon for use by an apparatus, which when executed by a processor, causes the apparatus to perform the method according to any of the above embodiments.
These and other embodiment of the invention will become apparent in view of the detailed disclosure.
List of drawings In the following, various embodiments of the invention will be described in more detail with reference to the appended drawings, in which
Fig. la shows an example of a multi -camera system as a simplified block diagram, in accordance with an embodiment;
Fig. lb shows a perspective view of a multi-camera system, in accordance with an embodiment;
Fig. 2 shows an example of a video playback apparatus as a simplified block diagram, in accordance with an embodiment;
Fig. 3 shows a flow chart of an image transformation process according to an embodiment of the invention;
Figs. 4a, 4b illustrate an example of moving the display virtually closer to a viewer according to an embodiment of the invention;
Fig. 5 illustrates an example of geometric analysis for determining a modification ratio according to an embodiment of the invention;
Fig. 6 is a graph illustrating an example of image element movements according to an embodiment of the invention;
Fig. 7 is a graph illustrating an example of the magnification as a function of the original location of the image element according to an embodiment of the invention;
Fig. 8 shows a qualitative explanation of the results of the image transformation process;
and
Figs. 9a, 9b show the basic principle of changing the image properties as a function of an angle in a cylindrical display and in a spherical display, respectively.
Description of embodiments
The following embodiments are exemplary. Although the specification may refer to "an", "one", or "some" embodiment(s) in several locations, this does not necessarily mean that each such reference is to the same embodiment(s), or that the feature only applies to a single embodiment. Single features of different embodiments may also be combined to provide other embodiments.
When 360 degree stereo panorama video is viewed in an immersive multimedia display unit, for example, a head mounted display with a video player software, a video player may be able to create the similar effect of the viewer moving in the immersed space which is present in the real world. The forward and/or backward motion of a head of the user may make the video more realistic looking as the objects in the foreground appear to move slightly in relation to background objects when the head is moved. However, considering stereo panorama videos based on filmed virtual reality content that is captured e.g. by a stereo- or multi-camera device, the effect of the movement of foreground and background objects relative to each other is difficult to achieve, at least in a computationally lightweight manner.
Figure la illustrates an example of a multi -camera system 100, which may be able to capture and produce 360 degree stereo panorama video. The multi-camera system 100 comprises two or more camera units 102. In this example, the number of camera units 102 is eight, but may also be less than eight or more than eight. Each camera unit 102 is located at a different location in the multi- camera system, and may have a different orientation with respect to other camera units 102, so that they may capture a part of the 360 degree scene from different viewpoints substantially simultaneously. A pair of camera units 102 of the multi-camera system 100 may correspond with left and right eye viewpoints at a time. As an example, the camera units 102 may have an omnidirectional constellation, so that it has a 360° viewing angle in a 3D-space. In other words, such multi-camera system 100 may be able to see each direction of a scene so that each spot of the scene around the multi-camera system 100 can be viewed by at least one camera unit 102 or a pair of camera units 102. Without losing generality, any two camera units 102 of the multi-camera system 100 may be regarded as a pair of camera units 102. Hence, a multi-camera system of two cameras may have only one pair of camera units, a multi-camera system of three cameras may have three pairs of camera units, a multi-camera system of four cameras may have six pairs of camera units, etc. Generally, a multi-camera system 100 comprising N camera units 102, where N is an integer greater than one, may have N(N-l)/2 pairs of camera units 102. Accordingly, images captured by the camera units 102 at a certain time may be considered as N(N-l)/2 pairs of captured images.
The multi-camera system 100 of Figure la may also comprise a processor 104 for controlling operations of the multi-camera system 100. There may also be a memory 106 for storing data and computer code to be executed by the processor 104, and a transceiver 108 for communicating with, for example, a communication network and/or other devices in a wireless and/or wired manner. The multi-camera systemlOO may further comprise a user interface (UI) 110 for displaying information to the user, for generating audio signals, and/or for receiving user inputs. However, the multi-camera system 100 need not comprise each feature mentioned above, or may comprise other features as well. For example, there may be electric and/or mechanical elements for adjusting and/or controlling optics of the camera units 102.
Figure la also illustrates some operational elements which may be implemented, for example, as a computer code which can be executed in the processor 104, in hardware, or both to perform a desired function. An optical flow estimation 114 may perform optical flow estimation for pair of images of different camera units 102. Transform vectors or other information indicative of an amount interpolation/extrapolation to be applied to different parts of a viewport may have been stored into the memory 106 or they may be calculated e.g. as a function of the location of a pixel
in question. The operation of the elements will be described later in more detail. It should be noted that there may also be other operational elements in the multi-camera system 100 than those depicted in Figure la. Figure lb shows a perspective view of the multi-camera system 100, in accordance with an embodiment. In Figure lb seven camera units 102a— 102g can be seen, but the multi-camera system 100 may comprise even more camera units which are not visible from this perspective view. Figure lb also shows two microphones 112a, 112b, but the apparatus may also comprise one microphone or more than two microphones.
In accordance with an embodiment, the multi-camera system 100 may be controlled by another device, wherein the multi-camera system 100 and the other device may communicate with each other and a user may use a user interface of the other device for entering commands, parameters, etc. and the user may be provided with information from the multi-camera system 100 via the user interface of the other device.
Some terminology regarding the multi-camera system 100 will now be shortly described. A viewport is a part of the scene which is displayed by a head mounted display at a time. Both left and right eye images may have overlapping, but slightly different viewports. A camera space, or camera coordinates, stands for a coordinate system of an individual camera unit 102 whereas a world space, or world coordinates, stands for a coordinate system of the multi-camera system 100 as a whole. An optical flow may be used to describe how objects, surfaces, and edges in a visual scene move or transform, when an observing point moves from a location of one camera to a location of another camera. In some embodiments, there need not be any actual movement but it may virtually be determined how the view of the scene might change when a viewing point is moved from one camera unit to another camera unit.
Figure 2 shows an example of a video playback apparatus 200 as a simplified block diagram, in accordance with an embodiment. A non-limiting example of video playback apparatus 200 includes, an immersive display unit. An example of the immersive display unit includes, but is not limited to, a head mounted display. The video playback apparatus 200 may comprise, for example, one or two displays 202 for video playback. When two displays are used a first display 202a may display images for a left eye and a second display 202b may display images for a right eye, in accordance with an embodiment. In case of only one display 202, that display 202 may be used to display images for the left eye on the left side of the display 202 and to display images for the right eye on the right side of the display 202. While describing various embodiments further below, it is assumed that the one or two displays are configured to create one or more images surrounding the viewer partly or completely, such as a head-mounted display with head tracking
system or a cylindrical or a spherical display curving in 2D or in 3D at least partly, but possibly 360° around the viewer.
The video playback apparatus 200 may be provided with encoded data streams via a communication interface 204 and a processor 206 may perform control operations for the video playback apparatus 200 and may also perform operations to reconstruct video streams for displaying on the basis of received encoded data streams. The video playback apparatus 200 may further comprise a processor 206 for reconstructing video streams for displaying on the basis of received encoded data streams. There may also be a decoder 208 for decoding received data streams and a memory 210 for storing data and computer code. In an embodiment, the decoding 208 is implemented, for example, as a software code, which can be executed by the processor 206 to perform the desired function, in hardware, or in both. The video playback apparatus 200 may further comprise a user input 212 for receiving e.g. user instructions or inputs. The video playback apparatus 200 may comprise an image modification unit 214 which may perform image modification on the basis of modification information provided by a user input and at least one image modification function and transform image elements as will be described later in this specification. In an embodiment, the image modification unit 214 can be implemented, for example, as a software code, which can be executed by the processor to perform the desired function, in hardware, or in both.
It should be noted that the video playback device 200 need not comprise each of the above elements or may also comprise other elements. For example, the decoding element 208 may be a separate device wherein that device may perform decoding operations and provide decoded data stream to the video playback device 200 for further processing and displaying decoded video streams.
In the following, the image transformation method in accordance with an embodiment is described in more detail with reference to the flow diagram of Figure 3. Video information to be processed may have been captured and processed by two or more camera units 102 to obtain a panorama video, for example a 360 degree panorama video. From this panorama video, a first stream of images representing e.g. left eye views and a second stream of images representing e.g. right eye views of the scene may be encoded. In the method, an immersive display unit, for example a head mounted display, arranged to create an image surrounding a user obtains (300) at least one image to be displayed on the immersive display, and an input indicating a main viewing direction of the immersive display unit is obtained (302). In an embodiment, the input indicating the main viewing direction of the
immersive display unit is given by a user. In another embodiment, the input indicating the main viewing direction is received automatically. A first value for at least one image modification parameter for an image element residing in the main viewing direction is determined (304). A second value is determined (306) for at least one image modification parameter for image elements deviating from the main viewing direction as a function of an angle between the main viewing direction and a direction from the user towards the image location deviating from the main viewing direction. Finally, the at least one image is displayed (308) according to the determined values of the at least one image modification parameter on the display. According to an embodiment, the immersive display unit may be a head mounted device (HMD) with a head tracking system or a curved display at least partly surrounding the head of the user.
Hence, a straightforward and computationally light method for providing changes in the surrounding display image properties is provided herein, which increases the feeling of the user of the HMD to be present and immersed in the surrounding 3D space. When a change in the image properties is desired, the user of the HMD may give a simple user input for indicating the main viewing direction. Alternatively, the input for indicating the main viewing direction may be obtained automatically in response to the user positioning him/herself within the immersive display unit. A value of at least one image modification parameter in the main viewing direction is used as a reference value, and the values of the image modification parameter in other directions of the 3D image are calculated as a function of the angle deviating from the main viewing direction.
According to an embodiment, the user input is determined on the basis of a movement of the head of the user. When the user has the glasses on (i.e. the user is using the HMD), a movement of the head, such as a nod of the head forward, is an intuitive manner for the user to determine how the image is modified. For example, moving the head forward can modify the image so that the viewer has the impression that s/he moves closer to the image in front of him/her. According to an embodiment, the user input is obtained via a user interface of the HMD. Herein, it is not necessary that the user has the glasses on, but the main viewing direction may be submitted via the user interface before the user starts wearing the HMD. It is also possible that there is an external user interface device connected to the HMD, and the user may give the user input via the external user interface device even if wearing the HMD.
According to an embodiment, the user input further indicates strength of the user input in the main viewing direction. The strength value of the user input may be used as the basis for adjusting the value of the image modification parameter in the main viewing direction. The strength may be
obtained e.g. by a movement sensor connected to the HMD, which determines the acceleration and/or distance of the head movement. If an external user interface device connected to the HMD is used, the strength may be given as a numerical value or otherwise proportional to a predefined scale.
According to an embodiment, the at least one image modification parameter is a zoom ratio of the image element. Herein, in response to the user input, the element of the display screen residing in the main viewing direction may be enlarged, while the image parts on the left and right side of the viewer do not change in size, but just move a little backwards. This may be illustrated by Figures 4a and 4b which show top views of cylindrical displays and a viewer in the center of the display. In the initial state of Figure 4a, the viewer wears the HMD and sees a cylindrical display around him/her. In response to the user input triggering the modification, the part of the display screen residing in the main viewing direction (i.e. in front of the viewer) seems to get closer to the viewer. Thus, the viewer has the feeling that the display around him has moved to the dashed line position of Fig. 4b, although the physical display stays all the time in the original location (solid line).
In the following, a geometrical analysis of the zoom ratio modification is discussed more in detail by referring to Figure 5. For clarity, the analysis is for a cylindrical physical display around the user, but the principles presented herein can be extended to spherical displays and head-mounted displays, and other form of immersive display units, as well. The arc 500 represents the physical cylindrical display around the viewer 502 at a radius r. Let us take an example image element 504 on the cylindrical display 500 around the viewer, where the image element 504 is represented by an arc, which in its initial state starts at angle a from the main viewing direction (i.e. to the left from the upright direction on the image). The viewer sees the width of the arc of the image element 504 in angle β. It is noted that for illustrative purposes of the geometrical analysis, the angle β is here shown as much larger than 1 degree. In practical implementations, β would be « 1 degree. Initially, the viewer 502 stays in the center of the cylindrical display. In response to a user input (e.g. a nod of the head), the dimensions of the image around the viewer change so that viewer feels that the circle of the display has moved distance k closer to the viewer. The virtual position of the "moved" display circle is marked in dashed line 500'. The viewer would see the original arc of the image element 504 now in a new location, in angle γ, as the image element 504' .
Naturally, with respect to the viewer 502, the actual physical cylindrical display 500 around the viewer remains at the same location. Thus, in order to show the arc of the image element 504 in
the new location, in angle γ, to the viewer, the physical cylindrical display 500 should show the original image element 504 as the black arc 504' ' on the display circle.
Now, analyzing the grey triangle OAB, where the sides OA=r, AB=k are known, and OB=s is unknown, and the known angle a is between the sides OA and AB and the angle Θ between the sides OA and OB is unknown, we have: s = (k2 + r2 - 2kr cosaj (1) k/sinO = s/sina (2)
Equations (1) and (2) can be combined as:
Θ = asin ( k sina / (k2 + r2 - 2kr cosa f2 ) (3)
Equation (3) now defines how the original angle a (corresponding to the start edge of the chosen image element) changes to a+ Θ when the circle display of radius r is virtually moved by distance k. For defining the zoom ratio, i.e. how much each image element is stretched or squeezed, a magnification factor M, corresponding to γ/β, can be calculated as follows:
The 360 degree display circle may be divided into m image elements of equal width, whereby the angle of one element from the viewer's perspective corresponds to 360°/m. Considering the n'h image element, its starting angle (i.e. the angle of the first side of the image element from the main viewing direction) is a(n) =(n-l)*360°/m.
With the image modification, the starting angle of each image element changes from a(n) to a '(n ), which corresponds to a+ Θ, as shown above. The angle difference between a(n) and a(n+l) is a(n+l) - a(n) = 360°/m. The angle difference in the modified image between a '(n) and a '(n+l) is a '(n+1 ) - a '(n). Thus, the magnification factor M (i.e. zoom ratio) of the image element n is
M = [a '(n+l) - a '(n)] /[a(n+l) - a(n)] =>
M = [a '(n+l) - a '(n)] / (360o/m) (4) where m is the number of image elements of equal width around a 360 degree display circle with the viewer at the circle center, a(n) is the angle deviating from the main viewing direction,
corresponding to the n image element before zooming, and a '(n) is the angle of a modified n image element after zooming, and a '(n) being a function of a(n) and the user input value for image magnification in the main viewing direction. The analysis above with Eq. (3) and Eq. (4) applies to the width of the image elements around a cylindrical display on a horizontal plane that is perpendicular to the axis of the cylinder. In zooming, the height of each image element changes with the same magnification factor M as the width of the elements changes so that the aspect ratio of the image element remains unchanged. Figure 6 shows an example of image element movements according to Eq. (3). The graph of Figure 6 shows an example, where k/r=0.2, i.e. the curved display is virtually moved one fifth closer to the viewer in the main viewing direction. The parameter m is selected as 100, which is large enough to obtain smooth output curves. The dashed line shows the original location of the image elements, whereas the solid line shows how the location of the image elements after the image modification. It can be seen that, for example, at 0 and 180 degrees there is no movement, whereas at around 90 and 270 degrees (i.e. directly left and right from the main viewing direction) the movements of the image elements are at maximum.
Figure 7 shows another example depicting the magnification as a function of the original location of the image element. Again, k/r=0.2. The dashed line shows the original magnification of the image elements, which equals to 1. The solid line shows the magnification ratio M according to Eq. (4); i.e. image element width in the new location vs. the element width in the original situation. The magnification is maximum (1.25) at 0 degrees (i.e. in the main viewing direction straight ahead from the viewer) and minimum (0.83) in the opposite direction. At around 90 and 270 degrees (left/right sides of the viewer), there is no magnification.
Herein, if k/r = -0.2, it would describe a situation where the viewer gives a user input in backward direction, e.g. moves his/her head backwards, in order to create the virtual feeling of moving backwards in the 3D space. In that situation, the solid magnification line of Figure 7 would be a mirror image relative to the dashed line with magnification= 1.
Figure 8 shows a qualitative explanation of the above embodiments. In the original situation, such as in Figure 4a, the viewer sees the image elements around the sphere such that each image element has the same angular slot marked with evenly spaced black lines. After the virtual movement forward, the size of image elements is modified such that the new element borders are marked with grey lines and black knobs. As a result of the modification, in main the direction selected by the viewer the image elements have stretched, the image elements just on the sides of the viewer have not changed their size but a moved slightly backwards, and the image element
just behind the viewer have squeezed. Altogether, the viewer experiences an illusion of a slight movement forward as the images in the main viewing direction grow and the other image elements change their size accordingly. The above equations (3) and (4) of the analysis apply for the case where the display is of circular shape around the viewer. If the shape of the display is different, for example oval or polygon, the display shape needs to be projected on a virtual circle inside the shape, whereupon the calculations are performed in relation to the virtual surface of the circle, and then the transformed elements are projected back to the original display form.
According to an embodiment, the at least one image modification parameter is at least one of: a blurring strength of the image element;
a color tone strength of the image element;
a color selection of the image element.
Thus, alternatively or additionally to modifying the zoom ratio of the image elements as a function of an angle between the main viewing direction and the direction towards an image location deviating from the main viewing direction, the above embodiments may be applied to other image modification parameters, as well. For example, the blurring strength can be applied such that the image elements close to the main viewing direction are shown sharp in focus, whereas the image elements deviating from the main viewing direction are shown blurred (out of focus), and the blurring strength depends on the deviating angle according to a pre-defined formula. In a similar manner, the tone of the color (brighter/darker) may be adjusted in the image elements deviating from the main viewing direction. Moreover, the color of the image elements may change according to a predefined scheme (e.g. red-purple-blue) in the image elements deviating from the main viewing direction.
In addition to video, the embodiments may be applied to still images, even in GIF format, presented on displays showing images which surround the viewer.
In addition to images, the embodiments may be applied to directional audio. For example, similarly to adjusting zoom ratio, the audio level may be raised in the direction where the image is enlarged. According to an embodiment, two or more image modification parameters may be adjusted simultaneously.
The above embodiments have been mainly described in relation to cylindrical 2D 360 degree displays. Figure 9a shows the main direction vector r and the deviated direction vector s and the angle a between the vectors in a cylindrical display. However, the same principles may be applied also in. spherical displays. Figure 9b shows the main direction vector r and the deviated direction vector s and the angle a between the vectors inside a spherical display.
According to an embodiment, the viewer inside a spherical display or wearing a 360 degree head- mounted display may be watching a virtual reality video where s/he is moving, e.g. flying, in direction r. As a response to a user input, the surrounding image elements may change: the view in front of the viewer may remain in focus and in the original color, possibly provided with magnification at the same time, while the views by the sides of the viewer, i.e. on left, right, top and below, may become blurry and change colors or color tones. This would provide the viewer with the feeling that s/he is moving through more and less blurry rings of different colors. As becomes evident from the above, the various embodiments may be applied when a user views the immersive video content and interacts with it in real-time by the applying the various image transformations (zooming, blurring etc.). Alternatively or additionally, the various embodiments may be applied in a post-production (e.g. editing) phase of the content. Herein, the editor (e.g. a person or a computer program) applies one or more of the image transformations to add desired effects to the capture footage. The footage with the transformations implemented may then be recorded for further use.
For 3D 360 degree displays, the image transformation procedure needs to be performed for the two display views, for the two eyes, separately. For 3D 360 degree displays, it is preferable to apply parallax correction procedures for enhancing the 3D effect by further adjusting the relative movement of foreground and background objects, thereby making the scene more realistic looking. Various parallax corrections are known, as such, in the zooming of 3D images.
For example, parallax motion effect may be achieved by embedding optical flow data from stereo images into the video stream and at the playback time slightly warping left and right eye images according to the viewport and optical flow data. The warping may be done so that the video quality or stereo effect does not degrade at all in the middle of the viewport, which is the area that may be critical for the perceived video quality. Instead the video quality and stereo effect may get degraded gradually towards the edges of the viewport, but human stereo vision ability is naturally already reduced in that direction.
In addition to the illusory linear movement of the viewer in the 3D (x,y,z) space, the image transformation may relate to creating the feeling of image rotation around one of the axis in the
x,y,z, space. For example, the viewer may look straight ahead at one point in the horizon and starts to see the horizon of the image to shake around the point.
The above embodiments may provide various advantages. The feeling of immersiveness and interactivity is increased to the filmed virtual reality (VR) content that usually have image content wherein the viewer cannot move around and has no interaction possibility. The image modification algorithm is simple with a short processing time, thereby enabling the transformation to be performed even on live streaming VR content. In general, the various embodiments may be implemented in hardware or special purpose circuits or any combination thereof. While various embodiments may be illustrated and described as block diagrams or using some other pictorial representation, it is well understood that these blocks, apparatus, systems, techniques or methods described herein may be implemented in, as non- limiting examples, hardware, software, firmware, special purpose circuits or logic, general purpose hardware or controller or other computing devices, or some combination thereof.
Embodiments of the inventions may be practiced in various components such as integrated circuit modules. A skilled man appreciates that any of the embodiments described above may be implemented as a combination with one or more of the other embodiments, unless there is explicitly or implicitly stated that certain embodiments are only alternatives to each other.
The various embodiments can be implemented with the help of computer program code that resides in a memory and causes the relevant apparatuses to carry out the invention. Thus, the implementation may include a computer readable storage medium stored with code thereon for use by an apparatus, which when executed by a processor, causes the apparatus to perform the various embodiments or a subset of them. Additionally or alternatively, the implementation may include a computer program embodied on a non-transitory computer readable medium, the computer program comprising instructions causing, when executed on at least one processor, at least one apparatus to apparatus to perform the various embodiments or a subset of them. For example, an apparatus may comprise circuitry and electronics for handling, receiving and transmitting data, computer program code in a memory, and a processor that, when running the computer program code, causes the terminal device to carry out the features of an embodiment.
The above-presented embodiments are not limiting, but it can be modified within the scope of the appended claims.
Claims
1. A method comprising:
obtaining, by an immersive display unit, at least one image to be displayed on the immersive display unit;
obtaining an input indicating a main viewing direction of a user of the immersive display unit;
determining a first value for at least one image modification parameter for an image element residing in the main viewing direction;
determining a second value for the at least one image modification parameter for image elements deviating from said main viewing direction, the second value being a function of an angle between the main viewing direction and a direction from the user towards an image location deviating from the main viewing direction; and
displaying the at least one image according to the determined values of the at least one image modification parameter on the immersive display unit.
2. The method according to claim 1, wherein the input is determined based on a head movement of the user.
3. The method according to claim 1, wherein the input is obtained via a user interface of the immersive display unit.
4. The method according to claim 1, wherein the input is obtained automatically as a response to the user positioning within the immersive display unit.
5. The method according to any preceding claim, wherein the input further indicates strength of the input in the main viewing direction.
6. The method according to any preceding claim, wherein the at least one image modification parameter is a zoom ratio of the image element.
7. The method according to claim 6, wherein the zoom ratio of the nth equally spaced image element around a 360 degree display circle is calculated as:
M = [α ' (η+1) - α ' (n)] / (360°/m),
where m is the total number of image elements of equal width around a 360 degree display circle, a(n) is the angle deviating from the main viewing direction, corresponding to the nth image element before zooming, and a '(n) is the angle of a modified nth image element after
zooming, and a '(n) being a function of a(n) and the user input value for image magnification in the main viewing direction.
8. The method according to any preceding claim, wherein the at least one image modification parameter is at least one of:
a blurring strength of the image element;
a color tone strength of the image element;
a color selection of the image element.
9. The method according to any preceding claim, wherein the immersive display unit is a head mounted device (HMD) with head tracking system or a curved display at least partly surrounding the head of the user.
10. The method according to any of the preceding claims, wherein the at least one image comprises video image data.
11. A computer program embodied on a non-transitory computer readable medium, the computer program comprising instructions causing, when executed on at least one processor, at least one apparatus to:
obtain, by an immersive display unit, at least one image to be displayed on the immersive display unit;
obtain an input indicating a main viewing direction of a user of the immersive display unit;
determine a first value for at least one image modification parameter for an image element residing in the main viewing direction;
determine a second value for the at least one image modification parameter for image elements deviating from said main viewing direction, the second value being a function of an angle between the main viewing direction and a direction from the user towards an image location deviating from the main viewing direction; and
display the at least one image according to the determined values of the at least one image modification parameter on the immersive display unit.
12. The computer program according to claim 11, further comprising instructions causing the at least one apparatus to determine the input based on a head movement of the user.
13. The computer program according to claim 11, further comprising instructions causing the at least one apparatus to obtain the input via a user interface of the immersive display unit.
14. The computer program according to claim 11, further comprising instructions causing the at least one apparatus to obtain the input automatically as a response to the user positioning within the immersive display unit.
15. The computer program according to any of claims 11 - 14, further comprising instructions causing the at least one apparatus to obtain, from the input, an indication of the strength of the input in the main viewing direction.
16. The computer program according to any of claims 11 - 15, wherein the at least one image modification parameter is a zoom ratio of the image element.
17. The computer program according to claim 16, further comprising instructions causing the at least one apparatus to calculate the zoom ratio of the nth equally spaced image element around a 360 degree display circle as:
M = [α ' (η+1) - α ' (n)] / (360°/m),
where m is the total number of image elements of equal width around a 360 degree display circle, a(n) is the angle deviating from the main viewing direction, corresponding to the nth image element before zooming, and a '(n) is the angle of a modified nth image element after zooming, and a '(n) being a function of a(n) and the user input value for image magnification in the main viewing direction.
18. The computer program according to any of claims 11 - 17, wherein the at least one image modification parameter is at least one of:
- a blurring strength of the image element;
a color tone strength of the image element;
a color selection of the image element.
19. The computer program according to any of claims 11 - 18, wherein the immersive display unit is a head mounted device (HMD) with head tracking system or a curved display at least partly surrounding the head of the user.
20. The computer program according to any of claims 11 - 19, wherein the at least one image comprises video image data.
21. An apparatus comprising an immersive display unit comprising at least one processor, memory including computer program code, the memory and the computer program code configured to, with the at least one processor, cause the apparatus to at least:
obtain at least one image to be displayed on the immersive display unit;
obtain an input indicating a main viewing direction of a user of the immersive display unit;
determine a first value for at least one image modification parameter for an image element residing in the main viewing direction;
determine a second value for the at least one image modification parameter for image elements deviating from said main viewing direction, the second value being a function of an angle between the main viewing direction and a direction from the user towards an image location deviating from the main viewing direction; and
display the at least one image according to the determined values of the at least one image modification parameter on the immersive display unit.
22. The apparatus according to claim 21, further comprising computer program code configured to, with the at least one processor, cause the apparatus to
determine the input based on a head movement of the user.
23. The apparatus according to claim 21, further comprising computer program code configured to, with the at least one processor, cause the apparatus to
obtain the input via a user interface of the immersive display unit.
24. The apparatus according to claim 21, further comprising computer program code configured to, with the at least one processor, cause the apparatus to
obtain the input automatically as a response to the user positioning within the immersive display unit.
25. The apparatus according to any of claims 21 - 24, wherein the input further indicates strength of the input in the main viewing direction.
26. The apparatus according to any of claims 21 - 25, wherein the at least one image modification parameter is a zoom ratio of the image element.
27. The apparatus according to claim 26, wherein the zoom ratio of the nth equally spaced image element around a 360 degree display circle is calculated as:
M = [α ' (η+1) - α ' (n)] / (360°/m),
where m is the total number of image elements of equal width around a 360 degree display circle, a(n) is the angle deviating from the main viewing direction, corresponding to the nth image element before zooming, and a '(n) is the angle of a modified nth image element after
zooming, and a '(n) being a function of a(n) and the user input value for image magnification in the main viewing direction.
28. The apparatus according to any of claims 21 - 27, wherein the at least one image modification parameter is at least one of:
a blurring strength of the image element;
a color tone strength of the image element;
a color selection of the image element.
29. The apparatus according to any of claims 21 - 28, wherein the immersive display unit is a head mounted device (HMD) with head tracking system or a curved display at least partly surrounding the head of the user.
30. The apparatus according to any of claims 21 - 29, wherein the at least one image comprises video image data.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
GB1602903.5 | 2016-02-19 | ||
GB1602903.5A GB2548080B (en) | 2016-02-19 | 2016-02-19 | A method for image transformation |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2017141139A1 true WO2017141139A1 (en) | 2017-08-24 |
Family
ID=55752888
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/IB2017/050688 WO2017141139A1 (en) | 2016-02-19 | 2017-02-08 | A method for image transformation |
Country Status (2)
Country | Link |
---|---|
GB (1) | GB2548080B (en) |
WO (1) | WO2017141139A1 (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2020201571A1 (en) * | 2019-04-05 | 2020-10-08 | Psholix Ag | Method for an immersive display of stereoscopic images and image sequences |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2001095061A2 (en) * | 1999-12-07 | 2001-12-13 | Frauenhofer Institut Fuer Graphische Datenverarbeitung | The extended virtual table: an optical extension for table-like projection systems |
US20040109022A1 (en) * | 2002-12-04 | 2004-06-10 | Bennett Daniel H | System and method for three-dimensional imaging |
US20110254914A1 (en) * | 2010-04-14 | 2011-10-20 | Alcatel-Lucent Usa, Incorporated | Immersive viewer, a method of providing scenes on a display and an immersive viewing system |
US20160042567A1 (en) * | 2014-08-05 | 2016-02-11 | Utherverse Digital Inc. | Immersive displays |
US20160093105A1 (en) * | 2014-09-30 | 2016-03-31 | Sony Computer Entertainment Inc. | Display of text information on a head-mounted display |
US20160191893A1 (en) * | 2014-12-05 | 2016-06-30 | Warner Bros. Entertainment, Inc. | Immersive virtual reality production and playback for storytelling content |
US20160378177A1 (en) * | 2015-06-29 | 2016-12-29 | Beijing Zhigu Rui Tuo Tech Co., Ltd | Visualized content transmission control method, sending method and apparatuses thereof |
US20160378178A1 (en) * | 2015-06-29 | 2016-12-29 | Beijing Zhigu Rui Tuo Tech Co., Ltd | Visualized content transmission control method, sending method and apparatuses thereof |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20050156817A1 (en) * | 2002-08-30 | 2005-07-21 | Olympus Corporation | Head-mounted display system and method for processing images |
-
2016
- 2016-02-19 GB GB1602903.5A patent/GB2548080B/en not_active Expired - Fee Related
-
2017
- 2017-02-08 WO PCT/IB2017/050688 patent/WO2017141139A1/en active Application Filing
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2001095061A2 (en) * | 1999-12-07 | 2001-12-13 | Frauenhofer Institut Fuer Graphische Datenverarbeitung | The extended virtual table: an optical extension for table-like projection systems |
US20040109022A1 (en) * | 2002-12-04 | 2004-06-10 | Bennett Daniel H | System and method for three-dimensional imaging |
US20110254914A1 (en) * | 2010-04-14 | 2011-10-20 | Alcatel-Lucent Usa, Incorporated | Immersive viewer, a method of providing scenes on a display and an immersive viewing system |
US20160042567A1 (en) * | 2014-08-05 | 2016-02-11 | Utherverse Digital Inc. | Immersive displays |
US20160093105A1 (en) * | 2014-09-30 | 2016-03-31 | Sony Computer Entertainment Inc. | Display of text information on a head-mounted display |
US20160191893A1 (en) * | 2014-12-05 | 2016-06-30 | Warner Bros. Entertainment, Inc. | Immersive virtual reality production and playback for storytelling content |
US20160378177A1 (en) * | 2015-06-29 | 2016-12-29 | Beijing Zhigu Rui Tuo Tech Co., Ltd | Visualized content transmission control method, sending method and apparatuses thereof |
US20160378178A1 (en) * | 2015-06-29 | 2016-12-29 | Beijing Zhigu Rui Tuo Tech Co., Ltd | Visualized content transmission control method, sending method and apparatuses thereof |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2020201571A1 (en) * | 2019-04-05 | 2020-10-08 | Psholix Ag | Method for an immersive display of stereoscopic images and image sequences |
Also Published As
Publication number | Publication date |
---|---|
GB201602903D0 (en) | 2016-04-06 |
GB2548080A (en) | 2017-09-13 |
GB2548080B (en) | 2021-07-14 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11575876B2 (en) | Stereo viewing | |
KR102535947B1 (en) | Apparatus and method for generating images | |
US11099392B2 (en) | Stabilized and tracked enhanced reality images | |
US11218681B2 (en) | Apparatus and method for generating an image | |
WO2019043025A1 (en) | Zooming an omnidirectional image or video | |
WO2020166376A1 (en) | Image processing device, image processing method, and program | |
KR20200128661A (en) | Apparatus and method for generating a view image | |
WO2017141139A1 (en) | A method for image transformation | |
WO2009109804A1 (en) | Method and apparatus for image processing | |
KR102659115B1 (en) | Image generating device and method therefor | |
JP2022525526A (en) | Image signal representing the scene |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 17752754 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 17752754 Country of ref document: EP Kind code of ref document: A1 |