WO2013173548A2 - Adaptation d'un contenu stéréoscopique grand format à d'autres plateformes - Google Patents

Adaptation d'un contenu stéréoscopique grand format à d'autres plateformes Download PDF

Info

Publication number
WO2013173548A2
WO2013173548A2 PCT/US2013/041286 US2013041286W WO2013173548A2 WO 2013173548 A2 WO2013173548 A2 WO 2013173548A2 US 2013041286 W US2013041286 W US 2013041286W WO 2013173548 A2 WO2013173548 A2 WO 2013173548A2
Authority
WO
WIPO (PCT)
Prior art keywords
picture
destination
source
selected area
resolution
Prior art date
Application number
PCT/US2013/041286
Other languages
English (en)
Other versions
WO2013173548A3 (fr
Inventor
Pierre Hughes ROUTHIER
Brian J. Dorini
Original Assignee
Thomson Licensing
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Thomson Licensing filed Critical Thomson Licensing
Publication of WO2013173548A2 publication Critical patent/WO2013173548A2/fr
Publication of WO2013173548A3 publication Critical patent/WO2013173548A3/fr

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • G06T3/4007Scaling of whole images or parts thereof, e.g. expanding or contracting based on interpolation, e.g. bilinear interpolation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • H04N13/139Format conversion, e.g. of frame-rate or size
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/30Image reproducers
    • H04N13/398Synchronisation thereof; Control thereof

Definitions

  • Implementations are described that relate to digital pictures. Various particular implementations relate to processing 3D digital pictures.
  • a source picture having a source resolution is accessed.
  • a destination resolution for a destination picture is accessed.
  • a lowe bound is determined.
  • the lower bound is on a size of an area in the source picture that can be selected for use in generating the destination picture.
  • the lower bound is based on a disparity metric associated with the destination resolution.
  • a selected area of the source picture is scaled. The selected area is at least as large as the lower bound.
  • the destination picture is based on the scaled selected area.
  • At least a portion of a source picture having a source resolution is displayed.
  • a minimum size for cropping is identified on the display of at least the portion.
  • Input is accepted that identifies a selected area o the source picture for cropping.
  • the selected area is at least as large as the identified minimum size.
  • the selected area is scaled to form a scaled picture.
  • an implementation may be performed as a method, or embodied as an apparatus, such as, for example, an apparatus configured to perform a set of operations or an apparatus storing instructions for performing a set of operations, or embodied in a signal.
  • an apparatus such as, for example, an apparatus configured to perform a set of operations or an apparatus storing instructions for performing a set of operations, or embodied in a signal.
  • FIG. 1 provides a pictorial diagram depicting an example of a particular parallax and disparity situation.
  • FIG. 2 provides a pictorial diagram depicting an example of a 2k window in an 8I picture.
  • FIG. 3 provides a pictorial diagram depicting an example of a large format pictur that includes a maximum crop window and a pan and scan window.
  • FIG. 4 provides a pictorial diagram depicting an example of the pan and scan window of FIG. 3 being scaled down to a smaller format screen.
  • FIG. 5 provides a pictorial diagram depicting an example of an 8k picture overlai with a maximum crop window and a pan and scan window.
  • FIG. 6 provides a pictorial diagram depicting an example of the pan and scan window of FIG. 5 scaled down to a 2k picture.
  • FIG. 7 provides a flow diagram depicting an example of a process for scaling a picture based on a disparity metric.
  • FIG. 8 provides a flow diagram depicting another example of a process for scaling a picture.
  • FIG. 9 provides a block diagram depicting an example of a system for scaling a picture based on a disparity metric.
  • composition and aspect ratio often leads, in 2D, to a pan and scan process whic cuts out a significant portion of the source frame.
  • Using the same approach for stereoscopic content is frequently unsuitable, because of the excessive parallax this would cause.
  • At least one implementation described in this application provides a methodolog that attempts to ensure that excessive parallax is not possible when resizing large screen items to smaller viewing platforms.
  • Parallax is the angular difference between two sight-lines to an object.
  • the sight-lines originate at each of a viewer's left- and right-eyes and proceed to corresponding left- and right-eye image views of the object (or portion thereof).
  • Disparity is the linear difference in the positions of an object (or portion thereof) i each of the left- and right-eye images of a stereoscopic image pair. Disparity may be expressed as a physical measure (for example, in centimeters), or in an image-relative measure, (for example, in pixels or as a percentage of image width). A conversion between the two forms is possible when the size of the images as displayed is known. Depth, as perceived in a stereoscopic presentation, can be determined from parallax. However, depth is trigonometrically, not linearly, related to disparity an parallax. Depth is shown by the apparent distance (D) in FIG. 1 , as discussed below, and can in theory range from zero to infinity.
  • D apparent distance
  • FIG. 1 shows a situation 100 where a viewer 1 10 is watching a stereoscopic presentation on a screen 120.
  • the viewer 1 10 perceives a particular object 130 whose apparent distance (D) from the viewer results from parallax ( ⁇ ), which is induced by the combination of physical disparity (dp), viewing distance (V), and the viewer's interocular spacing (t E ).
  • the situation 100 is shown with left and right sightlines 131 , 132 forming a right triangle with a line between the viewer's left and right eyes 1 1 1 , 1 12, the line having a length of (t E ). Further, an again for simplicity, that line is considered to be parallel to the screen 120.
  • the physical disparity (d P ) is by some conventions, and herein, considered to be negative.
  • the physical disparity (dp) is negative whenever the left-eye image of the object is to the left of the right-eye image of the object.
  • the parallax angle ( ⁇ ), in the situation 100 is positive, and is positive for all values of (dp) greater than (-t E ).
  • the parallax ( ⁇ ) can be displayed, by presenting a stereoscopic image pair in which the left- and right-eye images of the object 130 have a disparity of less than (-t E ).
  • the parallax becomes negative, at which point the presentation of the object 130 by the screen 120 ceases to have a meaningful interpretation and the parallax ( ⁇ ) becomes negative.
  • the sightlines 131 , 132 would intersect behind the viewer 1 10, but the images of the object 130 still appear on the scree 120. This produces a contradiction that the human visual system typically does not enjoy.
  • interocular spacing (t E ), distance V from the viewer to : screen, physical disparity (dp), and an object's apparent distance (D) from the viewer can be described in an equation obtained by inspecting the similar triangles of FIG. 1 (and recalling that the sign of dp in situation 100 is negative):
  • Physical disparity (d P ) is proportional to the disparity (d,) intrinsic to the images c a stereoscopic image pair and the size of the screen (S). As the size (S) of the screen grows larger (and with it the size of the stereoscopic images), a particula intrinsic disparity (d,) will produce a larger physical disparity (dp):
  • the apparent distance (D) of the object from the viewer is altered b varying the viewer's distance from the screen (V) and/or size of the screen (S).
  • interocular spacing (t E ) is essentially constant for an individua over long spans of time, individuals within a population may differ substantially, especially if comparing adults to children.
  • default values examples include, for example, interocular distance (t E ), screen size (S), and/or viewing distance (V).
  • t E interocular distance
  • S screen size
  • V viewing distance
  • the use of default values can allow other values (for example, distance D) to be presented in a contingent fashion, in which final values might depend upon how actual values differ from the default assumptions.
  • any head turning could result in a deviation in either directior with square-facing being the median condition.
  • t E is small compared to V ; that is, theta is typically small, the approximation is generally considered acceptable because both sin() and tan() are typically considered to be fairly linei in that region.
  • parallax value that is to ⁇ large for comfort or safety. It should be understood, however, that all of the implementations can be used to increase parallax values if, for example, the parallax is considered to be too small. Frequently, an implementation will provid safety and/or comfort levels for both hyperdivergence and hyperconvergence.
  • a large format image frame (such as an Imax ® asset, for example) has a composition that is typically very different from theatrical and home viewing compositions.
  • Typical conditions include a very large screen, and a very high ratio between the width c the screen and the average distance of the viewer to the screen.
  • th ⁇ viewer typically only actively perceives a very small portion of the image, often located in the lower third of the screen.
  • the rest of the frame is displayed, for example, to immerse the viewer in the image, filling the viewer's peripheral vision.
  • HDTV High Definition Television
  • Croping refers generally to selecting a portion of ⁇ source picture and discarding (or ignoring) the remainder of the source picture, with the selected portion to be converted into a destination picture.
  • the selectee portion is typically less than the entire source picture.
  • the selected portion (rather than the unselected, or discarded, portion) is typically referred to as the cropped portion.
  • the destination picture can have a different resolution, including a different aspect ratio, than the source picture.
  • the cropped portion is typically scaled to adapt the cropped portion to the resolution of the destination picture.
  • cropping and scaling are generally referred to as separate processes, although various implementations combine the two operations of cropping (selecting) and scaling.
  • disparity (sometimes referred to as parallax, in a general sense) is often expressed as a percentage c a screen's width, and this value (expressible, for example, as a ratio of disparity over screen width) will change if the image is simply cropped instead of being scaled down.
  • FIG. 2 includes an 8k (8192 pixels wide) picture 200 that includes a 2k (2048 pixels wide) window 210.
  • the window 210 shows a tree 220. If the window 210 is used as a cropping window to generate an output 2k picture, then no scaling will be needed. As a result, the 2048 columns of pixels in the window 210 will become the pixels of the output 2k picture. Accordingly, the disparity present in, for example, the tree 220 will be completely preserved.
  • the disparity present in the tree 220 in the 8 picture 200 is 82 pixels, then the disparity present in the tree 220 in the output 2 picture will also be 82 pixels.
  • post production companies typically perform a manual combination of cropping and scaling. Unfortunately, because such cropping and scaling is a manual operation, the result is that a high percentage ⁇ pictures (or shots) typically still exceed the maximum thresholds for disparity. Consequently, the pictures (or shots) that exceed the maximum thresholds will generally be reformatted, until a final result is attained that meets the thresholds This often results in an iterative and time-consuming process.
  • FIG. 3 provides a pictorial diagram depicting an example of a large format picture that includes a maximum crop window and a pan and scan window.
  • FIG. 4 provides a pictorial diagram depicting an example of the pan and scan window of FIG. 3 being scaled down to a smaller format screen.
  • a large format picture is shown as a large screen source picture 310.
  • a measure of maximum observed negative and positive disparity is calculated for the picture 310.
  • the disparity calculations are performed using, for example, a disparity estimator, feature point analysis, or any other suitable method, as is known in th art.
  • MRR Minimum Reduction Ratio
  • MRR Largest of MRR (negative) and MRR (positive)
  • the source has an observed maximum negative disparity of 60 pixels
  • the destination specification is 15 pixels.
  • the MRR in negative disparity is 60 px / 15 px, producing an MRR (negative) of 4.
  • the specification is provided, for example, in various implementations, by a device manufacturer, by a content producer, or by a user's viewing preferences.
  • the source has an observed maximum positive disparity of 100 pixels, and the destination specification is 30 pixels. Then the MRR in positive disparity is 100 px / 30 px, producing an MRR (positiv* of 3.33.
  • the MRR is the largest of MRR (negative) and MRR (positive). Therefore, in thi case, the MRR is 4. This means that the source image has to be scaled down b a minimum factor of 4. A scaling down of at least 4 results in a minimum reduction in size of 75% because the image is reduced to no more than 25% of its original size.
  • a Maximun Crop Window (MCW) 320 appears on the screen, shown in dashed lines in FIG. 3, overlaying the picture 310.
  • This window 320 is an indication of the smallest possible cropping size in order to be over the MRR. That is, if the cropping size is less than the MCW 320, then the required scaling to achieve the destination resolution will be less than the MRR.
  • the size of the MCW 320 is calculated in the following way:
  • MCW Width Source Frame Width / MRR
  • the dimensions of the MCW 320 will be:
  • the MCW 320 is then locked in the pan and scan software used by an operator that is performing a conversion of the source content to a new destination resolution. This prevents the operator from selecting a final cropped window tha is smaller than the MCW 320, which would result in a final picture (or a final shol with excessive disparity on the destination device.
  • the operator selects the final, desired Pan and Scan Window (PSW) 330 fc the picture.
  • the PSW 330 can be static or dynamic (for example, one dynamic implementation changes the location and/or size of the PSW 330 in each picture), but it is automatically locked to be at least as large as the MCW 330.
  • This lock ensures that the final picture in the destination resolution does not exceed the destination device's parallax or disparity specifications.
  • FIG. 3 show this lock by depicting the MCW 320 as being smaller than the PSW 330, with the arrows in FIG. 3 indicating that the PSW 330 has a range of allowable sizes but that all allowable sizes are at least as large as the MCW 320.
  • the final operation is to apply the proper scale down ratio to the PSW 330 so th ⁇ the final picture (or, in various implementations, the final shot) is in the destination's resolution and aspect ratio.
  • This is depicted in FIG. 4 by showing the PSW 330 being scaled down into a final picture 340.
  • the final picture 340 will typically b smaller than the MCW 320. This is because typical implementations will have ai MRR that is greater than 1 , which will have the effect of reducing the size of the PSW 330 so that it is smaller than the MCW 320.
  • FIG. 5 provides a pictorial diagram depicting an example of an 8k picture overlaid with a maximum crop window and a pan and scan window.
  • FIG. 6 provides a pictorial diagram depicting an example of the pan and scan window of FIG. 5 scaled down to a 2k picture.
  • FIG. 5 includes an 8k picture 510 that includes a tree 515.
  • FIG. 5 also includes an MCW 520 that does not include the entire tree 515.
  • the MCW 520 is 4096 pixels wide. If the MCW is calculated using the equation discussed above with respect to FIGS. 3-4, then the MCW is calculated as:
  • MCW Width Source Frame Width / MRR
  • the MRR can be solved for as Source Frame Width / MCW Width.
  • the MRR is 2.
  • FIG. 5 also includes a PSW 530 that does include the entire tree 515. The operator is presumed to have selected the PSW 530 to include the entire tree 515.
  • the PSW 530 is scaled down, by at least the value of the MRR, to produce the final (destination) picture 540 having a width of 2048 pixels. It is clear that the PSW 530 is scaled down by more than the value of MRR, which is 2, but the exact scale down value is not known because the size of the PSW 530 is not indicated.
  • MCW Width MRR * Final Resolution Width
  • a flow diagram depicts an example of a process 700 for scaling a picture based on a disparity metric.
  • a disparity metric is used in this application to refer, for example, to any variable or quantity that is based on, or derivable from, disparity.
  • a disparity metric also includes any variable or quantit that is based on, or derivable from, a value that is derivable from or based on disparity such as, for example, parallax or depth.
  • Disparity metrics include, for example, an MRR whether based on disparity or parallax, and an MCW size.
  • the process 700 includes accessing a source picture having a source resolution (710).
  • the process 700 also includes accessing a destination resolution for a destination picture (720).
  • the process 700 includes determining a lower bound, based on a disparity metr associated with the destination resolution (730).
  • the operation 730 includes determining a lower bound on a size of an area in the source picture that can be selected for use in generating the destination picture, the lower bound being based on a disparity metric associated with the destinatio resolution.
  • Various implementations determine a lower bound by determining a single dimension, with the second dimension being implicit. Additionally, various implementations provide a display of the lower bound in two dimensions (for example, an MCW). However, other implementations provide a display of only ⁇ single dimension, such as, for example, the width of the MCW.
  • the process 700 includes scaling a selected area (740).
  • the operation 740 includes scaling a selected area of the source picture, the selected area being at least as large as the lower bound, wherein thi destination picture is based on the scaled selected area.
  • the process 700 also includes an optional operation (not shown in FIG. 7) of receiving or accessing an indication of the selected area
  • the indication is provided, for example, by a user choosing an area on a screen that is displaying the source picture.
  • the process 700 does not specifically recite the operation of generating the destination picture.
  • the scaled selected area is the destination picture exactly and has the destination resolution.
  • the process 700 does generate the destination picture by virtui of generating the scaled selected area. In such implementations in which the destination picture is the scaled selected area, the destination picture is clearly based on the selected scaled area.
  • the process 700 further includes an optional operation (not shown in FIG. 7) of performing additional processing in order to generate th destination picture.
  • additional processing can include, for example, one or more of (i) truncating part of the scaled selected area, (ii) padding the scaled selected area to increase the number of pixels, (iii) adapting the color or luminance of at least some of the pixels in the scaled selected area, or (iv) performing some other filtering operation on at least some of the pixel values in the scaled selected area.
  • the disparity metric is further associated with the source picture, - the disparity metric is a minimum reduction ratio,
  • the minimum reduction ratio is based on (i) a maximum disparity for the source picture and (ii) a maximum disparity for the destination picture,
  • the maximum disparity for the source picture is measured or calculated, and th maximum disparity for the destination picture is specified, - the minimum reduction ratio is the ratio of the maximum disparity for the source picture over the maximum disparity for the destination picture,
  • the lower bound is further based on the source resolution, - the lower bound is expressed or displayed with respect to only one dimension,
  • the lower bound is based on a ratio of the source resolution to a minimum
  • the lower bound is displayed by being overlaid on the source picture
  • receiving the selected area includes accepting input from a user identifying the selected area, and/or - the destination picture is transmitted, encoded, and/or displayed.
  • a flow diagram depicts an example of a process 800 for scaling a picture.
  • the process 800 includes displaying a source picture (810).
  • the operation 810 includes displaying at least a portion of a source picture having a source resolution.
  • the process 800 includes identifying a minimum size for cropping (820).
  • the operation 820 includes identifying on a display of ; least a portion of the source picture having a source resolution, a minimum size for cropping.
  • the process 800 includes accepting input identifying a selected area at least as large as the minimum size (830).
  • the operation 830 includes accepting input identifying a selected area of the source picture for cropping, the selected area being at least as large as the identified minimum size.
  • Input can identify the selected area by, for example, (i) indicating one or more coordinates of the selected area (for example, corner or center
  • Input can be accepted b; for example, (i) receiving input from a touch screen, a mouse, or a keyboard, and/or (ii) accessing input from a stored memory location (for example, in using default value for the input, or accessing a stored profile for the input, or accessin input used and stored for another picture).
  • the process 800 includes scaling the selected area to form a scaled picture (840).
  • the scaled selected area need not be a final destination picture.
  • the scaled picture is a destination picture having a destination resolution.
  • the process 800 ends without generating the destination picture.
  • the process 800 includes further processing of the scaled selected area to generate the destination picture. Such further processing is, in various implementations, for example, as described above with respect to the process 700.
  • Various implementations of the process 800 include one or more of the followinc additional features: - displaying an entire source picture,
  • - identifying the minimum size includes displaying an outline of a rectangle over display of at least a portion of the source picture, the rectangle indicating the minimum size for cropping,
  • - accepting input identifying the selected area includes allowing a user to adjust size of a window to identify the selected area, and/or
  • a block diagram depicts an example of a system 900 for scaling a picture based on a disparity metric.
  • the system 900 includes a processor 910 that is communicatively coupled to a display 920 for displaying, fc example, digital pictures. Pictures are displayed, in various implementations, before, during, and/or after being processed by the processor 910.
  • the display 920 is also able to send signals back to the processor 910 in the event, for example, that the display 920 is a touch-screen and provides input information from the touch-screen to the processor 910.
  • the processor 910 is also communicatively coupled to a user interface 930 for accepting input from, for example, an operator.
  • the user interface 930 is also communicatively coupled to the display 920 to allow, for example, operator inpul to be displayed on the display 920 directly without intervention by the processor 910.
  • the processor 910 is communicatively coupled a storage device 940, an encode 950, and a transmitter 960.
  • the processor 910 provides digital pictures to one or more of the storage device 940 for storing the pictures, the encoder 950 for encoding of the pictures, or the transmitter 960 for transmitting the pictures.
  • the encoder 950 is also communicatively coupled to the storage device 940 and/or the transmitter 960. This allows, for example, encoded pictures from the encoder 950 to be (i) stored in the storage device 940 and/or (ii) transmitted using the transmitter 960.
  • the communications is two-way (not shown in FIG. ⁇ between the processor 910 on the one hand, and one or more of the storage device 940, the encoder 950, or the transmitter 960, on the other hand.
  • Such two-way communication allows, for example, stored pictures to be retrieved by the processor 910, encoded pictures to be provided directly to the processor 91 (without going through the storage device 940, and transmitted pictures also to b provided to the processor 910. Additionally, parameters or other information car be provided to the processor 910 from one or more of the storage device 940, th encoder 950, or the transmitter 960.
  • the system 900 is used, in various implementations, to perform the process 70C as well as any of the additional features described with respect to the process 700.
  • the processor 910 accesses a source picture having a source resolution (710) from the storage device 940.
  • the processor 910 also accesses a destination resolution for a destination picture (720) from the user interface 930 or the storage device 940. - The processor 910 determines a lower bound on a size of an area in the souro picture that can be selected for use in generating the destination picture, the lower bound being based on a disparity metric associated with the destinatioi resolution (730). The processor 910 uses, in various implementations, any o the equations described in this application for calculating an MCW as the lower bound.
  • the processor 910 scales a selected area of the source picture, the selected area being at least as large as the lower bound, wherein the destination picture is based on the scaled selected area (740). In at least one
  • the processor 910 receives the selected area from the user interface 930.
  • the system 900 is used, in various implementations, to perform the process 80C as well as any of the additional features described with respect to the process 800. In one such implementation:
  • the processor 910 displays a source picture (810) on the display 920 after accessing the source picture from the storage device 940.
  • the processor 910 identifies a minimum size for cropping (820).
  • the processor 910 accepts input, from the user interface 930, identifying a selected area at least as large as the minimum size (830).
  • the processor 910 scales the selected area to form a scaled picture (840).
  • the display 920 includes, in various implementations, one or more of a compute display, a laptop display, a tablet display, a cell phone display, a television display, or any of the other displays mentioned in this application or known in th( art, including projected displays that may be visible on any surface, such as, for example, a wall, a ceiling, a floor, or a sidewalk.
  • the user interface 930 includes, in various implementations, one or more of a mouse, a track pad, a keyboard, a touch screen, a microphone for accepting voice commands that are interpreted by the processor 910, a remote control, a cell phone, a separate computer whether remote or local, or any other input device mentioned in this application or known in the art.
  • the storage device 940 includes, in various implementations, any of the storag devices mentioned in this application or known in the art.
  • the encoder 950 includes, in various implementations, an AVC or H.264 (as defined elsewhere in this application) encoder, an encoder for any other standard, or any other encoding device mentioned in this application or known ir the art.
  • the transmitter 960 includes, in various implementations, an output pin of any integrated circuit, a Universal Asynchronous Receiver/Transmitter (UART), a broadcast transmitter, a satellite transmitter, a cable transmitter, or any other transmitting device mentioned in this application or known in the art.
  • the transmitter 960 may be, for example, adapted to transmit a program signal having one or more bitstreams representing encoded pictures and/or informatior related thereto.
  • Typical transmitters perform functions such as, for example, on ⁇ or more of providing error-correction coding (which may alternatively, or additionally, be performed in the encoder 950), interleaving the data in the signa (which may alternatively, or additionally, be performed in the encoder 950), randomizing the energy in the signal, and modulating the signal onto one or mor carriers using a modulator.
  • the transmitter 960 may include, or interface with, a antenna (not shown). Further, implementations of the transmitter 960 may be limited to a modulator.
  • FIG. 9 This application provides multiple figures, including the block diagram of FIG. 9, the flow diagrams of FIGS. 7-8, and the pictorial diagrams of FIGS. 1 -6. Each ⁇ these figures provides disclosure for a variety of implementations.
  • the block diagram certainly describes an interconnection of
  • FIG. 9 also presents a flow diagram for performing various processes that include the functions of the blocks of FIG. 9.
  • the block for the encoder 950 also represents the operation of encoding pictures
  • the block for the transmitter 960 also represents the operation of transmitting pictures
  • the interconnection between the encoder 950 and the transmitter 960 represents a process in which pictures are encoded and then transmitted.
  • Other blocks of FIG. 9 are similarly interpreted in describing this flow process.
  • the flow diagram certainly describes a flow process.
  • FIG. 7 also presents a block diagram for performin the functions of the process 700.
  • reference element 720 als represents a block for performing the function of accessing a destination resolution
  • reference element 730 represents a block for performing the function of determining a lower bound
  • the interconnection between elements 720 and 730 represents an apparatus having a component for accessing a destination resolution coupled to a component for determining a lower bound.
  • Other blocks of FIG. 7 are similarly interpreted in describing th system/apparatus.
  • FIG. 4 also describes a process of accepting a PSW 330 from ; user and then scaling the PSW 330 to generate the final output 340.
  • FIGS. 2-3 and 5-6 can also be interpreted in a similar fashion to describe respective flow processes.
  • Various implementations are described with respect to a picture. Many implementations perform one or more of the described processes on every picture in a series of pictures. Other implementations, however, apply more consistency across pictures that belong to the same shot, or scene, or movie, fo example. It is often advantageous, for example, to apply the same cropping anc scaling to every picture in a shot. Additionally, it is often advantageous for an operator to view an entire shot, for example, before selecting the pan and scan window for that shot or for an individual picture in the shot. Various implementations calculate the maximum disparity of a source picture. Typically the maximum disparity will be in a region of the source picture that an operator desires to preserve in the pan and scan window.
  • the "effective" MRR will be different.
  • Various implementations take this into consideration when creating the maximum crop window, and can vary the size of the maximum crop window depending on the location, selected by an operator, of the pan and scan window.
  • AVC refers to the existing International Organization for Standardization/International Electrotechnical Commission (“ISO/IEC”) Moving Picture Experts Group-4 (“MPEG-4") Part 10 Advanced Video Coding (“AVC”) standard/International Telecommunication Union, Telecommunication Sector (“ITU-T”) H.264 Recommendation (variously referred to throughout this document as the “H.264/MPEG-4 AVC Standard” or variations thereof, such as the "AVC standard”, the “H.264 standard”, or simply "AVC” or “H.264”).
  • ISO/IEC International Organization for Standardization/International Electrotechnical Commission
  • MPEG-4 Moving Picture Experts Group-4
  • AVC Advanced Video Coding
  • ITU-T International Telecommunication Union, Telecommunication Sector
  • references to "one embodiment” or “an embodiment” or “one implementation” or “an implementation” of the present principles, as well as other variations thereof mean that a particular feature, structure, characteristic, and so forth described in connection with the embodiment is included in at least one embodiment of the present principles.
  • Determining the information may include one or more of, for example, estimating the information, calculating the information, predicting th information, or retrieving the information from memory.
  • Accessing the information may include one or more of, for example receiving the information, retrieving the information (for example, memory), storing the information, processing the information, transmitting the information, moving the information, copying the information, erasing the information, calculating the information, determining the information, predicting the information.
  • Receiving is, as with “accessing”, intended to be a broad term.
  • Receiving the information may include one or more of, for example, accessing the information, or retrieving the information (for example, from memory).
  • receiving is typically involved, in one way or another, during operation: such as, for example, storing the information, processing the information, transmitting the information, moving the information, copying the information, erasing the information, calculating the information, determining the information, predicting the information, or estimating the information.
  • images and/or “pictures”.
  • image and/or “picture” are used interchangeably throughout this document, and are intended to be broad terms.
  • An “image” or a “picture” may be, for example, all o part of a frame or of a field.
  • video refers to a sequence of images (or pictures).
  • An image, or a picture may include, for example, any of various videc components or their combinations.
  • Such components include, for example, luminance, chrominance, Y (of YUV or YCbCr or YPbPr), I (of YUV), V (of YUV), Cb (of YCbCr), Cr (of YCbCr), Pb (of YPbPr), Pr (of YPbPr), red (of RGB), green (of RGB), blue (of RGB), S-Video, and negatives oi positives of any of these components.
  • An “image” or a “picture” may also, or alternatively, refer to various different types of content, including, for example, typical two-dimensional video, a disparity map for a 2D video picture, a depth map that corresponds to a 2D video picture, or an edge map.
  • implementations may refer to a "frame”. However, such implementations are assumed to be equally applicable to a "picture” or "image”.
  • a “depth map”, or “disparity map”, or “edge map”, or similar terms are also intended to be broad terms.
  • a map generally refers, for example, to a picture that includes a particular type of information. However, a map may include othe types of information not indicated by its name. For example, a depth map typically includes depth information, but may also include other information such as, for example, video or edge information. This application refers to "encoders” and “decoders” in a variety of
  • an encoder can include, for example, one or more (or no) source encoders and/or one or more (or no) channel encoders, as well as one or more (or no) modulators.
  • a decoder can include, for example, one or more (or no) modulators as well as one or more (or no) channel encoders and/or one or more (or no) source encoders.
  • any of the following 7", “and/or”, and “at least one of, for example, in the cases of “A/B”, “A and/or B” and “at least one o A and B”, is intended to encompass the selection of the first listed option (A) onh or the selection of the second listed option (B) only, or the selection of both options (A and B).
  • such phrasing is intended to encompass the selection of the first listed option (A) only, or the selection of the second listed option (B) only, or the selection of the third listed option (C) only, or the selection of the first and the second listed options (A and B) only, or the selection of the first and third listed options (A and C) only, or the selection of the second and third listed options (B and C) only, or the selection o all three options (A and B and C).
  • This may be extended, as readily apparent b) one of ordinary skill in this and related arts, for as many items listed.
  • processors such as, for example, a post-processor or a pre-processor.
  • the processors discussei in this application do, in various implementations, include multiple processors (sub-processors) that are collectively configured to perform, for example, a process, a function, or an operation.
  • the processor 910 as well as other processing components such as, for example, the encoder 950 and the transmitter 960, are, in various implementations, composed of multiple sub- processors that are collectively configured to perform the operations of that component.
  • the implementations described herein may be implemented in, for example, a method or a process, an apparatus, a software program, a data stream, or a signal.
  • An apparatus may be implemented in, for example, appropriate hardware, software, and firmware.
  • the methods may be implemented in, for example, an apparatus such as, for example, a processor, which refers to processing devices in general, including, for example, a computer, a
  • microprocessor an integrated circuit, or a programmable logic device.
  • Processors also include communication devices, such as, for example, computers, cell phones, tablets, portable/personal digital assistants ("PDAs”), and other devices that facilitate communication of information between end- users.
  • communication devices such as, for example, computers, cell phones, tablets, portable/personal digital assistants (“PDAs”), and other devices that facilitate communication of information between end- users.
  • PDAs portable/personal digital assistants
  • Implementations of the various processes and features described herein may be embodied in a variety of different equipment or applications.
  • equipment include an encoder, a decoder, a post-processor, a pre-processor, a video coder, a video decoder, a video codec, a web server, a television, a set-to box, a router, a gateway, a modem, a laptop, a personal computer, a tablet, a ce phone, a PDA, and other communication devices.
  • the equipment may be mobile and even installed in a mobile vehicle.
  • the methods may be implemented by instructions being performed by a processor, and such instructions (and/or data values produced by an implementation) may be stored on a processor-readable medium such as, for example, an integrated circuit, a software carrier or other storage device such as for example, a hard disk, a compact diskette (“CD”), an optical disc (such as, for example, a DVD, often referred to as a digital versatile disc or a digital video dis ⁇ or a Blu-Ray disc), a random access memory (“RAM”), a read-only memory (“ROM”), a USB thumb drive, or some other storage device.
  • the instructions may form an application program tangibly embodied on a processor-readable medium. Instructions may be, for example, in hardware, firmware, software, or ⁇ combination.
  • a processor may be characterized, therefore, as, for example, both a device configured to carry out ⁇ process and a device that includes a processor-readable medium (such as a storage device) having instructions for carrying out a process. Further, a processor-readable medium may store, in addition to or in lieu of instructions, data values produced by an implementation.
  • implementations may produce a variel of signals formatted to carry information that may be, for example, stored or transmitted.
  • the information may include, for example, instructions for performing a method, or data produced by one of the described implementations
  • a signal may be formatted to carry as data the rules for writing or reading syntax, or to carry as data the actual syntax-values generated using the syntax rules.
  • Such a signal may be formatted, for example, as an
  • the formatting may include, for example, encoding a data stream and modulating a carrier with the encoded data stream.
  • the information that the signal carries may be, for example, analog or digital information.
  • the signal may be transmitted over a variety of different wired or wireless links, as is known.
  • the signal may be stored on a processor-readable medium.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
  • Image Processing (AREA)
  • Processing Or Creating Images (AREA)

Abstract

Selon divers modes de réalisation, l'invention porte sur la mise à l'échelle d'une image pour lui donner une résolution différente. Dans un mode de réalisation, l'accès à une image source ayant une résolution source est obtenu, ainsi que l'accès à une résolution cible destinée à une image cible. Une limite inférieure est déterminée pour la taille d'une zone dans l'image source pouvant être sélectionnée afin de générer l'image cible. Cette limite inférieure est basée sur une mesure d'écart associée à la résolution cible. Une zone sélectionnée de l'image source, atteignant au moins la limite inférieure, est mise à l'échelle. L'image cible est basée sur la zone sélectionnée mise à l'échelle. Dans un autre mode de réalisation, une partie de l'image source est affichée. Une taille minimale pour le recadrage est identifiée à l'affichage. Une entrée est acceptée, cette entrée identifiant une zone sélectionnée de l'image source affichée pour le recadrage, et la zone sélectionnée a au moins la taille minimale.
PCT/US2013/041286 2012-05-17 2013-05-16 Adaptation d'un contenu stéréoscopique grand format à d'autres plateformes WO2013173548A2 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US201261688587P 2012-05-17 2012-05-17
US61/688,587 2012-05-17

Publications (2)

Publication Number Publication Date
WO2013173548A2 true WO2013173548A2 (fr) 2013-11-21
WO2013173548A3 WO2013173548A3 (fr) 2014-02-27

Family

ID=48626599

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2013/041286 WO2013173548A2 (fr) 2012-05-17 2013-05-16 Adaptation d'un contenu stéréoscopique grand format à d'autres plateformes

Country Status (1)

Country Link
WO (1) WO2013173548A2 (fr)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2015100490A1 (fr) * 2014-01-06 2015-07-09 Sensio Technologies Inc. Reconfiguration et distribution de contenu stereoscopique dans une configuration convenant pour un environnement de visualisation a distance
CN110314376A (zh) * 2019-06-04 2019-10-11 福建天晴数码有限公司 背景图片及背景光效的适配方法及计算机可读存储介质
WO2020000393A1 (fr) * 2018-06-29 2020-01-02 深圳市大疆创新科技有限公司 Procédé et appareil de traitement d'image, premier dispositif électronique et système de traitement d'image
CN111448587A (zh) * 2018-08-16 2020-07-24 华为技术有限公司 一种广告图片的显示方法、上传方法及装置
CN112684966A (zh) * 2017-12-13 2021-04-20 创新先进技术有限公司 一种图片缩放方法、装置以及电子设备
WO2022089076A1 (fr) * 2020-10-29 2022-05-05 海信视像科技股份有限公司 Dispositif d'affichage, terminal mobile et procédé de mise à l'échelle synchrone d'image
US20220181001A1 (en) * 2020-11-17 2022-06-09 Trumpf Medizin Systeme Gmbh + Co. Kg Operating room control and communication system

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2004349736A (ja) * 2003-05-08 2004-12-09 Sharp Corp 立体画像処理装置、立体画像処理プログラムおよびそのプログラムを記録した記録媒体
US20110050857A1 (en) * 2009-09-03 2011-03-03 Electronics And Telecommunications Research Institute Apparatus and method for displaying 3d image in 3d image system

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2004349736A (ja) * 2003-05-08 2004-12-09 Sharp Corp 立体画像処理装置、立体画像処理プログラムおよびそのプログラムを記録した記録媒体
US20110050857A1 (en) * 2009-09-03 2011-03-03 Electronics And Telecommunications Research Institute Apparatus and method for displaying 3d image in 3d image system

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2015100490A1 (fr) * 2014-01-06 2015-07-09 Sensio Technologies Inc. Reconfiguration et distribution de contenu stereoscopique dans une configuration convenant pour un environnement de visualisation a distance
CN112684966A (zh) * 2017-12-13 2021-04-20 创新先进技术有限公司 一种图片缩放方法、装置以及电子设备
WO2020000393A1 (fr) * 2018-06-29 2020-01-02 深圳市大疆创新科技有限公司 Procédé et appareil de traitement d'image, premier dispositif électronique et système de traitement d'image
CN111448587A (zh) * 2018-08-16 2020-07-24 华为技术有限公司 一种广告图片的显示方法、上传方法及装置
CN111448587B (zh) * 2018-08-16 2023-11-10 花瓣云科技有限公司 一种广告图片的显示方法、上传方法及装置
CN110314376A (zh) * 2019-06-04 2019-10-11 福建天晴数码有限公司 背景图片及背景光效的适配方法及计算机可读存储介质
WO2022089076A1 (fr) * 2020-10-29 2022-05-05 海信视像科技股份有限公司 Dispositif d'affichage, terminal mobile et procédé de mise à l'échelle synchrone d'image
US20220181001A1 (en) * 2020-11-17 2022-06-09 Trumpf Medizin Systeme Gmbh + Co. Kg Operating room control and communication system

Also Published As

Publication number Publication date
WO2013173548A3 (fr) 2014-02-27

Similar Documents

Publication Publication Date Title
US9986258B2 (en) Efficient encoding of multiple views
KR101810845B1 (ko) 스케일-독립적인 맵
Zinger et al. Free-viewpoint depth image based rendering
RU2554465C2 (ru) Комбинирование 3d видео и вспомогательных данных
WO2013173548A2 (fr) Adaptation d'un contenu stéréoscopique grand format à d'autres plateformes
JP5575778B2 (ja) 信号に含まれる視差情報を処理する方法
AU2011234163B2 (en) 3D disparity maps
US20090284584A1 (en) Image processing device
US9501815B2 (en) Processing panoramic pictures
EP2553932B1 (fr) Indications de valeur de disparité
US20130251241A1 (en) Applying Perceptually Correct 3D Film Noise
EP2875636A1 (fr) Métadonnées pour filtrage en profondeur
JP6148154B2 (ja) 画像処理装置及び画像処理プログラム
US10063845B2 (en) Guided 3D display adaptation
RU2632404C2 (ru) Данные сигнализации глубины
Ti et al. ROI-preserving 3D video compression method utilizing depth information

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 13729131

Country of ref document: EP

Kind code of ref document: A2

REG Reference to national code

Ref country code: BR

Ref legal event code: B01A

Ref document number: 112014025864

Country of ref document: BR

122 Ep: pct application non-entry in european phase

Ref document number: 13729131

Country of ref document: EP

Kind code of ref document: A2

ENP Entry into the national phase in:

Ref document number: 112014025864

Country of ref document: BR

Kind code of ref document: A2

Effective date: 20141016