EP2172025A1 - Enhancing image quality - Google Patents

Enhancing image quality

Info

Publication number
EP2172025A1
EP2172025A1 EP07796523A EP07796523A EP2172025A1 EP 2172025 A1 EP2172025 A1 EP 2172025A1 EP 07796523 A EP07796523 A EP 07796523A EP 07796523 A EP07796523 A EP 07796523A EP 2172025 A1 EP2172025 A1 EP 2172025A1
Authority
EP
European Patent Office
Prior art keywords
image
decoding
resolution
lower resolution
encoded
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Ceased
Application number
EP07796523A
Other languages
German (de)
French (fr)
Inventor
Rajan Laxman Joshi
James Arthur Fancher
Ana Belen Benitez
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Thomson Licensing SAS
Original Assignee
Thomson Licensing SAS
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Thomson Licensing SAS filed Critical Thomson Licensing SAS
Publication of EP2172025A1 publication Critical patent/EP2172025A1/en
Ceased legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/30Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/59Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial sub-sampling or interpolation, e.g. alteration of picture size or resolution
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/30Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
    • H04N19/33Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability in the spatial domain
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/46Embedding additional information in the video signal during the compression process
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/48Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using compressed domain processing techniques other than decoding, e.g. modification of transform coefficients, variable length coding [VLC] data or run-length data
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • H04N19/63Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding using sub-band based transform, e.g. wavelets

Definitions

  • the present principles relate to image processing. More particularly, at least one implementation relates to resolution scalability for imaging systems.
  • Imaging system is used to encompass systems which process and render still images as well as moving images or videos. Many of the known imaging systems provide access to a video frame at different resolutions. For example, in 4K workflows for Digital Cinema (DC) or Digital Intermediates (DI), lower resolution versions (proxies) of the 4K resolution images are available.
  • the video frames are stored in a compressed format, and a lower resolution version of the video frame (in compressed format) can be obtained from the higher resolution compressed code-stream by simple truncation or parsing.
  • parsing is used in the sense of accessing non-contiguous parts of the compressed code-stream without performing decompression.
  • a system may decode the high resolution video frame and perform downscaling to the lower resolution. Downscaling allows system providers to use different algorithms for downscaling.
  • a digital image is encoded into an encoded image.
  • the encoded image is capable of being processed to provide a higher resolution decoding of the digital image and to provide a lower resolution decoding of the digital image. Additional information is encoded for enhancing the lower resolution decoding to provide an enhanced lower resolution image.
  • encoded image data is accessed.
  • the accessed encoded image data is capable of being processed to provide a higher resolution decoding of an image and to provide a lower resolution decoding of the image.
  • Additional encoded information is accessed for enhancing the lower resolution decoding of the image to provide an enhanced lower resolution image.
  • an apparatus includes a processor that is configured to receive encoded image data and additional encoded data.
  • the apparatus includes a high resolution processing group, coupled to the processor, configured to generate low resolution extracted image data from high resolution compressed image data contained within the encoded image data.
  • the apparatus also includes a low resolution processing group, coupled to the processor, configured to generate an enhanced low resolution image using (a) the low resolution extracted image data and (b) the additional encoded data.
  • a signal representing information includes a first signal and a second signal portion.
  • the first signal portion represents encoded data for an encoded image, the encoded data capable of being processed to produce a higher resolution decoding of the encoded image and to produce a lower resolution decoding of the encoded image.
  • the second signal portion represents additional encoded information for enhancing the lower resolution decoding.
  • Figure 1 is a flow diagram of a method for creating a low resolution enhancement layer tc be used in conjunction with a resolution scalable encoding of a high resolution image.
  • Figure 2 is a flow diagram of a method for creating an enhanced low resolution image using the created low resolution enhancement layer of Figure 1.
  • Figure 3 is a flow diagram of a method for producing a 2K enhancement layer compressed code-stream.
  • Figure 4 is a flow diagram of a method for producing an enhanced 2K resolution image using the produced 2K enhancement layer compressed code-stream of Figure 3.
  • Figure 5 is a flow diagram of another implementation of a method for producing a 2K enhancement layer compressed code-stream.
  • Figure 6 is a flow diagram of another implementation of a method for producing an enhanced 2K resolution image using the produced 2K resolution enhancement layer compressed- code stream of Figure 5.
  • Figure 7 is a block diagram of an encoder for producing a low resolution enhancement layer.
  • Figure 8 is a block diagram of a decoder for producing an enhanced low resolution image.
  • Figure 9 is a block diagram of an encoder for producing a 2K enhancement layer compressed code-stream according to another implementation.
  • Figure 10 is a block diagram of a decoder for producing an enhanced 2K resolution image according to another implementation.
  • Figure 11 is a flow diagram of an encoding method for producing an enhanced low resolution image according to an implementation.
  • Figure 12 is a flow diagram of a decoding method for producing an enhanced low resolution image according to an implementation.
  • Figure 13 is a flow diagram of a DCP-Specification-compliant method for producing an enhanced 2K resolution image.
  • Figure 14 is a flow diagram of a DCP-Specification-compliant method for producing an enhanced 2K resolution image.
  • the compression method and file format used may place restrictions on the lower resolution image.
  • DCI Digital Cinema Initiative
  • Vl .0 June 2005
  • Digital Cinema Initiatives, LLC mandates that a 2K version should be contained in a 4K version and be accessible by simple truncation.
  • the downscaling filter is restricted to be the 9-tap low- pass filter used by JPEG2000.
  • a 2K projector has to use the 2K extracted code-stream to reconstruct the 2K resolution video, it may be placed at a disadvantage compared to 2K projectors which have access to 2K compressed material which was produced by first downscaling the 4K resolution video and then compressing the downscaled 2K version.
  • performing the downscaling operation is typically computationally complex. It would be desirable to provide a method to enhance the quality of a low resolution image extracted from a scalable high resolution code-stream.
  • Enhancement layers are commonly used to provide quality scalability.
  • An example of this is MPEG-4 Fine Grain Scalability in the MPEG-4 video standard.
  • the purpose of the enhancement layer is to produce a reconstructed image that is closer to the original image.
  • an enhancement layer is used to produce a reconstructed image that is closer to a downscaled version of the original image—and not necessarily closer to the original image itself.
  • the inventors have recognized that such a counter-intuitive operation provides for technical advantages such as improved quality within the context of, for example, a DCI-like system or standard.
  • a DCI-like standard may be defined as a standard that provides a compressed code-stream from which both a higher- resolution compressed bitstream and a lower-resolution compressed bitstream can be extracted before decompressing the compressed code-stream.
  • the enhancement layer implementation referred to above also uses different coding processes to produce the original encoding (layer) and the additional (enhancement) encoding (layer). This is different from systems that use a common algorithm to gradually create and send more granular information for the iterative encoding of a given image. In such systems, each of the successive encodings (layers) is related to each other. An example is a system that sends, in each layer, increasingly fine quantizations of DCT (discrete cosine transform) coefficients.
  • DCT discrete cosine transform
  • At least one implementation provides a method for increasing or enhancing the quality of a low resolution image extracted from the scalable high resolution code-stream.
  • the downscaling method of the present principles utilizes an enhancement layer to produce an enhanced low resolution image by directly operating on the high resolution image.
  • the high resolution image is a 4K (4096 * 2160) image and the low resolution image is a 2K image (2048 * 1080).
  • the high resolution image is a 4K (4096 * 2160) image and the low resolution image is a 2K image (2048 * 1080).
  • the decoder extracts a compressed code-stream corresponding to the 2K resolution and then decodes it to produce a 2K resolution extracted image.
  • this 2K resolution image is constrained to be the compressed version of the LL band (as known, the LL band is low frequency in both the horizontal and vertical directions) at the next lower resolution from the resolution hierarchy produced by the wavelet transform used in JPEG2000. As mentioned above, this may not be the best choice for a downscaling operation. For example, a downscaling method may be able to use a much better filter and adaptive processing to produce an improved or enhanced 2K image.
  • the projector is a 2K projector having a 2K decoder
  • the 2K image can be compressed at the maximum rate allowed by the DCI specification to produce a 2K distribution which is fed to the 2K projector.
  • the single inventory being the 4K distribution. This compares to having to provide in inventory the 4K distribution as well as the improved or enhanced 2K image.
  • the 2K projectors may be disadvantaged.
  • Figure 1 shows one implementation of creating a low resolution enhancement layer to be used in conjunction with a resolution scalable encoding of a high resolution image to produce an enhanced low resolution image.
  • a high resolution original image 100 (i.e., 4K image) is accessed and undergoes resolution scalable encoding 102 to produce a high resolution compressed code-stream 114.
  • Accessing is a broad term, including, for example, obtaining, retrieving, receiving, manipulating, or processing in various manners.
  • encoding and “compression”, are used herein interchangeably.
  • the DCI Spec (JPEG2000) encoding essentially comprises DC level- shifting, inter-component transform, wavelet transform, quantization, and entropy coding.
  • the high resolution original image 100 also undergoes downscaling 104 to produce a low resolution downscaled image 105.
  • the downscaling could be any desired downscaling operation, including a proprietary operation, to produce a downscaled image having a desired quality or a desired characteristic.
  • a low resolution compressed code-stream 107 is extracted (106) from the high resolution compressed code-stream 114, and the low resolution compressed code-stream 107 ⁇ undergoes low resolution decoding 108 to produce a low resolution extracted image 109. Taking the difference (110) between the low resolution downscaled image 105 and the low resolution extracted image 109 produces a low resolution enhancement layer 111.
  • the enhancement layer 111 is encoded 112 to produce a low resolution enhancement layer compressed code-stream 116.
  • This additional information 111 is, in a particular implementation of Figure 1, referred to as the enhancement layer 111.
  • the additional information is exactly the same as an enhancement layer.
  • the additional information is not referred to as an enhancement layer until after the additional information is formatted into a particular format of an enhancement layer.
  • the size of the enhancement layer compressed code-stream 116 may not exceed a fixed number of bytes. The byte limit may be determined, for example, based on user input.
  • Figure 2 shows how a low resolution enhancement layer compressed code-stream 116 can be combined with a corresponding high resolution compressed code-stream 114 to produce an enhanced low resolution image.
  • a low resolution compressed code-stream 204 is extracted 202 from the high resolution compressed code-stream 114.
  • the low resolution compressed code-stream 204 is then low resolution decoded 206 to obtain a low resolution extracted image 208 (expected to be the same as image 109).
  • a low resolution enhancement layer 212 is produced by decoding (210) the low resolution enhancement layer compressed code-stream 116.
  • the decoded low resolution enhancement layer 212 is added 214 to the low resolution extracted image 208 to produce an enhanced low resolution image 216.
  • the enhanced low resolution image 216 is expected to be the same as the low resolution downscaled image 105 if lossless encoding is used to encode the low resolution enhancement layer (112).
  • FIGS 7 and 8 show block diagrams of an encoder 700 and decoder 800 according to the implementation shown in Figures 1 and 2, respectively.
  • the encoder 700 includes a processor 702 configured to receive high resolution original image data, a memory 703 in communication with the processor and a high resolution processing group 720 and low resolution processing group 730.
  • the processor 702 is directly coupled to the memory 703, the encoder 704, and the downscaler 706.
  • the processor 702 is indirectly coupled to the extractor 708, the decoder 709, and other units within the encoder 700.
  • the term “coupled” refers to both direct (no intervening units) and indirect (intervening units) connections, and such connections may be, for example, wired or wireless, and permanent or transient.
  • the processor 702 sends the high resolution image data to both the high resolution processing group 720 and the low resolution processing group 730.
  • the high resolution processing group 720 inputs the high resolution original image data to a resolution scalable encoder 704.
  • the encoder 704 outputs the high resolution compressed code-stream which is both input to the extractor 708 and to the transmitter 712.
  • the extractor 708 extracts a low resolution compressed code-stream from the high resolution compressed code-stream, and inputs the same to a low resolution decoder 709 to produce a low resolution extracted image.
  • the low resolution processing group 730 inputs the high resolution original image data to a downscaler 706 which downscales the high resolution original image data to a low resolution downscaled image.
  • the subtractor 710 outputs the difference between the low resolution downscaled image and the low resolution extracted image from the high resolution processing group 720 to produce a low resolution enhancement layer.
  • the low resolution enhancement layer is encoded 711 to produce a low resolution enhancement layer compressed code-stream, which is input to the transmitter 712 for transmission.
  • implementation of the transmit/send step 712 may include, for example, storing the high resolution compressed code-stream and the enhancement layer compressed code-stream on a hard drive or other physical media and transporting it to another location.
  • the enhancement layer is generated by the combined action of both the high resolution processing group 720 and the low resolution processing group 730. Accordingly, the combination of the two groups 720 and 730 is referred to as an enhancement layer generator.
  • Figure 8 shows a decoder side 800 according to an implementation.
  • the data received from the encoder 700 which includes both the high resolution compressed code- stream and the low resolution enhancement layer compressed code stream, is received by a processor 802 or other processing device which may include a memory 803.
  • the processor 802 directs the high resolution compressed code stream to the high resolution processing group 820 and directs the low resolution enhancement layer compressed code stream to the low resolution processing group 830.
  • the high resolution processing group 820 inputs the high resolution compressed code stream to an extractor 804 to extract a low resolution compressed code stream.
  • the low resolution compressed code stream output from the extractor 804 is low resolution decoded 806 to produce a low resolution extracted image.
  • the low resolution processing group 830 directs the low resolution enhancement layer compressed code stream to an enhancement layer decoder 808.
  • the decoded low resolution enhancement layer is added 810 to the low resolution extracted image (generated from the high resolution image data) to produce an enhanced low resolution image.
  • the image could then be, for example, used for display on a
  • the enhanced image is generated by the combined action of both the high resolution processing group 820 and the low resolution processing group 830. Accordingly, the combination of the two groups 820 and 830 is referred to as an enhanced image generator.
  • the above implementation of Figures 2-3 and 7-8 illustrate a spatial-domain enhancement layer.
  • the enhancement layer includes data determined by, for example, subtracting (110) two images (109, 105) from each other.
  • the images are all in the spatial domain, and typically include pixel values.
  • the enhancement layer includes data determined by, for example, subtracting (322) two sets (314, 320) of wavelet coefficients from each other. The coefficients are part of frequency- domain representations of the underlying images.
  • FIG. 3 A flowchart for producing a 2K enhancement layer compressed code-stream in accordance with an implementation is shown in Figure 3.
  • An original 4K image 300 is encoded 302 as per the DCI specification (JPEG2000) to produce a DCI compliant 4K compressed code-stream 306.
  • a 2K resolution compressed code-stream 310 is extracted 308 from the DCI compliant 4K code-stream 306 by simple truncation.
  • the 2K resolution extracted compressed code-stream 310 undergoes entropy decoding and dequantization 312 to produce reconstructed wavelet coefficients 314 for the extracted 2K image.
  • Operation 312 does not include, in this implementation, inverse wavelet transformation because the wavelet coefficients, rather than for example the spatial- domain pixel data, are desired.
  • the original 4K image undergoes downscaling by a factor of 2 in the horizontal and vertical direction 304 to produce a downscaled 2K image 316.
  • the downscaled 2K image 316 is DC level-shifted (318) to produce pixel values centered around zero; and then undergoes irreversible color transform (ICT) (318) as specified in the JPEG2000 standard, and wavelet transformation (318) with the JPEG2000 (9,7) filters to produce wavelet coefficients 320 for the downscaled 2K image.
  • ICT irreversible color transform
  • the ICT (318) is applied only if the DCI 4K encoding step (302) uses ICT.
  • the number of decomposition levels for the wavelet transformation is one less than the decomposition levels used for the DCI 4K encoding.
  • the reconstructed wavelet coefficients 314 for the extracted 2K image are subtracted 322 from the wavelet coefficients 320 for the downscaled 2K image to produce wavelet coefficients 324 for the 2K enhancement layer.
  • the wavelet coefficients 324 for the 2K enhancement layer are quantized and entropy encoded using JPEG2000 (326) to produce a 2K enhancement layer compressed code-stream 328.
  • the quantization level can be set to produce, at most, a fixed number of bytes, which can be user-specified.
  • Figure 4 shows how a 2K enhancement layer compressed code-stream can be combined with a corresponding DCI-compliant 4K compressed code-stream to produce the enhanced 2K image.
  • a 2K resolution compressed code-stream 404 is extracted 402 from a DCI-compliant 4K compressed code-stream 306.
  • the 2K resolution compressed code- stream 404 undergoes entropy decoding and dequantization 406 in the JPEG2000 framework to produce reconstructed wavelet coefficients for the extracted 2K image 408.
  • a corresponding 2K enhancement layer compressed code-stream 328 undergoes entropy decoding and dequantization 410 in the JPEG2000 framework to produce reconstructed wavelet coefficients for the 2K enhancement layer 412.
  • the reconstructed wavelet coefficients for the extracted 2K image 408 and for the 2K enhancement layer 412 are added 414 to produce the wavelet coefficients for the enhanced 2K image 416.
  • the wavelet coefficients for the enhanced extracted 2K image 416 undergo wavelet synthesis and inverse ICT; and are then DC level-shifted and clipped 418 to the appropriate bit-depth to produce an enhanced 2K image 420.
  • the inverse ICT is applied only if the DCI-compliant 4K compressed code-stream 306 was produced using ICT.
  • the wavelet synthesis refers to a reverse wavelet transformation.
  • Figures 5 and 6 show an alternative implementation for producing an enhanced 2k image from an original 4k image according to the present principles.
  • the implementation of Figures 5 and 6 is also DCI-compliant, but works with images data (spatial domain) rather than wavelet coefficients (frequency domain) or other descriptors of the image data.
  • an original 4K image 502 is DCI 4K encoded 504 as per the DCI specification to produce a DCI compliant 4K compressed code-stream 506.
  • a 2K resolution compressed code-stream 510 is extracted 508 from the DCI compliant 4K code- stream by simple truncation.
  • the 2K resolution extracted compressed code-stream 510 undergoes JPEG2000 decoding (including inverse wavelet transformation) 512 to produce a reconstructed extracted 2K image 514,
  • JPEG2000 decoding includes entropy decoding, dequantization, inverse ICT (if needed), DC level-shifting, and clipping to the appropriate bit-depth.
  • the original 4K image also undergoes downscaling by a factor of 2 in the horizontal and vertical direction (516) to produce a downscaled 2K image 518.
  • the reconstructed extracted 2K image 514 is subtracted 520 from the downscaled 2K image 518 to produce a 2K enhancement layer 522.
  • the 2K enhancement layer 522 is encoded 524 using JPEG2000 to produce a 2K enhancement layer compressed code-stream 526.
  • the quantization level can be set to produce at most a fixed number of bytes, which can be user-specified.
  • a 2K resolution compressed code-stream 604 is extracted from a DCI- compliant 4K compressed code-stream 506.
  • the 2K resolution compressed code-stream 604 undergoes JPEG2000 decoding 606 to produce a reconstructed extracted 2K image 608.
  • a corresponding 2K enhancement layer compressed code-stream 526 undergoes JPEG2000 decoding 610 to produce a reconstructed 2K enhancement layer 612.
  • the reconstructed extracted 2K image 608 and the reconstructed 2K enhancement layer 612 are added 614 and clipped 616 to the appropriate bit-depth, to produce an enhanced 2K image 618.
  • the enhanced 2K image 618 is expected to correspond to the downscaled 2K image 518 if lossless encoding is used to encode 524 the 2K enhancement layer 522.
  • Figure 9 shows an implementation of an encoder 900 that can be used to implement the method of Figure 3.
  • Figure 10 shows an implementation of a decoder 1000 that can be used to implement the method of Figure 4.
  • the original 4k image is received by a processor 902 or other input device.
  • the processor 902 is coupled to a memory 903 and is configured to input the original 4k image to a high resolution processing group 920 and a low resolution processing group 930.
  • the high resolution processing group 920 receives the original 4K image and inputs it to a DCI 4K encoder 904.
  • the DCI 4k encoder 904 produces a DCI compliant 4K compressed code stream, which is input to a transmitter 918 for transmission and is also input to an extractor 906 to extract a 2K resolution compressed code stream from the DCI compliant 4K compressed code stream.
  • the 2K resolution compressed code stream is decoded 908 (entropy decoding and dequantization) to produce reconstructed wavelet coefficients for the extracted 2K image.
  • the decoding in operation 908 may also be referred to as a partial decoding because the code stream is only decoded to the point of providing the wavelet coefficients and no inverse wavelet transformation is performed.
  • extraction (906) and decoding (908) can be generally referred to collectively as processing or even decoding.
  • the low resolution processing group 930 receives the original 4K image and downscales (910) the same to produce a downscaled 2K image.
  • Wavelet coefficients for the downscaled 2K image are generated by level shifting and wavelet transformation (912).
  • the difference between reconstructed wavelet coefficients for the extracted 2K image (generated by the high resolution processing group) and the wavelet coefficients for the downscaled 2K image is determined by a subtractor 914 to produce wavelet coefficients for the 2K enhancement layer.
  • These wavelet coefficients are enhancement layer encoded (916) and sent to the transmitter 918 for transmission.
  • the transmit/send unit 918 may combine the encoded enhancement layer (916) and the DCI compliant 4K compressed code stream (904) into a single transmission or storage unit.
  • unit 918 may form a DC package that includes that includes the compressed code-stream and at least one of (1) the enhancement-layer compressed code-stream or (2) a pointer to the enhancement-layer compressed code-stream. Accordingly, unit 918 may be referred to as a "combiner".
  • the enhancement layer is generated by the combined action of both the high resolution processing group 920 and the low resolution processing group 930. Accordingly, the combination of the two groups 920 and 930 is referred to as an enhancement layer generator.
  • FIG. 10 shows a decoder 1000 according to an implementation.
  • the decoder 1000 includes a processor 1002 and memory 1003 where the processor is configured to receive the data from an encoder.
  • the data received from the encoder includes the DCI compliant 4K compressed code-stream and the 2K resolution enhancement layer compressed code-stream.
  • the processor 1002 sends the 4K resolution compressed code-stream to a high resolution processing group 1020 and the 2K resolution enhancement layer compressed code-stream to a low resolution processing group 1030.
  • the extractor 1004 extracts the 2K resolution compressed code-stream from the 4K resolution image data and the decoder (e.g., JPEG2000 decoder) 1006 decodes the extracted 2K resolution compressed code-stream to produce reconstructed wavelet coefficients for the extracted 2K resolution image.
  • the decoder e.g., JPEG2000 decoder
  • the low resolution processing group 1030 includes an enhancement layer decoder 1008 that is configured to receive the 2K resolution enhancement layer compressed code stream, and to produce reconstructed wavelet coefficients for the 2K resolution enhancement layer.
  • the reconstructed wavelet coefficients for the extracted 2K resolution image (generated by the high resolution processing group) are then added (1010) with the reconstructed wavelet coefficients for the 2K resolution enhancement layer (generated by the low resolution processing group) to produce wavelet coefficients for the enhanced extracted low resolution image.
  • These wavelet coefficients are then synthesized, undergo inverse ICT, and are then level shifted and clipped (1012) to produce the enhanced 2K resolution image.
  • the inverse ICT step is performed only if the DCI compliant 4K compressed code-stream is generated using ICT. As before, these images may be displayed on a display 1014.
  • the enhanced image is generated by the combined action of both the high resolution processing group 1020 and the low resolution processing group 1030. Accordingly, the combination of the two groups 1020 and 1030 is referred to as an enhanced image generator.
  • Figures 9 and 10 may be adapted to implement the methods of Figures 5 and 6.
  • One of ordinary skill will readily understand how to make such modifications, particularly in light of the disclosure and discussion of Figures 7 and 8.
  • bit-depth for the 2K enhancement layer should be chosen so as to avoid unnecessary clipping.
  • the 4K and 2K images are 12-bit.
  • the 2K enhancement layer has to be encoded as a signed 13-bit image. If the subtraction is going to take place in the wavelet domain, as shown in Figure 3, then the analysis of the bit-depth needed may be more complex.
  • both the analysis filters (low pass and high pass) are implemented with (1,1) normalization as described by Taubman and Marcellin ("JPEG2000 image compression fundamentals, standards and practice" Kluwer Academic Publishers, ISBN 0-7923-7519-x, 2002).
  • N bit-depth of the original 4K image
  • (N+ 1) bits are sufficient to represent each subband without any clipping.
  • the downscaled 2K image also has a bit-depth of N
  • its wavelet coefficients can be represented by (N+ J) bits.
  • the wavelet coefficients for the 2K enhancement layer which are calculated as a difference, can be represented by (N+2) bits.
  • the wavelet coefficients of the enhancement layer have been generated by wavelet decomposition of a (N+ J) bit signed image.
  • the 2k enhancement layer may be compressed using any other compression method instead of JPEG2000 encoding.
  • the wavelet coefficients for the 2K enhancement layer may be compressed using any other compression method instead of JPEG2000 encoding.
  • JPEG2000 encoding techniques can be used for compressing the 2K enhancement layer without producing a JPEG2000 compliant code-stream. This is accomplished, for example, by forcing the 2K enhancement layer encoding engine to use the equivalent parameters as the 4K JPEG2000 encoding engine. In that case, it is not necessary to store a number of marker segments. But the resulting gain in compression efficiency is typically small.
  • a DC composition is defined as an ordered sequence of reels. Each reel can contain multiple media track files to be reproduced in parallel during the presentation. A reel can have a single main picture, main sound, and main subtitle track files. However, the current standard specification also supports additional track files for future uses.
  • a DC composition can be put into one or more DC packages for distribution to theaters or other venues.
  • the low resolution enhancement can be performed by maintaining the enhancement data separate from the DCP Stream.
  • Figures 11 and 12 show an exemplary implementation of this concept from the transmitter (i.e., creation) and receiving (i.e., playback) sides, respectively.
  • a DCI compliant 4K compressed code-stream 1102 is used to create DC picture track files 1106 and to create the 2K enhancement layer compressed code stream 1104.
  • the 2K enhancement layer compressed code stream 1104 is placed in a separate file 1108.
  • One or more DC packages are created 1110 using the previously created DC picture track files and which include the filename or a pointer to the enhancement layer file.
  • pointer is a broad term that includes, for example, an address, a name (e.g. a filename), or other descriptor that indicates where the enhancement layer file is located.
  • the filename, or pointer is included in a metadata field that the standard defines.
  • the metadata field may be, for example, a free-form text field, effectively allowing an implementation to use the field for a variety of purposes.
  • the one or more DC packages and the enhancement layer file can be sent 1112 (either manually via mail or electronically via communication network of any suitable type).
  • the enhancement layer file can be transmitted completely separate from the DC package information.
  • the creation of the DCI compliant 4K compressed code stream and the 2K enhancement layer compressed code-stream may be, for example, as described above.
  • Figure 12 shows the method 1200 for receiving and processing the enhancement file data in accordance with an implementation. As shown, one or more DC packages are received 1202. A determination 1204 is then made as to whether the DC packages include a filename or other pointer to an enhancement layer file.
  • the DC picture track is decoded and the low resolution (2K) image is generated in accordance with the DCI spec 1206. If there is an enhancement filename or pointer contained within the DC package information, the enhancement layer file named in the metadata is identified, located, and accessed 1208. Once identified and accessed, the DC picture track file(s) is decoded and the enhanced low resolution (2K) image is generated 1210 using the enhancement layer file data. On the receiving end of such transmissions, for example, movie theaters or other playback locations that are not aware of the enhancement layer data would playback the extracted 2K version directly from the DC picture track file(s).
  • Playback systems that are aware of the enhancement layer data would produce the enhanced 2K version from the DC main picture track file(s) and the enhancement layer picture track file(s).
  • this implementation is backward compatible with legacy and other existing systems that implement the DCI specification. Such systems will typically ignore a field that is not expected to have data, or if the system does not know what to do with data in such a field.
  • Figures 13 and 14 show a representative implementation of a DCP Specification compliant implementation of the present principles.
  • the method 1300 of Figure 13 shows that both the high resolution compressed code-stream and the low resolution enhancement layer compressed code-stream are combined such that the low resolution enhancement layer compressed code-stream is included in the created DC packages. That is, the low resolution enhancement layer compressed code-stream is included in an additional picture track within the DC packages.
  • the high resolution compressed code-stream is included in the DC main picture track (1302) and the low resolution enhancement layer compressed code-stream is included into DC additional picture track file(s) (1304).
  • the DC package(s) are created 1306.
  • the created DC packages can then be sent (i.e., manually in reels or other storage media) or transmitted electronically 1308 to the respective receivers (e.g., movie theaters) of the image data.
  • FIG 14 shows the receiving end 1400 of the DC package data.
  • the DC package data is received 1402, and a determination is then made (1404) as to whether the DC package data includes the low resolution enhancement layer data as an additional picture track file. If it does, the DC main and additional picture tracks are decoded (1408) and the enhanced low resolution image is generated using the extracted 2K image and the enhancement layer data contained in the DC main and additional picture track file (1408). If there is no enhancement layer data contained as an additional picture track in the DC package(s), the DC package(s) are decoded (1406) and the low resolution (2K) image is generated in accordance with standard DCI Specification requirements.
  • the additional picture track file or is aware, but does not know what to do with the additional picture track file
  • the system typically ignores the additional picture track file.
  • a system may include an additional picture track file (in a DC package) for a first enhancement layer, as well as a pointer to a second enhancement layer.
  • a system may include in any given DC package only one of an additional picture track file or a pointer, but the system may allow a user to choose either mechanism for each DC package. Such a choice may be made, for example, in real time based on the size of the enhancement layer for the given DC package.
  • a system could make the choice without user-intervention by, for example, consulting a look-up table that indicates the preferred mechanism for a given movie or a given intended-recipient of the DC package.
  • Figure 4 describes a decoder that requires at least one inverse wavelet transformation (unit 418), and Figure 6 describes a decoder that requires at least two inverse wavelet transformations (units 606 and 610).
  • Apps include, for example, a pre-processor or an encoder for creating DC packages or other data packages, or a decoder or other video receiving/processing apparatus, such as described above.
  • various units may be integrated, such as, for example, a pre-processor and an encoder.
  • the features and aspects herein described may, however, be adapted for other application areas.
  • the implementations described herein may be implemented in, for example, a method or process, an apparatus, or a software program. Even if only discussed in the context of a single form of implementation (for example, discussed only as a method), the implementation of features discussed may also be implemented in other forms (for example, an apparatus or program).
  • An apparatus may be implemented in, for example, appropriate hardware, software, and firmware.
  • the methods may be implemented in, for example, an apparatus such as, for example, a processor, which refers to processing devices in general, including, for example, a computer, a microprocessor, an integrated circuit, or a programmable logic device.
  • Implementations of the various processes and features described herein may be . embodied in a variety of different equipment or applications, particularly, for example, equipment or applications associated with video transmission.
  • equipment include video coders, video decoders, video codecs, web servers, and personal computers.
  • encodings may be sent or received over a variety of paths, including, for example, wireless or wired paths, the Internet, cable television lines, telephone lines, and Ethernet connections. Additionally, as should be clear, the equipment may be mobile and even installed in a mobile vehicle.
  • the methods may be implemented by instructions being performed by a processor, and such instructions may be stored on a processor readable medium such as, for example, an integrated circuit, a software carrier, or other storage device such as, for example, a hard disk, a compact diskette, a random access memory (“RAM”), or a read-only memory (“ROM").
  • a processor may also include a processor readable medium having, for example, instructions for carrying out a process.
  • implementations may also produce a signal formatted to carry information that may be, for example, stored or transmitted.
  • the signal may be transmitted as, for example, an electromagnetic wave, and may carry information by, for example, modulating one or more carrier frequencies.
  • the information may include, for example, instructions for performing a method, or data produced by one of the described implementations.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)
  • Editing Of Facsimile Originals (AREA)
  • Compression Of Band Width Or Redundancy In Fax (AREA)

Abstract

At least one particular method and apparatus provide an enhanced quality low resolution image extracted from a scalable high resolution bit stream in a DCP bit stream. In various implementations, an enhancement layer is used to produce the enhanced low resolution image. The enhancement layer may include, for example, spatial-domain or frequency-domain information to be used to enhance the typical low resolution image extracted from a DCP bit stream. One specific process includes encoding (102, 302) a digital image into an encoded image. The encoded image is capable of being processed to provide a higher resolution decoding of the digital image and to provide a lower resolution decoding of the digital image. The process further includes encoding (112, 326) additional information for enhancing the lower resolution decoding to provide an enhanced lower resolution image. A complimentary signal and decoding process are also provided, as well as structures for performing the processes.

Description

ENHANCING IMAGE QUALITY
BACKGROUND
Technical Field The present principles relate to image processing. More particularly, at least one implementation relates to resolution scalability for imaging systems.
Description of the prior art
Resolution scalability is a feature in a number of imaging systems. The term "imaging system" is used to encompass systems which process and render still images as well as moving images or videos. Many of the known imaging systems provide access to a video frame at different resolutions. For example, in 4K workflows for Digital Cinema (DC) or Digital Intermediates (DI), lower resolution versions (proxies) of the 4K resolution images are available. The video frames are stored in a compressed format, and a lower resolution version of the video frame (in compressed format) can be obtained from the higher resolution compressed code-stream by simple truncation or parsing. The term parsing is used in the sense of accessing non-contiguous parts of the compressed code-stream without performing decompression. In the absence of resolution scalability, to obtain a lower resolution, a system may decode the high resolution video frame and perform downscaling to the lower resolution. Downscaling allows system providers to use different algorithms for downscaling. SUMMARY
According to one general aspect, a digital image is encoded into an encoded image. The encoded image is capable of being processed to provide a higher resolution decoding of the digital image and to provide a lower resolution decoding of the digital image. Additional information is encoded for enhancing the lower resolution decoding to provide an enhanced lower resolution image.
According to another general aspect, encoded image data is accessed. The accessed encoded image data is capable of being processed to provide a higher resolution decoding of an image and to provide a lower resolution decoding of the image. Additional encoded information is accessed for enhancing the lower resolution decoding of the image to provide an enhanced lower resolution image.
According to another general aspect, an apparatus includes a processor that is configured to receive encoded image data and additional encoded data. The apparatus includes a high resolution processing group, coupled to the processor, configured to generate low resolution extracted image data from high resolution compressed image data contained within the encoded image data. The apparatus also includes a low resolution processing group, coupled to the processor, configured to generate an enhanced low resolution image using (a) the low resolution extracted image data and (b) the additional encoded data.
According to another general aspect, a signal representing information includes a first signal and a second signal portion. The first signal portion represents encoded data for an encoded image, the encoded data capable of being processed to produce a higher resolution decoding of the encoded image and to produce a lower resolution decoding of the encoded image. The second signal portion represents additional encoded information for enhancing the lower resolution decoding. The details of one or more implementations are set forth in the accompanying drawings and the description below. Even if described in one particular manner, it should be clear that implementations may be configured or embodied in various manners. For example, an implementation may be performed as a method, or embodied as an apparatus configured to perform a set of operations or an apparatus storing instructions for performing a set of operations. Other aspects and features will become apparent from the following detailed description considered in conjunction with the accompanying drawings and the claims.
BRIEF DESCRIPTION OF THE DRAWINGS In the drawings like reference numerals generally denote similar components throughout the views.
Figure 1 is a flow diagram of a method for creating a low resolution enhancement layer tc be used in conjunction with a resolution scalable encoding of a high resolution image.
Figure 2 is a flow diagram of a method for creating an enhanced low resolution image using the created low resolution enhancement layer of Figure 1.
Figure 3 is a flow diagram of a method for producing a 2K enhancement layer compressed code-stream.
Figure 4 is a flow diagram of a method for producing an enhanced 2K resolution image using the produced 2K enhancement layer compressed code-stream of Figure 3. Figure 5 is a flow diagram of another implementation of a method for producing a 2K enhancement layer compressed code-stream.
Figure 6 is a flow diagram of another implementation of a method for producing an enhanced 2K resolution image using the produced 2K resolution enhancement layer compressed- code stream of Figure 5. Figure 7 is a block diagram of an encoder for producing a low resolution enhancement layer.
Figure 8 is a block diagram of a decoder for producing an enhanced low resolution image. Figure 9 is a block diagram of an encoder for producing a 2K enhancement layer compressed code-stream according to another implementation.
Figure 10 is a block diagram of a decoder for producing an enhanced 2K resolution image according to another implementation.
Figure 11 is a flow diagram of an encoding method for producing an enhanced low resolution image according to an implementation.
Figure 12 is a flow diagram of a decoding method for producing an enhanced low resolution image according to an implementation.
Figure 13 is a flow diagram of a DCP-Specification-compliant method for producing an enhanced 2K resolution image. Figure 14 is a flow diagram of a DCP-Specification-compliant method for producing an enhanced 2K resolution image.
DETAILED DESCRIPTION
In the case of resolution scalability, the compression method and file format used may place restrictions on the lower resolution image. As an example, the Digital Cinema
Initiative (DCI) specification Vl .0, July 2005, issued by Digital Cinema Initiatives, LLC, mandates that a 2K version should be contained in a 4K version and be accessible by simple truncation. However, this means that the downscaling filter is restricted to be the 9-tap low- pass filter used by JPEG2000. If a 2K projector has to use the 2K extracted code-stream to reconstruct the 2K resolution video, it may be placed at a disadvantage compared to 2K projectors which have access to 2K compressed material which was produced by first downscaling the 4K resolution video and then compressing the downscaled 2K version. However, performing the downscaling operation is typically computationally complex. It would be desirable to provide a method to enhance the quality of a low resolution image extracted from a scalable high resolution code-stream.
Enhancement layers are commonly used to provide quality scalability. An example of this is MPEG-4 Fine Grain Scalability in the MPEG-4 video standard. Typically, the purpose of the enhancement layer is to produce a reconstructed image that is closer to the original image.
However, in at least one implementation, an enhancement layer is used to produce a reconstructed image that is closer to a downscaled version of the original image—and not necessarily closer to the original image itself. The inventors have recognized that such a counter-intuitive operation provides for technical advantages such as improved quality within the context of, for example, a DCI-like system or standard. A DCI-like standard may be defined as a standard that provides a compressed code-stream from which both a higher- resolution compressed bitstream and a lower-resolution compressed bitstream can be extracted before decompressing the compressed code-stream.
The enhancement layer implementation referred to above also uses different coding processes to produce the original encoding (layer) and the additional (enhancement) encoding (layer). This is different from systems that use a common algorithm to gradually create and send more granular information for the iterative encoding of a given image. In such systems, each of the successive encodings (layers) is related to each other. An example is a system that sends, in each layer, increasingly fine quantizations of DCT (discrete cosine transform) coefficients. However, the enhancement layer implementation referred to above uses two separate and independent encoding processes, as well as starts with different images.
At least one implementation provides a method for increasing or enhancing the quality of a low resolution image extracted from the scalable high resolution code-stream. In accordance with an implementation, the downscaling method of the present principles utilizes an enhancement layer to produce an enhanced low resolution image by directly operating on the high resolution image. An implementation of the present principles is now described with reference to Digital Cinema System Specification V 1.0, July 2005 issued by Digital Cinema Initiatives, LLC.
By way of example, the high resolution image is a 4K (4096 * 2160) image and the low resolution image is a 2K image (2048 * 1080). Consider an original 4K image that is encoded according to the DCI specification using JPEG2000 to produce a DCI compliant 4K compressed code-stream. Those of skill in the art will recognize that the terminology used herein is from the JPEG2000 standard.
If the 4K compressed code-stream is fed to a 2K decoder, the decoder extracts a compressed code-stream corresponding to the 2K resolution and then decodes it to produce a 2K resolution extracted image. However, this 2K resolution image is constrained to be the compressed version of the LL band (as known, the LL band is low frequency in both the horizontal and vertical directions) at the next lower resolution from the resolution hierarchy produced by the wavelet transform used in JPEG2000. As mentioned above, this may not be the best choice for a downscaling operation. For example, a downscaling method may be able to use a much better filter and adaptive processing to produce an improved or enhanced 2K image. Thus, if the projector is a 2K projector having a 2K decoder, it may be better to perform the downscaling operation on the higher resolution 4K original image to produce the downscaled 2K image. Then, the 2K image can be compressed at the maximum rate allowed by the DCI specification to produce a 2K distribution which is fed to the 2K projector. However, there is a desire to have a single inventory on the sender (e.g. studio) side to reduce costs. The single inventory being the 4K distribution. This compares to having to provide in inventory the 4K distribution as well as the improved or enhanced 2K image. In the case of a single inventory, the 2K projectors may be disadvantaged.
One way to reduce this disadvantage is through an implementation of the present principles. Figure 1 shows one implementation of creating a low resolution enhancement layer to be used in conjunction with a resolution scalable encoding of a high resolution image to produce an enhanced low resolution image.
A high resolution original image 100 (i.e., 4K image) is accessed and undergoes resolution scalable encoding 102 to produce a high resolution compressed code-stream 114. "Accessing" is a broad term, including, for example, obtaining, retrieving, receiving, manipulating, or processing in various manners. The terms "encoding" and "compression", are used herein interchangeably.
For example, the DCI Spec (JPEG2000) encoding essentially comprises DC level- shifting, inter-component transform, wavelet transform, quantization, and entropy coding. As mentioned earlier, the high resolution original image 100 also undergoes downscaling 104 to produce a low resolution downscaled image 105. The downscaling could be any desired downscaling operation, including a proprietary operation, to produce a downscaled image having a desired quality or a desired characteristic.
A low resolution compressed code-stream 107 is extracted (106) from the high resolution compressed code-stream 114, and the low resolution compressed code-stream 107 δ undergoes low resolution decoding 108 to produce a low resolution extracted image 109. Taking the difference (110) between the low resolution downscaled image 105 and the low resolution extracted image 109 produces a low resolution enhancement layer 111. The enhancement layer 111 is encoded 112 to produce a low resolution enhancement layer compressed code-stream 116.
Note that, more generally, the implementation produces additional information 1 11. This additional information 111 is, in a particular implementation of Figure 1, referred to as the enhancement layer 111. In various implementations the additional information is exactly the same as an enhancement layer. In other implementations the additional information is not referred to as an enhancement layer until after the additional information is formatted into a particular format of an enhancement layer.
Those skilled in the art will recognize that it is possible to encode the low resolution enhancement layer at any quality, including lossless. This affects the size of the low resolution layer compressed code-stream 116. In at least one implementation, the size of the enhancement layer compressed code-stream 116 may not exceed a fixed number of bytes. The byte limit may be determined, for example, based on user input.
Figure 2 shows how a low resolution enhancement layer compressed code-stream 116 can be combined with a corresponding high resolution compressed code-stream 114 to produce an enhanced low resolution image. A low resolution compressed code-stream 204 is extracted 202 from the high resolution compressed code-stream 114. The low resolution compressed code-stream 204 is then low resolution decoded 206 to obtain a low resolution extracted image 208 (expected to be the same as image 109). Meanwhile, a low resolution enhancement layer 212 is produced by decoding (210) the low resolution enhancement layer compressed code-stream 116. The decoded low resolution enhancement layer 212 is added 214 to the low resolution extracted image 208 to produce an enhanced low resolution image 216. The enhanced low resolution image 216 is expected to be the same as the low resolution downscaled image 105 if lossless encoding is used to encode the low resolution enhancement layer (112).
Those skilled in the art will also recognize that it is possible to create multiple low resolution enhancement layers, each corresponding to a different resolution. As an example, you could have a 2K enhancement layer and another IK enhancement layer. Each layer, when combined with the corresponding extracted resolution would create an enhanced image at that lower resolution. Figures 7 and 8 show block diagrams of an encoder 700 and decoder 800 according to the implementation shown in Figures 1 and 2, respectively. The encoder 700 includes a processor 702 configured to receive high resolution original image data, a memory 703 in communication with the processor and a high resolution processing group 720 and low resolution processing group 730.
The processor 702 is directly coupled to the memory 703, the encoder 704, and the downscaler 706. The processor 702 is indirectly coupled to the extractor 708, the decoder 709, and other units within the encoder 700. As should be clear, the term "coupled" refers to both direct (no intervening units) and indirect (intervening units) connections, and such connections may be, for example, wired or wireless, and permanent or transient.
The processor 702 sends the high resolution image data to both the high resolution processing group 720 and the low resolution processing group 730. The high resolution processing group 720 inputs the high resolution original image data to a resolution scalable encoder 704. The encoder 704 outputs the high resolution compressed code-stream which is both input to the extractor 708 and to the transmitter 712. The extractor 708 extracts a low resolution compressed code-stream from the high resolution compressed code-stream, and inputs the same to a low resolution decoder 709 to produce a low resolution extracted image. The low resolution processing group 730 inputs the high resolution original image data to a downscaler 706 which downscales the high resolution original image data to a low resolution downscaled image. The subtractor 710 outputs the difference between the low resolution downscaled image and the low resolution extracted image from the high resolution processing group 720 to produce a low resolution enhancement layer. The low resolution enhancement layer is encoded 711 to produce a low resolution enhancement layer compressed code-stream, which is input to the transmitter 712 for transmission. It will be understood by those skilled in the art that implementation of the transmit/send step 712 may include, for example, storing the high resolution compressed code-stream and the enhancement layer compressed code-stream on a hard drive or other physical media and transporting it to another location.
As should be clear, the enhancement layer is generated by the combined action of both the high resolution processing group 720 and the low resolution processing group 730. Accordingly, the combination of the two groups 720 and 730 is referred to as an enhancement layer generator.
Figure 8 shows a decoder side 800 according to an implementation. The data received from the encoder 700, which includes both the high resolution compressed code- stream and the low resolution enhancement layer compressed code stream, is received by a processor 802 or other processing device which may include a memory 803. The processor 802 directs the high resolution compressed code stream to the high resolution processing group 820 and directs the low resolution enhancement layer compressed code stream to the low resolution processing group 830. The high resolution processing group 820 inputs the high resolution compressed code stream to an extractor 804 to extract a low resolution compressed code stream. The low resolution compressed code stream output from the extractor 804 is low resolution decoded 806 to produce a low resolution extracted image. The low resolution processing group 830 directs the low resolution enhancement layer compressed code stream to an enhancement layer decoder 808. The decoded low resolution enhancement layer is added 810 to the low resolution extracted image (generated from the high resolution image data) to produce an enhanced low resolution image. The image could then be, for example, used for display on a display 812.
Analogous to the discussion of Figure 7, it is clear that the enhanced image is generated by the combined action of both the high resolution processing group 820 and the low resolution processing group 830. Accordingly, the combination of the two groups 820 and 830 is referred to as an enhanced image generator.
The above implementation of Figures 2-3 and 7-8 illustrate a spatial-domain enhancement layer. In the spatial-domain implementation, the enhancement layer includes data determined by, for example, subtracting (110) two images (109, 105) from each other. The images are all in the spatial domain, and typically include pixel values.
In contrast, an implementation below, described in the context of Figures 3-4, illustrates a frequency-domain enhancement layer. In the frequency-domain implementation, the enhancement layer includes data determined by, for example, subtracting (322) two sets (314, 320) of wavelet coefficients from each other. The coefficients are part of frequency- domain representations of the underlying images.
Another implementation of the present principles is now discussed in the framework of DCI compliant 4K encoding. As mentioned earlier, in this scenario, the high resolution image is a 4K image and the low resolution image is a 2K image. A flowchart for producing a 2K enhancement layer compressed code-stream in accordance with an implementation is shown in Figure 3. An original 4K image 300 is encoded 302 as per the DCI specification (JPEG2000) to produce a DCI compliant 4K compressed code-stream 306. A 2K resolution compressed code-stream 310 is extracted 308 from the DCI compliant 4K code-stream 306 by simple truncation. The 2K resolution extracted compressed code-stream 310 undergoes entropy decoding and dequantization 312 to produce reconstructed wavelet coefficients 314 for the extracted 2K image. Operation 312 does not include, in this implementation, inverse wavelet transformation because the wavelet coefficients, rather than for example the spatial- domain pixel data, are desired.
The original 4K image undergoes downscaling by a factor of 2 in the horizontal and vertical direction 304 to produce a downscaled 2K image 316. The downscaled 2K image 316 is DC level-shifted (318) to produce pixel values centered around zero; and then undergoes irreversible color transform (ICT) (318) as specified in the JPEG2000 standard, and wavelet transformation (318) with the JPEG2000 (9,7) filters to produce wavelet coefficients 320 for the downscaled 2K image. The ICT (318) is applied only if the DCI 4K encoding step (302) uses ICT. The number of decomposition levels for the wavelet transformation is one less than the decomposition levels used for the DCI 4K encoding.
The reconstructed wavelet coefficients 314 for the extracted 2K image are subtracted 322 from the wavelet coefficients 320 for the downscaled 2K image to produce wavelet coefficients 324 for the 2K enhancement layer. The wavelet coefficients 324 for the 2K enhancement layer are quantized and entropy encoded using JPEG2000 (326) to produce a 2K enhancement layer compressed code-stream 328. As before, the quantization level can be set to produce, at most, a fixed number of bytes, which can be user-specified. Figure 4 shows how a 2K enhancement layer compressed code-stream can be combined with a corresponding DCI-compliant 4K compressed code-stream to produce the enhanced 2K image. A 2K resolution compressed code-stream 404 is extracted 402 from a DCI-compliant 4K compressed code-stream 306. The 2K resolution compressed code- stream 404 undergoes entropy decoding and dequantization 406 in the JPEG2000 framework to produce reconstructed wavelet coefficients for the extracted 2K image 408. Similarly, a corresponding 2K enhancement layer compressed code-stream 328 undergoes entropy decoding and dequantization 410 in the JPEG2000 framework to produce reconstructed wavelet coefficients for the 2K enhancement layer 412. The reconstructed wavelet coefficients for the extracted 2K image 408 and for the 2K enhancement layer 412 are added 414 to produce the wavelet coefficients for the enhanced 2K image 416. The wavelet coefficients for the enhanced extracted 2K image 416 undergo wavelet synthesis and inverse ICT; and are then DC level-shifted and clipped 418 to the appropriate bit-depth to produce an enhanced 2K image 420. The inverse ICT is applied only if the DCI-compliant 4K compressed code-stream 306 was produced using ICT. The wavelet synthesis refers to a reverse wavelet transformation.
Figures 5 and 6 show an alternative implementation for producing an enhanced 2k image from an original 4k image according to the present principles. The implementation of Figures 5 and 6 is also DCI-compliant, but works with images data (spatial domain) rather than wavelet coefficients (frequency domain) or other descriptors of the image data.
Referring to Figure 5, an original 4K image 502 is DCI 4K encoded 504 as per the DCI specification to produce a DCI compliant 4K compressed code-stream 506. A 2K resolution compressed code-stream 510 is extracted 508 from the DCI compliant 4K code- stream by simple truncation. The 2K resolution extracted compressed code-stream 510 undergoes JPEG2000 decoding (including inverse wavelet transformation) 512 to produce a reconstructed extracted 2K image 514, Here JPEG2000 decoding includes entropy decoding, dequantization, inverse ICT (if needed), DC level-shifting, and clipping to the appropriate bit-depth. The original 4K image also undergoes downscaling by a factor of 2 in the horizontal and vertical direction (516) to produce a downscaled 2K image 518. The reconstructed extracted 2K image 514 is subtracted 520 from the downscaled 2K image 518 to produce a 2K enhancement layer 522. The 2K enhancement layer 522 is encoded 524 using JPEG2000 to produce a 2K enhancement layer compressed code-stream 526. As before, the quantization level can be set to produce at most a fixed number of bytes, which can be user-specified.
In Figure 6 a 2K resolution compressed code-stream 604 is extracted from a DCI- compliant 4K compressed code-stream 506. The 2K resolution compressed code-stream 604 undergoes JPEG2000 decoding 606 to produce a reconstructed extracted 2K image 608. Similarly, a corresponding 2K enhancement layer compressed code-stream 526 undergoes JPEG2000 decoding 610 to produce a reconstructed 2K enhancement layer 612. The reconstructed extracted 2K image 608 and the reconstructed 2K enhancement layer 612 are added 614 and clipped 616 to the appropriate bit-depth, to produce an enhanced 2K image 618. The enhanced 2K image 618 is expected to correspond to the downscaled 2K image 518 if lossless encoding is used to encode 524 the 2K enhancement layer 522.
Figure 9 shows an implementation of an encoder 900 that can be used to implement the method of Figure 3. Similarly, Figure 10 shows an implementation of a decoder 1000 that can be used to implement the method of Figure 4.
Referring to Figure 9, the original 4k image is received by a processor 902 or other input device. The processor 902 is coupled to a memory 903 and is configured to input the original 4k image to a high resolution processing group 920 and a low resolution processing group 930. The high resolution processing group 920 receives the original 4K image and inputs it to a DCI 4K encoder 904. The DCI 4k encoder 904 produces a DCI compliant 4K compressed code stream, which is input to a transmitter 918 for transmission and is also input to an extractor 906 to extract a 2K resolution compressed code stream from the DCI compliant 4K compressed code stream. Once extracted, the 2K resolution compressed code stream is decoded 908 (entropy decoding and dequantization) to produce reconstructed wavelet coefficients for the extracted 2K image. The decoding in operation 908 may also be referred to as a partial decoding because the code stream is only decoded to the point of providing the wavelet coefficients and no inverse wavelet transformation is performed.
Further, the operations of extraction (906) and decoding (908) can be generally referred to collectively as processing or even decoding.
The low resolution processing group 930 receives the original 4K image and downscales (910) the same to produce a downscaled 2K image. Wavelet coefficients for the downscaled 2K image are generated by level shifting and wavelet transformation (912). The difference between reconstructed wavelet coefficients for the extracted 2K image (generated by the high resolution processing group) and the wavelet coefficients for the downscaled 2K image is determined by a subtractor 914 to produce wavelet coefficients for the 2K enhancement layer. These wavelet coefficients are enhancement layer encoded (916) and sent to the transmitter 918 for transmission.
The transmit/send unit 918 may combine the encoded enhancement layer (916) and the DCI compliant 4K compressed code stream (904) into a single transmission or storage unit. For example, as discussed below with respect to Figures 11 and 13, unit 918 may form a DC package that includes that includes the compressed code-stream and at least one of (1) the enhancement-layer compressed code-stream or (2) a pointer to the enhancement-layer compressed code-stream. Accordingly, unit 918 may be referred to as a "combiner".
Analogous to the discussion of Figure 7, it is clear that the enhancement layer is generated by the combined action of both the high resolution processing group 920 and the low resolution processing group 930. Accordingly, the combination of the two groups 920 and 930 is referred to as an enhancement layer generator.
Figure 10 shows a decoder 1000 according to an implementation. The decoder 1000 includes a processor 1002 and memory 1003 where the processor is configured to receive the data from an encoder. The data received from the encoder includes the DCI compliant 4K compressed code-stream and the 2K resolution enhancement layer compressed code-stream. The processor 1002 sends the 4K resolution compressed code-stream to a high resolution processing group 1020 and the 2K resolution enhancement layer compressed code-stream to a low resolution processing group 1030. The extractor 1004 extracts the 2K resolution compressed code-stream from the 4K resolution image data and the decoder (e.g., JPEG2000 decoder) 1006 decodes the extracted 2K resolution compressed code-stream to produce reconstructed wavelet coefficients for the extracted 2K resolution image.
The low resolution processing group 1030 includes an enhancement layer decoder 1008 that is configured to receive the 2K resolution enhancement layer compressed code stream, and to produce reconstructed wavelet coefficients for the 2K resolution enhancement layer. The reconstructed wavelet coefficients for the extracted 2K resolution image (generated by the high resolution processing group) are then added (1010) with the reconstructed wavelet coefficients for the 2K resolution enhancement layer (generated by the low resolution processing group) to produce wavelet coefficients for the enhanced extracted low resolution image. These wavelet coefficients are then synthesized, undergo inverse ICT, and are then level shifted and clipped (1012) to produce the enhanced 2K resolution image. The inverse ICT step is performed only if the DCI compliant 4K compressed code-stream is generated using ICT. As before, these images may be displayed on a display 1014.
Analogous to the discussion of Figure 8, it is clear that the enhanced image is generated by the combined action of both the high resolution processing group 1020 and the low resolution processing group 1030. Accordingly, the combination of the two groups 1020 and 1030 is referred to as an enhanced image generator.
Figures 9 and 10 may be adapted to implement the methods of Figures 5 and 6. One of ordinary skill will readily understand how to make such modifications, particularly in light of the disclosure and discussion of Figures 7 and 8.
Apart from the restriction on the wavelet decomposition levels in the case of the implementation of Figures 3 and 4, most other parameters for encoding the 2K enhancement layer can be chosen freely. The bit-depth for the 2K enhancement layer should be chosen so as to avoid unnecessary clipping. In the DCI framework, the 4K and 2K images are 12-bit. Then in Figure 5, the 2K enhancement layer has to be encoded as a signed 13-bit image. If the subtraction is going to take place in the wavelet domain, as shown in Figure 3, then the analysis of the bit-depth needed may be more complex. Suppose that both the analysis filters (low pass and high pass) are implemented with (1,1) normalization as described by Taubman and Marcellin ("JPEG2000 image compression fundamentals, standards and practice" Kluwer Academic Publishers, ISBN 0-7923-7519-x, 2002). Then, if the bit-depth of the original 4K image is N, (N+ 1) bits are sufficient to represent each subband without any clipping. Assuming, the downscaled 2K image also has a bit-depth of N, its wavelet coefficients can be represented by (N+ J) bits. Thus, the wavelet coefficients for the 2K enhancement layer, which are calculated as a difference, can be represented by (N+2) bits. Thus, for JPEG2000 encoding purposes, it can be assumed that the wavelet coefficients of the enhancement layer have been generated by wavelet decomposition of a (N+ J) bit signed image.
Those skilled in the art will recognize that in Figure 5, the 2k enhancement layer may be compressed using any other compression method instead of JPEG2000 encoding. Similarly, in Figure 3, the wavelet coefficients for the 2K enhancement layer may be compressed using any other compression method instead of JPEG2000 encoding. Furthermore, JPEG2000 encoding techniques can be used for compressing the 2K enhancement layer without producing a JPEG2000 compliant code-stream. This is accomplished, for example, by forcing the 2K enhancement layer encoding engine to use the equivalent parameters as the 4K JPEG2000 encoding engine. In that case, it is not necessary to store a number of marker segments. But the resulting gain in compression efficiency is typically small.
A DC composition is defined as an ordered sequence of reels. Each reel can contain multiple media track files to be reproduced in parallel during the presentation. A reel can have a single main picture, main sound, and main subtitle track files. However, the current standard specification also supports additional track files for future uses. A DC composition can be put into one or more DC packages for distribution to theaters or other venues.
In accordance with another implementation, the low resolution enhancement can be performed by maintaining the enhancement data separate from the DCP Stream. Figures 11 and 12 show an exemplary implementation of this concept from the transmitter (i.e., creation) and receiving (i.e., playback) sides, respectively.
Referring to Figure 11, a DCI compliant 4K compressed code-stream 1102 is used to create DC picture track files 1106 and to create the 2K enhancement layer compressed code stream 1104. Once completed, the 2K enhancement layer compressed code stream 1104 is placed in a separate file 1108. One or more DC packages are created 1110 using the previously created DC picture track files and which include the filename or a pointer to the enhancement layer file. Note that "pointer" is a broad term that includes, for example, an address, a name (e.g. a filename), or other descriptor that indicates where the enhancement layer file is located. In one implementation, the filename, or pointer, is included in a metadata field that the standard defines. The metadata field may be, for example, a free-form text field, effectively allowing an implementation to use the field for a variety of purposes.
At this stage, the one or more DC packages and the enhancement layer file can be sent 1112 (either manually via mail or electronically via communication network of any suitable type). Those skilled in the art will also recognize that the enhancement layer file can be transmitted completely separate from the DC package information. The creation of the DCI compliant 4K compressed code stream and the 2K enhancement layer compressed code-stream may be, for example, as described above. Figure 12 shows the method 1200 for receiving and processing the enhancement file data in accordance with an implementation. As shown, one or more DC packages are received 1202. A determination 1204 is then made as to whether the DC packages include a filename or other pointer to an enhancement layer file. If there is no enhancement layer filename or pointer, the DC picture track is decoded and the low resolution (2K) image is generated in accordance with the DCI spec 1206. If there is an enhancement filename or pointer contained within the DC package information, the enhancement layer file named in the metadata is identified, located, and accessed 1208. Once identified and accessed, the DC picture track file(s) is decoded and the enhanced low resolution (2K) image is generated 1210 using the enhancement layer file data. On the receiving end of such transmissions, for example, movie theaters or other playback locations that are not aware of the enhancement layer data would playback the extracted 2K version directly from the DC picture track file(s). Playback systems that are aware of the enhancement layer data would produce the enhanced 2K version from the DC main picture track file(s) and the enhancement layer picture track file(s). Thus, this implementation is backward compatible with legacy and other existing systems that implement the DCI specification. Such systems will typically ignore a field that is not expected to have data, or if the system does not know what to do with data in such a field.
Figures 13 and 14 show a representative implementation of a DCP Specification compliant implementation of the present principles. The method 1300 of Figure 13 shows that both the high resolution compressed code-stream and the low resolution enhancement layer compressed code-stream are combined such that the low resolution enhancement layer compressed code-stream is included in the created DC packages. That is, the low resolution enhancement layer compressed code-stream is included in an additional picture track within the DC packages. The high resolution compressed code-stream is included in the DC main picture track (1302) and the low resolution enhancement layer compressed code-stream is included into DC additional picture track file(s) (1304). Once included, the DC package(s) are created 1306. The created DC packages can then be sent (i.e., manually in reels or other storage media) or transmitted electronically 1308 to the respective receivers (e.g., movie theaters) of the image data.
Figure 14 shows the receiving end 1400 of the DC package data. The DC package data is received 1402, and a determination is then made (1404) as to whether the DC package data includes the low resolution enhancement layer data as an additional picture track file. If it does, the DC main and additional picture tracks are decoded (1408) and the enhanced low resolution image is generated using the extracted 2K image and the enhancement layer data contained in the DC main and additional picture track file (1408). If there is no enhancement layer data contained as an additional picture track in the DC package(s), the DC package(s) are decoded (1406) and the low resolution (2K) image is generated in accordance with standard DCI Specification requirements. Thus, as in the earlier discussion of Figures 11-12, if a system is not aware of the additional picture track file (or is aware, but does not know what to do with the additional picture track file), then the system typically ignores the additional picture track file.
Note that the above implementations may be combined in various ways. For example, a system may include an additional picture track file (in a DC package) for a first enhancement layer, as well as a pointer to a second enhancement layer. As another example, a system may include in any given DC package only one of an additional picture track file or a pointer, but the system may allow a user to choose either mechanism for each DC package. Such a choice may be made, for example, in real time based on the size of the enhancement layer for the given DC package. Alternatively, a system could make the choice without user-intervention by, for example, consulting a look-up table that indicates the preferred mechanism for a given movie or a given intended-recipient of the DC package.
Implementations have different levels of computational complexity. For example, the implementations of Figures 4 and 6 may be compared. Figure 4 describes a decoder that requires at least one inverse wavelet transformation (unit 418), and Figure 6 describes a decoder that requires at least two inverse wavelet transformations (units 606 and 610).
Features and aspects of described implementations may be applied to various applications. Applications include, for example, a pre-processor or an encoder for creating DC packages or other data packages, or a decoder or other video receiving/processing apparatus, such as described above. Of course, various units may be integrated, such as, for example, a pre-processor and an encoder. The features and aspects herein described may, however, be adapted for other application areas.
The implementations described herein may be implemented in, for example, a method or process, an apparatus, or a software program. Even if only discussed in the context of a single form of implementation (for example, discussed only as a method), the implementation of features discussed may also be implemented in other forms (for example, an apparatus or program). An apparatus may be implemented in, for example, appropriate hardware, software, and firmware. The methods may be implemented in, for example, an apparatus such as, for example, a processor, which refers to processing devices in general, including, for example, a computer, a microprocessor, an integrated circuit, or a programmable logic device.
Implementations of the various processes and features described herein may be . embodied in a variety of different equipment or applications, particularly, for example, equipment or applications associated with video transmission. Examples of equipment include video coders, video decoders, video codecs, web servers, and personal computers. As should be clear from these examples, encodings may be sent or received over a variety of paths, including, for example, wireless or wired paths, the Internet, cable television lines, telephone lines, and Ethernet connections. Additionally, as should be clear, the equipment may be mobile and even installed in a mobile vehicle.
Additionally, the methods may be implemented by instructions being performed by a processor, and such instructions may be stored on a processor readable medium such as, for example, an integrated circuit, a software carrier, or other storage device such as, for example, a hard disk, a compact diskette, a random access memory ("RAM"), or a read-only memory ("ROM"). A processor may also include a processor readable medium having, for example, instructions for carrying out a process.
As should be evident to one of skill in the art, implementations may also produce a signal formatted to carry information that may be, for example, stored or transmitted. The signal may be transmitted as, for example, an electromagnetic wave, and may carry information by, for example, modulating one or more carrier frequencies. The information may include, for example, instructions for performing a method, or data produced by one of the described implementations.
A number of implementations have been described. Nevertheless, it will be understood that various modifications may be made. For example, elements of different implementations may be combined, supplemented, modified, or removed to produce other implementations. Additionally, one of ordinary skill will understand that other structures and processes may be substituted for those disclosed and the resulting implementations will perform at least substantially the same function(s), in at least substantially the same way(s), to achieve at least substantially the same result(s) as the implementations disclosed.
Accordingly, these and other implementations are contemplated by this application and are within the scope of the following claims.

Claims

1. A method comprising: encoding (102, 302) a digital image into an encoded image capable of being processed to provide a higher resolution decoding of the digital image and to provide a lower resolution decoding of the digital image; and encoding (112, 326) additional information for enhancing the lower resolution decoding to provide an enhanced lower resolution image.
2. The method of claim 1, wherein the additional information comprises spatial-domain data related to the enhanced lower resolution image for combining with the lower resolution decoding of the digital image.
3. The method of -claim 1, wherein the additional information comprises frequency-domain data related to the enhanced lower resolution image, and the frequency-domain data is for combining with other frequency-domain data related to the lower resolution decoding to provide the enhanced lower resolution image.
4. The method of claim 3, wherein the frequency-domain data related to the enhanced lower resolution image comprises a wavelet coefficient related to the enhanced lower resolution image.
5. The method of claim 1, wherein said encoding is performed according to DCI Specification.
6. The method of claim 1, further comprising: determining a difference between (a) the lower resolution decoding decoded from a low resolution compressed-code stream extracted from the encoded image and (b) an image downscaled from the digital image; determining the additional information based on the difference; and forming an enhancement layer based on the encoded additional information.
7. The method of claim 1, wherein the encoding of the digital image comprises resolution scalable encoding of the digital image to produce a higher resolution compressed code stream, and the method further comprises: downscaling the digital image to produce a downscaled image; and deriving the additional information using the downscaled image and the lower resolution decoding decoded from a lower resolution compressed code-stream extracted from' the higher resolution compressed code stream.
8. The method of claim 7, wherein said deriving comprises: determining a difference between the downscaled image and the lower resolution decoding; and forming the additional information based on said difference.
9. The method of claim 1, wherein the encoding of the digital image comprises resolution scalable encoding of the digital image to produce a higher resolution compressed code stream, and the method further comprises: downscaling the digital image to produce a downscaled image; deriving wavelet coefficients from the downscaled image; deriving reconstructed wavelet coefficients by decoding a lower resolution compressed code-stream extracted from the encoded image; and determining a difference between the wavelet coefficients and the reconstructed wavelet coefficients to produce the additional information.
10. The method of claim 1, wherein encoding the additional information comprises: downscaling the digital image into a downscaled image; extracting a lower resolution compressed code-stream from the encoded image; decoding the lower resolution compressed code-stream to produce the lower resolution decoding; determining a difference between the lower resolution decoding and the downscaled image; determining the additional information based on the difference; forming an enhancement layer that includes the additional information; and encoding the enhancement layer to produce an enhancement layer code-stream.
11. An apparatus comprising: an encoder (704) to encode a digital image into an encoded image capable of being processed to provide a higher resolution decoding of the digital image and to provide a lower resolution decoding of the digital image; and a second encoder (711) to encode additional information for enhancing the lower resolution decoding to provide an enhanced lower resolution image.
12. The apparatus of claim 11, further comprising: a downscaler (706, 910) configured to downscale the digital image into a downscaled image; an extractor (708, 906) for extracting a low resolution compressed code stream from the encoded image; a decoder (709) for decoding the low resolution compressed code stream and producing a low resolution extracted image; and a combiner (710, 914) for combining the downscaled image with the extracted low resolution image to produce the additional information as an enhancement layer, wherein the second encoder encodes the enhancement layer to produce an enhancement layer compressed code stream.
13. An apparatus comprising: means for encoding a digital image into an encoded image capable of being processed to provide a higher resolution decoding of the digital image and to provide a lower resolution decoding of the digital image; and means for encoding additional information for enhancing the lower resolution decoding to provide an enhanced lower resolution image.
14. A program storage device having an application program tangibly embodied thereon, the application program including instructions for performing at least the following: encoding a digital image into an encoded image capable of being processed to provide a higher resolution decoding of the digital image and to provide a lower resolution decoding of the digital image; and encoding additional information for enhancing the lower resolution decoding to provide an enhanced lower resolution image.
15. A method comprising: accessing (202-206) encoded image data capable of being processed to provide a higher resolution decoding of an image and to provide a lower resolution decoding of the image; and accessing (210) additional encoded information for enhancing the lower resolution decoding of the image to provide an enhanced lower resolution image.
16. The method of claim 15, wherein the encoded image data comprises a high resolution compressed code-stream, and the additional encoded information comprises a low resolution enhancement layer compressed code-stream.
17. The method of claim 15, further comprising receiving a pointer to the additional encoded information in metadata contained in a digital cinema package received with the encoded image data.
18. The method of claim 15, further comprising generating the enhanced lower resolution image by performing at least the following: extracting a low resolution compressed code-stream from the encoded image data; decoding the extracted low resolution compressed code-stream to produce the lower resolution decoding; decoding the additional encoded information to produce a decoded enhancement layer; and combining the lower resolution decoding with the decoded enhancement layer to provide the enhanced lower resolution image.
19. The method of claim 15, further comprising generating the enhanced lower resolution image by performing at least the following: extracting a low resolution compressed code-stream from the encoded image data; decoding the extracted low resolution compressed-code stream to produce reconstructed wavelet coefficients for the lower resolution decoding; decoding the additional encoded information to produce reconstructed wavelet coefficients for enhancing the lower resolution decoding; combining (a) the reconstructed wavelet coefficients for the lower resolution decoding and (b) the reconstructed wavelet coefficients for enhancing the lower resolution decoding, to generate wavelet coefficients for the enhanced lower resolution image; and processing the generated wavelet coefficients to produce the enhanced lower resolution image.
20. An apparatus comprising: a processor (802) configured to receive encoded image data and additional encoded data; a high resolution processing group (820,1020), coupled to the processor, configured to generate low resolution extracted image data from high resolution compressed image data contained within the encoded image data; and a low resolution processing group (830, 1030), coupled to the processor, configured to generate an enhanced low resolution image using (a) the low resolution extracted image data and (b) the additional encoded data.
21. The apparatus of claim 20, wherein said high resolution processing group (820, 1020) further comprises: an extractor (804) configured to receive the high resolution compressed image data and extract a low resolution compressed code-stream; and a decoder (806) configured to generate the low resolution extracted image from the low resolution compressed code-stream.
22. The apparatus of claim 20, wherein said low resolution processing group (830, 1030) further comprises: an enhancement layer decoder (808) configured to generate a decoded low resolution enhancement layer from the additional encoded data; and a combiner (810) configured to combine the decoded low resolution enhancement layer with the low resolution extracted image data to produce the enhanced low resolution image.
23. An apparatus comprising: means for accessing encoded image data capable of being processed to provide a higher resolution decoding of an image and to provide a lower resolution decoding of the image; and means for accessing additional encoded information for enhancing the lower resolution decoding of the image to provide an enhanced lower resolution image.
24. The apparatus of claim 15, further comprising a generating means for generating the enhanced lower resolution image, the generating means comprising: means for extracting a low resolution compressed code-stream from the encoded image data; means for decoding the low resolution compressed code-stream to produce reconstructed wavelet coefficients for the lower resolution decoding; means for decoding the additional encoded information to produce reconstructed wavelet coefficients for enhancing the lower resolution decoding; means for combining (a) the reconstructed wavelet coefficients for the lower resolution decoding and (b) the reconstructed wavelet coefficients for enhancing the lower resolution decoding, to generate wavelet coefficients for the enhanced lower resolution image; and means for processing the generated wavelet coefficients to produce the enhanced lower resolution image.
25. A program storage device having an application program tangibly embodied thereon, the application program including instructions for performing at least the following: accessing encoded image data capable of being processed to provide a higher resolution decoding of an image and to provide a lower resolution decoding of the image; and accessing additional encoded information for enhancing the lower resolution decoding of the image to provide an enhanced lower resolution image.
26. A signal representing information, the signal comprising: a first signal portion representing encoded data for an encoded image, the encoded data capable of being processed to produce a higher resolution decoding of the encoded image and to produce a lower resolution decoding of the encoded image; and a second signal portion representing additional encoded information for enhancing the lower resolution decoding.
27. The signal of claim 26 wherein the second signal portion represents additional encoded information that includes metadata.
28. The signal of claim 26 wherein the second signal portion represents additional encoded information that includes one or more of spatial-domain data or frequency-domain data.
29. The signal of claim 26 wherein: the first signal portion is formatted to represent the encoded data by modulating one or more carrier frequencies, and the second signal portion is formatted to represent the additional encoded information by modulating one or more carrier frequencies.
EP07796523A 2007-06-27 2007-06-27 Enhancing image quality Ceased EP2172025A1 (en)

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/US2007/014961 WO2009002313A1 (en) 2007-06-27 2007-06-27 Enhancing image quality

Publications (1)

Publication Number Publication Date
EP2172025A1 true EP2172025A1 (en) 2010-04-07

Family

ID=39316395

Family Applications (1)

Application Number Title Priority Date Filing Date
EP07796523A Ceased EP2172025A1 (en) 2007-06-27 2007-06-27 Enhancing image quality

Country Status (7)

Country Link
US (1) US8644632B2 (en)
EP (1) EP2172025A1 (en)
JP (1) JP4920110B2 (en)
KR (1) KR101376666B1 (en)
CN (1) CN101682759B (en)
BR (1) BRPI0721640A2 (en)
WO (1) WO2009002313A1 (en)

Families Citing this family (21)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2009051704A1 (en) * 2007-10-16 2009-04-23 Thomson Licensing Methods and apparatus for artifact removal for bit depth scalability
US9699242B2 (en) * 2007-12-07 2017-07-04 Dan Atsmon Multimedia file upload
US8243824B2 (en) * 2008-01-16 2012-08-14 Sony Corporation Information processing apparatus and method
EP2428042B1 (en) * 2009-05-05 2013-05-01 Telefonaktiebolaget LM Ericsson (publ) Scalable video coding method, encoder and computer program
US8589999B1 (en) * 2009-08-11 2013-11-19 Arris Solutions, Inc. Methods and systems for splicing between media streams
EP2315445A1 (en) * 2009-10-09 2011-04-27 Vidast Oy Video compression
US20110317755A1 (en) * 2010-06-24 2011-12-29 Worldplay (Barbados) Inc. Systems and methods for highly efficient compression of video
KR101343527B1 (en) 2010-11-17 2013-12-19 한국전자통신연구원 Method for Producing and playing Digital Cinema Contents and Apparatus for producing and playing digital cinema contents using the method
CA2807545C (en) * 2011-02-22 2018-04-10 Panasonic Corporation Image coding method, image decoding method, image coding apparatus, image decoding apparatus, and image coding and decoding apparatus
KR102030977B1 (en) 2011-02-22 2019-10-10 타지반 투 엘엘씨 Filter method, dynamic image encoding device, dynamic image decoding device, and dynamic image encoding/decoding device
CN102724522B (en) * 2011-03-29 2014-07-09 雷欧尼斯(北京)信息技术有限公司 Method and system for realizing 4D digital film based on DCI standard
EP3843394B1 (en) 2011-07-19 2024-04-03 Tagivan Ii Llc Coding method
CN103718562B (en) * 2011-07-26 2017-06-20 Lg电子株式会社 The device for sending video flowing, the device, the method for transmission video flowing, the method for reception video flowing that receive video flowing
WO2013090120A1 (en) 2011-12-15 2013-06-20 Dolby Laboratories Licensing Corporation Backwards-compatible delivery of digital cinema content with extended dynamic range
JP2014093569A (en) * 2012-10-31 2014-05-19 Toshiba Corp Information processing device, control method, and control program
US20150016502A1 (en) * 2013-07-15 2015-01-15 Qualcomm Incorporated Device and method for scalable coding of video information
US8879858B1 (en) * 2013-10-01 2014-11-04 Gopro, Inc. Multi-channel bit packing engine
CN104090709B (en) * 2014-06-11 2017-11-21 小米科技有限责任公司 Picture switching method and device
US10880566B2 (en) * 2015-08-28 2020-12-29 Boe Technology Group Co., Ltd. Method and device for image encoding and image decoding
CN108681988B (en) * 2018-04-17 2022-03-29 中国科学院光电技术研究所 Robust image resolution enhancement method based on multiple images
US11804166B2 (en) 2019-09-24 2023-10-31 Lg Electronics Inc. Signal processing device and image display apparatus including same

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060133484A1 (en) * 2004-12-06 2006-06-22 Park Seung W Method for decoding image block
WO2006083661A1 (en) * 2005-02-04 2006-08-10 Dts (Bvi) Az Research Limited Digital intermediate (di) processing and distribution with scalable compression in the post-production of motion pictures

Family Cites Families (24)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5621660A (en) * 1995-04-18 1997-04-15 Sun Microsystems, Inc. Software-based encoder for a software-implemented end-to-end scalable video delivery system
US6801665B1 (en) * 1998-09-15 2004-10-05 University Of Maryland Method and apparatus for compressing and decompressing images
US6639943B1 (en) * 1999-11-23 2003-10-28 Koninklijke Philips Electronics N.V. Hybrid temporal-SNR fine granular scalability video coding
US6907073B2 (en) 1999-12-20 2005-06-14 Sarnoff Corporation Tweening-based codec for scaleable encoders and decoders with varying motion computation capability
JP2004509574A (en) * 2000-09-22 2004-03-25 コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ Preferred Transmission / Streaming Order for Fine Granular Scalability
US7463683B2 (en) * 2000-10-11 2008-12-09 Koninklijke Philips Electronics N.V. Method and apparatus for decoding spatially scaled fine granular encoded video signals
US6937772B2 (en) * 2000-12-20 2005-08-30 Eastman Kodak Company Multiresolution based method for removing noise from digital images
JP2003032496A (en) * 2001-07-12 2003-01-31 Sanyo Electric Co Ltd Image coding device and method
AUPS271002A0 (en) 2002-05-31 2002-06-20 Canon Kabushiki Kaisha Embedding a multiresolution compressed thumbnail image in a compressed image file
US7200277B2 (en) * 2003-07-01 2007-04-03 Eastman Kodak Company Method for transcoding a JPEG2000 compressed image
JP4081680B2 (en) * 2003-11-10 2008-04-30 ソニー株式会社 Recording apparatus, recording method, recording medium, reproducing apparatus, reproducing method, and content transmission method
JP4448714B2 (en) * 2004-03-03 2010-04-14 パイオニア株式会社 Digital image signal decoding apparatus and decoding method
US20060015927A1 (en) * 2004-04-09 2006-01-19 Darcy Antonellis Motion picture distribution system and related method
US7400653B2 (en) * 2004-06-18 2008-07-15 Dolby Laboratories Licensing Corporation Maintaining synchronization of streaming audio and video using internet protocol
KR20070037488A (en) * 2004-07-13 2007-04-04 코닌클리케 필립스 일렉트로닉스 엔.브이. Method of spatial and snr picture compression
CN101951509B (en) 2004-11-19 2012-05-23 松下电器产业株式会社 Moving-image encoding method, and moving-image decoding method
WO2006058921A1 (en) 2004-12-03 2006-06-08 Thomson Licensing Method for scalable video coding
JP2006197074A (en) 2005-01-12 2006-07-27 Fuji Xerox Co Ltd Image processor, image processing method, and its program
KR101092539B1 (en) * 2005-02-18 2011-12-14 삼성전자주식회사 Image apparatus for controlling white-balance automatically and method for controlling white-balance thereof
JP2007013706A (en) * 2005-06-30 2007-01-18 Sony Corp Image data generating apparatus, image data generating method, and image data generating program
US20090172028A1 (en) * 2005-07-14 2009-07-02 Ana Belen Benitez Method and Apparatus for Providing an Auxiliary Media In a Digital Cinema Composition Playlist
WO2007032760A1 (en) * 2005-09-09 2007-03-22 Thomson Licensing Coefficient modification for video watermarking
US20070273798A1 (en) * 2006-05-26 2007-11-29 Silverstein Barry D High efficiency digital cinema projection system with increased etendue
US8170334B2 (en) * 2011-10-13 2012-05-01 University Of Dayton Image processing systems employing image compression and accelerated image decompression

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060133484A1 (en) * 2004-12-06 2006-06-22 Park Seung W Method for decoding image block
WO2006083661A1 (en) * 2005-02-04 2006-08-10 Dts (Bvi) Az Research Limited Digital intermediate (di) processing and distribution with scalable compression in the post-production of motion pictures

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
OHM J-R ET AL: "Interframe wavelet coding-motion picture representation for universal scalability", SIGNAL PROCESSING. IMAGE COMMUNICATION, ELSEVIER SCIENCE PUBLISHERS, AMSTERDAM, NL, vol. 19, no. 9, 1 October 2004 (2004-10-01), pages 877 - 908, XP004607152, ISSN: 0923-5965, DOI: 10.1016/J.IMAGE.2004.06.004 *

Also Published As

Publication number Publication date
US8644632B2 (en) 2014-02-04
US20100142836A1 (en) 2010-06-10
WO2009002313A1 (en) 2008-12-31
CN101682759A (en) 2010-03-24
BRPI0721640A2 (en) 2013-01-29
JP4920110B2 (en) 2012-04-18
JP2010532133A (en) 2010-09-30
KR101376666B1 (en) 2014-03-21
KR20100024436A (en) 2010-03-05
CN101682759B (en) 2013-01-09

Similar Documents

Publication Publication Date Title
US8644632B2 (en) Enhancing image quality
Taubman et al. JPEG2000: Standard for interactive imaging
US8896717B2 (en) Methods for deploying video monitoring applications and services across heterogeneous networks
KR100681168B1 (en) System and method for encoding and decoding residual signals for fine granular scalable video
US6510177B1 (en) System and method for layered video coding enhancement
KR101751221B1 (en) Efficient transcoding for backward-compatible vdr codec
JP2003503975A (en) System and method for scalable video coding
CA2543947A1 (en) Method and apparatus for adaptively selecting context model for entropy coding
JP2010263657A (en) Apparatus and method for multiple description encoding
US8243798B2 (en) Methods and apparatus for scalable video bitstreams
KR20070085316A (en) Mobile imaging application, device architecture, and service platform architecture
US20140368672A1 (en) Methods for Deploying Video Monitoring Applications and Services Across Heterogeneous Networks
WO2009002321A1 (en) Enhancing image quality
KR20070085317A (en) Video monitoring application, device architectures, and system architecture
Kobayashi et al. Extension of JPEG XS for two-layer lossless coding
Richardson Video compression codecs: a survival guide
JP4385928B2 (en) Scalable video encoding / decoding method and system
Skodras The JPEG2000 image compression standard in mobile health
JP3617358B2 (en) Video encoding method and apparatus
Furht Image presentation and compression
Smith et al. Simple video format for mobile applications
Demaude et al. Using interframe correlation in a low-latency and lightweight video codec
JP2006246241A (en) Scalable transcoding method, deice, and program
JP2005136873A (en) Resolution scalable decoding method and device

Legal Events

Date Code Title Description
PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

17P Request for examination filed

Effective date: 20100118

AK Designated contracting states

Kind code of ref document: A1

Designated state(s): AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IS IT LI LT LU LV MC MT NL PL PT RO SE SI SK TR

AX Request for extension of the european patent

Extension state: AL BA HR MK RS

DAX Request for extension of the european patent (deleted)
17Q First examination report despatched

Effective date: 20120801

REG Reference to a national code

Ref country code: DE

Ref legal event code: R003

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE APPLICATION HAS BEEN REFUSED

18R Application refused

Effective date: 20150317