US20030235250A1 - Video deblocking - Google Patents

Video deblocking Download PDF

Info

Publication number
US20030235250A1
US20030235250A1 US10/179,825 US17982502A US2003235250A1 US 20030235250 A1 US20030235250 A1 US 20030235250A1 US 17982502 A US17982502 A US 17982502A US 2003235250 A1 US2003235250 A1 US 2003235250A1
Authority
US
United States
Prior art keywords
discontinuities
recited
image
discontinuity
smoothing
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US10/179,825
Inventor
Ankur Varma
Shankar Moni
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Microsoft Technology Licensing LLC
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Priority to US10/179,825 priority Critical patent/US20030235250A1/en
Assigned to MICROSOFT CORPORATION reassignment MICROSOFT CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: MONI, SHANKAR, VARMA, ANKUR
Publication of US20030235250A1 publication Critical patent/US20030235250A1/en
Priority to US11/191,131 priority patent/US7567618B2/en
Priority to US11/191,220 priority patent/US7567619B2/en
Priority to US11/191,204 priority patent/US7660351B2/en
Assigned to MICROSOFT TECHNOLOGY LICENSING, LLC reassignment MICROSOFT TECHNOLOGY LICENSING, LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: MICROSOFT CORPORATION
Abandoned legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/85Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression
    • H04N19/86Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression involving reduction of coding artifacts, e.g. of blockiness

Definitions

  • This disclosure relates in general to video deblocking and in particular, by way of example but not limitation, to post-processing of video information to reduce blockiness in the display thereof.
  • TV-based entertainment systems are expanding the programming and services that they offer.
  • television service providers are adding on-demand video, as well as other interactive services, features, and applications.
  • Such content and additional information are downloaded over a network for display, use, and/or storage on client-side set-top boxes or similar devices.
  • downloads that include audio and/or video information are transmitted in a coded or compressed format. In other words, to reduce the amount of data that is transmitted, the information is typically compressed from a first size to a second smaller size.
  • the data is decompressed/decoded into a semblance of the original audio and video information.
  • the compression-decompression cycle is typically lossy, the decompressed version of the audio and video information differs from the original version of the audio and video information. With video information, for example, such differences can result in the presentation quality of the video information being visibly decreased.
  • Video deblocking can be implemented in video content processing and delivery environments when displaying decompressed/decoded video information. Discontinuities are identified and smoothed to reduce blockiness, for example at macroblock boundaries in conjunction with Moving Picture Expert Group (MPEG)-based schemes.
  • MPEG Moving Picture Expert Group
  • a client device detects discontinuities at or near block boundaries, determines which discontinuities are artificial, and smooths at least those discontinuities that are determined to be artificial.
  • These actions may be accomplished using (i) one or more Laplacian of Gaussian (LoG) operations as applied to both an image and a quantization matrix and (ii) windowed comparisons between values and a threshold resulting from the LoG operations. Alternatively, these actions may be accomplished using a spatio-temporally varying filter.
  • intra image units may be deblocked using the former approach while non-intra image units may be deblocked using the latter approach.
  • FIG. 1 illustrates an exemplary television system architecture in which the systems and methods for video deblocking can be implemented.
  • FIG. 2 illustrates an exemplary client device, a television, and various input devices that interact with the client device.
  • FIG. 3 is a block diagram that illustrates components of the exemplary client devices shown in FIGS. 1 and 2.
  • FIG. 4 is a set of graphs that illustrate an exemplary discontinuity and smoothing thereof.
  • FIG. 5 is a flow diagram that illustrates an exemplary method for video deblocking.
  • FIG. 6 is a flow diagram that illustrates an exemplary process for directly deblocking video.
  • FIG. 7 illustrates an exemplary window for deblocking analysis around a macroblock boundary.
  • FIG. 8 illustrates an exemplary Laplacian of Gaussian (LoG) operation in diagrammatic form.
  • FIG. 9 is a graph that illustrates exemplary values resulting from an LoG operation on a quantization matrix.
  • FIG. 10 illustrates pixels around a macroblock boundary for an exemplary visual adjustment procedure.
  • FIGS. 11A, 11B, and 11 C are exemplary diagrams of a portion of a video information presentation across multiple frames of a group of pictures (GOP).
  • GOP group of pictures
  • FIG. 12 is a graph that illustrates filters across macroblocks and macroblock boundaries in an exemplary approach for indirectly deblocking video using spatial filtering.
  • FIGS. 13A and 13B illustrate an exemplary approach for indirectly deblocking video using temporal filtering across multiple frames of a GOP.
  • FIG. 14 is a flow diagram of an exemplary approach to block-type-dependent filtering under a Moving Pictures Expert Group (MPEG)-compliant compression/coding scheme.
  • MPEG Moving Pictures Expert Group
  • video content processing and delivery systems such as interactive TV networks, cable/satellite networks that utilize electronic program guides and other applications, and Web-enabled TV networks.
  • Client devices in such systems range from full-resource clients with substantial memory and processing resources, such as TV-enabled personal computers and TV recorders equipped with hard-disks, to low-resource clients with limited memory and/or processing resources, such as traditional set-top boxes and personal digital assistants (PDAs) or mobile phones.
  • PDAs personal digital assistants
  • video deblocking as described herein may additionally be used in other environments such as in streaming (e.g., over the Internet), generally in post-processing of compression and decompression cycles, and so forth. While aspects of the described systems and methods can be used in any of these environments and for any types of client devices, they are described primarily in the context of the following exemplary environment.
  • FIG. 1 illustrates an exemplary television entertainment system 100 that is an architecture in which video deblocking may be implemented.
  • System 100 facilitates distribution of content and other information to multiple viewers.
  • System 100 includes one or more content providers 102 , one or more other information providers 104 , a content distribution system 106 , and multiple client devices 108 ( 1 ), 108 ( 2 ), . . . , 108 (N) coupled to content distribution system 106 via a broadcast network 110 .
  • Content provider 102 includes a content server 112 and stored content 114 , such as movies, television programs, commercials, music, and similar audio and/or video content.
  • Content server 112 controls distribution of stored content 114 from content provider 102 to content distribution system 106 . Additionally, content server 112 may control distribution of live content (e.g., content that was not previously stored, such as live feeds) and/or content stored at other locations to content distribution system 106 .
  • live content e.g., content that was not previously stored, such as live feeds
  • Other information provider 104 includes other information database 116 and other information server 118 .
  • Other information database 116 stores information that may be provided to client devices 108 . Such information includes software modules, files, images, text, executable programs, moving video, gaming or other interactive information, and so forth. The information may also include content, especially content of an irregular, one-of-a-kind, or similar nature, or content from smaller independent providers. Part or all of the information from other information database 116 may ultimately be displayed to subscribers after traversing one or more low-bandwidth transmission mediums, regardless of whether such transmission medium(s) are located upstream or downstream of content distribution system 106 .
  • Other information server 118 processes the other information from other information database 116 prior to distribution to generate one or more files that are optimized for, or at least capable of, transmission to content distribution system 106 .
  • Content distribution system 106 includes a broadcast transmitter 128 , one or more content processors 130 , and one or more other information processors 132 .
  • Broadcast transmitter 128 can alternatively be a transceiver if bi-directional communication is desired (e.g., in situations in which broadcast network 110 is a bi-directional network 110 ).
  • Broadcast transmitter 128 transmits (e.g., broadcasts) signals, such as cable/satellite television signals, across broadcast network 110 .
  • Broadcast network 110 can include a cable television network, RF, microwave, satellite, and/or data network, such as the Internet, and may also include wired or wireless media using any transmission format or protocol. Additionally, broadcast network 110 can be any type of network (including a bi-directional network), using any type of network topology and any network communication protocol, and can be represented or otherwise implemented as a combination of two or more networks.
  • Content processor 130 processes the content received from content provider 102 prior to transmitting the content across broadcast network 110 .
  • other information processor 132 processes the other information that is received from other information provider 104 prior to transmission of the other information across broadcast network 110 .
  • a particular content processor 130 may encode, or otherwise process, the received content into a format that is understood by the multiple client devices 108 ( 1 ), 108 ( 2 ), . . . , 108 (N) that are coupled to broadcast network 110 .
  • FIG. 1 shows a single content provider 102 , a single other information provider 104 , and a single content distribution system 106
  • the exemplary system 100 can include any number of content providers and/or other information providers coupled to any number of content distribution systems.
  • content distribution system 106 , content provider 102 , and/or other information provider 104 are individually or jointly representative of a headend service that provides content and other information to multiple subscribers.
  • Client devices 108 can be implemented in a number of ways. For example, a client device 108 ( 1 ) receives content and other information from a satellite-based transmitter via a satellite dish 134 . Client device 108 ( 1 ) is also referred to as a set-top box or a satellite receiving device. Client device 108 ( 1 ) is coupled to a television 136 ( 1 ) for presenting the content and other information (e.g., audio information, video information, and/or data information) that are received by the client device 108 ( 1 ), as well as for presenting a graphical user interface. A particular client device 108 can be coupled to any number of televisions 136 and/or similar devices that can be implemented to display or otherwise render content. Similarly, any number of client devices 108 can be coupled to a single television 136 .
  • a television 136 receives content and other information from a satellite-based transmitter via a satellite dish 134 .
  • Client device 108 ( 1 ) is also referred to as a set-top box
  • Client device 108 ( 2 ) is also coupled to receive content and other information from broadcast network 110 and to provide the received content and other information to associated television 136 ( 2 ).
  • Client device 108 (N) is an example of a combination television 138 and integrated set-top box 140 .
  • Set-top box 140 that is integrated into television 138 can receive signals (e.g., broadcast signals) via a satellite dish (similar to satellite dish 134 ) and/or directly via
  • broadcast network 110 may receive signals via the Internet or any other network, especially those network mediums that are broadcast-capable. As is further described below, client devices 108 may also engage in video deblocking prior to displaying video information (whether content video information or other video information) that is received from a low bandwidth medium, such as a memory storage, other client devices, possibly broadcast network 110 itself, and so forth.
  • a low bandwidth medium such as a memory storage
  • the exemplary system 100 also includes information from other networks/network providers 142 , which may provide information such as information streamed over the Internet, information received directly from a provider of the information, and so forth.
  • Information from other networks/network providers 142 may be accessible over broadcast network 110 (i.e., a network that also provides content information and other information from content distribution system 106 ).
  • broadcast network 110 i.e., a network that also provides content information and other information from content distribution system 106
  • information from other networks/network providers 142 may be accessible over a different network, including a wide area network (WAN), the Internet, a public or private telecommunications network, and so forth.
  • WAN wide area network
  • the Internet a public or private telecommunications network
  • FIG. 2 illustrates an exemplary implementation 200 of a client device 108 shown as a standalone unit that connects to a television 136 and communicates with various input devices 204 , 206 , and 208 .
  • Client device 108 can be implemented in any number of embodiments, including as a set-top box, a satellite receiver, a TV recorder with a hard disk, a digital video record (DVR) and playback system, a game console, an information appliance, and so forth.
  • DVR digital video record
  • Client device 108 includes a wireless port 202 , such as an infrared (IR) or Bluetooth wireless port, for receiving wireless communications from a remote control device 204 , a handheld input device 206 , or any other wireless device, such as a wireless keyboard.
  • Handheld input device 206 can be a personal digital assistant (PDA), handheld computer, wireless phone, or the like.
  • PDA personal digital assistant
  • a wired keyboard 208 can be coupled to communicate with client device 108 .
  • remote control device 204 , handheld device 206 , and/or keyboard 208 may use an RF communication link or other mode of transmission to communicate with client device 108 .
  • Client device 108 receives one or more (e.g., broadcast) signals 210 from one or more broadcast sources, such as from a satellite or a cable or a broadcast network, including a broadcast implementation of network 110 (of FIG. 1).
  • Client device 108 includes hardware and/or software for receiving and decoding a broadcast signal 210 , such as an NTSC, PAL, SECAM or other TV system video signal.
  • Client device 108 also includes hardware and/or software for providing the user with a graphical user interface by which the user can, for example, access various network services, configure client device 108 , and perform other functions, including requesting video delivery.
  • Client device 108 can communicate with other devices via one or more connections including a conventional telephone line 212 , an ISDN link 214 , a cable link 216 , an Ethernet link 218 , a DSL link 220 , and the like. Client device 108 may use any one or more of the various communication links 212 - 220 at a particular instant to communicate with any number of other devices. For example, in addition to receiving information over any one of communication links 212 - 220 , client device 108 may provide (e.g., transmit) information over communication links 212 - 220 or any other type of communication link.
  • Such other communication links may include links capable of interfacing with a local network such as a local area network (LAN), a Bluetooth® network, an IEEE 802.11b-compliant network, or other wired or wireless network type.
  • a client device 108 that receives information from broadcast network 110 may forward the information over the local network to one or more other client devices 108 .
  • the forwarding may be effectuated, for example, over a lower-bandwidth transmission medium that introduces blockiness.
  • Client device 108 generates video signal(s) 222 and audio signal(s) 224 , both of which are communicated to television 136 .
  • Video signals 222 and audio signals 224 can be communicated from client device 108 to television 136 via an RF (radio frequency) link, S-video link, composite video link, component video link, co-axial cable link, or other communication link.
  • the video signals 222 may include deblocked video signals, for example.
  • client device 108 may include one or more lights or other indicators identifying the current status of the device. Additionally, the client device may include one or more control buttons, switches, or other selectable controls for controlling operation of the device.
  • FIG. 3 illustrates selected components of exemplary client device 108 shown in FIGS. 1 and 2.
  • Client device 108 includes a first tuner 300 and an optional second tuner 302 .
  • the tuners 300 and 302 are representative of one or more in-band tuners that tune to various frequencies or channels to receive television signals, as well as at least one out-of-bound (OOB) tuner that tunes to the broadcast channel(s) over which data information is broadcast (e.g., carouseled or otherwise transmitted) to client device 108 .
  • OOB out-of-bound
  • Client device 108 also includes one or more processors 304 which process various instructions to control the operation of client device 108 and to communicate with other electronic and computing devices.
  • Client device 108 can be implemented with one or more memory components, examples of which include a random access memory (RAM) 306 , a disk drive 308 , another mass storage component 310 , and a non-volatile memory 312 (e.g., ROM, Flash, EPROM, EEPROM, etc.).
  • the memory components e.g., RAM 306 , disk drive 308 , mass storage 310 , and non-volatile memory 312 ) store various instructions and/or information such as received content, programs, configuration information for client device 108 , graphical user interface information, and/or video data or information.
  • client device 108 can include a range of processing and memory capabilities, and may include more or fewer types of memory components than those illustrated in FIG. 3.
  • full-resource clients can be implemented with substantial memory and processing resources, including the disk drive 308 to store content for replay by the viewer.
  • Low-resource clients may have limited processing and memory capabilities, such as a limited amount of RAM 306 , no disk drive 308 , and limited processing capabilities of a processor 304 .
  • An operating system 314 and one or more programs as represented by general instructions 316 may be stored in non-volatile memory 312 (and/or other memory component(s)) and executed on processor 304 to provide a runtime environment.
  • a runtime environment facilitates extensibility of client device 108 by allowing various interfaces to be defined that, in turn, allow the programs to interact with client device 108 .
  • these programs may be installed when client device 108 is manufactured, they may also be received via broadcast network 110 from content distribution system 106 (of FIG. 1).
  • Also stored in non-volatile memory 312 (and/or other memory component(s)) are electronically-executable instructions for video deblocking 318 , as described further herein.
  • Client device 108 also includes a decoder 320 to decode a broadcast video signal, such as an NTSC, PAL, SECAM or other TV system video signal.
  • Processor 304 along with tuner(s) 300 and 302 and/or decoder 320 , also enables client device 108 to reconstruct audio and video from an MPEG-2 stream or other digital packet signal, whether compressed or uncompressed.
  • Client device 108 can also include other components pertaining to a television entertainment system which are not illustrated in this example. For instance, client device 108 can include a user interface application and user interface lights, buttons, controls, and the like to facilitate viewer interaction with the device.
  • Client device 108 further includes a wireless interface 322 , a network interface 324 , a serial and/or parallel interface 326 , and a modem 328 .
  • Wireless interface 322 allows client device 108 to receive input commands and other information from a user-operated input device, such as from a remote control device or from another IR, Bluetooth, or similar RF input device.
  • Network interface 324 and serial and/or parallel interface 326 allows client device 108 to interact and communicate with other electronic and computing devices via various communication links, including local network communication links to other client devices 108 .
  • client device 108 may also include other types of data communication interfaces to communicate with other devices.
  • Modem 328 facilitates communication by client device 108 with other electronic and computing devices via a conventional telephone line.
  • Client device 108 also includes an audio output 330 and a video output 332 that provide signals to a television or other device that processes and/or displays or otherwise renders the audio and video information, including deblocked video information.
  • client device 108 may be implemented together in an application specific integrated circuit (ASIC).
  • ASIC application specific integrated circuit
  • electronically-executable instructions ( 318 ) and associated processing abilities for video deblocking may be integrated together onto one or more ASICs instead of using more-general memory (e.g., 306 , 312 , etc.) and/or processing (e.g., 304 , 320 , etc.) resources of a client device ( 108 ).
  • a system bus typically connects the various components within client device 108 .
  • a system bus can be implemented as one or more of any of several types of bus structures, including a memory bus or memory controller, a peripheral bus, an accelerated graphics port, or a local bus using any of a variety of bus architectures.
  • bus architectures can include an Industry Standard Architecture (ISA) bus, a Micro Channel Architecture (MCA) bus, an Enhanced ISA (EISA) bus, a Video Electronics Standards Association (VESA) local bus, and a Peripheral Component Interconnects (PCI) bus also known as a Mezzanine bus.
  • ISA Industry Standard Architecture
  • MCA Micro Channel Architecture
  • EISA Enhanced ISA
  • VESA Video Electronics Standards Association
  • PCI Peripheral Component Interconnects
  • Video information is often compressed or otherwise coded to facilitate the handling thereof.
  • the coding of the video information into video data can reduce the bandwidth required to display, store, transmit, or otherwise manipulate the video information. For example, it can be beneficial to reduce the bit rate of video information for storing a video stream to disk or streaming it over a noisy, bandwidth-limited channel.
  • compression and decompression of the video information typically causes visual artifacts that degrade the visual quality.
  • One artifact that degrades visual quality is blockiness. In other words, low bit rate video information is often plagued with blockiness that seriously degrades visual quality.
  • Video deblocking usually entails reducing such blocky artifacts to enhance the visual presentation quality of the video information. Implementing a good deblocking mechanism helps to mitigate the deleterious byproducts of a lower bit rate video stream and thereby justifies using a lower bit rate that enables the aforementioned low-bandwidth applications.
  • Deblocking is implemented after previously encoded video data is decoded into at least a semblance of the original video information.
  • One set of coding standards that may be used on the original video information are Moving Picture Expert Group (MPEG)-compliant standards.
  • MPEG-based encoding/decoding algorithms involve performing a discrete cosine transform (DCT) on original video information and quantizing the DCT coefficients thereof.
  • the resulting video data is de-quantized and an inverse DCT is performed to reacquire at least a semblance of the original video information.
  • Reducing the bit rate of the video data stream may be achieved by more coarsely quantizing the DCT coefficients. In doing so, however, blockiness that is introduced along macroblock boundaries tends to seriously degrade the visual quality.
  • the visual presentation quality may be at least partially restored by using deblocking during post-processing. Specifically, the deblocking may be used to reduce the blockiness around macroblock boundaries.
  • FIG. 4 is a set of graphs 400 that illustrate an exemplary discontinuity and smoothing thereof.
  • Graphs 400 include graphs 400 A, 400 B, 400 C, and 400 D.
  • Each of graphs 400 A- 400 D includes two macroblocks of a video picture. Although the macroblocks are shown as being rectangular, actual macroblocks may instead be square.
  • Graph 400 A illustrates original video information 402 (as a solid line). There is no break in original video information 402 between the left macroblock and the right macroblock.
  • Graph 400 B illustrates original video information 402 and adds thereto an illustration of discontinuity 404 (as a large dashed line).
  • Discontinuity 404 represents an exemplary discontinuity or break in the seamlessness of original video information 402 . This discontinuity arises at the macroblock boundary of the left and right macroblocks as a result of the encoding/decoding process. Discontinuity 404 may be visible as a blocky artifact. In fact, discontinuity 404 along with other discontinuities in the overall picture may be visible as a blockiness that reduces the overall visual quality.
  • Graph 400 C illustrates original video information 402 and adds thereto an illustration of smoothed video information 406 (as a small dashed line). Smoothed video information 406 results from application of video deblocking to discontinuity 404 . While smoothed video information 406 is not likely to be an exact replica of original video information 402 , smoothed video information 406 reduces the blocky artifact of discontinuity 404 and generally more-closely approximates original video information 402 . It should be noted that smoothed video information 406 is representative of deblocking mechanisms/approaches in general, including visual adjustment, filter smoothing, spatial-temporal filtering, and so forth.
  • Graph 400 D includes original video information 402 , discontinuity 404 , and smoothed video information 406 . All three are included together to illustrate (i) the discontinuity that may be formed from coarse quantization of original video information or other deleterious coding/decoding factors and (ii) the resulting video information after the discontinuity has been smoothed in post-processing. This post-processing video deblocking is described further herein.
  • deblocking in an MPEG-based environment in which discontinuities appear at macroblock boundaries.
  • deblocking may be employed after decompression/decoding in accordance with other standards and approaches and along other boundaries. It should therefore be understood that deblocking may be employed with video data/information in general and regardless of the size/type of blocks that are employed to segregate the video picture.
  • discontinuity is used herein, other such terms include “edge” and “boundary”.
  • An edge, boundary, or discontinuity in a video picture may be “artificial” or “real”.
  • Artificial discontinuities arise, for example, from coarse quantization of DCT coefficients.
  • Other terms for artificial discontinuities include, but are not limited to, quantization discontinuities, false discontinuities, fake discontinuities, and quantization noise discontinuities.
  • Real discontinuities arise from true boundaries in a video picture from one element to another element within the picture.
  • Other terms for real discontinuities include, but are not limited to, actual discontinuities, true discontinuities, and natural discontinuities. These real discontinuities may also lie along a macroblock boundary.
  • Video deblocking may be described in the general context of electronically-executable instructions.
  • electronically-executable instructions include routines, programs, objects, components, data structures, and the like that perform particular functions or implement particular abstract data types.
  • Video deblocking may also be practiced in distributed computing environments where functions are performed by remote processing devices that are linked through a communications network.
  • electronically-executable instructions may be located in both local and remote storage media.
  • FIGS. 5, 6, and 14 are illustrated in flow diagrams divided into multiple blocks. However, the order in which the methods and processes are described is not intended to be construed as a limitation, and any number of the described blocks can be combined in any order to implement one or more methods or processes for video deblocking. Furthermore, although the methods and processes are described below with reference to the television entertainment environments 100 and 200 and client devices 108 where applicable, the methods and processes can be implemented in any suitable hardware, software, firmware, or combination thereof and using any suitable mathematical alternatives.
  • FIG. 5 is a flow diagram 500 that illustrates an exemplary method for video deblocking.
  • Flow diagram 500 includes blocks 502 , 504 , and 506 that may be implemented, for instance, by client devices 108 (of FIGS. 1 - 3 ).
  • Blocks 502 , 504 , and 506 may correspond to electronically-executable instructions that are stored in one or more memories of a client device 108 and executed using one or more processors thereof.
  • a discontinuity is detected. For example, macroblock boundaries are analyzed to detect whether a discontinuity exists for video information between a first macroblock boundary and a second macroblock boundary. An exemplary window around a macroblock boundary for such an analysis is described further below with reference to FIG. 7.
  • the discontinuity is examined to determine whether it is a real discontinuity or an artificial discontinuity.
  • Real edges of actual picture elements may fall on a macroblock boundary, but smoothing real edges usually decreases video presentation quality. Exemplary mechanisms for differentiating between real discontinuities and artificial discontinuities are described further below especially with reference to FIGS. 8 and 9.
  • the artificial discontinuity or discontinuities are smoothed.
  • the smoothing of the artificial discontinuities reduces visually-apparent blockiness. It should be understood that smoothing may encompass a variety of mechanisms/approaches designed to reduce or eliminate the visibility of discontinuities. Exemplary mechanisms for smoothing artificial discontinuities are described further below especially with reference to FIGS. 10 et seq.
  • Flow diagram 500 is directed to both direct deblocking and indirect deblocking.
  • Direct deblocking of video is described especially with reference to FIGS. 6 - 10 .
  • Indirect deblocking of video is described especially with reference to FIGS. 11 A- 13 B.
  • the description of FIG. 14 below is directed to a video deblocking approach that utilizes both direct and indirect deblocking mechanisms.
  • FIG. 6 is a flow diagram 600 that illustrates an exemplary process for directly deblocking video.
  • Flow diagram 600 includes six (6) blocks 602 - 612 .
  • Blocks 602 and 604 pertain to input data for the process of flow diagram 600 .
  • Block 602 provides the scaled quantization matrix that was used to quantize the original video information during the coding/compressing process.
  • This scaled quantization matrix of block 602 may be related to, for example, the “quant_matrix” parameter of MPEG-compliant coding schemes.
  • the quant_matrix is determined by the encoder and encoded into the bit stream.
  • a scaled quant matrix may be created as the product of the quant matrix and the quantization scale, which is another parameter that is encoded into the bit stream by the encoder.
  • Block 604 provides the image that is formed from the decoded video information.
  • the image may include artificial discontinuities (e.g., at macroblock boundaries) and real discontinuities (e.g., either at or away from macroblock boundaries).
  • the artificial discontinuities at the macroblock boundaries can arise from noise attributable to the quantization matrix during the quantization process.
  • Block 606 performs a Laplacian of Gaussian (LoG) operation on each of the scaled quantization matrix of block 602 and the image video information of block 604 to produce two different results.
  • the LoG operation may be defined as convolving the argument with the LoG kernel. From the two different results of applying the LoG operation, the type of discontinuity is determined at block 608 .
  • the type of discontinuity (if any) at the macroblock boundary under consideration is determined by comparing the result of the LoG operation as performed on the image of block 604 with a (quantization) threshold determined responsive to the result of the LoG operation as performed on the scaled quantization matrix of block 602 .
  • the LoG operation of block 606 and the discontinuity type determination of block 608 are described further below especially with reference to FIGS. 8 and 9.
  • block 610 may optionally perform a visual adjustment procedure at the macroblock boundary.
  • An exemplary visual adjustment procedure is described further below especially with reference to FIG. 10.
  • block 612 performs a filter smoothing procedure at the macroblock boundary.
  • An exemplary filter smoothing procedure is described further below primarily after the description of FIG. 10.
  • FIG. 7 illustrates an exemplary window 702 for deblocking analysis around a macroblock boundary 706 .
  • Two macroblocks 708 L and 708 R are illustrated. Each macroblock 708 is surrounded by four (4) macroblock boundaries 706 . Within each macroblock 708 are a set of pixels. Although macroblocks 708 are shown as being blocks of eight-by-eight (8 ⁇ 8) pixels, they may alternatively be blocks of 16 ⁇ 16 pixels, 24 ⁇ 24 pixels, 30 ⁇ 30 pixels, or any other size. Because the illustrated macroblocks are of an eight-by-eight (8 ⁇ 8) size, window 702 includes sixteen (16) pixels to be analyzed for possible discontinuities at the macroblock boundary 706 that window 702 surrounds. However, the width of window 702 may alternatively be greater than two pixels (i.e., the width can be more than one pixel within each macroblock 708 ).
  • Window 702 is located between the left macroblock 708 L and the right macroblock 708 R. Similar “vertical” windows 702 are used at the other “vertical” macroblock boundaries 706 . Furthermore, “horizontal” windows (not shown) are used at macroblock boundaries between upper macroblocks and adjacent lower macroblocks. In other words, to address a two-dimensional (2-D) image, video deblocking is performed in both the vertical direction and the horizontal direction.
  • the following algorithm may be employed: Let Q represent the quantization matrix used in quantizing an image I. Because quantization is usually performed in the DCT domain, the effect of using Q in the time domain is first determined. Thus, an inverse DCT is performed on Q to generate Q t . Next, the discontinuities (e.g., edges, boundaries, etc.) are detected in the image I. Any such discontinuities may be detected using QO the Laplacian of Gaussian (LoG) operator to filter I and generate I LG .
  • LiG Laplacian of Gaussian
  • G ⁇ ( x , y ) 1 2 ⁇ ⁇ ⁇ ⁇ 1 2 ⁇ ⁇ 2 ⁇ ( x 2 + y 2 )
  • FIG. 8 illustrates an exemplary Laplacian of Gaussian (LoG) operation 800 in diagrammatic form.
  • An image 802 is filtered using LoG operation 804 to produce an edge-biased image 806 .
  • discontinuities e.g., edges
  • each discontinuity in each window 702 is a real discontinuity (e.g., from an edge of an image element) or an artificial discontinuity (e.g., from quantization or other deleterious compression/decompression factors).
  • the LoG operation detects both kinds of discontinuities, but it does not clearly differentiate between them by itself.
  • quant_threshold A variable that may be termed “quant_threshold” is created to determine which discontinuities are artificial.
  • the amount of blockiness attributable to quantization noise is effectively assessed.
  • the time-domain quantization matrix divided by two Q t /2) is passed through the LoG filter to obtain Q tLG .
  • the quantization matrix Q t is divided by 2 because half the quantization matrix Q t represents the actual loss of data due to round-off/truncation.
  • the maximum and minimum values of Q tLG are examined to determine the quant_threshold.
  • FIG. 9 is a graph 900 that illustrates exemplary values Q tLG resulting from a LoG operation on a time-domain quantization matrix Q t .
  • a plot 902 graphs different quantization matrix values (along the abscissa axis) versus corresponding LoG operation output values Q tLG (along the ordinate axis).
  • Maximum absolute values for the Q tLG values are denoted as one hundred percent (100%). Any portion of the 100% value may be selected as the threshold between real discontinuities and artificial discontinuities.
  • the threshold may be selected, for example, such that the majority of the visually displeasing artificial discontinuities are subject to smoothing while few if any of the real discontinuities are subject to smoothing (therein mitigating “fuzzying” of the original image).
  • An optimum value for the threshold for any given type of video information may be selected through experimental viewing of deblocked video. In the illustrated graph 900 , eighty-five percent (85%) is selected as the threshold between real discontinuities and artificial discontinuities. Furthermore, this threshold may be tunable subject to user/viewer preferences.
  • the threshold is used in conjunction with an analysis of macroblock boundaries 706 in windows 702 .
  • the analysis is focused in windows 702 because blockiness that is due to quantization noise is centered on macroblock boundaries 706 .
  • the values of I LG are compared to the threshold within each window 702 . If more than a certain fraction, ⁇ , of the I LG values in a window 702 are above the selected threshold, then that discontinuity is labeled as a real discontinuity. If not, then that discontinuity is labeled as an artificial discontinuity.
  • the fraction, ⁇ may also be selected experimentally, for example by viewing different types of video that has been deblocked using different values for the fraction, ⁇ , until a visually-appealing value is selected.
  • the parameter, ⁇ may be experimentally determined and fine tuned to maximize visual quality.
  • discontinuities After discontinuities have been (i) detected and (ii) determined to be (and “labeled” as) either real discontinuities or artificial discontinuities, appropriate deblocking procedures can be implemented. As described above with reference to blocks 608 , 610 , and 612 (of FIG. 6), real discontinuities may be subjected to a visual adjustment procedure, and artificial discontinuities may be subjected to a filter smoothing procedure. In an exemplary visual adjustment procedure, pixel values are adjusted slightly by moving them towards each other by an amount proportional to the threshold so as to reduce the effects of quantization.
  • FIG. 10 illustrates pixels 704 around a macroblock boundary 706 for an exemplary visual adjustment procedure.
  • Eight (8) adjacent pixels 704 from p 0 to p 7 are divided by the macroblock boundary 706 that is between left macroblock 708 L and right macroblock 708 R.
  • Pixels p 0 , p 1 , p 2 , and p 3 are in left macroblock 708 L.
  • Pixels p 4 , p 5 , p 6 , and p 7 are in right macroblock 708 R.
  • the values of the pixels p 0 to p 7 are adjusted slightly by moving them towards each other by an amount proportional to the selected threshold so as to reduce the effects of quantization. This adjustment is performed recursively to the adjacent pixels.
  • time-domain pixels are designated by “p” (p without the prime) and the adjusted pixels are represented by “p′” (p with the prime).
  • a filter smoothing procedure is performed.
  • pixels 704 (of FIG. 7) of a window 702 are filtered around boundary 706 with a smoothing filter.
  • a smoothing filter such as stretched and/or scaled version(s) of the Gaussian or tent filter.
  • any smoothing filter can be used with the degree of smoothing set to accommodate the desired amount of blurring.
  • a less rigid and more indirect mechanism is used to detect discontinuities, to differentiate between real and artificial discontinuities, and to smooth the detected discontinuities.
  • This second mechanism includes both spatial and temporal aspects.
  • the mechanism entails associating an increased likelihood of needing deblocking (e.g., of a discontinuity being an artificial discontinuity) with increased proximity to a macroblock boundary.
  • the mechanism entails associating the likelihood of blockiness at a macroblock boundary to the position of the picture within a GOP. For example, it is assumed that an I-frame of the GOP is more susceptible to blockiness at the macroblock boundary. On the other hand, a B-frame downstream in the GOP is less likely to experience blockiness at the macroblock boundary.
  • FIGS. 11A, 11B, and 11 C are exemplary diagrams 1100 of a portion of video information that is presented across multiple frames of a GOP.
  • frames are organized into GOPs that start with an intra (I) frame and are followed by predicted (P) frames and bidirectional (B) frames/modes.
  • I-frames are usually more susceptible to blockiness at macroblock boundaries than those frames that are downstream in the GOP as is explained by the description of the diagrams 1100 A-C.
  • diagram 1100 A represents an intra-frame with many macroblocks 708 and macroblock boundaries 706 .
  • Diagrams 1100 B and 1100 C represent non-intra frames. Each non-intra block in non-intra frames includes two components: the predicted term and the difference term. Typically, the predicted term dominates, and the difference term is relatively small. Because the predicted term is usually not macroblock aligned in the reference frame (as shown in diagram 1100 B), any macroblocking effects in the reference frame are seen in the interior of the reconstructed macroblock in the current frame. Moreover, as this effect occurs over several generations of prediction, the magnitude and location of blockiness tends to be diffused (as shown in diagram 1100 C). In other words, downstream in a GOP, images tend to be both less crisp as well as less blocky.
  • the position of a frame in a GOP tends to serve as an indicator of the extent of blockiness. This tendency is reflected in the temporal aspect of the second mechanism.
  • the extent of quantization-caused blockiness tends to increase as pixel location approaches a macroblock boundary.
  • the second mechanism thus involves filtering the image with a spatio-temporally varying filter.
  • the degree of filtering increases for pixels near macroblock boundaries and decreases for pixels away from macroblock boundaries.
  • This spatial filtering is described further below with reference to FIG. 12.
  • the degree of filtering is higher at the beginning of a GOP while the degree of filtering decreases downstream in the GOP.
  • This temporal filtering is described further below with reference to FIGS. 13A and 13B.
  • Combining the spatial filtering with the temporal filtering creates a spatio-temporally varying filter that indirectly detects discontinuities, differentiates between discontinuity types, and smooths the detected discontinuities.
  • FIG. 12 is a graph 1200 that illustrates filters 1202 across macroblocks 708 and macroblock boundaries 706 in an exemplary approach for indirectly deblocking video using spatial filtering.
  • Filters 1202 include three (3) different filtering zones labeled zone # 1 , zone # 2 , and zone # 3 . While each of the three different filtering zones may have filters of similar or identical energy levels (to avoid changing image intensity), each of the filters of the three different filtering zones filters pixels of the macroblocks 708 differently.
  • Filtering zone # 1 filters to the greatest degree in an area around macroblock boundaries 706 .
  • Filtering zone # 3 filters the least in an area around the center of macroblocks 708 .
  • Filters of filtering zones # 3 may be implemented as an impulse function so that the pixels are changed very slightly or not at all.
  • the centers of macroblocks 708 are the least likely to exhibit blockiness due to quantization noise.
  • Located spatially between the filtering zones # 1 and # 3 are filtering zones # 2 .
  • the filters of filtering zones # 2 filter the corresponding physically-intermediate pixels to a degree that is also intermediate to that of the filters of filtering zones # 1 and # 3 .
  • the pixels may be averaged, for example, to a greater extent near macroblock boundaries and to gradually lesser extents for pixels that are gradually nearer the middle of macroblocks.
  • filter 1202 includes an exemplary three filtering zones, four or more or less than three filtering zones may alternatively be employed.
  • the filters can be normalized in order to achieve a gain of 1 (or any desired value).
  • Such spatial filtering thus concentrates filtering near macroblock boundaries and away from the center of macroblocks.
  • Using a spatial filter designed accordingly effectively tends to indirectly filter blockiness due to quantization noise.
  • FIGS. 13A and 13B illustrate an exemplary approach for indirectly deblocking video using temporal filtering across multiple frames of a GOP.
  • the extent of spatial filtering that is necessary and/or desirable around macroblock boundaries tends to decrease as frame position number proceeds through a GOP, as is explained above with reference to FIGS. 11 A- 11 C.
  • Chart 1300 A illustrates a GOP and exemplary temporal filtering coefficients designated as ⁇ .
  • the GOP position number (GPN) runs from 1 to n, where n may be any natural number, but is often set to fifteen (15).
  • the MPEG frames are designated by I, B, or P in an exemplary sequence that runs: I, B, B, B, P, B, B, B, P, B, B, B, P, B, B, B, P, B, B, B.
  • Other sequences may alternatively be used.
  • Three different sets of temporal filtering coefficients ⁇ are included and are designated as ⁇ 1 , ⁇ 2 , and ⁇ 3 .
  • the temporal filtering coefficient a may be applied to the spatial filtering as described above with reference to FIG. 12.
  • the temporal filtering coefficient ⁇ serves to reduce the filtering by, and therefore the impact of, the spatial filter 1202 as frames progress downward through the GOP. Consequently, each ⁇ may be determined responsive to the GPN of the frame under consideration.
  • Table 1300 B provides a general GPN-dependent formula for determining an ⁇ for each of the three temporal filtering coefficient ⁇ examples. Each of the ⁇ examples decrease in value as the GPN increases to reduce the impact of the spatial filter.
  • the rapidity at which each ⁇ decreases in value as the GPN increases from 1 to n scales from a low with ⁇ 1 to a high with ⁇ 3 .
  • other formulas for determining ⁇ may alternatively be employed. Such other formulas, or even other sets of ⁇ values that are not necessarily formulaic, need not use an actual GPN numeral to set the ⁇ .
  • the number K is chosen to normalize the energy in the filter. Also, if the filtering ability of the spatial filter 1202 is inversely proportional to ⁇ , then the ⁇ formula may be set up so as to increase with increasing GPN.
  • ⁇ 1 is the set of numbers ⁇ 1, ⁇ fraction (14/15) ⁇ , ⁇ fraction (13/15) ⁇ , ⁇ fraction (12/15) ⁇ , ⁇ fraction (11/15) ⁇ , . . . , ⁇ fraction (2/15) ⁇ , ⁇ fraction (1/15) ⁇ ;
  • ⁇ 2 is the set of numbers ⁇ 11 ⁇ 2, 1 ⁇ 3, 1 ⁇ 4, 1 ⁇ 5, . . .
  • ⁇ fraction (1/14) ⁇ ⁇ fraction (1/15) ⁇
  • ⁇ 3 is the set of numbers ⁇ 11 ⁇ 4, ⁇ fraction (1/9) ⁇ , ⁇ fraction (1/16) ⁇ , ⁇ fraction (1/25) ⁇ , . . . , ⁇ fraction (1/196) ⁇ , ⁇ fraction (1/225) ⁇ .
  • An ⁇ formula or set of values that results in visually-pleasing video may be determined experimentally for any given type of video information.
  • Use of the temporal filtering coefficient ⁇ as a modifier to a spatial filter 1202 enables application of a spatio-temporally varying filter that indirectly detects, differentiates between types of, and smoothes discontinuities.
  • FIG. 14 is a flow diagram 1400 of an exemplary approach to block-type-dependent filtering under an MPEG-compliant compression/coding scheme.
  • the second mechanism as described above with particular reference to FIGS. 11 A- 13 B, may be applied to all frames of a GOP.
  • the approach of flow diagram 1400 may be employed in which the video deblocking mechanism that is being applied depends on the frame/block type.
  • Flow diagram 1400 is described in terms of blocks such as macroblocks, but it may instead be implemented on a frame-only basis. More generally, blocks, frames, macroblocks, etc. may be considered units of an image.
  • each MPEG GOP starts with an I-frame.
  • This I-frame has relatively hard and certain macroblock boundaries.
  • Non-I-frames e.g., B-frames and P-frames
  • the first mechanism (which is more direct) may be applied to I-frames while the second mechanism (which is more indirect) may be applied to non-I-frames.
  • the deblocking under such a scheme is frame-type-dependent.
  • individual macroblocks in a non-I-frame may be equivalent to an I-frame.
  • a determination as to which mechanism is to be used under an MPEG-compliant coding scheme may be block-type-dependent.
  • a flow diagram block 1402 determines whether a block under consideration is (i) an intra block or (ii) a predicted or a bi-directional block. For example, prior to actual smoothing, each macroblock of a frame of video information may be analyzed to determine whether it is an I-block or a P-block/B-block. If it is determined to be an I-block, then a visual adjustment procedure or a filter smoothing procedure is implemented at flow diagram block 1404 . Whether a visual adjustment procedure or a filter smoothing procedure is implemented is dependent on discontinuity type. As described further above with reference especially to FIGS. 6 - 10 , the visual adjustment procedure is applied to real discontinuities and the filter smoothing procedure is applied to artificial discontinuities.
  • the block under consideration is determined to be a P-block or a B-block (e.g., a non-I-block)
  • a spatio-temporally varying filtering procedure is implemented for the block under consideration at flow diagram block 1406 .
  • the extent of filtering for any given pixel or set of pixels is dependent on the spatial position of the pixel or pixels within the block.
  • the extent of filtering for the given pixel or set of pixels is also dependent on the temporal-based GPN of the frame of which the block forms a part.
  • Such a spafio-temporally varying filtering procedure is described further above with reference especially to FIGS. 11 A- 13 B.
  • the different types of MPEG frames or blocks may therefore be handled using different deblocking mechanisms.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

Video deblocking can be implemented in video content processing and delivery environments when displaying decompressed/decoded video information. Discontinuities are identified and smoothed to reduce blockiness, for example at macroblock boundaries in conjunction with Moving Picture Expert Group (MPEG)-based schemes. In exemplary implementation(s), a client device detects discontinuities at or near block boundaries, determines which discontinuities are artificial, and smooths at least those discontinuities that are determined to be artificial. These actions may be accomplished using (i) one or more Laplacian of Gaussian (LoG) operations as applied to both an image and a quantization matrix and (ii) windowed comparisons between values and a threshold resulting from the LoG operations. Alternatively, these actions may be accomplished using a spatio-temporally varying filter. Additionally, in an MPEG context, intra image units may be deblocked using the former approach while non-intra image units may be deblocked using the latter approach.

Description

    TECHNICAL FIELD
  • This disclosure relates in general to video deblocking and in particular, by way of example but not limitation, to post-processing of video information to reduce blockiness in the display thereof. [0001]
  • BACKGROUND
  • Television-based entertainment systems are expanding the programming and services that they offer. In addition to television program content such as that found on broadcast and traditional cable networks, television service providers are adding on-demand video, as well as other interactive services, features, and applications. Such content and additional information are downloaded over a network for display, use, and/or storage on client-side set-top boxes or similar devices. Because of bandwidth constraints, downloads that include audio and/or video information, for example, are transmitted in a coded or compressed format. In other words, to reduce the amount of data that is transmitted, the information is typically compressed from a first size to a second smaller size. [0002]
  • After the compressed/coded data is transmitted, the data is decompressed/decoded into a semblance of the original audio and video information. Unfortunately, because the compression-decompression cycle is typically lossy, the decompressed version of the audio and video information differs from the original version of the audio and video information. With video information, for example, such differences can result in the presentation quality of the video information being visibly decreased. [0003]
  • Accordingly, for television-based entertainment systems, there is a need for schemes and techniques to enable post-processing improvements to video information that has been adulterated in a compression-decompression cycle. [0004]
  • SUMMARY
  • Video deblocking can be implemented in video content processing and delivery environments when displaying decompressed/decoded video information. Discontinuities are identified and smoothed to reduce blockiness, for example at macroblock boundaries in conjunction with Moving Picture Expert Group (MPEG)-based schemes. In exemplary implementation(s), a client device detects discontinuities at or near block boundaries, determines which discontinuities are artificial, and smooths at least those discontinuities that are determined to be artificial. These actions may be accomplished using (i) one or more Laplacian of Gaussian (LoG) operations as applied to both an image and a quantization matrix and (ii) windowed comparisons between values and a threshold resulting from the LoG operations. Alternatively, these actions may be accomplished using a spatio-temporally varying filter. Additionally, in an MPEG context, intra image units may be deblocked using the former approach while non-intra image units may be deblocked using the latter approach. [0005]
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The same numbers are used throughout the drawings to reference like and/or corresponding aspects, features, and components. [0006]
  • FIG. 1 illustrates an exemplary television system architecture in which the systems and methods for video deblocking can be implemented. [0007]
  • FIG. 2 illustrates an exemplary client device, a television, and various input devices that interact with the client device. [0008]
  • FIG. 3 is a block diagram that illustrates components of the exemplary client devices shown in FIGS. 1 and 2. [0009]
  • FIG. 4 is a set of graphs that illustrate an exemplary discontinuity and smoothing thereof. [0010]
  • FIG. 5 is a flow diagram that illustrates an exemplary method for video deblocking. [0011]
  • FIG. 6 is a flow diagram that illustrates an exemplary process for directly deblocking video. [0012]
  • FIG. 7 illustrates an exemplary window for deblocking analysis around a macroblock boundary. [0013]
  • FIG. 8 illustrates an exemplary Laplacian of Gaussian (LoG) operation in diagrammatic form. [0014]
  • FIG. 9 is a graph that illustrates exemplary values resulting from an LoG operation on a quantization matrix. [0015]
  • FIG. 10 illustrates pixels around a macroblock boundary for an exemplary visual adjustment procedure. [0016]
  • FIGS. 11A, 11B, and [0017] 11C are exemplary diagrams of a portion of a video information presentation across multiple frames of a group of pictures (GOP).
  • FIG. 12 is a graph that illustrates filters across macroblocks and macroblock boundaries in an exemplary approach for indirectly deblocking video using spatial filtering. [0018]
  • FIGS. 13A and 13B illustrate an exemplary approach for indirectly deblocking video using temporal filtering across multiple frames of a GOP. [0019]
  • FIG. 14 is a flow diagram of an exemplary approach to block-type-dependent filtering under a Moving Pictures Expert Group (MPEG)-compliant compression/coding scheme.[0020]
  • DETAILED DESCRIPTION
  • The following discussion is directed to video content processing and delivery systems, such as interactive TV networks, cable/satellite networks that utilize electronic program guides and other applications, and Web-enabled TV networks. Client devices in such systems range from full-resource clients with substantial memory and processing resources, such as TV-enabled personal computers and TV recorders equipped with hard-disks, to low-resource clients with limited memory and/or processing resources, such as traditional set-top boxes and personal digital assistants (PDAs) or mobile phones. However, video deblocking as described herein may additionally be used in other environments such as in streaming (e.g., over the Internet), generally in post-processing of compression and decompression cycles, and so forth. While aspects of the described systems and methods can be used in any of these environments and for any types of client devices, they are described primarily in the context of the following exemplary environment. [0021]
  • Exemplary System Architecture
  • FIG. 1 illustrates an exemplary [0022] television entertainment system 100 that is an architecture in which video deblocking may be implemented. System 100 facilitates distribution of content and other information to multiple viewers. System 100 includes one or more content providers 102, one or more other information providers 104, a content distribution system 106, and multiple client devices 108(1), 108(2), . . . , 108(N) coupled to content distribution system 106 via a broadcast network 110.
  • [0023] Content provider 102 includes a content server 112 and stored content 114, such as movies, television programs, commercials, music, and similar audio and/or video content. Content server 112 controls distribution of stored content 114 from content provider 102 to content distribution system 106. Additionally, content server 112 may control distribution of live content (e.g., content that was not previously stored, such as live feeds) and/or content stored at other locations to content distribution system 106.
  • [0024] Other information provider 104 includes other information database 116 and other information server 118. Other information database 116 stores information that may be provided to client devices 108. Such information includes software modules, files, images, text, executable programs, moving video, gaming or other interactive information, and so forth. The information may also include content, especially content of an irregular, one-of-a-kind, or similar nature, or content from smaller independent providers. Part or all of the information from other information database 116 may ultimately be displayed to subscribers after traversing one or more low-bandwidth transmission mediums, regardless of whether such transmission medium(s) are located upstream or downstream of content distribution system 106. Other information server 118 processes the other information from other information database 116 prior to distribution to generate one or more files that are optimized for, or at least capable of, transmission to content distribution system 106.
  • [0025] Content distribution system 106 includes a broadcast transmitter 128, one or more content processors 130, and one or more other information processors 132. Broadcast transmitter 128 can alternatively be a transceiver if bi-directional communication is desired (e.g., in situations in which broadcast network 110 is a bi-directional network 110). Broadcast transmitter 128 transmits (e.g., broadcasts) signals, such as cable/satellite television signals, across broadcast network 110. Broadcast network 110 can include a cable television network, RF, microwave, satellite, and/or data network, such as the Internet, and may also include wired or wireless media using any transmission format or protocol. Additionally, broadcast network 110 can be any type of network (including a bi-directional network), using any type of network topology and any network communication protocol, and can be represented or otherwise implemented as a combination of two or more networks.
  • [0026] Content processor 130 processes the content received from content provider 102 prior to transmitting the content across broadcast network 110. Similarly, other information processor 132 processes the other information that is received from other information provider 104 prior to transmission of the other information across broadcast network 110. A particular content processor 130 may encode, or otherwise process, the received content into a format that is understood by the multiple client devices 108(1), 108(2), . . . , 108(N) that are coupled to broadcast network 110. Although FIG. 1 shows a single content provider 102, a single other information provider 104, and a single content distribution system 106, the exemplary system 100 can include any number of content providers and/or other information providers coupled to any number of content distribution systems. Thus, content distribution system 106, content provider 102, and/or other information provider 104 are individually or jointly representative of a headend service that provides content and other information to multiple subscribers.
  • [0027] Client devices 108 can be implemented in a number of ways. For example, a client device 108(1) receives content and other information from a satellite-based transmitter via a satellite dish 134. Client device 108(1) is also referred to as a set-top box or a satellite receiving device. Client device 108(1) is coupled to a television 136(1) for presenting the content and other information (e.g., audio information, video information, and/or data information) that are received by the client device 108(1), as well as for presenting a graphical user interface. A particular client device 108 can be coupled to any number of televisions 136 and/or similar devices that can be implemented to display or otherwise render content. Similarly, any number of client devices 108 can be coupled to a single television 136.
  • Client device [0028] 108(2) is also coupled to receive content and other information from broadcast network 110 and to provide the received content and other information to associated television 136(2). Client device 108(N) is an example of a combination television 138 and integrated set-top box 140. In this example, the various components and functionality of the set-top box are incorporated into the television, rather than using two separate devices. Set-top box 140 that is integrated into television 138 can receive signals (e.g., broadcast signals) via a satellite dish (similar to satellite dish 134) and/or directly via
  • [0029] broadcast network 110. In alternate implementations, client devices 108 may receive signals via the Internet or any other network, especially those network mediums that are broadcast-capable. As is further described below, client devices 108 may also engage in video deblocking prior to displaying video information (whether content video information or other video information) that is received from a low bandwidth medium, such as a memory storage, other client devices, possibly broadcast network 110 itself, and so forth.
  • The [0030] exemplary system 100 also includes information from other networks/network providers 142, which may provide information such as information streamed over the Internet, information received directly from a provider of the information, and so forth. Information from other networks/network providers 142 may be accessible over broadcast network 110 (i.e., a network that also provides content information and other information from content distribution system 106). Alternatively, information from other networks/network providers 142 may be accessible over a different network, including a wide area network (WAN), the Internet, a public or private telecommunications network, and so forth.
  • Exemplary Client Device
  • FIG. 2 illustrates an [0031] exemplary implementation 200 of a client device 108 shown as a standalone unit that connects to a television 136 and communicates with various input devices 204, 206, and 208. Client device 108 can be implemented in any number of embodiments, including as a set-top box, a satellite receiver, a TV recorder with a hard disk, a digital video record (DVR) and playback system, a game console, an information appliance, and so forth.
  • [0032] Client device 108 includes a wireless port 202, such as an infrared (IR) or Bluetooth wireless port, for receiving wireless communications from a remote control device 204, a handheld input device 206, or any other wireless device, such as a wireless keyboard. Handheld input device 206 can be a personal digital assistant (PDA), handheld computer, wireless phone, or the like. Additionally, a wired keyboard 208 can be coupled to communicate with client device 108. In alternate embodiments, remote control device 204, handheld device 206, and/or keyboard 208 may use an RF communication link or other mode of transmission to communicate with client device 108.
  • [0033] Client device 108 receives one or more (e.g., broadcast) signals 210 from one or more broadcast sources, such as from a satellite or a cable or a broadcast network, including a broadcast implementation of network 110 (of FIG. 1). Client device 108 includes hardware and/or software for receiving and decoding a broadcast signal 210, such as an NTSC, PAL, SECAM or other TV system video signal. Client device 108 also includes hardware and/or software for providing the user with a graphical user interface by which the user can, for example, access various network services, configure client device 108, and perform other functions, including requesting video delivery.
  • [0034] Client device 108 can communicate with other devices via one or more connections including a conventional telephone line 212, an ISDN link 214, a cable link 216, an Ethernet link 218, a DSL link 220, and the like. Client device 108 may use any one or more of the various communication links 212-220 at a particular instant to communicate with any number of other devices. For example, in addition to receiving information over any one of communication links 212-220, client device 108 may provide (e.g., transmit) information over communication links 212-220 or any other type of communication link. Such other communication links may include links capable of interfacing with a local network such as a local area network (LAN), a Bluetooth® network, an IEEE 802.11b-compliant network, or other wired or wireless network type. In such a local network environment, a client device 108 that receives information from broadcast network 110 may forward the information over the local network to one or more other client devices 108. The forwarding may be effectuated, for example, over a lower-bandwidth transmission medium that introduces blockiness.
  • [0035] Client device 108 generates video signal(s) 222 and audio signal(s) 224, both of which are communicated to television 136. Video signals 222 and audio signals 224 can be communicated from client device 108 to television 136 via an RF (radio frequency) link, S-video link, composite video link, component video link, co-axial cable link, or other communication link. The video signals 222 may include deblocked video signals, for example. Although not shown in FIG. 2, client device 108 may include one or more lights or other indicators identifying the current status of the device. Additionally, the client device may include one or more control buttons, switches, or other selectable controls for controlling operation of the device.
  • FIG. 3 illustrates selected components of [0036] exemplary client device 108 shown in FIGS. 1 and 2. Client device 108 includes a first tuner 300 and an optional second tuner 302. The tuners 300 and 302 are representative of one or more in-band tuners that tune to various frequencies or channels to receive television signals, as well as at least one out-of-bound (OOB) tuner that tunes to the broadcast channel(s) over which data information is broadcast (e.g., carouseled or otherwise transmitted) to client device 108.
  • [0037] Client device 108 also includes one or more processors 304 which process various instructions to control the operation of client device 108 and to communicate with other electronic and computing devices. Client device 108 can be implemented with one or more memory components, examples of which include a random access memory (RAM) 306, a disk drive 308, another mass storage component 310, and a non-volatile memory 312 (e.g., ROM, Flash, EPROM, EEPROM, etc.). The memory components (e.g., RAM 306, disk drive 308, mass storage 310, and non-volatile memory 312) store various instructions and/or information such as received content, programs, configuration information for client device 108, graphical user interface information, and/or video data or information.
  • Alternative implementations of [0038] client device 108 can include a range of processing and memory capabilities, and may include more or fewer types of memory components than those illustrated in FIG. 3. For example, full-resource clients can be implemented with substantial memory and processing resources, including the disk drive 308 to store content for replay by the viewer. Low-resource clients, however, may have limited processing and memory capabilities, such as a limited amount of RAM 306, no disk drive 308, and limited processing capabilities of a processor 304.
  • An [0039] operating system 314 and one or more programs as represented by general instructions 316 may be stored in non-volatile memory 312 (and/or other memory component(s)) and executed on processor 304 to provide a runtime environment. A runtime environment facilitates extensibility of client device 108 by allowing various interfaces to be defined that, in turn, allow the programs to interact with client device 108. Although these programs may be installed when client device 108 is manufactured, they may also be received via broadcast network 110 from content distribution system 106 (of FIG. 1). Also stored in non-volatile memory 312 (and/or other memory component(s)) are electronically-executable instructions for video deblocking 318, as described further herein.
  • [0040] Client device 108 also includes a decoder 320 to decode a broadcast video signal, such as an NTSC, PAL, SECAM or other TV system video signal. Processor 304, along with tuner(s) 300 and 302 and/or decoder 320, also enables client device 108 to reconstruct audio and video from an MPEG-2 stream or other digital packet signal, whether compressed or uncompressed. Client device 108 can also include other components pertaining to a television entertainment system which are not illustrated in this example. For instance, client device 108 can include a user interface application and user interface lights, buttons, controls, and the like to facilitate viewer interaction with the device.
  • [0041] Client device 108 further includes a wireless interface 322, a network interface 324, a serial and/or parallel interface 326, and a modem 328. Wireless interface 322 allows client device 108 to receive input commands and other information from a user-operated input device, such as from a remote control device or from another IR, Bluetooth, or similar RF input device. Network interface 324 and serial and/or parallel interface 326 allows client device 108 to interact and communicate with other electronic and computing devices via various communication links, including local network communication links to other client devices 108. Although not shown, client device 108 may also include other types of data communication interfaces to communicate with other devices. Modem 328 facilitates communication by client device 108 with other electronic and computing devices via a conventional telephone line.
  • [0042] Client device 108 also includes an audio output 330 and a video output 332 that provide signals to a television or other device that processes and/or displays or otherwise renders the audio and video information, including deblocked video information. Although shown separately, some of the components of client device 108 may be implemented together in an application specific integrated circuit (ASIC). For example, electronically-executable instructions (318) and associated processing abilities for video deblocking may be integrated together onto one or more ASICs instead of using more-general memory (e.g., 306, 312, etc.) and/or processing (e.g., 304, 320, etc.) resources of a client device (108).
  • Additionally, a system bus (not shown) typically connects the various components within [0043] client device 108. A system bus can be implemented as one or more of any of several types of bus structures, including a memory bus or memory controller, a peripheral bus, an accelerated graphics port, or a local bus using any of a variety of bus architectures. By way of example, such architectures can include an Industry Standard Architecture (ISA) bus, a Micro Channel Architecture (MCA) bus, an Enhanced ISA (EISA) bus, a Video Electronics Standards Association (VESA) local bus, and a Peripheral Component Interconnects (PCI) bus also known as a Mezzanine bus.
  • Video Deblocking
  • Video information is often compressed or otherwise coded to facilitate the handling thereof. The coding of the video information into video data can reduce the bandwidth required to display, store, transmit, or otherwise manipulate the video information. For example, it can be beneficial to reduce the bit rate of video information for storing a video stream to disk or streaming it over a noisy, bandwidth-limited channel. However, compression and decompression of the video information typically causes visual artifacts that degrade the visual quality. One artifact that degrades visual quality is blockiness. In other words, low bit rate video information is often plagued with blockiness that seriously degrades visual quality. Video deblocking, on the other hand, usually entails reducing such blocky artifacts to enhance the visual presentation quality of the video information. Implementing a good deblocking mechanism helps to mitigate the deleterious byproducts of a lower bit rate video stream and thereby justifies using a lower bit rate that enables the aforementioned low-bandwidth applications. [0044]
  • Deblocking is implemented after previously encoded video data is decoded into at least a semblance of the original video information. One set of coding standards that may be used on the original video information are Moving Picture Expert Group (MPEG)-compliant standards. MPEG-based encoding/decoding algorithms involve performing a discrete cosine transform (DCT) on original video information and quantizing the DCT coefficients thereof. The resulting video data is de-quantized and an inverse DCT is performed to reacquire at least a semblance of the original video information. Reducing the bit rate of the video data stream may be achieved by more coarsely quantizing the DCT coefficients. In doing so, however, blockiness that is introduced along macroblock boundaries tends to seriously degrade the visual quality. The visual presentation quality may be at least partially restored by using deblocking during post-processing. Specifically, the deblocking may be used to reduce the blockiness around macroblock boundaries. [0045]
  • FIG. 4 is a set of [0046] graphs 400 that illustrate an exemplary discontinuity and smoothing thereof. Graphs 400 include graphs 400A, 400B, 400C, and 400D. Each of graphs 400A-400D includes two macroblocks of a video picture. Although the macroblocks are shown as being rectangular, actual macroblocks may instead be square. Graph 400A illustrates original video information 402 (as a solid line). There is no break in original video information 402 between the left macroblock and the right macroblock.
  • [0047] Graph 400B illustrates original video information 402 and adds thereto an illustration of discontinuity 404 (as a large dashed line). Discontinuity 404 represents an exemplary discontinuity or break in the seamlessness of original video information 402. This discontinuity arises at the macroblock boundary of the left and right macroblocks as a result of the encoding/decoding process. Discontinuity 404 may be visible as a blocky artifact. In fact, discontinuity 404 along with other discontinuities in the overall picture may be visible as a blockiness that reduces the overall visual quality.
  • [0048] Graph 400C illustrates original video information 402 and adds thereto an illustration of smoothed video information 406 (as a small dashed line). Smoothed video information 406 results from application of video deblocking to discontinuity 404. While smoothed video information 406 is not likely to be an exact replica of original video information 402, smoothed video information 406 reduces the blocky artifact of discontinuity 404 and generally more-closely approximates original video information 402. It should be noted that smoothed video information 406 is representative of deblocking mechanisms/approaches in general, including visual adjustment, filter smoothing, spatial-temporal filtering, and so forth.
  • [0049] Graph 400D includes original video information 402, discontinuity 404, and smoothed video information 406. All three are included together to illustrate (i) the discontinuity that may be formed from coarse quantization of original video information or other deleterious coding/decoding factors and (ii) the resulting video information after the discontinuity has been smoothed in post-processing. This post-processing video deblocking is described further herein.
  • Specifically, the description herein focuses on deblocking in an MPEG-based environment in which discontinuities appear at macroblock boundaries. However, deblocking may be employed after decompression/decoding in accordance with other standards and approaches and along other boundaries. It should therefore be understood that deblocking may be employed with video data/information in general and regardless of the size/type of blocks that are employed to segregate the video picture. [0050]
  • Although the term “discontinuity” is used herein, other such terms include “edge” and “boundary”. An edge, boundary, or discontinuity in a video picture may be “artificial” or “real”. Artificial discontinuities arise, for example, from coarse quantization of DCT coefficients. Other terms for artificial discontinuities include, but are not limited to, quantization discontinuities, false discontinuities, fake discontinuities, and quantization noise discontinuities. Real discontinuities, on the other hand, arise from true boundaries in a video picture from one element to another element within the picture. Other terms for real discontinuities include, but are not limited to, actual discontinuities, true discontinuities, and natural discontinuities. These real discontinuities may also lie along a macroblock boundary. Consequently, if smoothing is performed on all discontinuities at all macroblock boundaries, some real discontinuities will be smoothed. This causes the picture to be fuzzy and less distinct. However, if discontinuities at macroblock boundaries can be separated into real discontinuities and artificial discontinuities, then smoothing may be performed on the artificial discontinuities. [0051]
  • Methods for Video Deblocking
  • Video deblocking may be described in the general context of electronically-executable instructions. Generally, electronically-executable instructions include routines, programs, objects, components, data structures, and the like that perform particular functions or implement particular abstract data types. Video deblocking may also be practiced in distributed computing environments where functions are performed by remote processing devices that are linked through a communications network. In a distributed computing environment, electronically-executable instructions may be located in both local and remote storage media. [0052]
  • The methods and processes of FIGS. 5, 6, and [0053] 14 are illustrated in flow diagrams divided into multiple blocks. However, the order in which the methods and processes are described is not intended to be construed as a limitation, and any number of the described blocks can be combined in any order to implement one or more methods or processes for video deblocking. Furthermore, although the methods and processes are described below with reference to the television entertainment environments 100 and 200 and client devices 108 where applicable, the methods and processes can be implemented in any suitable hardware, software, firmware, or combination thereof and using any suitable mathematical alternatives.
  • FIG. 5 is a flow diagram [0054] 500 that illustrates an exemplary method for video deblocking. Flow diagram 500 includes blocks 502, 504, and 506 that may be implemented, for instance, by client devices 108 (of FIGS. 1-3). Blocks 502, 504, and 506 may correspond to electronically-executable instructions that are stored in one or more memories of a client device 108 and executed using one or more processors thereof. At block 502, a discontinuity is detected. For example, macroblock boundaries are analyzed to detect whether a discontinuity exists for video information between a first macroblock boundary and a second macroblock boundary. An exemplary window around a macroblock boundary for such an analysis is described further below with reference to FIG. 7.
  • At [0055] block 504, the discontinuity is examined to determine whether it is a real discontinuity or an artificial discontinuity. Real edges of actual picture elements may fall on a macroblock boundary, but smoothing real edges usually decreases video presentation quality. Exemplary mechanisms for differentiating between real discontinuities and artificial discontinuities are described further below especially with reference to FIGS. 8 and 9.
  • At [0056] block 506, the artificial discontinuity or discontinuities are smoothed. The smoothing of the artificial discontinuities reduces visually-apparent blockiness. It should be understood that smoothing may encompass a variety of mechanisms/approaches designed to reduce or eliminate the visibility of discontinuities. Exemplary mechanisms for smoothing artificial discontinuities are described further below especially with reference to FIGS. 10 et seq.
  • Flow diagram [0057] 500 is directed to both direct deblocking and indirect deblocking. Direct deblocking of video is described especially with reference to FIGS. 6-10. Indirect deblocking of video is described especially with reference to FIGS. 11A-13B. The description of FIG. 14 below is directed to a video deblocking approach that utilizes both direct and indirect deblocking mechanisms.
  • FIG. 6 is a flow diagram [0058] 600 that illustrates an exemplary process for directly deblocking video. Flow diagram 600 includes six (6) blocks 602-612. Blocks 602 and 604 pertain to input data for the process of flow diagram 600. Block 602 provides the scaled quantization matrix that was used to quantize the original video information during the coding/compressing process. This scaled quantization matrix of block 602 may be related to, for example, the “quant_matrix” parameter of MPEG-compliant coding schemes. The quant_matrix is determined by the encoder and encoded into the bit stream. A scaled quant matrix may be created as the product of the quant matrix and the quantization scale, which is another parameter that is encoded into the bit stream by the encoder. Block 604 provides the image that is formed from the decoded video information. As such, the image may include artificial discontinuities (e.g., at macroblock boundaries) and real discontinuities (e.g., either at or away from macroblock boundaries). The artificial discontinuities at the macroblock boundaries can arise from noise attributable to the quantization matrix during the quantization process.
  • The scaled quantization matrix of [0059] block 602 and the image video information of block 604 are submitted to block 606. Block 606 performs a Laplacian of Gaussian (LoG) operation on each of the scaled quantization matrix of block 602 and the image video information of block 604 to produce two different results. The LoG operation may be defined as convolving the argument with the LoG kernel. From the two different results of applying the LoG operation, the type of discontinuity is determined at block 608. For example, the type of discontinuity (if any) at the macroblock boundary under consideration is determined by comparing the result of the LoG operation as performed on the image of block 604 with a (quantization) threshold determined responsive to the result of the LoG operation as performed on the scaled quantization matrix of block 602. The LoG operation of block 606 and the discontinuity type determination of block 608 are described further below especially with reference to FIGS. 8 and 9.
  • If the discontinuity at the macroblock boundary is determined to be a real discontinuity (at block [0060] 608), block 610 may optionally perform a visual adjustment procedure at the macroblock boundary. An exemplary visual adjustment procedure is described further below especially with reference to FIG. 10. If, on the other hand, the discontinuity at the macroblock boundary is determined to be an artificial discontinuity (at block 608), block 612 performs a filter smoothing procedure at the macroblock boundary. An exemplary filter smoothing procedure is described further below primarily after the description of FIG. 10.
  • FIG. 7 illustrates an [0061] exemplary window 702 for deblocking analysis around a macroblock boundary 706. Two macroblocks 708L and 708R are illustrated. Each macroblock 708 is surrounded by four (4) macroblock boundaries 706. Within each macroblock 708 are a set of pixels. Although macroblocks 708 are shown as being blocks of eight-by-eight (8×8) pixels, they may alternatively be blocks of 16×16 pixels, 24×24 pixels, 30×30 pixels, or any other size. Because the illustrated macroblocks are of an eight-by-eight (8×8) size, window 702 includes sixteen (16) pixels to be analyzed for possible discontinuities at the macroblock boundary 706 that window 702 surrounds. However, the width of window 702 may alternatively be greater than two pixels (i.e., the width can be more than one pixel within each macroblock 708).
  • [0062] Window 702 is located between the left macroblock 708L and the right macroblock 708R. Similar “vertical” windows 702 are used at the other “vertical” macroblock boundaries 706. Furthermore, “horizontal” windows (not shown) are used at macroblock boundaries between upper macroblocks and adjacent lower macroblocks. In other words, to address a two-dimensional (2-D) image, video deblocking is performed in both the vertical direction and the horizontal direction.
  • To detect a discontinuity within a [0063] window 702, the following algorithm may be employed: Let Q represent the quantization matrix used in quantizing an image I. Because quantization is usually performed in the DCT domain, the effect of using Q in the time domain is first determined. Thus, an inverse DCT is performed on Q to generate Qt. Next, the discontinuities (e.g., edges, boundaries, etc.) are detected in the image I. Any such discontinuities may be detected using QO the Laplacian of Gaussian (LoG) operator to filter I and generate ILG. The LoG operator may be considered a concatenation of the Laplacian operator, which is given by: L ( x , y ) = 2 f ( x , y ) = 2 f ( x , y ) x 2 + 2 f ( x , y ) y 2
    Figure US20030235250A1-20031225-M00001
  • and the Gaussian operator, which is given by: [0064] G ( x , y ) = 1 2 π 1 2 σ 2 ( x 2 + y 2 )
    Figure US20030235250A1-20031225-M00002
  • thereby yielding: [0065] LoG ( x , y ) = - 1 πσ 4 [ 1 - x 2 + y 2 2 σ 2 ] - x 2 + y 2 2 σ 2 .
    Figure US20030235250A1-20031225-M00003
  • The filtered image is therefore given by:[0066]
  • I LG=LoG*I.
  • FIG. 8 illustrates an exemplary Laplacian of Gaussian (LoG) [0067] operation 800 in diagrammatic form. An image 802 is filtered using LoG operation 804 to produce an edge-biased image 806. After applying the LoG operation to image I, discontinuities (e.g., edges) can be detected in each window 702 surrounding every analyzed macroblock boundary 706. However, it is still to be determined whether each discontinuity in each window 702 is a real discontinuity (e.g., from an edge of an image element) or an artificial discontinuity (e.g., from quantization or other deleterious compression/decompression factors). The LoG operation detects both kinds of discontinuities, but it does not clearly differentiate between them by itself.
  • A variable that may be termed “quant_threshold” is created to determine which discontinuities are artificial. The amount of blockiness attributable to quantization noise is effectively assessed. In order to assess the amount of blockiness attributable to quantization, the time-domain quantization matrix divided by two (Q[0068] t/2) is passed through the LoG filter to obtain QtLG. The quantization matrix Qt is divided by 2 because half the quantization matrix Qt represents the actual loss of data due to round-off/truncation. After obtaining QtLG from passing (Qt/2) through the LoG filter, the maximum and minimum values of QtLG are examined to determine the quant_threshold.
  • FIG. 9 is a [0069] graph 900 that illustrates exemplary values QtLG resulting from a LoG operation on a time-domain quantization matrix Qt. Thus, a plot 902 graphs different quantization matrix values (along the abscissa axis) versus corresponding LoG operation output values QtLG (along the ordinate axis). Maximum absolute values for the QtLG values are denoted as one hundred percent (100%). Any portion of the 100% value may be selected as the threshold between real discontinuities and artificial discontinuities. The threshold may be selected, for example, such that the majority of the visually displeasing artificial discontinuities are subject to smoothing while few if any of the real discontinuities are subject to smoothing (therein mitigating “fuzzying” of the original image). An optimum value for the threshold for any given type of video information may be selected through experimental viewing of deblocked video. In the illustrated graph 900, eighty-five percent (85%) is selected as the threshold between real discontinuities and artificial discontinuities. Furthermore, this threshold may be tunable subject to user/viewer preferences.
  • After the threshold between real discontinuities and artificial discontinuities is selected, the threshold is used in conjunction with an analysis of [0070] macroblock boundaries 706 in windows 702. The analysis is focused in windows 702 because blockiness that is due to quantization noise is centered on macroblock boundaries 706. The values of ILG are compared to the threshold within each window 702. If more than a certain fraction, η, of the ILG values in a window 702 are above the selected threshold, then that discontinuity is labeled as a real discontinuity. If not, then that discontinuity is labeled as an artificial discontinuity. The fraction, η, may also be selected experimentally, for example by viewing different types of video that has been deblocked using different values for the fraction, η, until a visually-appealing value is selected. The parameter, η, may be experimentally determined and fine tuned to maximize visual quality.
  • After discontinuities have been (i) detected and (ii) determined to be (and “labeled” as) either real discontinuities or artificial discontinuities, appropriate deblocking procedures can be implemented. As described above with reference to [0071] blocks 608, 610, and 612 (of FIG. 6), real discontinuities may be subjected to a visual adjustment procedure, and artificial discontinuities may be subjected to a filter smoothing procedure. In an exemplary visual adjustment procedure, pixel values are adjusted slightly by moving them towards each other by an amount proportional to the threshold so as to reduce the effects of quantization.
  • FIG. 10 illustrates [0072] pixels 704 around a macroblock boundary 706 for an exemplary visual adjustment procedure. Eight (8) adjacent pixels 704 from p0 to p7 are divided by the macroblock boundary 706 that is between left macroblock 708L and right macroblock 708R. Pixels p0, p1, p2, and p3 are in left macroblock 708L. Pixels p4, p5, p6, and p7 are in right macroblock 708R. The values of the pixels p0 to p7 are adjusted slightly by moving them towards each other by an amount proportional to the selected threshold so as to reduce the effects of quantization. This adjustment is performed recursively to the adjacent pixels.
  • The exemplary visual adjustment procedure for real discontinuities, when performed, may be effectuated as follows: Let p=[p[0073] 0p1 . . . p7] represent the pixels 704 around a macroblock boundary 706 as illustrated in FIG. 10. The following series of operations are performed: d = min ( p 3 - p 4 2 , quant_threshold ) p 3 = p 3 - d p 4 = p 4 + d p 2 = p 2 + p 3 2 p 5 = p 5 + p 4 2
    Figure US20030235250A1-20031225-M00004
  • where decoded, time-domain pixels are designated by “p” (p without the prime) and the adjusted pixels are represented by “p′” (p with the prime). [0074]
  • If, on the other hand, a detected discontinuity has been determined to be, and “labeled” as, an artificial discontinuity, then a filter smoothing procedure is performed. In an exemplary filter smoothing procedure, pixels [0075] 704 (of FIG. 7) of a window 702 are filtered around boundary 706 with a smoothing filter. In other words, a pixel array, p, is filtered with a smoothing filter such as stretched and/or scaled version(s) of the Gaussian or tent filter. An example of a Gaussian/tent filter is: G [ n ] = 1 2 π - 1 2 σ 2 n 2 T [ n ] = { 1 - n 5 - 5 n 5 0 otherwise
    Figure US20030235250A1-20031225-M00005
  • In general, any smoothing filter can be used with the degree of smoothing set to accommodate the desired amount of blurring. [0076]
  • In the second approach to video deblocking, a less rigid and more indirect mechanism is used to detect discontinuities, to differentiate between real and artificial discontinuities, and to smooth the detected discontinuities. This second mechanism includes both spatial and temporal aspects. In a spatial aspect, the mechanism entails associating an increased likelihood of needing deblocking (e.g., of a discontinuity being an artificial discontinuity) with increased proximity to a macroblock boundary. In a temporal aspect, the mechanism entails associating the likelihood of blockiness at a macroblock boundary to the position of the picture within a GOP. For example, it is assumed that an I-frame of the GOP is more susceptible to blockiness at the macroblock boundary. On the other hand, a B-frame downstream in the GOP is less likely to experience blockiness at the macroblock boundary. [0077]
  • FIGS. 11A, 11B, and [0078] 11C are exemplary diagrams 1100 of a portion of video information that is presented across multiple frames of a GOP. In an MPEG-compliant coding scheme, frames are organized into GOPs that start with an intra (I) frame and are followed by predicted (P) frames and bidirectional (B) frames/modes. I-frames are usually more susceptible to blockiness at macroblock boundaries than those frames that are downstream in the GOP as is explained by the description of the diagrams 1100A-C. Specifically, diagram 1100A represents an intra-frame with many macroblocks 708 and macroblock boundaries 706.
  • Diagrams [0079] 1100B and 1100C represent non-intra frames. Each non-intra block in non-intra frames includes two components: the predicted term and the difference term. Typically, the predicted term dominates, and the difference term is relatively small. Because the predicted term is usually not macroblock aligned in the reference frame (as shown in diagram 1100B), any macroblocking effects in the reference frame are seen in the interior of the reconstructed macroblock in the current frame. Moreover, as this effect occurs over several generations of prediction, the magnitude and location of blockiness tends to be diffused (as shown in diagram 1100C). In other words, downstream in a GOP, images tend to be both less crisp as well as less blocky. Thus, the position of a frame in a GOP, or GOP position number (GPN), tends to serve as an indicator of the extent of blockiness. This tendency is reflected in the temporal aspect of the second mechanism. Similarly, for the spatial aspect of the second mechanism, the extent of quantization-caused blockiness tends to increase as pixel location approaches a macroblock boundary.
  • The second mechanism thus involves filtering the image with a spatio-temporally varying filter. The degree of filtering increases for pixels near macroblock boundaries and decreases for pixels away from macroblock boundaries. This spatial filtering is described further below with reference to FIG. 12. Also, the degree of filtering is higher at the beginning of a GOP while the degree of filtering decreases downstream in the GOP. This temporal filtering is described further below with reference to FIGS. 13A and 13B. Combining the spatial filtering with the temporal filtering creates a spatio-temporally varying filter that indirectly detects discontinuities, differentiates between discontinuity types, and smooths the detected discontinuities. [0080]
  • FIG. 12 is a [0081] graph 1200 that illustrates filters 1202 across macroblocks 708 and macroblock boundaries 706 in an exemplary approach for indirectly deblocking video using spatial filtering. Filters 1202 include three (3) different filtering zones labeled zone # 1, zone # 2, and zone # 3. While each of the three different filtering zones may have filters of similar or identical energy levels (to avoid changing image intensity), each of the filters of the three different filtering zones filters pixels of the macroblocks 708 differently. Filtering zone # 1 filters to the greatest degree in an area around macroblock boundaries 706. Filtering zone # 3 filters the least in an area around the center of macroblocks 708. Filters of filtering zones # 3 may be implemented as an impulse function so that the pixels are changed very slightly or not at all. The centers of macroblocks 708 are the least likely to exhibit blockiness due to quantization noise. Located spatially between the filtering zones # 1 and #3 are filtering zones # 2. The filters of filtering zones # 2 filter the corresponding physically-intermediate pixels to a degree that is also intermediate to that of the filters of filtering zones # 1 and #3.
  • Using multiple filtering zones, the pixels may be averaged, for example, to a greater extent near macroblock boundaries and to gradually lesser extents for pixels that are gradually nearer the middle of macroblocks. Although [0082] filter 1202 includes an exemplary three filtering zones, four or more or less than three filtering zones may alternatively be employed. An example of a suitable spatial filter is: G σ [ n ] = 1 2 π - 1 2 σ 2 n 2
    Figure US20030235250A1-20031225-M00006
  • where we can distinguish the various zones on the basis of the chosen value for σ, for instance, as: [0083]
  • Zone [0084] 1: σ=1;
  • Zone [0085] 2: σ=0.707; and
  • Zone [0086] 3: σ=0.5.
  • Furthermore, the filters can be normalized in order to achieve a gain of 1 (or any desired value). Such spatial filtering thus concentrates filtering near macroblock boundaries and away from the center of macroblocks. Using a spatial filter designed accordingly effectively tends to indirectly filter blockiness due to quantization noise. [0087]
  • FIGS. 13A and 13B illustrate an exemplary approach for indirectly deblocking video using temporal filtering across multiple frames of a GOP. The extent of spatial filtering that is necessary and/or desirable around macroblock boundaries tends to decrease as frame position number proceeds through a GOP, as is explained above with reference to FIGS. [0088] 11A-11C. Chart 1300A illustrates a GOP and exemplary temporal filtering coefficients designated as α. The GOP position number (GPN) runs from 1 to n, where n may be any natural number, but is often set to fifteen (15). The MPEG frames are designated by I, B, or P in an exemplary sequence that runs: I, B, B, B, P, B, B, B, P, B, B, B, P, B, B. Other sequences may alternatively be used. Three different sets of temporal filtering coefficients α are included and are designated as α1, α2, and α3.
  • The temporal filtering coefficient a may be applied to the spatial filtering as described above with reference to FIG. 12. The temporal filtering coefficient α serves to reduce the filtering by, and therefore the impact of, the [0089] spatial filter 1202 as frames progress downward through the GOP. Consequently, each α may be determined responsive to the GPN of the frame under consideration. Table 1300B provides a general GPN-dependent formula for determining an α for each of the three temporal filtering coefficient α examples. Each of the α examples decrease in value as the GPN increases to reduce the impact of the spatial filter. The rapidity at which each α decreases in value as the GPN increases from 1 to n scales from a low with α1 to a high with α3. Besides the formulas provided in table 1300B, other formulas for determining α may alternatively be employed. Such other formulas, or even other sets of α values that are not necessarily formulaic, need not use an actual GPN numeral to set the α.
  • The temporally varying parameter, α, may be used to vary the degree of filtering using, for example, the following approach: [0090] G σ [ α , n ] = K · 1 2 π - 1 2 ( ασ ) 2 n 2
    Figure US20030235250A1-20031225-M00007
  • where the number K is chosen to normalize the energy in the filter. Also, if the filtering ability of the [0091] spatial filter 1202 is inversely proportional to α, then the α formula may be set up so as to increase with increasing GPN.
  • An application of the α formulas of table [0092] 1300B are provided under the appropriate GPNs of chart 1300A. For example, given an n=15, α1 is the set of numbers {1, {fraction (14/15)}, {fraction (13/15)}, {fraction (12/15)}, {fraction (11/15)}, . . . , {fraction (2/15)}, {fraction (1/15)}}; α2 is the set of numbers {1½, ⅓, ¼, ⅕, . . . , {fraction (1/14)}, {fraction (1/15)}}; and α3 is the set of numbers {1¼, {fraction (1/9)}, {fraction (1/16)}, {fraction (1/25)}, . . . , {fraction (1/196)}, {fraction (1/225)}}. An α formula or set of values that results in visually-pleasing video may be determined experimentally for any given type of video information. Use of the temporal filtering coefficient α as a modifier to a spatial filter 1202 enables application of a spatio-temporally varying filter that indirectly detects, differentiates between types of, and smoothes discontinuities.
  • FIG. 14 is a flow diagram [0093] 1400 of an exemplary approach to block-type-dependent filtering under an MPEG-compliant compression/coding scheme. The second mechanism, as described above with particular reference to FIGS. 11A-13B, may be applied to all frames of a GOP. Alternatively, the approach of flow diagram 1400 may be employed in which the video deblocking mechanism that is being applied depends on the frame/block type. Flow diagram 1400 is described in terms of blocks such as macroblocks, but it may instead be implemented on a frame-only basis. More generally, blocks, frames, macroblocks, etc. may be considered units of an image.
  • As noted above especially in the descriptions of FIGS. [0094] 11A-11C and 13A, each MPEG GOP starts with an I-frame. This I-frame has relatively hard and certain macroblock boundaries. Non-I-frames (e.g., B-frames and P-frames), on the other hand, may have discontinuities that are not macroblock aligned in the current frame. To handle this difference, the first mechanism (which is more direct) may be applied to I-frames while the second mechanism (which is more indirect) may be applied to non-I-frames. Thus, the deblocking under such a scheme is frame-type-dependent. Under MPEG coding, however, individual macroblocks in a non-I-frame may be equivalent to an I-frame. In other words, some blocks in P-frames and B-frames are unpredicted, have no inherent discontinuity blurring, and are macroblock aligned in the current frame. These I-blocks in P-frames and B-frames can thus be treated using the more direct first mechanism.
  • Accordingly, a determination as to which mechanism is to be used under an MPEG-compliant coding scheme may be block-type-dependent. In flow diagram [0095] 1400, a flow diagram block 1402 determines whether a block under consideration is (i) an intra block or (ii) a predicted or a bi-directional block. For example, prior to actual smoothing, each macroblock of a frame of video information may be analyzed to determine whether it is an I-block or a P-block/B-block. If it is determined to be an I-block, then a visual adjustment procedure or a filter smoothing procedure is implemented at flow diagram block 1404. Whether a visual adjustment procedure or a filter smoothing procedure is implemented is dependent on discontinuity type. As described further above with reference especially to FIGS. 6-10, the visual adjustment procedure is applied to real discontinuities and the filter smoothing procedure is applied to artificial discontinuities.
  • If, on the other hand, the block under consideration is determined to be a P-block or a B-block (e.g., a non-I-block), then a spatio-temporally varying filtering procedure is implemented for the block under consideration at [0096] flow diagram block 1406. The extent of filtering for any given pixel or set of pixels is dependent on the spatial position of the pixel or pixels within the block. The extent of filtering for the given pixel or set of pixels is also dependent on the temporal-based GPN of the frame of which the block forms a part. Such a spafio-temporally varying filtering procedure is described further above with reference especially to FIGS. 11A-13B. In manners in accordance with flow diagram 1400, the different types of MPEG frames or blocks may therefore be handled using different deblocking mechanisms.
  • Conclusion
  • Although systems and methods have been described in language specific to structural features and/or methods, it is to be understood that the invention defined in the appended claims is not necessarily limited to the specific features or methods described. Rather, the specific features and methods are disclosed as exemplary forms of implementing the claimed invention. [0097]

Claims (99)

1. A method for deblocking video, comprising:
detecting one or more discontinuities in proximity to block boundaries of an image;
determining whether any of the one or more discontinuities are artificial discontinuities using the image and a quantization matrix; and
smoothing the one or more discontinuities that are determined to be artificial discontinuities from the action of determining.
2. The method as recited in claim 1, wherein the action of detecting one or more discontinuities in proximity to block boundaries of an image comprises:
detecting the one or more discontinuities in proximity to the block boundaries of the image by applying a Laplacian of Gaussian (LoG) operation on the image.
3. The method as recited in claim 1, wherein the action of determining whether any of the one or more discontinuities are artificial discontinuities using the image and a quantization matrix comprises:
determining whether any of the one or more discontinuities are artificial discontinuities using the image and a scaled quantization matrix.
4. The method as recited in claim 1, wherein the action of detecting one or more discontinuities in proximity to block boundaries of an image comprises:
detecting the one or more discontinuities in proximity to the block boundaries of the image by applying a Laplacian of Gaussian (LoG) operation on the image to produce a plurality of LoG image values; and
wherein the action of determining whether any of the one or more discontinuities are artificial discontinuities using the image and a quantization matrix comprises:
applying the LoG operation on the quantization matrix to produce a plurality of LoG quantization matrix values;
determining a threshold responsive to the plurality of LoG quantization matrix values;
comparing at least a subset of the plurality of LoG image values to the threshold; and
labeling a discontinuity of the one or more discontinuities as an artificial discontinuity if a number of the at least a subset of the plurality of LoG image values that are greater than the threshold is not greater than a predetermined fraction.
5. The method as recited in claim 1, wherein the action of smoothing the one or more discontinuities that are determined to be artificial discontinuities from the action of determining comprises:
filtering the artificial discontinuities using at least one of a Gaussian filter and a tent filter.
6. The method as recited in claim 1, wherein the actions of detecting, determining, and smoothing are performed by a client device of a television-based entertainment environment.
7. The method as recited in claim 1, wherein the actions of detecting and determining are performed with respect to windows of a predetermined pixel width that surround the block boundaries of the image.
8. The method as recited in claim 1, further comprising:
determining whether any of the one or more discontinuities are real discontinuities using the image and the quantization matrix; and
smoothing the one or more discontinuities that are determined to be real discontinuities.
9. The method as recited in claim 8, wherein the action of detecting one or more discontinuities in proximity to block boundaries of an image comprises:
detecting the one or more discontinuities in proximity to the block boundaries of the image by applying a Laplacian of Gaussian (LoG) operation on the image to produce a plurality of LoG image values; and
wherein the action of determining whether any of the one or more discontinuities are real discontinuities using the image and the quantization matrix comprises:
applying the LoG operation on the quantization matrix to produce a plurality of LoG quantization matrix values;
determining a threshold responsive to the plurality of LoG quantization matrix values;
comparing at least a subset of the plurality of LoG image values to the threshold; and
labeling a discontinuity of the one or more discontinuities as a real discontinuity if a number of the at least a subset of the plurality of LoG image values that are greater than the threshold is greater than a predetermined fraction.
10. The method as recited in claim 8, wherein the action of smoothing the one or more discontinuities that are determined to be real discontinuities comprises:
visually adjusting the real discontinuities by moving pixel values toward each other.
11. A method for providing video deblocking, comprising:
applying a Laplacian of Gaussian (LoG) operation on an image to produce a first plurality of values, the image being previously encoded using a quantization matrix;
applying the LoG operation on the quantization matrix to produce a second plurality of values;
determining a threshold responsive to the second plurality of values;
comparing at least a portion of the first plurality of values to the threshold;
ascertaining a number of the at least a portion of the first plurality of values that exceed the threshold; and
if the number is less than a predetermined fraction, then labeling a discontinuity that is associated with the at least a portion of the first plurality of values as an artificial discontinuity.
12. The method as recited in claim 11, further comprising:
if the number is greater than the predetermined fraction, then labeling the discontinuity that is associated with the at least a portion of the first plurality of values as a real discontinuity.
13. The method as recited in claim 12, further comprising:
performing a visual adjustment procedure on the real discontinuity.
14. The method as recited in claim 11, wherein the at least a portion of the first plurality of values corresponds to pixels of the image that are proximate to at least one macroblock boundary.
15. The method as recited in claim 11, wherein the quantization matrix comprises a scaled quantization matrix.
16. The method as recited in claim 11, wherein the action of determining a threshold responsive to the second plurality of values comprises:
selecting a percentage of an absolute maximum value of the second plurality of values as the threshold.
17. The method as recited in claim 11, further comprising:
performing a filter smoothing procedure on the artificial discontinuity.
18. One or more electronically-accessible media comprising electronically-executable instructions that, when executed, direct an electronic device to perform the method as recited in claim 11.
19. An arrangement capable of deblocking video, comprising:
detecting means for detecting one or more discontinuities at macroblock boundaries of a video image;
determining means for determining whether the one or more discontinuities include one or more artificial discontinuities and whether the one or more discontinuities include one or more real discontinuities; and
smoothing means for smoothing the one or more discontinuities in dependence on whether each discontinuity of the one or more discontinuities comprises an artificial discontinuity or a real discontinuity.
20. The arrangement as recited in claim 19, wherein the arrangement comprises a client device of a television-based entertainment environment.
21. The arrangement as recited in claim 19, wherein the detecting means utilizes a Laplacian of Gaussian (LoG) operation on the video image and the determining means utilizes the LoG operation on a quantization matrix.
22. The arrangement as recited in claim 21, wherein the determining means includes selecting means for selecting a discontinuity threshold from a plurality of results of the LoG operation on the quantization matrix.
23. The arrangement as recited in claim 21, wherein the quantization matrix comprises a scaled quantization matrix.
24. The arrangement as recited in claim 19, wherein the smoothing means includes visual adjustment means for visually adjusting each discontinuity that comprises a real discontinuity and filter smoothing means for filter smoothing each discontinuity that comprises an artificial discontinuity with at least one smoothing filter.
25. The arrangement as recited in claim 24, wherein the at least one smoothing filter comprises at least one of a tent filter and a Gaussian filter.
26. A client device for a television-based entertainment environment, the client device configured to perform actions comprising:
detecting one or more discontinuities at macroblock boundaries of a video image;
determining whether the one or more discontinuities include one or more artificial discontinuities and whether the one or more discontinuities include one or more real discontinuities;
smoothing the one or more artificial discontinuities using a filter smoothing procedure; and
smoothing the one or more real discontinuities using a visual adjustment procedure.
27. The client device as recited in claim 26, wherein the client device is further configured to detect the one or more discontinuities by applying a Laplacian of Gaussian (LoG) operation on the video image.
28. The client device as recited in claim 26, wherein the client device is further configured to determine whether the one or more discontinuities include one or more artificial discontinuities and whether the one or more discontinuities include one or more real discontinuities using a Laplacian of Gaussian (LoG) operation on a quantization matrix.
29. One or more electronically-accessible media comprising instructions that, when executed, direct an electronic device to:
detect one or more discontinuities in proximity to block boundaries of an image, the image having been previously quantized using a quantization matrix;
determine whether any of the one or more discontinuities are artificial discontinuities using the image and the quantization matrix; and
smooth the one or more discontinuities that are determined to be artificial discontinuities using a filter smoothing procedure;
wherein the instructions for the detection and the determination entail performing at least one Laplacian of Gaussian (LoG) operation.
30. The one or more electronically-accessible media as recited in claim 29, wherein the one or more electronically-accessible media comprise at least one memory of a client device of a television-based entertainment environment.
31. The one or more electronically-accessible media as recited in claim 29, wherein the quantization matrix comprises a scaled quantization matrix.
32. The one or more electronically-accessible media as recited in claim 29, further comprising instructions that, when executed, direct the electronic device to:
smooth one or more discontinuities that are real discontinuities using a visual adjustment procedure.
33. A method for deblocking video, comprising:
detecting one or more discontinuities in proximity to block boundaries of an image;
determining whether any of the one or more discontinuities are artificial discontinuities; and
smoothing the one or more discontinuities that are determined to be artificial discontinuities;
wherein the actions of detecting, determining, and smoothing are effectuated using a spatio-temporally varying filter.
34. The method as recited in claim 33, wherein the spatio-temporally varying filter smooths artificial discontinuities to a greater extent in the vicinity of the block boundaries.
35. The method as recited in claim 33, wherein the image is one image in a group of images; and wherein the spatio-temporally varying filter smooths artificial discontinuities to a greater extent in images that are relatively upstream in the group of images and smooths artificial discontinuities to a lesser extent in images that are relatively downstream in the group of images.
36. The method as recited in claim 35, wherein the image comprises a picture that has been compressed and decompressed in accordance with a Moving Pictures Expert Group (MPEG)-compliant scheme; and wherein the group of images comprises a group of pictures (GOP).
37. The method as recited in claim 33, wherein the spatio-temporally varying filter includes at least one temporal variable that is dependent on a group of pictures (GOP) position number (GPN).
38. The method as recited in claim 33, wherein the spatio-temporally varying filter includes a plurality of filter zones between two block boundaries of the image.
39. A method for deblocking video, comprising:
ascertaining block boundaries of an image unit;
ascertaining a position of the image unit in a group of images; and
smoothing one or more discontinuities of the image unit responsive to individual locations of the one or more discontinuities relative to the block boundaries of the image unit and responsive to the position of the image unit in the group of images.
40. The method as recited in claim 39, wherein the image unit comprises at least one of a frame and a macroblock.
41. The method as recited in claim 39, wherein the action of smoothing one or more discontinuities of the image unit responsive to individual locations of the one or more discontinuities relative to the block boundaries of the image unit comprises smoothing the one or more discontinuities in increasing amounts as the individual locations of the one or more discontinuities approach the block boundaries.
42. The method as recited in claim 39, wherein the action of smoothing one or more discontinuities of the image unit responsive to the position of the image unit in the group of images comprises smoothing the one or more discontinuities in decreasing amounts as the position of the image unit in the group of images increases.
43. The method as recited in claim 39, wherein the actions of ascertaining and smoothing are performed by a client device of a television-based entertainment environment.
44. One or more electronically-accessible media comprising electronically-executable instructions that, when executed, direct an electronic device to perform the method as recited in claim 39.
45. An arrangement that is capable of deblocking video, comprising:
ascertaining means for ascertaining block boundaries of an image unit and a position of the image unit in a group of images; and
smoothing means for smoothing one or more discontinuities of the image unit (i) responsive to individual locations of the one or more discontinuities relative to the block boundaries of the image unit and (ii) responsive to the position of the image unit in the group of images.
46. The arrangement as recited in claim 45, wherein the ascertaining means comprises a Moving Pictures Expert Group (MPEG)-compliant decoder.
47. The arrangement as recited in claim 45, wherein the smoothing means comprises a spatial filtering means for spatial filtering that includes a plurality of filtering zones for application across the image unit; and wherein filtering zones of the plurality of filtering zones that overlap the block boundaries of the image unit filter more strongly than filtering zones of the plurality of filtering zones that are located away from the block boundaries.
48. The arrangement as recited in claim 45, wherein the smoothing means comprises a temporal filtering means for temporal filtering that includes a filtering modifier that varies depending on how far downstream the position of the image unit is from an intra image unit in the group of images.
49. The arrangement as recited in claim 45, wherein the arrangement comprises a client device for a television-based entertainment environment.
50. A client device for a television-based entertainment environment, the client device configured to perform actions comprising:
spatially filtering an image having a plurality of blocks, each block of the plurality of blocks having a middle and at least two boundaries; the spatial filtering orchestrated such that filtering increases near the at least two boundaries and decreases near the middle of each block of the plurality of blocks; and
modifying the spatial filtering using a temporally-varying variable, the temporally-varying variable established such that the spatial filtering is greater for images near the beginning of groups of images and reduced for images near the end of groups of images.
51. The client device as recited in claim 50, wherein the spatial filtering smooths at least artificial discontinuities.
52. The client device as recited in claim 50, wherein the temporally-varying variable may be determined from a formula using a group of images position number.
53. The client device as recited in claim 50, wherein the temporally-varying variable may be determined from a set of temporally-varying variables using a group of images position number.
54. The client device as recited in claim 50, wherein the client device is further configured to perform actions comprising:
receiving video data via an interface; and
decoding the video data into video information that includes the image.
55. The client device as recited in claim 50, wherein the client device is further configured to perform an action comprising:
forwarding the image, which has been spatially filtered as modified by the temporally-varying variable, to a video screen.
56. One or more electronically-accessible media comprising instructions that, when executed, direct an electronic device to:
spatially filter an image having a plurality of blocks, each block of the plurality of blocks having a middle and at least two boundaries; the spatial filtering orchestrated such that filtering increases near the at least two boundaries and decreases near the middle of each block of the plurality of blocks; and
modify the spatial filtering using a temporally-varying variable, the temporally-varying variable established such that the spatial filtering is greater for images near the beginning of groups of images than for images near the end of groups of images.
57. The one or more electronically-accessible media as recited in claim 56, further comprising instructions that, when executed, direct the electronic device to:
apply the spatial filtering on discontinuities of the image.
58. The one or more electronically-accessible media as recited in claim 56, further comprising instructions that, when executed, direct the electronic device to:
receive video data via an interface;
decode the video data into video information that includes the image; and
forward the image, which has been spatially filtered as modified by the temporally-varying variable, to a video screen.
59. A method for providing video deblocking, comprising:
detecting a discontinuity in proximity to a block boundary of video information;
determining whether the discontinuity is an artificial discontinuity; and
if so, smoothing the artificial discontinuity on both sides of the block boundary.
60. The method as recited in claim 59, wherein the action of detecting a discontinuity in proximity to a block boundary of video information comprises:
detecting one or more discontinuities at four different block boundaries of a block of the video information.
61. The method as recited in claim 59, wherein the action of detecting a discontinuity in proximity to a block boundary of video information comprises:
detecting one or more discontinuities in a window around the block boundary of a block of the video information.
62. The method as recited in claim 59, wherein the block boundary comprises a macroblock boundary; and
wherein the action of detecting a discontinuity in proximity to a block boundary of video information comprises:
detecting one or more discontinuities in proximity to the macroblock boundary of a macroblock of the video information.
63. The method as recited in claim 59, wherein the action of detecting a discontinuity in proximity to a block boundary of video information comprises:
applying a Laplacian of Gaussian (LoG) operation on an image of the video information.
64. The method as recited in claim 59, wherein the action of determining whether the discontinuity is an artificial discontinuity comprises:
applying a Laplacian of Gaussian (LoG) operation on a quantization matrix that has been used to de-quantize the video information.
65. The method as recited in claim 59, wherein the action of determining whether the discontinuity is an artificial discontinuity comprises:
applying a Laplacian of Gaussian (LoG) operation on a quantization matrix that is in the time domain.
66. The method as recited in claim 59, wherein the action of determining whether the discontinuity is an artificial discontinuity comprises:
applying a Laplacian of Gaussian (LoG) operation on a quantization matrix to produce a plurality of values;
analyzing the plurality of values to determine a real discontinuity threshold responsive to at least one maximum absolute value of the plurality of values.
67. The method as recited in claim 59, wherein the action of determining whether the discontinuity is an artificial discontinuity comprises:
comparing a first plurality of values that result from applying a Laplacian of Gaussian (LoG) operation on an image of the video information to a threshold that is determined responsive to a second plurality of values that result from applying the LoG operation on a quantization matrix.
68. The method as recited in claim 67, wherein the quantization matrix comprises a scaled quantization matrix.
69. The method as recited in claim 59, wherein the action of determining whether the discontinuity is an artificial discontinuity comprises:
determining that the discontinuity is an artificial discontinuity if a predetermined fraction of a first plurality of values that result from applying a Laplacian of Gaussian (LoG) operation on an image of the video information are less than a threshold that is determined responsive to a second plurality of values that result from applying the LoG operation on a quantization matrix.
70. The method as recited in claim 59, wherein the action of determining whether the discontinuity is an artificial discontinuity comprises:
determining that the discontinuity is not an artificial discontinuity if a predetermined fraction of a first plurality of values that result from applying a Laplacian of Gaussian (LoG) operation on an image of the video information are greater than a threshold that is determined responsive to a second plurality of values that result from applying the LoG operation on a quantization matrix.
71. The method as recited in claim 70, further comprising:
visually adjusting the discontinuity that is not an artificial discontinuity.
72. The method as recited in claim 59, wherein the action of smoothing the artificial discontinuity on both sides of the block boundary comprises:
applying at least one smoothing filter to the artificial discontinuity.
73. The method as recited in claim 72, wherein the action of applying at least one smoothing filter to the artificial discontinuity comprises:
applying at least one of a Gaussian filter and a tent filter to the artificial discontinuity.
74. The method as recited in claim 59, further comprising:
decoding compressed video data to produce the video information.
75. The method as recited in claim 59, wherein the action of smoothing the artificial discontinuity on both sides of the block boundary comprises:
smoothing pixels of the video information; and
further comprising:
forwarding the video information for display on a screen.
76. The method as recited in claim 59, wherein the actions of detecting, determining, and smoothing are performed by a client device of a television-based entertainment environment.
77. The method as recited in claim 59, wherein the actions of detecting, determining, and smoothing comprise applying a spatio-temporally varying filter.
78. The method as recited in claim 59, wherein the actions of detecting, determining, and smoothing comprise applying a spatio-temporally varying filter that varies spatially across a block of the video information and temporally along a group of pictures.
79. The method as recited in claim 78, wherein the spatial variation is such that filtering increases near block boundaries of the block and decreases near the middle of the block.
80. The method as recited in claim 78, wherein the temporal variation is such that filtering decreases as pictures progress downstream along the group of pictures.
81. The method as recited in claim 59, wherein, when an intra image unit of the video information is under consideration,
the action of detecting comprises applying a Laplacian of Gaussian (LoG) operation on the intra image unit to produce a first plurality of values;
the action of determining comprises comparing the first plurality of values to a threshold that is determined responsive to a second plurality of values that result from applying the LoG operation on a quantization matrix; and
the action of smoothing comprises applying at least one of a Gaussian filter and a tent filter to the artificial discontinuity on both sides of the block boundary.
82. The method as recited in claim 59, wherein, when a non-intra image unit of the video information is under consideration,
the actions of detecting, determining, and smoothing comprise applying a spatio-temporally varying filter that varies spatially across the non-intra image unit and temporally along a group of images of the video information.
83. One or more electronically-accessible media comprising electronically-executable instructions that, when executed, direct an electronic device to perform the method as recited in claim 59.
84. A method for deblocking video, comprising:
determining whether an image unit of video information is an intra image unit or a non-intra image unit, the image unit having at least two boundaries;
if the image unit is determined to be an intra image unit, then performing actions comprising:
detecting one or more discontinuities at the at least two boundaries of the image unit;
determining whether the one or more discontinuities include one or more artificial discontinuities and whether the one or more discontinuities include one or more real discontinuities;
smoothing the one or more artificial discontinuities, if any, using a filter smoothing procedure; and
smoothing the one or more real discontinuities, if any, using a visual adjustment procedure; and
if the image unit is determined to be a non-intra image unit, then performing actions comprising:
ascertaining a position of the image unit in a group of images; and
smoothing one or more discontinuities of the image unit responsive (i) to individual locations of the one or more discontinuities relative to the at least two boundaries of the image unit and (ii) to the position of the image unit in the group of images.
85. The method as recited in claim 84, wherein the image unit comprises at least one of a frame and a macroblock.
86. The method as recited in claim 84, wherein the method is performed by a client device of a television-based entertainment environment.
87. The method as recited in claim 84, wherein the group of images comprises a group of pictures (GOP) in accordance with a Moving Pictures Expert Group (MPEG)-compliant encoding/decoding scheme.
88. The method as recited in claim 84, wherein the action of determining whether the one or more discontinuities include one or more artificial discontinuities and whether the one or more discontinuities include one or more real discontinuities comprises applying a Laplacian of Gaussian (LoG) operation on a quantization matrix, the quantization matrix previously being used to quantize the image unit.
89. The method as recited in claim 84, wherein the action of smoothing one or more discontinuities of the image unit responsive (i) to individual locations of the one or more discontinuities relative to the at least two boundaries of the image unit and (ii) to the position of the image unit in the group of images comprises smoothing the one or more discontinuities to a greater extent the closer the individual locations are to a boundary of the at least two boundaries and to a lesser extent the further the non-intra image unit is from an immediately-previous intra image unit.
90. The method as recited in claim 84, wherein the non-intra image unit comprises at least one of a predicted image unit and a bi-directional image unit.
91. A method for deblocking video, comprising:
analyzing intra macroblocks of video information to detect one or more discontinuities and to determine whether the one or more discontinuities include one or more artificial discontinuities and whether the one or more discontinuities include one or more real discontinuities; and
applying a spatio-temporally varying filter to non-intra macroblocks of the video information.
92. The method as recited in claim 91, further comprising:
determining whether a macroblock is an intra macroblock or a non-intra macroblock based on whether a frame that contains the macroblock is an intra frame or a non-intra frame, respectively.
93. The method as recited in claim 91, further comprising:
determining whether a macroblock is an intra macroblock or a non-intra macroblock on a macroblock-by-macroblock basis.
94. The method as recited in claim 91, further comprising:
smoothing the one or more artificial discontinuities, if any, using a filter smoothing procedure.
95. The method as recited in claim 91, further comprising:
smoothing the one or more real discontinuities, if any, using a visual adjustment procedure.
96. The method as recited in claim 91, wherein the action of analyzing intra macroblocks of video information comprises:
detecting the one or more discontinuities by using a Laplacian of Gaussian (LoG) operation on an image of the video information.
97. The method as recited in claim 91, wherein the action of analyzing intra macroblocks of video information comprises:
determining whether the one or more discontinuities include one or more artificial discontinuities and whether the one or more discontinuities include one or more real discontinuities by using a Laplacian of Gaussian (LoG) operation on a quantization matrix, the quantization matrix previously used to quantize an image of the video information.
98. The method as recited in claim 97, wherein the action of determining whether the one or more discontinuities include one or more artificial discontinuities and whether the one or more discontinuities include one or more real discontinuities by using a Laplacian of Gaussian (LoG) operation on a quantization matrix comprises:
determining whether the one or more discontinuities include one or more real discontinuities by using the LoG operation on a scaled quantization matrix.
99. The method as recited in claim 91, wherein the action of applying a spatio-temporally varying filter comprises:
increasing filtering for pixels proximate to macroblock boundaries of the non-intra macroblocks of the video information; and
decreasing filtering for pixels distant from the macroblock boundaries of the non-intra macroblocks of the video information.
US10/179,825 2002-06-24 2002-06-24 Video deblocking Abandoned US20030235250A1 (en)

Priority Applications (4)

Application Number Priority Date Filing Date Title
US10/179,825 US20030235250A1 (en) 2002-06-24 2002-06-24 Video deblocking
US11/191,131 US7567618B2 (en) 2002-06-24 2005-07-27 Video deblocking
US11/191,220 US7567619B2 (en) 2002-06-24 2005-07-27 Video deblocking
US11/191,204 US7660351B2 (en) 2002-06-24 2005-07-27 Video deblocking

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US10/179,825 US20030235250A1 (en) 2002-06-24 2002-06-24 Video deblocking

Related Child Applications (3)

Application Number Title Priority Date Filing Date
US11/191,131 Division US7567618B2 (en) 2002-06-24 2005-07-27 Video deblocking
US11/191,220 Division US7567619B2 (en) 2002-06-24 2005-07-27 Video deblocking
US11/191,204 Division US7660351B2 (en) 2002-06-24 2005-07-27 Video deblocking

Publications (1)

Publication Number Publication Date
US20030235250A1 true US20030235250A1 (en) 2003-12-25

Family

ID=29734992

Family Applications (4)

Application Number Title Priority Date Filing Date
US10/179,825 Abandoned US20030235250A1 (en) 2002-06-24 2002-06-24 Video deblocking
US11/191,131 Expired - Fee Related US7567618B2 (en) 2002-06-24 2005-07-27 Video deblocking
US11/191,220 Expired - Fee Related US7567619B2 (en) 2002-06-24 2005-07-27 Video deblocking
US11/191,204 Expired - Fee Related US7660351B2 (en) 2002-06-24 2005-07-27 Video deblocking

Family Applications After (3)

Application Number Title Priority Date Filing Date
US11/191,131 Expired - Fee Related US7567618B2 (en) 2002-06-24 2005-07-27 Video deblocking
US11/191,220 Expired - Fee Related US7567619B2 (en) 2002-06-24 2005-07-27 Video deblocking
US11/191,204 Expired - Fee Related US7660351B2 (en) 2002-06-24 2005-07-27 Video deblocking

Country Status (1)

Country Link
US (4) US20030235250A1 (en)

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040208392A1 (en) * 2003-03-17 2004-10-21 Raveendran Vijayalakshmi R. Method and apparatus for improving video quality of low bit-rate video
US20050053288A1 (en) * 2003-09-07 2005-03-10 Microsoft Corporation Bitstream-controlled post-processing filtering
US20050084012A1 (en) * 2003-09-07 2005-04-21 Microsoft Corporation In-loop deblocking for interlaced video
US20060146941A1 (en) * 2005-01-04 2006-07-06 Samsung Electronics Co., Ltd. Deblocking control method considering intra BL mode and multilayer video encoder/decoder using the same
US20060197691A1 (en) * 2002-11-15 2006-09-07 Irvine Ann C Apparatus and method for multiple description encoding
US20070171980A1 (en) * 2006-01-26 2007-07-26 Yen-Lin Lee Method and Related Apparatus For Decoding Video Streams
US20080089600A1 (en) * 2006-09-28 2008-04-17 Arthur Mitchell Image processing
US20080187292A1 (en) * 2005-01-19 2008-08-07 Nxp B.V. Device for and Method of Providing Operating Data and/or Data Associated with Playback Data to a Remote Device
US8787443B2 (en) 2010-10-05 2014-07-22 Microsoft Corporation Content adaptive deblocking during video encoding and decoding
US20150085935A1 (en) * 2013-09-26 2015-03-26 Qualcomm Incorporated Sub-prediction unit (pu) based temporal motion vector prediction in hevc and sub-pu design in 3d-hevc
US9042458B2 (en) 2011-04-01 2015-05-26 Microsoft Technology Licensing, Llc Multi-threaded implementations of deblock filtering
US9762927B2 (en) 2013-09-26 2017-09-12 Qualcomm Incorporated Sub-prediction unit (PU) based temporal motion vector prediction in HEVC and sub-PU design in 3D-HEVC

Families Citing this family (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090064242A1 (en) * 2004-12-23 2009-03-05 Bitband Technologies Ltd. Fast channel switching for digital tv
US8340098B2 (en) * 2005-12-07 2012-12-25 General Instrument Corporation Method and apparatus for delivering compressed video to subscriber terminals
KR100771879B1 (en) * 2006-08-17 2007-11-01 삼성전자주식회사 Method of deblocking filtering decreasing inner memory storage and a video processing device using the method
WO2008085425A2 (en) * 2006-12-28 2008-07-17 Thomson Licensing Detecting block artifacts in coded images and video
US8700792B2 (en) * 2008-01-31 2014-04-15 General Instrument Corporation Method and apparatus for expediting delivery of programming content over a broadband network
JP5050158B2 (en) * 2008-06-02 2012-10-17 株式会社メガチップス Transcoder
US20090304086A1 (en) * 2008-06-06 2009-12-10 Apple Inc. Method and system for video coder and decoder joint optimization
US8752092B2 (en) 2008-06-27 2014-06-10 General Instrument Corporation Method and apparatus for providing low resolution images in a broadcast system
US9357244B2 (en) * 2010-03-11 2016-05-31 Arris Enterprises, Inc. Method and system for inhibiting audio-video synchronization delay
US8976856B2 (en) * 2010-09-30 2015-03-10 Apple Inc. Optimized deblocking filters
WO2012148238A2 (en) * 2011-04-28 2012-11-01 삼성전자 주식회사 Method and apparatus for adjusting a data transmission rate in a wireless communication system
US8982948B2 (en) 2011-12-21 2015-03-17 Sony Corporation Video system with quantization matrix coding mechanism and method of operation thereof

Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5668598A (en) * 1995-03-27 1997-09-16 International Business Machines Corporation Motion video compression system with guaranteed bit production limits
US5852475A (en) * 1995-06-06 1998-12-22 Compression Labs, Inc. Transform artifact reduction process
US5903673A (en) * 1997-03-14 1999-05-11 Microsoft Corporation Digital video signal encoder and encoding method
US6040861A (en) * 1997-10-10 2000-03-21 International Business Machines Corporation Adaptive real-time encoding of video sequence employing image statistics
US6181742B1 (en) * 1998-01-26 2001-01-30 International Business Machines Corporation Single pass target allocation for video encoding
US6278735B1 (en) * 1998-03-19 2001-08-21 International Business Machines Corporation Real-time single pass variable bit rate control strategy and encoder
US6281942B1 (en) * 1997-08-11 2001-08-28 Microsoft Corporation Spatial and temporal filtering mechanism for digital motion video signals
US6373482B1 (en) * 1998-12-23 2002-04-16 Microsoft Corporation Method, system, and computer program product for modified blending between clip-map tiles
US20030035586A1 (en) * 2001-05-18 2003-02-20 Jim Chou Decoding compressed image data
US6539060B1 (en) * 1997-10-25 2003-03-25 Samsung Electronics Co., Ltd. Image data post-processing method for reducing quantization effect, apparatus therefor
US6816166B2 (en) * 2000-02-25 2004-11-09 International Business Machines Corporation Image conversion method, image processing apparatus, and image display apparatus

Family Cites Families (36)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0381067A3 (en) 1989-01-31 1992-08-12 Schlumberger Technologies, Inc. A method for registration of cad model to video images with added clutter
US5142592A (en) * 1990-12-17 1992-08-25 Moler Keith E Method and apparatus for detection of parallel edges in image processing
KR960014201B1 (en) * 1993-02-08 1996-10-14 김재창 System for detecting a moving butt
US5512956A (en) 1994-02-04 1996-04-30 At&T Corp. Adaptive spatial-temporal postprocessing for low bit-rate coded image sequences
US5446804A (en) * 1994-04-14 1995-08-29 Hewlett-Packard Company Magnifying digital image using edge mapping
US5802213A (en) * 1994-10-18 1998-09-01 Intel Corporation Encoding video signals using local quantization levels
KR0159575B1 (en) * 1994-10-31 1999-01-15 배순훈 Discrete processing apparatus between neighbor region in region based coding
US5684894A (en) * 1994-11-28 1997-11-04 Eastman Kodak Company Scale specific and robust line/edge encoding of images
US5850294A (en) * 1995-12-18 1998-12-15 Lucent Technologies Inc. Method and apparatus for post-processing images
US5668698A (en) * 1996-01-22 1997-09-16 General Motors Corporation Smart connector for an electrical device
JPH09271002A (en) * 1996-03-29 1997-10-14 Mitsubishi Electric Corp Video data distribution system
US5995080A (en) 1996-06-21 1999-11-30 Digital Equipment Corporation Method and apparatus for interleaving and de-interleaving YUV pixel data
US5796875A (en) * 1996-08-13 1998-08-18 Sony Electronics, Inc. Selective de-blocking filter for DCT compressed images
JP3466032B2 (en) * 1996-10-24 2003-11-10 富士通株式会社 Video encoding device and decoding device
FI103003B1 (en) * 1997-06-13 1999-03-31 Nokia Mobile Phones Ltd Filtering procedure, filter and mobile terminal
US6178205B1 (en) * 1997-12-12 2001-01-23 Vtel Corporation Video postfiltering with motion-compensated temporal filtering and/or spatial-adaptive filtering
US6611503B1 (en) 1998-05-22 2003-08-26 Tandberg Telecom As Method and apparatus for multimedia conferencing with dynamic bandwidth allocation
US6285801B1 (en) 1998-05-29 2001-09-04 Stmicroelectronics, Inc. Non-linear adaptive image filter for filtering noise such as blocking artifacts
US6320905B1 (en) * 1998-07-08 2001-11-20 Stream Machine Company Postprocessing system for removing blocking artifacts in block-based codecs
AU717480B2 (en) * 1998-08-01 2000-03-30 Korea Advanced Institute Of Science And Technology Loop-filtering method for image data and apparatus therefor
GB9822094D0 (en) * 1998-10-09 1998-12-02 Snell & Wilcox Ltd Improvements in data compression
KR100366643B1 (en) 1998-11-25 2003-02-19 삼성전자 주식회사 Method and apparatus for de-blocking
US6236764B1 (en) * 1998-11-30 2001-05-22 Equator Technologies, Inc. Image processing circuit and method for reducing a difference between pixel values across an image boundary
US6282322B1 (en) * 1998-12-03 2001-08-28 Philips Electronics North America Corp. System and method for compressing and decompressing images
US6449255B1 (en) * 1999-04-26 2002-09-10 Cisco Technology, Inc. Method and apparatus for managing packets using a real-time feedback signal
FI117534B (en) * 2000-01-21 2006-11-15 Nokia Corp A method for filtering digital images and a filter
EP1193649A1 (en) * 2000-09-28 2002-04-03 Sony International (Europe) GmbH Quality rating function for a discrete decoded picture
US7006255B2 (en) * 2001-03-29 2006-02-28 Sharp Laboratories Of America Adaptive image filtering based on a distance transform
US7003174B2 (en) * 2001-07-02 2006-02-21 Corel Corporation Removal of block encoding artifacts
US6983079B2 (en) * 2001-09-20 2006-01-03 Seiko Epson Corporation Reducing blocking and ringing artifacts in low-bit-rate coding
US7440504B2 (en) * 2001-09-24 2008-10-21 Broadcom Corporation Method and apparatus for performing deblocking filtering with interlace capability
CN101448162B (en) * 2001-12-17 2013-01-02 微软公司 Method for processing video image
US6963613B2 (en) 2002-04-01 2005-11-08 Broadcom Corporation Method of communicating between modules in a decoding system
US6950473B2 (en) * 2002-06-21 2005-09-27 Seiko Epson Corporation Hybrid technique for reducing blocking and ringing artifacts in low-bit-rate coding
US7031392B2 (en) * 2002-09-20 2006-04-18 Seiko Epson Corporation Method and apparatus for video deblocking
US7227901B2 (en) * 2002-11-21 2007-06-05 Ub Video Inc. Low-complexity deblocking filter

Patent Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5668598A (en) * 1995-03-27 1997-09-16 International Business Machines Corporation Motion video compression system with guaranteed bit production limits
US5852475A (en) * 1995-06-06 1998-12-22 Compression Labs, Inc. Transform artifact reduction process
US5920356A (en) * 1995-06-06 1999-07-06 Compressions Labs, Inc. Coding parameter adaptive transform artifact reduction process
US5903673A (en) * 1997-03-14 1999-05-11 Microsoft Corporation Digital video signal encoder and encoding method
US6281942B1 (en) * 1997-08-11 2001-08-28 Microsoft Corporation Spatial and temporal filtering mechanism for digital motion video signals
US6040861A (en) * 1997-10-10 2000-03-21 International Business Machines Corporation Adaptive real-time encoding of video sequence employing image statistics
US6539060B1 (en) * 1997-10-25 2003-03-25 Samsung Electronics Co., Ltd. Image data post-processing method for reducing quantization effect, apparatus therefor
US6181742B1 (en) * 1998-01-26 2001-01-30 International Business Machines Corporation Single pass target allocation for video encoding
US6278735B1 (en) * 1998-03-19 2001-08-21 International Business Machines Corporation Real-time single pass variable bit rate control strategy and encoder
US6373482B1 (en) * 1998-12-23 2002-04-16 Microsoft Corporation Method, system, and computer program product for modified blending between clip-map tiles
US6816166B2 (en) * 2000-02-25 2004-11-09 International Business Machines Corporation Image conversion method, image processing apparatus, and image display apparatus
US20030035586A1 (en) * 2001-05-18 2003-02-20 Jim Chou Decoding compressed image data

Cited By (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7564382B2 (en) 2002-11-15 2009-07-21 Qualcomm Incorporated Apparatus and method for multiple description encoding
US20060197691A1 (en) * 2002-11-15 2006-09-07 Irvine Ann C Apparatus and method for multiple description encoding
US7561073B2 (en) 2002-11-15 2009-07-14 Qualcomm Incorporated Apparatus and method for multiple description encoding
US20040208392A1 (en) * 2003-03-17 2004-10-21 Raveendran Vijayalakshmi R. Method and apparatus for improving video quality of low bit-rate video
US7995849B2 (en) * 2003-03-17 2011-08-09 Qualcomm, Incorporated Method and apparatus for improving video quality of low bit-rate video
US20050053288A1 (en) * 2003-09-07 2005-03-10 Microsoft Corporation Bitstream-controlled post-processing filtering
US20050084012A1 (en) * 2003-09-07 2005-04-21 Microsoft Corporation In-loop deblocking for interlaced video
US8687709B2 (en) * 2003-09-07 2014-04-01 Microsoft Corporation In-loop deblocking for interlaced video
US8625680B2 (en) 2003-09-07 2014-01-07 Microsoft Corporation Bitstream-controlled post-processing filtering
US20060146941A1 (en) * 2005-01-04 2006-07-06 Samsung Electronics Co., Ltd. Deblocking control method considering intra BL mode and multilayer video encoder/decoder using the same
US8542750B2 (en) 2005-01-04 2013-09-24 Samsung Electronics Co., Ltd. Deblocking control method considering intra BL mode and multilayer video encoder/decoder using the same
US20080187292A1 (en) * 2005-01-19 2008-08-07 Nxp B.V. Device for and Method of Providing Operating Data and/or Data Associated with Playback Data to a Remote Device
US9183489B2 (en) * 2005-01-19 2015-11-10 Nxp B.V. Device for and method of providing operating data and/or data associated with playback data to a remote device
US20070171980A1 (en) * 2006-01-26 2007-07-26 Yen-Lin Lee Method and Related Apparatus For Decoding Video Streams
US20080089600A1 (en) * 2006-09-28 2008-04-17 Arthur Mitchell Image processing
US8787443B2 (en) 2010-10-05 2014-07-22 Microsoft Corporation Content adaptive deblocking during video encoding and decoding
US10284868B2 (en) 2010-10-05 2019-05-07 Microsoft Technology Licensing, Llc Content adaptive deblocking during video encoding and decoding
US9042458B2 (en) 2011-04-01 2015-05-26 Microsoft Technology Licensing, Llc Multi-threaded implementations of deblock filtering
US10051290B2 (en) 2011-04-01 2018-08-14 Microsoft Technology Licensing, Llc Multi-threaded implementations of deblock filtering
US20150085935A1 (en) * 2013-09-26 2015-03-26 Qualcomm Incorporated Sub-prediction unit (pu) based temporal motion vector prediction in hevc and sub-pu design in 3d-hevc
US9667996B2 (en) * 2013-09-26 2017-05-30 Qualcomm Incorporated Sub-prediction unit (PU) based temporal motion vector prediction in HEVC and sub-PU design in 3D-HEVC
US9762927B2 (en) 2013-09-26 2017-09-12 Qualcomm Incorporated Sub-prediction unit (PU) based temporal motion vector prediction in HEVC and sub-PU design in 3D-HEVC

Also Published As

Publication number Publication date
US7660351B2 (en) 2010-02-09
US20050259745A1 (en) 2005-11-24
US7567618B2 (en) 2009-07-28
US20050262531A1 (en) 2005-11-24
US7567619B2 (en) 2009-07-28
US20050281340A1 (en) 2005-12-22

Similar Documents

Publication Publication Date Title
US7567618B2 (en) Video deblocking
US7620261B2 (en) Edge adaptive filtering system for reducing artifacts and method
US7283588B2 (en) Deblocking filter
EP1513349B1 (en) Bitstream-controlled post-processing video filtering
US10027966B2 (en) Apparatus and method for compressing pictures with ROI-dependent compression parameters
US8831111B2 (en) Decoding with embedded denoising
US20110299604A1 (en) Method and apparatus for adaptive video sharpening
US20090304086A1 (en) Method and system for video coder and decoder joint optimization
EP1564997A1 (en) Encoding and decoding of video images based on a quantization with an adaptive dead-zone size
US7787541B2 (en) Dynamic pre-filter control with subjective noise detector for video compression
US7848408B2 (en) Method and system for parameter generation for digital noise reduction based on bitstream properties
US20090080517A1 (en) Method and Related Device for Reducing Blocking Artifacts in Video Streams
US8077773B2 (en) Systems and methods for highly efficient video compression using selective retention of relevant visual detail
WO2013115972A1 (en) Video coding using eye tracking maps
US20090060368A1 (en) Method and System for an Adaptive HVS Filter
US20030031377A1 (en) Apparatus and method for removing block artifacts, and displaying device having the same apparatus
US6040875A (en) Method to compensate for a fade in a digital video input sequence
KR101051331B1 (en) Adaptive Fast Mode Decision Method Using Dynamic Threshold
US20100208800A1 (en) Method and decoder for decoding an image frame of an image frame sequence
US20090046783A1 (en) Method and Related Device for Decoding Video Streams
Davies A Modified Rate-Distortion Optimisation Strategy for Hybrid Wavelet Video Coding
Basavaraju et al. Modified pre and post processing methods for optimizing and improving the quality of VP8 video codec
Bodecek et al. Image compression in digital video broadcasting

Legal Events

Date Code Title Description
AS Assignment

Owner name: MICROSOFT CORPORATION, WASHINGTON

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:VARMA, ANKUR;MONI, SHANKAR;REEL/FRAME:013058/0150;SIGNING DATES FROM 20020620 TO 20020621

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION

AS Assignment

Owner name: MICROSOFT TECHNOLOGY LICENSING, LLC, WASHINGTON

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MICROSOFT CORPORATION;REEL/FRAME:034766/0001

Effective date: 20141014