US20200304837A1 - Adaptive syntax grouping and compression in video data - Google Patents

Adaptive syntax grouping and compression in video data Download PDF

Info

Publication number
US20200304837A1
US20200304837A1 US16/895,815 US202016895815A US2020304837A1 US 20200304837 A1 US20200304837 A1 US 20200304837A1 US 202016895815 A US202016895815 A US 202016895815A US 2020304837 A1 US2020304837 A1 US 2020304837A1
Authority
US
United States
Prior art keywords
video
video data
syntax
parameters
coded
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US16/895,815
Inventor
Yunfei Zheng
Dazhong ZHANG
Xiaosong ZHOU
Chris Y. Chung
Hsi-Jung Wu
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Apple Inc
Original Assignee
Apple Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Apple Inc filed Critical Apple Inc
Priority to US16/895,815 priority Critical patent/US20200304837A1/en
Assigned to APPLE INC. reassignment APPLE INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: WU, HSI-JUNG, CHUNG, CHRIS Y., ZHOU, XIAOSONG, ZHANG, DAZHONG, ZHENG, YUNFEI
Publication of US20200304837A1 publication Critical patent/US20200304837A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/70Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards

Definitions

  • video coding may include hierarchical picture partition techniques, such as quad-tree based partitioning, which may result in better adaptation to video content.
  • an image frame may be divided into many non-overlapping largest coding units (LCU's).
  • LCU non-overlapping largest coding units
  • Each LCU may be further partitioned into smaller coding units (CU's) in a quad-tree manner (each unit is divided into four smaller units).
  • the video coder may determine the quad-tree structure of an LCU.
  • each CU may have its own CU level syntax variables, such as skip flag, predMode, partMode, etc.
  • the CU's are encoded one by one with their own syntax variables.
  • a LCU may be considered a group of pixels, which includes a plurality of CU's as subgroups of pixels.
  • the video content or some specific video parameters may not vary significantly from CU to CU inside a LCU (or between some subgroups of pixels within a specific group).
  • the CU's may share similar encoding modes and syntax, which means there could be some redundancy among the syntax.
  • Context-adaptive binary arithmetic coding (CABAC), a form of entropy encoding, and some data compression processes may remove some of the redundancy, but may not be optimal.
  • FIG. 1 illustrates a communication system according to an embodiment of the present disclosure.
  • FIG. 2 illustrates a decoding system according to an embodiment of the present disclosure.
  • FIG. 3 illustrates a coding system according to an embodiment of the present disclosure.
  • FIG. 4 illustrates a decoding method according to an embodiment of the present disclosure.
  • FIG. 5 illustrates a coding method according to an embodiment of the present disclosure.
  • FIG. 1 illustrates a simplified block diagram of a communication system 100 according to an embodiment of the present invention.
  • the system 100 may include at least two terminals 110 - 120 interconnected via a network 150 .
  • a first terminal 110 may code video data at a local location for transmission to the other terminal 120 via the network 150 .
  • the second terminal 120 may receive the coded video data of the other terminal from the network 150 , decode the coded data and display the recovered video data.
  • Unidirectional data transmission may be common in media serving applications and the like.
  • FIG. 1 illustrates a second pair of terminals 130 , 140 provided to support bidirectional transmission of coded video that may occur, for example, during videoconferencing.
  • each terminal 130 , 140 may code video data captured at a local location for transmission to the other terminal via the network 150 .
  • Each terminal 130 , 140 also may receive the coded video data transmitted by the other terminal, may decode the coded data and may display the recovered video data at a local display device.
  • the terminals 110 - 140 may be illustrated as servers, personal computers and smart phones but the principles of the present invention may be not so limited. Embodiments of the present invention find application with laptop computers, tablet computers, media players and/or dedicated video conferencing equipment.
  • the network 150 represents any number of networks that convey coded video data among the terminals 110 - 140 , including for example wireline and/or wireless communication networks.
  • the communication network 150 may exchange data in circuit-switched and/or packet-switched channels.
  • Representative networks include telecommunications networks, local area networks, wide area networks and/or the Internet. For the purposes of the present discussion, the architecture and topology of the network 150 may be immaterial to the operation of the present invention unless explained herein below.
  • FIG. 2 may be a functional block diagram of a video decoding system 200 according to an embodiment of the present invention.
  • the video decoding system 200 may include a receiver 210 that receives encoded video data, a video decoder 220 , a controller 228 to manage operation of the system 200 and a display 234 to display the decoded video data.
  • the video decoder 220 may decode video sequence received.
  • the controller 228 may set the subgroup parameters for at least one of the subgroups of pixels in the video decoder, based upon at least one parameters corresponding to the group.
  • the parameters and the subgroup parameters in the encoded video may include parameters for groups of pixels, where each group includes a plurality of subgroups of pixels. Additional details of the parameters and the subgroup parameters will be described below.
  • the receiver 210 may receive video to be decoded by the system 200 .
  • the encoded video data may be received from a channel 212 , which may be a hardware/software link to a storage device which stores the encoded video data.
  • the receiver 210 may receive the encoded video data with other data, for example, coded audio data and/or ancillary data streams.
  • the receiver 210 may separate the encoded video data from the other data.
  • the video decoder 220 may perform decoding operation on the video sequence received from the receiver 210 .
  • the video decoder 220 may include a video decoder 222 , a reference picture cache 224 , and a prediction mode selection 226 operating under control of controller 228 .
  • the video decoder 222 may reconstruct coded video data received from the receiver 210 with reference to reference pictures stored in the reference picture cache 224 .
  • the video decoder 222 may output reconstructed video data to display 234 for display. Reconstructed video data of reference frames also may be stored to the reference picture cache 224 for use during decoding of subsequently received coded video data.
  • the video decoder 222 may perform decoding operations that invert coding operations performed by the video coder 330 (shown in FIG. 3 ).
  • the video decoder 222 may perform entropy decoding, dequantization, transform decoding, and filtering to generate recovered pixel block data. Quantization/dequantization operations may be lossy processes and, therefore, the recovered pixel block data likely will be a replica of the source pixel blocks that were coded by the video coder 330 (shown in FIG. 3 ) but may include some error.
  • the transform decoding may generate residual data; the video decoder 222 may use motion vectors associated with the pixel blocks to retrieve predicted pixel blocks from the reference picture cache 224 to be combined with the prediction residuals.
  • the prediction mode selector 226 may identify a temporal prediction mode being used for each pixel block of an encoded frame being decoded and request the needed data for the decoding to be read from the reference picture cache 224 .
  • the video decoder 220 may perform decoding operations according to a predetermined protocol, such as H.263, H.264, MPEG-2, HEVC. In its operation, the video decoder 220 may perform various decoding operations, including predictive decoding operations that exploit temporal and spatial redundancies in the encoded video sequence.
  • the coded video data therefore, may conform to a syntax specified by the protocol being used.
  • the parameters may be received as part of the syntax specified by the protocol in the coded video data, or appended as ancillary portion of the coded video data, to allow for backward compatibility.
  • the receiver 210 may receive additional data with the encoded video.
  • the additional data may be included as part of the encoded video frames.
  • the additional data may be used by the video decoder 220 to properly decode the data and/or to more accurately reconstruct the original video data.
  • FIG. 3 may be a functional block diagram of a video coding system 300 according to an embodiment of the present invention.
  • the system 300 may include a video source 310 that captures video image to be coded by the system 300 , a video coder 330 , a transmitter 340 , and a controller 350 to manage operation of the system 300 .
  • the video coder 330 may encode the video image into encoded video data using a plurality of subgroup parameters corresponding to a plurality of subgroups of pixels within a group.
  • the controller 350 may set the subgroup parameters for at least one of the subgroups of pixels in the video coder, based upon at least one parameters corresponding to the group.
  • the transmitter 340 may transmit the video data.
  • the video source 310 may provide video to be coded by the system 300 .
  • the video source 310 may be a storage device storing previously prepared video.
  • the video source 310 may be a camera that captures local image information as a video sequence.
  • Video data typically may be provided as a plurality of individual frames that impart motion when viewed in sequence. The frames themselves typically may be organized as a spatial array of pixels.
  • the system 300 may code and compress the image information for frames of the video sequence in real time, based upon one or more parameters.
  • the controller 350 may control the compression and coding in video coder 330 , based on the parameters.
  • the video coder 330 may perform motion compensated predictive coding, which codes an input frame predictively with reference to one or more previously-coded frames from the video sequence that were designated as “reference frames.” In this manner, the coding engine 332 codes differences between pixel blocks of an input frame and pixel blocks of reference frame(s) that may be selected as prediction reference(s) to the input frame.
  • the local video decoder 333 may decode coded video data of frames that may be designated as reference frames. Operations of the coding engine 332 typically may be lossy processes. When the coded video data may be decoded at a video decoder (not shown in FIG. 3 ), the recovered video sequence typically may be a replica of the source video sequence with some errors. The local video decoder 333 replicates decoding processes that will be performed by the video decoder on reference frames and may cause reconstructed reference frames to be stored in the reference picture cache 334 . In this manner, the system 300 may store copies of reconstructed reference frames locally that have common content as the reconstructed reference frames that will be obtained by a far-end video decoder (absent transmission errors).
  • the predictor 335 may perform prediction searches for the coding engine 332 . That is, for a new frame to be coded, the predictor 335 may search the reference picture cache 334 for image data (as candidate reference pixel blocks) that may serve as an appropriate prediction reference for the new frames. The predictor 335 may operate on a pixel block-by-pixel block basis to find appropriate prediction references. In some cases, as determined by search results obtained by the predictor 335 , an input frame may have prediction references drawn from multiple frames stored in the reference picture cache 334 .
  • the controller 350 may manage coding operations of the video coder 330 , including, for example, setting of parameters and subgroup parameters used for encoding the video data.
  • the transmitter 340 may buffer coded video data to prepare it for transmission via a communication channel 360 , which may be a hardware/software link to a storage device which would store the encoded video data.
  • the transmitter 340 may merge coded video data from the video coder 330 with other data to be transmitted, for example, coded audio data and/or ancillary data streams (sources not shown).
  • the controller 350 may manage operation of the system 300 .
  • the controller 350 may assign to each frame a certain frame type (either of its own accord or in cooperation with the controller 350 ), which may affect the coding techniques that may be applied to the respective frame. For example, frames often may be assigned as one of the following frame types:
  • An Intra Frame may be one that may be coded and decoded without using any other frame in the sequence as a source of prediction.
  • a Predictive Frame may be one that may be coded and decoded using intra prediction or inter prediction using at most one motion vector and reference index to predict the sample values of each block.
  • a Bi-directionally Predictive Frame may be one that may be coded and decoded using intra prediction or inter prediction using at most two motion vectors and reference indices to predict the sample values of each block.
  • Frames commonly may be parsed spatially into a plurality of pixel blocks (for example, blocks of 4 ⁇ 4, 8 ⁇ 8 or 16 ⁇ 16 pixels each) and coded on a pixel block-by-pixel block basis.
  • Pixel blocks may be coded predictively with reference to other coded pixel blocks as determined by the coding assignment applied to the pixel blocks' respective frames.
  • pixel blocks of I frames may be coded non-predictively or they may be coded predictively with reference to pixel blocks of the same frame (spatial prediction).
  • Pixel blocks of P frames may be coded non-predictively, via spatial prediction or via temporal prediction with reference to one previously coded reference frame.
  • Pixel blocks of B frames may be coded non-predictively, via spatial prediction or via temporal prediction with reference to one or two previously coded reference frames.
  • the video coder 330 may perform coding operations according to a predetermined protocol, such as H.263, H.264, MPEG-2, HEVC. In its operation, the video coder 330 may perform various compression operations, including predictive coding operations that exploit temporal and spatial redundancies in the input video sequence.
  • the coded video data therefore, may conform to a syntax specified by the protocol being used.
  • the transmitter 340 may transmit additional data with the encoded video.
  • the video coder 330 may include such data as part of the encoded video frames.
  • syntax of video data may be grouped and set by the controller 350 in video coder 330 to reduce redundancy.
  • PredMode of a CU is a flag that indicates whether a CU is encoded with inter mode or intra mode.
  • Each non-skip CU may expect to be encoded with a predMode value to indicate whether the CU will be predicted by inter or intra mode, and then related syntax may provide additional data for the prediction of the CU.
  • the predMode syntax for all the CU's in a LCU may be compressed by signaling a flag at the LCU level to indicate that it is an all inter-mode LCU. Then there would be no need to encode the predMode syntax for each individual CU inside a specific LCU.
  • encoding syntax and parameters may have redundancy reduced by similar implementation, saving even more bits.
  • Partition mode For example, Partition mode, skip mode, SAO parameters, residue flags, transform split, transform depth, etc.
  • more than one syntax can be grouped together with sharing the same signaling information at LCU level. For example, predMode and partMode can be grouped together.
  • the above example illustrates a syntax grouping that uses a “signaled default signaled exception” method. That is, the syntax signals a default setting of 1 for “all_inter_flag” for CU's within a LCU, and also signals an exception setting where CU's for a particular LCU are not all inter mode.
  • a “signaled default signaled exception” method that is, the syntax signals a default setting of 1 for “all_inter_flag” for CU's within a LCU, and also signals an exception setting where CU's for a particular LCU are not all inter mode.
  • many other possible methods for syntax grouping are possible.
  • a syntax grouping may use an “assumed default signaled exception” method.
  • the syntax may assume that all CU's are by default in inter 2N ⁇ 2N partition mode, and only signals partMode syntax for a CU if the CU is predicted in other partition modes (2N ⁇ N, N ⁇ 2N, or N ⁇ N).
  • the syntax at LCU e.g. all 2N ⁇ 2N flag
  • the syntax at LCU does not even need to be used, saving even more bits.
  • a syntax grouping may use a “signaled baseline signaled change” method, for more complex variables such as a numeric parameter (for example color values, coefficients).
  • a numeric parameter in coding may have some baseline or average value from which individual CU's may deviate only slightly from.
  • the LCU may be signaled by syntax for the baseline value of that numeric parameter, and individual CU's may be signaled by syntax for amount of deviation from the baseline value. If the deviations are small enough, the signaled deviation value may be coded with very few bits, saving significant number of bits in the encoded video data.
  • a syntax grouping may use a “signaled change hold until next change” method, for parameters that change infrequently.
  • the encoded video may start with a CU (a subgroup) that has syntax of an initial value for a parameter coded as a deviation from zero value. Then, the parameter would not need to be coded again until a CU (a subgroup) has a parameter value that deviates from the previous value. In other words, the video coder and the video decoder would assume that the parameter value is held from CU to CU, unless the parameter syntax is coded otherwise for a particular CU. Once the parameter value changes for a particular CU, that new parameter value holds until the next change.
  • the methods can be enabled or disabled by syntax at higher levels (ex. SPS, PPS, VPS, and/or slice header), by for example, including in the syntax of the higher levels, a control parameter that indicates whether all of the plurality of the subgroup parameters within the group are encoded in the encoded video data, or a control parameter that indicates a specific mode of compressing the plurality of the subgroup parameters within the group in the encoded video data.
  • the control parameter may correspond to controlling one group or a plurality of groups of pixels at the lower levels.
  • the different syntax grouping methods may have different data size efficiency depending on the nature of the various subgroup parameters intended for redundancy reduction. Thus, it may be necessary for the controller 350 to analyze the video data before encoding to determine and/or select an optimal method of syntax grouping for a specific set of subgroup parameters in a group. Additionally, the methods of syntax grouping may vary from one group of pixel to the next group of pixel.
  • the video coder 330 may encode additional syntax information to signal to the video decoder which method of syntax grouping is used to code a specific group of pixels.
  • the embodied syntax grouping may be implemented for non-contiguous groups of pixels or a group of non-contiguous subgroups of pixels.
  • FIG. 4 illustrates a decoding method 400 according to an embodiment of the present disclosure.
  • the system 200 may parse the syntax of the encoded video data.
  • the controller 228 may determine, from the parsed syntax, parameters and set subgroup parameters to be used for decoding. Some subgroup parameters may be absent for specific subgroups and the controller 228 may determine these absent subgroup parameters by calculations based on other subgroup parameters and parameters of group the group, in accordance with specific syntax grouping methods.
  • the controller 228 may control the video decoder 220 to decode video data using the set subgroup parameters. That is, some subgroups may not include corresponding subgroup parameters of specific types in the encoded video data.
  • the absent subgroup parameters may be derived from other subgroup parameters and/or parameters of the group by the video decoder 220 or the controller 228 in accordance to the syntax grouping methods.
  • FIG. 5 illustrates a coding method 500 according to an embodiment of the present disclosure.
  • the system 300 may analyze the video image from the video source 310 for subgroup parameters for encoding of the subgroups of pixels.
  • the controller 350 may determine if any subgroup parameters may be reduced for redundancy in encoding, according to predetermined syntax grouping methods.
  • the controller 350 may also select the optimal syntax grouping method for specific types of subgroup parameters based upon the analysis of the subgroup parameters.
  • the controller 350 may also set parameters for the group as part of the encoding of the group.
  • the parameters for the group may or may not be included in the encoded video data, and may be derived from some of the subgroup parameters.
  • the video coder 330 may encode video data using the reduced redundancy subgroup parameters. That is, some subgroups may not include corresponding subgroup parameters of specific types in the encoded video data.
  • the absent subgroup parameters may be derived later from other subgroup parameters and/or parameters of the group by a video decoder in accordance to the syntax grouping methods.
  • While the computer-readable medium may be described as a single medium, the term “computer-readable medium” includes a single medium or multiple media, such as a centralized or distributed database, and/or associated caches and servers that store one or more sets of instructions.
  • the term “computer-readable medium” shall also include any medium that is capable of storing, encoding or carrying a set of instructions for execution by a processor or that cause a computer system to perform any one or more of the embodiments disclosed herein.
  • the computer-readable medium may comprise a non-transitory computer-readable medium or media and/or comprise a transitory computer-readable medium or media.
  • the computer-readable medium can include a solid-state memory such as a memory card or other package that houses one or more non-volatile read-only memories.
  • the computer-readable medium can be a random access memory or other volatile re-writable memory.
  • the computer-readable medium can include a magneto-optical or optical medium, such as a disk or tapes or other storage device to capture carrier wave signals such as a signal communicated over a transmission medium. Accordingly, the disclosure is considered to include any computer-readable medium or other equivalents and successor media, in which data or instructions may be stored.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

An encoding system may include a video source that captures video image, a video coder, and a controller to manage operation of the system. The video coder may encode the video image into encoded video data using a plurality of subgroup parameters corresponding to a plurality of subgroups of pixels within a group. The controller may set the subgroup parameters for at least one of the subgroups of pixels in the video coder, based upon at least one parameters corresponding to the group. A decoding system may decode the video data based upon the motion prediction parameters.

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • This application is a continuation of U.S. patent application Ser. No. 14/289,082, filed May 28, 2014, now allowed, the disclosure of which is hereby incorporated by reference in its entirety.
  • BACKGROUND
  • In video recording, video coding may include hierarchical picture partition techniques, such as quad-tree based partitioning, which may result in better adaptation to video content. In such techniques, an image frame may be divided into many non-overlapping largest coding units (LCU's). Each LCU may be further partitioned into smaller coding units (CU's) in a quad-tree manner (each unit is divided into four smaller units). The video coder may determine the quad-tree structure of an LCU. Inside a LCU, each CU may have its own CU level syntax variables, such as skip flag, predMode, partMode, etc. The CU's are encoded one by one with their own syntax variables. For the purpose of discussion below, a LCU may be considered a group of pixels, which includes a plurality of CU's as subgroups of pixels.
  • Often the video content or some specific video parameters may not vary significantly from CU to CU inside a LCU (or between some subgroups of pixels within a specific group). The CU's may share similar encoding modes and syntax, which means there could be some redundancy among the syntax. Context-adaptive binary arithmetic coding (CABAC), a form of entropy encoding, and some data compression processes may remove some of the redundancy, but may not be optimal.
  • Thus, there is a need to reduce the redundancy by adaptively grouping the syntax in encoded video data.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 illustrates a communication system according to an embodiment of the present disclosure.
  • FIG. 2 illustrates a decoding system according to an embodiment of the present disclosure.
  • FIG. 3 illustrates a coding system according to an embodiment of the present disclosure.
  • FIG. 4 illustrates a decoding method according to an embodiment of the present disclosure.
  • FIG. 5 illustrates a coding method according to an embodiment of the present disclosure.
  • DETAILED DESCRIPTION
  • FIG. 1 illustrates a simplified block diagram of a communication system 100 according to an embodiment of the present invention. The system 100 may include at least two terminals 110-120 interconnected via a network 150. For unidirectional transmission of data, a first terminal 110 may code video data at a local location for transmission to the other terminal 120 via the network 150. The second terminal 120 may receive the coded video data of the other terminal from the network 150, decode the coded data and display the recovered video data. Unidirectional data transmission may be common in media serving applications and the like.
  • FIG. 1 illustrates a second pair of terminals 130, 140 provided to support bidirectional transmission of coded video that may occur, for example, during videoconferencing. For bidirectional transmission of data, each terminal 130, 140 may code video data captured at a local location for transmission to the other terminal via the network 150. Each terminal 130, 140 also may receive the coded video data transmitted by the other terminal, may decode the coded data and may display the recovered video data at a local display device.
  • In FIG. 1, the terminals 110-140 may be illustrated as servers, personal computers and smart phones but the principles of the present invention may be not so limited. Embodiments of the present invention find application with laptop computers, tablet computers, media players and/or dedicated video conferencing equipment. The network 150 represents any number of networks that convey coded video data among the terminals 110-140, including for example wireline and/or wireless communication networks. The communication network 150 may exchange data in circuit-switched and/or packet-switched channels. Representative networks include telecommunications networks, local area networks, wide area networks and/or the Internet. For the purposes of the present discussion, the architecture and topology of the network 150 may be immaterial to the operation of the present invention unless explained herein below.
  • FIG. 2 may be a functional block diagram of a video decoding system 200 according to an embodiment of the present invention.
  • The video decoding system 200 may include a receiver 210 that receives encoded video data, a video decoder 220, a controller 228 to manage operation of the system 200 and a display 234 to display the decoded video data. The video decoder 220 may decode video sequence received. The controller 228 may set the subgroup parameters for at least one of the subgroups of pixels in the video decoder, based upon at least one parameters corresponding to the group.
  • The parameters and the subgroup parameters in the encoded video may include parameters for groups of pixels, where each group includes a plurality of subgroups of pixels. Additional details of the parameters and the subgroup parameters will be described below.
  • The receiver 210 may receive video to be decoded by the system 200. The encoded video data may be received from a channel 212, which may be a hardware/software link to a storage device which stores the encoded video data. The receiver 210 may receive the encoded video data with other data, for example, coded audio data and/or ancillary data streams. The receiver 210 may separate the encoded video data from the other data.
  • The video decoder 220 may perform decoding operation on the video sequence received from the receiver 210. The video decoder 220 may include a video decoder 222, a reference picture cache 224, and a prediction mode selection 226 operating under control of controller 228. The video decoder 222 may reconstruct coded video data received from the receiver 210 with reference to reference pictures stored in the reference picture cache 224. The video decoder 222 may output reconstructed video data to display 234 for display. Reconstructed video data of reference frames also may be stored to the reference picture cache 224 for use during decoding of subsequently received coded video data.
  • The video decoder 222 may perform decoding operations that invert coding operations performed by the video coder 330 (shown in FIG. 3). The video decoder 222 may perform entropy decoding, dequantization, transform decoding, and filtering to generate recovered pixel block data. Quantization/dequantization operations may be lossy processes and, therefore, the recovered pixel block data likely will be a replica of the source pixel blocks that were coded by the video coder 330 (shown in FIG. 3) but may include some error. For pixel blocks coded predictively, the transform decoding may generate residual data; the video decoder 222 may use motion vectors associated with the pixel blocks to retrieve predicted pixel blocks from the reference picture cache 224 to be combined with the prediction residuals. The prediction mode selector 226 may identify a temporal prediction mode being used for each pixel block of an encoded frame being decoded and request the needed data for the decoding to be read from the reference picture cache 224.
  • The video decoder 220 may perform decoding operations according to a predetermined protocol, such as H.263, H.264, MPEG-2, HEVC. In its operation, the video decoder 220 may perform various decoding operations, including predictive decoding operations that exploit temporal and spatial redundancies in the encoded video sequence. The coded video data, therefore, may conform to a syntax specified by the protocol being used.
  • The parameters may be received as part of the syntax specified by the protocol in the coded video data, or appended as ancillary portion of the coded video data, to allow for backward compatibility.
  • In an embodiment, the receiver 210 may receive additional data with the encoded video.
  • The additional data may be included as part of the encoded video frames. The additional data may be used by the video decoder 220 to properly decode the data and/or to more accurately reconstruct the original video data.
  • FIG. 3 may be a functional block diagram of a video coding system 300 according to an embodiment of the present invention.
  • The system 300 may include a video source 310 that captures video image to be coded by the system 300, a video coder 330, a transmitter 340, and a controller 350 to manage operation of the system 300. The video coder 330 may encode the video image into encoded video data using a plurality of subgroup parameters corresponding to a plurality of subgroups of pixels within a group. The controller 350 may set the subgroup parameters for at least one of the subgroups of pixels in the video coder, based upon at least one parameters corresponding to the group. The transmitter 340 may transmit the video data.
  • The video source 310 may provide video to be coded by the system 300. In a media serving system, the video source 310 may be a storage device storing previously prepared video. In a videoconferencing system, the video source 310 may be a camera that captures local image information as a video sequence. Video data typically may be provided as a plurality of individual frames that impart motion when viewed in sequence. The frames themselves typically may be organized as a spatial array of pixels.
  • According to an embodiment, the system 300 may code and compress the image information for frames of the video sequence in real time, based upon one or more parameters. The controller 350 may control the compression and coding in video coder 330, based on the parameters.
  • As part of its operation, the video coder 330 may perform motion compensated predictive coding, which codes an input frame predictively with reference to one or more previously-coded frames from the video sequence that were designated as “reference frames.” In this manner, the coding engine 332 codes differences between pixel blocks of an input frame and pixel blocks of reference frame(s) that may be selected as prediction reference(s) to the input frame.
  • The local video decoder 333 may decode coded video data of frames that may be designated as reference frames. Operations of the coding engine 332 typically may be lossy processes. When the coded video data may be decoded at a video decoder (not shown in FIG. 3), the recovered video sequence typically may be a replica of the source video sequence with some errors. The local video decoder 333 replicates decoding processes that will be performed by the video decoder on reference frames and may cause reconstructed reference frames to be stored in the reference picture cache 334. In this manner, the system 300 may store copies of reconstructed reference frames locally that have common content as the reconstructed reference frames that will be obtained by a far-end video decoder (absent transmission errors).
  • The predictor 335 may perform prediction searches for the coding engine 332. That is, for a new frame to be coded, the predictor 335 may search the reference picture cache 334 for image data (as candidate reference pixel blocks) that may serve as an appropriate prediction reference for the new frames. The predictor 335 may operate on a pixel block-by-pixel block basis to find appropriate prediction references. In some cases, as determined by search results obtained by the predictor 335, an input frame may have prediction references drawn from multiple frames stored in the reference picture cache 334.
  • The controller 350 may manage coding operations of the video coder 330, including, for example, setting of parameters and subgroup parameters used for encoding the video data.
  • The transmitter 340 may buffer coded video data to prepare it for transmission via a communication channel 360, which may be a hardware/software link to a storage device which would store the encoded video data. The transmitter 340 may merge coded video data from the video coder 330 with other data to be transmitted, for example, coded audio data and/or ancillary data streams (sources not shown).
  • The controller 350 may manage operation of the system 300. During coding, the controller 350 may assign to each frame a certain frame type (either of its own accord or in cooperation with the controller 350), which may affect the coding techniques that may be applied to the respective frame. For example, frames often may be assigned as one of the following frame types:
  • An Intra Frame (I frame) may be one that may be coded and decoded without using any other frame in the sequence as a source of prediction.
  • A Predictive Frame (P frame) may be one that may be coded and decoded using intra prediction or inter prediction using at most one motion vector and reference index to predict the sample values of each block.
  • A Bi-directionally Predictive Frame (B frame) may be one that may be coded and decoded using intra prediction or inter prediction using at most
    Figure US20200304837A1-20200924-P00001
    two motion vectors and reference indices to predict the sample values of each block.
  • Frames commonly may be parsed spatially into a plurality of pixel blocks (for example, blocks of 4×4, 8×8 or 16×16 pixels each) and coded on a pixel block-by-pixel block basis. Pixel blocks may be coded predictively with reference to other coded pixel blocks as determined by the coding assignment applied to the pixel blocks' respective frames. For example, pixel blocks of I frames may be coded non-predictively or they may be coded predictively with reference to pixel blocks of the same frame (spatial prediction). Pixel blocks of P frames may be coded non-predictively, via spatial prediction or via temporal prediction with reference to one previously coded reference frame. Pixel blocks of B frames may be coded non-predictively, via spatial prediction or via temporal prediction with reference to one or two previously coded reference frames.
  • The video coder 330 may perform coding operations according to a predetermined protocol, such as H.263, H.264, MPEG-2, HEVC. In its operation, the video coder 330 may perform various compression operations, including predictive coding operations that exploit temporal and spatial redundancies in the input video sequence. The coded video data, therefore, may conform to a syntax specified by the protocol being used.
  • In an embodiment, the transmitter 340 may transmit additional data with the encoded video. The video coder 330 may include such data as part of the encoded video frames.
  • In an embodiment according the invention, syntax of video data may be grouped and set by the controller 350 in video coder 330 to reduce redundancy.
  • The syntax predMode is used as an illustrative example here. Similar methods of implementation may be used for other video data syntax. PredMode of a CU is a flag that indicates whether a CU is encoded with inter mode or intra mode. Each non-skip CU may expect to be encoded with a predMode value to indicate whether the CU will be predicted by inter or intra mode, and then related syntax may provide additional data for the prediction of the CU. In many cases, there is no intra mode inside any of the CU's in an entire LCU. In such a case, the predMode syntax for all the CU's in a LCU may be compressed by signaling a flag at the LCU level to indicate that it is an all inter-mode LCU. Then there would be no need to encode the predMode syntax for each individual CU inside a specific LCU.
  • In an example coding unit semantics, “all_inter_flag” may be coded for a LCU (or a group of pixels with multiple sub-groups). “all_inter_flag”=1 specifies that the current LCU includes CU's all predicted in inter mode, which means individual CU's in the current LCU no longer need their own predMode syntax. “all_inter_flag”=0 specifies that the current LCU includes CU's predicted in either intra or inter modes. If “all_inter_flag”=0, then individual CU's inside the current LCU may be further coded with their own predMode syntax. See the semantics below:
  • Descriptor
    coding_tree_unit( ) {
    xCtb = ( CtbAddrInRs % PicWidthInCtbsY ) <<
    CtbLog2SizeY
    yCtb = ( CtbAddrInRs / PicWidthInCtbsY ) <<
    CtbLog2SizeY
    ...
    all_inter_flag[ x0 ][ y0 ] ae(v)
    coding_quadtree( xCtb, yCtb, CtbLog2SizeY, 0,
    all_inter_flag )
    . . .
  • Descriptor
    coding_quadtree( x0, y0, log2CbSize,
    cqtDepth ,all_inter_flag ) {
    . . .
    Coding_unit(x0, y0, log2CbSize, all_inter_flag)
    . . .
  • Descriptor
    coding_unit( x0, y0, log2CbSize, all_inter_flag ) {
    . . .
    if( slice_type != I & !all_inter_flag )
    pred_mode_flag ae(v)
    . . .
  • Thus, significant number of bits may be saved in the encoded video data, which would reduce bandwidth during video data transmission and reduce storage space for the video file.
  • Similarly, other encoding syntax and parameters may have redundancy reduced by similar implementation, saving even more bits. For example, Partition mode, skip mode, SAO parameters, residue flags, transform split, transform depth, etc. Additionally, more than one syntax can be grouped together with sharing the same signaling information at LCU level. For example, predMode and partMode can be grouped together.
  • Additionally, while the above example illustrates the case in coding CU's and LCU's, other groupings of pixels comprising subgroups of pixels may implement similar methods in their syntax to reduce redundancy. For example, slice headers, SPS, PPS, region of interest (ROI), etc.
  • The above example illustrates a syntax grouping that uses a “signaled default signaled exception” method. That is, the syntax signals a default setting of 1 for “all_inter_flag” for CU's within a LCU, and also signals an exception setting where CU's for a particular LCU are not all inter mode. However, many other possible methods for syntax grouping are possible.
  • In an embodiment, a syntax grouping may use an “assumed default signaled exception” method. In this method, for example, the syntax may assume that all CU's are by default in inter 2N×2N partition mode, and only signals partMode syntax for a CU if the CU is predicted in other partition modes (2N×N, N×2N, or N×N). By this way, the syntax at LCU (e.g. all 2N×2N flag) does not even need to be used, saving even more bits.
  • In an embodiment, a syntax grouping may use a “signaled baseline signaled change” method, for more complex variables such as a numeric parameter (for example color values, coefficients). A numeric parameter in coding may have some baseline or average value from which individual CU's may deviate only slightly from. In such a case, the LCU may be signaled by syntax for the baseline value of that numeric parameter, and individual CU's may be signaled by syntax for amount of deviation from the baseline value. If the deviations are small enough, the signaled deviation value may be coded with very few bits, saving significant number of bits in the encoded video data.
  • In an embodiment, a syntax grouping may use a “signaled change hold until next change” method, for parameters that change infrequently. In such cases, the encoded video may start with a CU (a subgroup) that has syntax of an initial value for a parameter coded as a deviation from zero value. Then, the parameter would not need to be coded again until a CU (a subgroup) has a parameter value that deviates from the previous value. In other words, the video coder and the video decoder would assume that the parameter value is held from CU to CU, unless the parameter syntax is coded otherwise for a particular CU. Once the parameter value changes for a particular CU, that new parameter value holds until the next change.
  • As illustrated above in the various embodiments of methods for syntax grouping, the methods can be enabled or disabled by syntax at higher levels (ex. SPS, PPS, VPS, and/or slice header), by for example, including in the syntax of the higher levels, a control parameter that indicates whether all of the plurality of the subgroup parameters within the group are encoded in the encoded video data, or a control parameter that indicates a specific mode of compressing the plurality of the subgroup parameters within the group in the encoded video data. The control parameter may correspond to controlling one group or a plurality of groups of pixels at the lower levels.
  • As illustrated above in the various embodiments of methods for syntax grouping, the different syntax grouping methods may have different data size efficiency depending on the nature of the various subgroup parameters intended for redundancy reduction. Thus, it may be necessary for the controller 350 to analyze the video data before encoding to determine and/or select an optimal method of syntax grouping for a specific set of subgroup parameters in a group. Additionally, the methods of syntax grouping may vary from one group of pixel to the next group of pixel. The video coder 330 may encode additional syntax information to signal to the video decoder which method of syntax grouping is used to code a specific group of pixels.
  • While the above examples illustrate scenarios of implementations for a group of pixels containing contiguous subgroups of pixels, the embodied syntax grouping may be implemented for non-contiguous groups of pixels or a group of non-contiguous subgroups of pixels.
  • FIG. 4 illustrates a decoding method 400 according to an embodiment of the present disclosure.
  • At block 410, the system 200 may parse the syntax of the encoded video data.
  • At block 420, the controller 228 may determine, from the parsed syntax, parameters and set subgroup parameters to be used for decoding. Some subgroup parameters may be absent for specific subgroups and the controller 228 may determine these absent subgroup parameters by calculations based on other subgroup parameters and parameters of group the group, in accordance with specific syntax grouping methods.
  • At block 430, the controller 228 may control the video decoder 220 to decode video data using the set subgroup parameters. That is, some subgroups may not include corresponding subgroup parameters of specific types in the encoded video data. The absent subgroup parameters may be derived from other subgroup parameters and/or parameters of the group by the video decoder 220 or the controller 228 in accordance to the syntax grouping methods.
  • FIG. 5 illustrates a coding method 500 according to an embodiment of the present disclosure.
  • At block 510, the system 300 may analyze the video image from the video source 310 for subgroup parameters for encoding of the subgroups of pixels.
  • At block 520, the controller 350 may determine if any subgroup parameters may be reduced for redundancy in encoding, according to predetermined syntax grouping methods. The controller 350 may also select the optimal syntax grouping method for specific types of subgroup parameters based upon the analysis of the subgroup parameters. The controller 350 may also set parameters for the group as part of the encoding of the group. The parameters for the group may or may not be included in the encoded video data, and may be derived from some of the subgroup parameters.
  • At block 530, the video coder 330 may encode video data using the reduced redundancy subgroup parameters. That is, some subgroups may not include corresponding subgroup parameters of specific types in the encoded video data. The absent subgroup parameters may be derived later from other subgroup parameters and/or parameters of the group by a video decoder in accordance to the syntax grouping methods.
  • It is appreciated that the disclosure is not limited to the described embodiments, and that any number of scenarios and embodiments in which conflicting appointments exist may be resolved.
  • Although the disclosure has been described with reference to several exemplary embodiments, it is understood that the words that have been used are words of description and illustration, rather than words of limitation. Changes may be made within the purview of the appended claims, as presently stated and as amended, without departing from the scope and spirit of the disclosure in its aspects. Although the disclosure has been described with reference to particular means, materials and embodiments, the disclosure is not intended to be limited to the particulars disclosed; rather the disclosure extends to all functionally equivalent structures, methods, and uses such as are within the scope of the appended claims.
  • While the computer-readable medium may be described as a single medium, the term “computer-readable medium” includes a single medium or multiple media, such as a centralized or distributed database, and/or associated caches and servers that store one or more sets of instructions. The term “computer-readable medium” shall also include any medium that is capable of storing, encoding or carrying a set of instructions for execution by a processor or that cause a computer system to perform any one or more of the embodiments disclosed herein.
  • The computer-readable medium may comprise a non-transitory computer-readable medium or media and/or comprise a transitory computer-readable medium or media. In a particular non-limiting, exemplary embodiment, the computer-readable medium can include a solid-state memory such as a memory card or other package that houses one or more non-volatile read-only memories. Further, the computer-readable medium can be a random access memory or other volatile re-writable memory. Additionally, the computer-readable medium can include a magneto-optical or optical medium, such as a disk or tapes or other storage device to capture carrier wave signals such as a signal communicated over a transmission medium. Accordingly, the disclosure is considered to include any computer-readable medium or other equivalents and successor media, in which data or instructions may be stored.
  • Although the present application describes specific embodiments which may be implemented as code segments in computer-readable media, it is to be understood that dedicated hardware implementations, such as application specific integrated circuits, programmable logic arrays and other hardware devices, can be constructed to implement one or more of the embodiments described herein. Applications that may include the various embodiments set forth herein may broadly include a variety of electronic and computer systems. Accordingly, the present application may encompass software, firmware, and hardware implementations, or combinations thereof.
  • The present specification describes components and functions that may be implemented in particular embodiments with reference to particular standards and protocols, the disclosure is not limited to such standards and protocols. Such standards are periodically superseded by faster or more efficient equivalents having essentially the same functions. Accordingly, replacement standards and protocols having the same or similar functions are considered equivalents thereof.
  • The illustrations of the embodiments described herein are intended to provide a general understanding of the various embodiments. The illustrations are not intended to serve as a complete description of all of the elements and features of apparatus and systems that utilize the structures or methods described herein. Many other embodiments may be apparent to those of skill in the art upon reviewing the disclosure. Other embodiments may be utilized and derived from the disclosure, such that structural and logical substitutions and changes may be made without departing from the scope of the disclosure. Additionally, the illustrations are merely representational and may not be drawn to scale. Certain proportions within the illustrations may be exaggerated, while other proportions may be minimized. Accordingly, the disclosure and the figures are to be regarded as illustrative rather than restrictive.
  • One or more embodiments of the disclosure may be referred to herein, individually and/or collectively, by the term “disclosure” merely for convenience and without intending to voluntarily limit the scope of this application to any particular disclosure or inventive concept. Moreover, although specific embodiments have been illustrated and described herein, it should be appreciated that any subsequent arrangement designed to achieve the same or similar purpose may be substituted for the specific embodiments shown. This disclosure is intended to cover any and all subsequent adaptations or variations of various embodiments. Combinations of the above embodiments, and other embodiments not specifically described herein, will be apparent to those of skill in the art upon reviewing the description.
  • In addition, in the foregoing Detailed Description, various features may be grouped together or described in a single embodiment for the purpose of streamlining the disclosure. This disclosure is not to be interpreted as reflecting an intention that the claimed embodiments require more features than are expressly recited in each claim. Rather, as the following claims reflect, inventive subject matter may be directed to less than all of the features of any of the disclosed embodiments. Thus, the following claims are incorporated into the Detailed Description, with each claim standing on its own as defining separately claimed subject matter.
  • The above disclosed subject matter is to be considered illustrative, and not restrictive, and the appended claims are intended to cover all such modifications, enhancements, and other embodiments which fall within the true spirit and scope of the present disclosure. Thus, to the maximum extent allowed by law, the scope of the present disclosure is to be determined by the broadest permissible interpretation of the following claims and their equivalents, and shall not be restricted or limited by the foregoing detailed description.

Claims (2)

1. A system comprising:
a receiver receiving encoded video data;
a video decoder decoding the encoded video data using a plurality of subgroup parameters corresponding to a plurality of subgroups of pixels within a group; and
a controller setting the subgroup parameters for at least one of the subgroups of pixels in the video decoder, based upon at least one parameters corresponding to the group,
wherein the encoded video data does not include the subgroup parameters for the at least one of the subgroups of pixels but includes the subgroup parameters for at least one other of the subgroups of pixels.
2.-23. (canceled)
US16/895,815 2014-05-28 2020-06-08 Adaptive syntax grouping and compression in video data Abandoned US20200304837A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US16/895,815 US20200304837A1 (en) 2014-05-28 2020-06-08 Adaptive syntax grouping and compression in video data

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US14/289,082 US10715833B2 (en) 2014-05-28 2014-05-28 Adaptive syntax grouping and compression in video data using a default value and an exception value
US16/895,815 US20200304837A1 (en) 2014-05-28 2020-06-08 Adaptive syntax grouping and compression in video data

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US14/289,082 Continuation US10715833B2 (en) 2014-05-28 2014-05-28 Adaptive syntax grouping and compression in video data using a default value and an exception value

Publications (1)

Publication Number Publication Date
US20200304837A1 true US20200304837A1 (en) 2020-09-24

Family

ID=54703318

Family Applications (2)

Application Number Title Priority Date Filing Date
US14/289,082 Active 2035-08-09 US10715833B2 (en) 2014-05-28 2014-05-28 Adaptive syntax grouping and compression in video data using a default value and an exception value
US16/895,815 Abandoned US20200304837A1 (en) 2014-05-28 2020-06-08 Adaptive syntax grouping and compression in video data

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US14/289,082 Active 2035-08-09 US10715833B2 (en) 2014-05-28 2014-05-28 Adaptive syntax grouping and compression in video data using a default value and an exception value

Country Status (1)

Country Link
US (2) US10715833B2 (en)

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR102660119B1 (en) 2018-09-18 2024-04-22 후아웨이 테크놀러지 컴퍼니 리미티드 Video encoders, video decoders and methods
CN112740668B (en) 2018-09-18 2022-05-24 华为技术有限公司 Decoding method, device and system
EP3841752A4 (en) * 2018-11-19 2021-10-20 Huawei Technologies Co., Ltd. Method and apparatus for hierarchical signaling of enabling and disabling prediction mechanisms for video coding
EP3987803A1 (en) * 2019-06-24 2022-04-27 InterDigital VC Holdings, Inc. Method and apparatus for signaling decoding data using high level syntax elements
CN111918142B (en) * 2020-07-29 2023-04-11 杭州叙简科技股份有限公司 Smoothing method, device, equipment and medium for converting national standard video code stream into RTP stream

Family Cites Families (67)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6092041A (en) * 1996-08-22 2000-07-18 Motorola, Inc. System and method of encoding and decoding a layered bitstream by re-applying psychoacoustic analysis in the decoder
JP2901560B2 (en) * 1996-12-24 1999-06-07 新潟日本電気株式会社 Color image forming equipment
US6687384B1 (en) * 2000-03-27 2004-02-03 Sarnoff Corporation Method and apparatus for embedding data in encoded digital bitstreams
US20020036990A1 (en) * 2000-09-11 2002-03-28 Ben Chodor Still picture capture, editing and distribution system utilizing a publicly accessible wide area network
US7151778B2 (en) * 2001-04-18 2006-12-19 Brocade Communications Systems, Inc. Frame filtering of fibre channel packets
US20040066466A1 (en) * 2002-10-08 2004-04-08 Macinnis Alexander Progressive conversion of interlaced video based on coded bitstream analysis
JP2006525722A (en) * 2003-05-06 2006-11-09 コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ Video encoding and decoding method and corresponding encoding and decoding apparatus
US7848432B2 (en) * 2003-07-02 2010-12-07 Broadcom Corporation System and method for efficiently storing macroblocks in SD-RAM
US6894628B2 (en) * 2003-07-17 2005-05-17 Fraunhofer-Gesellschaft Zur Forderung Der Angewandten Forschung E.V. Apparatus and methods for entropy-encoding or entropy-decoding using an initialization of context variables
FI115589B (en) * 2003-10-14 2005-05-31 Nokia Corp Encoding and decoding redundant images
US20080123947A1 (en) 2005-07-22 2008-05-29 Mitsubishi Electric Corporation Image encoding device, image decoding device, image encoding method, image decoding method, image encoding program, image decoding program, computer readable recording medium having image encoding program recorded therein
US20070050149A1 (en) * 2005-08-23 2007-03-01 Michael Raskin Method for Modeling, Analyzing, and Predicting Disjunctive Systems
CN101507280B (en) * 2006-08-25 2012-12-26 汤姆逊许可公司 Methods and apparatus for reduced resolution partitioning
WO2008057308A2 (en) * 2006-11-08 2008-05-15 Thomson Licensing Methods and apparatus for in-loop de-artifact filtering
JP5377287B2 (en) * 2007-03-02 2013-12-25 パナソニック株式会社 Post filter, decoding device, and post filter processing method
US8121966B2 (en) * 2008-06-05 2012-02-21 International Business Machines Corporation Method and system for automated integrated server-network-storage disaster recovery planning
US8780986B2 (en) * 2008-06-06 2014-07-15 Apple Inc. Refresh pixel group selection and coding adjustment
US8494058B2 (en) * 2008-06-23 2013-07-23 Mediatek Inc. Video/image processing apparatus with motion estimation sharing, and related method and machine readable medium
CN101790096B (en) 2009-01-24 2013-03-13 华为技术有限公司 Encoding and decoding method and device based on double prediction
US8266604B2 (en) * 2009-01-26 2012-09-11 Microsoft Corporation Transactional memory compatibility management
JP5798556B2 (en) * 2010-07-15 2015-10-21 シャープ株式会社 Decoding device, encoding device
HRP20231669T1 (en) * 2010-09-02 2024-03-15 Lg Electronics, Inc. Inter prediction
US8538193B2 (en) * 2010-09-28 2013-09-17 Sharp Laboratories Of America, Inc. Methods and systems for image enhancement and estimation of compression noise
US20120114034A1 (en) * 2010-11-08 2012-05-10 Mediatek Inc. Method and Apparatus of Delta Quantization Parameter Processing for High Efficiency Video Coding
WO2012096614A2 (en) * 2011-01-14 2012-07-19 Telefonaktiebolaget L M Ericsson (Publ) A method for filter control and a filtering control device
WO2012094750A1 (en) * 2011-01-14 2012-07-19 Ebrisk Video Inc. Adaptive loop filtering using multiple filter shapes
WO2012110482A2 (en) * 2011-02-14 2012-08-23 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Noise generation in audio codecs
US9066110B2 (en) * 2011-03-08 2015-06-23 Texas Instruments Incorporated Parsing friendly and error resilient merge flag coding in video coding
JP5875236B2 (en) * 2011-03-09 2016-03-02 キヤノン株式会社 Image encoding device, image encoding method and program, image decoding device, image decoding method and program
EP2702765A1 (en) * 2011-04-28 2014-03-05 Koninklijke Philips N.V. Method and apparatus for generating an image coding signal
US9338465B2 (en) * 2011-06-30 2016-05-10 Sharp Kabushiki Kaisha Context initialization based on decoder picture buffer
CN102857755B (en) * 2011-07-01 2016-12-14 华为技术有限公司 The method and apparatus determining transform block size
KR101287786B1 (en) * 2011-09-22 2013-07-18 엘지전자 주식회사 Method for displaying stereoscopic image and display apparatus thereof
US11184623B2 (en) * 2011-09-26 2021-11-23 Texas Instruments Incorporated Method and system for lossless coding mode in video coding
EP2769550A4 (en) * 2011-10-14 2016-03-09 Mediatek Inc Method and apparatus for loop filtering
US9838692B2 (en) * 2011-10-18 2017-12-05 Qualcomm Incorporated Detecting availabilities of neighboring video units for video coding
US8977478B2 (en) * 2011-10-28 2015-03-10 Chrysler Group Llc Method of setting a control parameter for emissions robustness
KR101965388B1 (en) * 2011-11-04 2019-04-04 주식회사 골드피크이노베이션즈 Methods of coding quantization parameter and appratuses using the same
US20140286406A1 (en) * 2011-11-08 2014-09-25 Samsung Electronics Co., Ltd. Method for determining quantization parameters on basis of size of conversion block, and device for same
US9304857B2 (en) * 2011-12-12 2016-04-05 Cleversafe, Inc. Retrieving data from a distributed storage network
US20130188691A1 (en) * 2012-01-20 2013-07-25 Sony Corporation Quantization matrix design for hevc standard
KR20130086004A (en) * 2012-01-20 2013-07-30 삼성전자주식회사 Method and apparatus for parallel entropy encoding, method and apparatus for parallel entropy decoding
US9172608B2 (en) * 2012-02-07 2015-10-27 Cloudera, Inc. Centralized configuration and monitoring of a distributed computing cluster
EP2822276B1 (en) * 2012-02-29 2018-11-07 LG Electronics Inc. Inter-layer prediction method and apparatus using same
US9591331B2 (en) * 2012-03-28 2017-03-07 Qualcomm Incorporated Merge signaling and loop filter on/off signaling
US9491458B2 (en) 2012-04-12 2016-11-08 Qualcomm Incorporated Scalable video coding prediction with non-causal information
CN103379333B (en) * 2012-04-25 2018-12-04 浙江大学 The decoding method and its corresponding device of decoding method, video sequence code stream
CN104396252B (en) * 2012-04-25 2018-05-04 三星电子株式会社 Use the multi-view point video decoding method and its device of the reference picture collection predicted for multi-view point video
GB2501535A (en) * 2012-04-26 2013-10-30 Sony Corp Chrominance Processing in High Efficiency Video Codecs
EP2868090A4 (en) * 2012-07-01 2016-02-17 Sharp Kk Device for signaling a long-term reference picture in a parameter set
US9361693B2 (en) * 2012-07-06 2016-06-07 Navico Holding As Adjusting parameters of marine electronics data
US20140044162A1 (en) * 2012-08-08 2014-02-13 Qualcomm Incorporated Adaptive inference mode information derivation in scalable video coding
CA2807404C (en) * 2012-09-04 2017-04-04 Research In Motion Limited Methods and devices for inter-layer prediction in scalable video compression
US9374583B2 (en) * 2012-09-20 2016-06-21 Qualcomm Incorporated Video coding with improved random access point picture behaviors
CN103310411B (en) * 2012-09-25 2017-04-12 中兴通讯股份有限公司 Image local reinforcement method and device
GB2509901A (en) * 2013-01-04 2014-07-23 Canon Kk Image coding methods based on suitability of base layer (BL) prediction data, and most probable prediction modes (MPMs)
US9445130B2 (en) * 2013-01-09 2016-09-13 Qualcomm Incorporated Blockiness metric for large HEVC block artifacts
CN105164370B (en) * 2013-02-25 2019-11-01 开拓工程股份有限公司 Integrated downhole system with multiple telemetry subsystems
JP5751455B2 (en) * 2013-03-04 2015-07-22 株式会社デンソー Rotating electrical machine control device
WO2014137175A1 (en) * 2013-03-06 2014-09-12 삼성전자 주식회사 Method and apparatus for encoding scalable video using selective denoising filtering, and method and apparatus for decoding scalable video using selective denoising filtering
AU2013202653A1 (en) * 2013-04-05 2014-10-23 Canon Kabushiki Kaisha Method, apparatus and system for generating intra-predicted samples
US20140321532A1 (en) * 2013-04-26 2014-10-30 Devadutta Ghat Techniques for coordinating parallel video transcoding
JP2014236264A (en) * 2013-05-31 2014-12-15 ソニー株式会社 Image processing apparatus, image processing method and program
US9787989B2 (en) * 2013-06-11 2017-10-10 Blackberry Limited Intra-coding mode-dependent quantization tuning
US9955163B2 (en) * 2013-06-17 2018-04-24 Google Llc Two pass quantization of video data
US20150206177A1 (en) * 2013-10-18 2015-07-23 Google Inc. Using content selection urls
EP3138293A4 (en) * 2014-04-29 2017-05-24 Microsoft Technology Licensing, LLC Encoder-side decisions for sample adaptive offset filtering

Also Published As

Publication number Publication date
US10715833B2 (en) 2020-07-14
US20150350646A1 (en) 2015-12-03

Similar Documents

Publication Publication Date Title
US11991386B2 (en) Methods and apparatus of video coding using history-based motion vector prediction
US20200304837A1 (en) Adaptive syntax grouping and compression in video data
CN113727101B (en) Encoder, decoder and corresponding inter-frame prediction method
US20180332298A1 (en) Bidirectional Prediction In Video Compression
AU2020246735B2 (en) An encoder, a decoder and corresponding methods for intra prediction
US20170078683A1 (en) Palette predictor initialization and merge for video coding
US10212434B2 (en) Palette entries coding in video coding
US20200014946A1 (en) Method and apparatus for block partition with non-uniform quad split
US20130336395A1 (en) Unification of signaling lossless coding mode and pulse code modulation (pcm) mode in video coding
CN114710679B9 (en) Small chroma block size restriction in video coding
US20110249742A1 (en) Coupled video pre-processor and codec including reference picture filter that minimizes coding expense during pre-processing mode transitions
US20230068657A1 (en) Selecting a coding method for suffix values for displacement vector differences based on value intervals
US20210409746A1 (en) Signaling of lossless coding in video coding
US11985311B2 (en) Prediction mode signaling in video coding
US11683514B2 (en) Method and apparatus for video coding for machine
US20140321528A1 (en) Video encoding and/or decoding method and video encoding and/or decoding apparatus
US20220150471A1 (en) Decoder Side Displacement Vector Prediction
CN111263147A (en) Inter-frame prediction method and related device
EP4320861A1 (en) Video coding with dynamic groups of pictures
CN115550666A (en) Encoding method, decoding method, computing device, and medium for video data
CN114342374A (en) Method and apparatus for video encoding and decoding using palette mode

Legal Events

Date Code Title Description
AS Assignment

Owner name: APPLE INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:ZHENG, YUNFEI;ZHANG, DAZHONG;ZHOU, XIAOSONG;AND OTHERS;SIGNING DATES FROM 20140522 TO 20140528;REEL/FRAME:053151/0722

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE AFTER FINAL ACTION FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: ADVISORY ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION