WO2023222521A1 - Sei conçues pour de multiples points de conformité - Google Patents

Sei conçues pour de multiples points de conformité Download PDF

Info

Publication number
WO2023222521A1
WO2023222521A1 PCT/EP2023/062663 EP2023062663W WO2023222521A1 WO 2023222521 A1 WO2023222521 A1 WO 2023222521A1 EP 2023062663 W EP2023062663 W EP 2023062663W WO 2023222521 A1 WO2023222521 A1 WO 2023222521A1
Authority
WO
WIPO (PCT)
Prior art keywords
conformance
points
point
video stream
syntax structure
Prior art date
Application number
PCT/EP2023/062663
Other languages
English (en)
Inventor
Philippe Bordes
Edouard Francois
Philippe DE LAGRANGE
Franck Galpin
Original Assignee
Interdigital Ce Patent Holdings, Sas
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Interdigital Ce Patent Holdings, Sas filed Critical Interdigital Ce Patent Holdings, Sas
Publication of WO2023222521A1 publication Critical patent/WO2023222521A1/fr

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/70Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards

Definitions

  • At least one of the present embodiments generally relates to a method and a device for checking the conformance of a video stream and in particular, a method and a device for providing metadata representative of at least one conformance point among a plurality of conformance points, a conformance point specifying which data reconstructed from the video stream is to be tested for conformance.
  • video coding schemes usually employ predictions and transforms to leverage spatial and temporal redundancies in a video content.
  • pictures of the video content are divided into blocks of samples (i.e. Pixels), these blocks being then partitioned into one or more sub-blocks, called original sub-blocks in the following.
  • An intra or inter prediction is then applied to each sub-block to exploit intra or inter image correlations.
  • a predictor sub-block is determined for each original subblock.
  • a sub-block representing a difference between the original sub-block and the predictor sub-block is transformed, quantized and entropy coded to generate an encoded video stream.
  • the compressed data is decoded by inverse processes corresponding to the transform, quantization and entropic coding.
  • Bitstream conformance' it is a requirement that the structure of the bitstream, as defined in a standard specification, should be respected.
  • Hypothetical stream scheduler (HSS) conformance A hypothetical delivery mechanism used for checking the conformance of a bitstream or of a decoder with regards to the timing and data flow of the input of a bitstream into the hypothetical reference decoder, including buffers management.
  • Decoded picture conformance all the samples of a decoded picture should be identical to corresponding samples of a decoded picture generated by a reference standard decoder.
  • the checking of the conformance of a picture decoded from a bitstream portion by a nonreference decoder could involve for example metadata defining a conformance cropping window and comprising, for each component of the picture, a hash.
  • the conformance cropping window defines a subset of luma samples and associated chroma samples of the decoded picture.
  • Each hash comprised in the metadata is computed using samples of the conformance cropping window resulting from the decoding by the reference standard decoder of the same bitstream portion.
  • a hash is computed from the samples of the conformance cropping window and compared to the corresponding hash comprised in the metadata. For each color component, the computed hash should be equal to the corresponding hash provided by the metadata.
  • the conformance is generally tested at a given point of a codec, called a conformance point.
  • the decoded picture conformance is generally checked at the output of the decoder.
  • a decoder can generate several different outputs with the same bitstream. This is the case for instance with a bitstream representing a scalable video that can be decoded at different temporal, quality, and spatial levels.
  • a decoder is not necessarily the last step for generating a displayed picture. Indeed, a decoder can be followed by at least one postprocess that modifies the decoded picture.
  • one or more of the present embodiments provide a method comprising: obtaining a video stream; obtaining metadata associated with the video stream comprising an information representative of at least one conformance point among a plurality of conformance points, a conformance point specifying which data reconstructed from the video stream is to be tested for conformance; decoding the video stream, and; checking a conformance of data of the decoded video stream specified by the at least one conformance point.
  • the conformance points are predefined.
  • conformance points are defined for profiles, tiers or levels of a plurality of profiles, tiers and levels.
  • the metadata is a syntax structure comprising at least one syntax element a value of which indicates that the syntax structure comprises information representative of several conformance points.
  • the syntax structure comprises at least one syntax element indicating for each conformance point a position of the conformance point in a decoding process.
  • the syntax structure comprises for each conformance point at least one syntax element indicating a type of hash associated to the conformance point or a single syntax element indicating a type of hash associated to all conformance points.
  • the syntax structure comprises a syntax element indicating a number of conformance points represented by the syntax structure.
  • the syntax structure is a SEI message.
  • the information representative of at least one conformance point among a plurality of conformance points is indicated in a payload extension of the SEI message.
  • at least two conformance points are associated to two different cropping windows delimiting samples of at least one decoded picture to be considered for conformance checks.
  • At least two conformance points are associated to two different numbers of most significant bits to consider for each sample for conformance checks.
  • one or more of the present embodiments provide a method comprising: encoding video data in a video stream; generating metadata comprising an information representative of at least one conformance point among a plurality of conformance points, a conformance point specifying which data reconstructed from the video stream is to be tested for conformance; associating the metadata to the video stream; and, providing the video stream along with the metadata to allow checking a conformance of data decoded from the video stream specified by the at least one conformance point.
  • the conformance points are predefined.
  • conformance points are defined for profiles, tiers or levels of a plurality of profiles, tiers and levels.
  • the metadata is a syntax structure comprising at least one syntax element a value of which indicates that the syntax structure comprises information representative of several conformance points.
  • the syntax structure comprises at least one syntax element indicating for each conformance point a position of the conformance point in a decoding process.
  • the syntax structure comprises for each conformance point at least one syntax element indicating a type of hash associated to the conformance point or a single syntax element indicating a type of hash associated to all conformance points.
  • the syntax structure comprises a syntax element indicating a number of conformance points represented by the syntax structure.
  • the syntax structure is a SEI message.
  • the information representative of at least one conformance point among a plurality of conformance points is indicated in a payload extension of the SEI message.
  • At least two conformance points are associated to two different cropping windows delimiting samples of at least one decoded picture to be considered for conformance checks.
  • At least two conformance points are associated to two different numbers of most significant bits to consider for each sample for conformance checks.
  • one or more of the present embodiments provide a device comprising electronic circuitry configured for: obtaining a video stream; obtaining metadata associated with the video stream comprising an information representative of at least one conformance point among a plurality of conformance points, a conformance point specifying which data reconstructed from the video stream is to be tested for conformance; decoding the video stream, and; checking a conformance of data of the decoded video stream specified by the at least one conformance point.
  • the conformance points are predefined.
  • conformance points are defined for profiles, tiers or levels of a plurality of profiles, tiers and levels.
  • the metadata is a syntax structure comprising at least one syntax element a value of which indicates that the syntax structure comprises information representative of several conformance points.
  • the syntax structure comprises at least one syntax element indicating for each conformance point a position of the conformance point in a decoding process.
  • the syntax structure comprises for each conformance point at least one syntax element indicating a type of hash associated to the conformance point or a single syntax element indicating a type of hash associated to all conformance points.
  • the syntax structure comprises a syntax element indicating a number of conformance points represented by the syntax structure.
  • the syntax structure is a SEI message.
  • the information representative of at least one conformance point among a plurality of conformance points is indicated in a payload extension of the SEI message.
  • At least two conformance points are associated to two different cropping windows delimiting samples of at least one decoded picture to be considered for conformance checks.
  • At least two conformance points are associated to two different numbers of most significant bits to consider for each sample for conformance checks.
  • one or more of the present embodiments provide a device comprising electronic circuitry configured for: encoding video data in a video stream; generating metadata comprising an information representative of at least one conformance point among a plurality of conformance points, a conformance point specifying which data reconstructed from the video stream is to be tested for conformance; associating the metadata to the video stream; and, providing the video stream along with the metadata to allow checking a conformance of data decoded from the video stream specified by the at least one conformance point.
  • the conformance points are predefined.
  • conformance points are defined for profiles, tiers or levels of a plurality of profiles, tiers and levels.
  • the metadata is a syntax structure comprising at least one syntax element a value of which indicates that the syntax structure comprises information representative of several conformance points.
  • the syntax structure comprises at least one syntax element indicating for each conformance point a position of the conformance point in a decoding process.
  • the syntax structure comprises for each conformance point at least one syntax element indicating a type of hash associated to the conformance point or a single syntax element indicating a type of hash associated to all conformance points.
  • the syntax structure comprises a syntax element indicating a number of conformance points represented by the syntax structure.
  • the syntax structure is a SEI message.
  • the information representative of at least one conformance point among a plurality of conformance points is indicated in a payload extension of the SEI message.
  • At least two conformance points are associated to two different cropping windows delimiting samples of at least one decoded picture to be considered for conformance checks.
  • At least two conformance points are associated to two different numbers of most significant bits to consider for each sample for conformance checks.
  • one or more of the present embodiments provide a signal comprising metadata associated with a video stream comprising an information representative of at least one conformance point among a plurality of conformance points, a conformance point specifying which data reconstructed from the video stream is to be tested for conformance.
  • one or more of the present embodiments provide a computer program comprising program code instructions for implementing the method according to the first and the second aspect.
  • one or more of the present embodiments provide a non- transitory information storage medium storing program code instructions for implementing the method according to the first or the second aspect. 4. BRIEF SUMMARY OF THE DRAWINGS
  • FIG. 1 illustrates schematically a context in which embodiments are implemented
  • Fig. 2 illustrates schematically an example of partitioning undergone by a picture of pixels of an original video
  • Fig. 3 depicts schematically a method for encoding a video stream
  • Fig. 4 depicts schematically a method for decoding an encoded video stream
  • Fig. 5 A illustrates schematically an example of hardware architecture of a processing module able to implement an encoding module or a decoding module in which various aspects and embodiments are implemented;
  • Fig. 5B illustrates a block diagram of an example of a first system in which various aspects and embodiments are implemented
  • Fig. 5C illustrates a block diagram of an example of a second system in which various aspects and embodiments are implemented
  • Fig. 6A illustrates schematically an example of process for associating metadata comprising information representative of at least one conformance point among a plurality of conformance points to a video stream;
  • Fig. 6B illustrates schematically an example of process for obtaining a video stream along with metadata comprising information representative of at least one conformance point among a plurality of conformance points
  • Fig.7 illustrates an example of cropping window defined implicitly with the position of the virtual boundary of the gradual decoding refresh (GDR) coding tool.
  • GDR gradual decoding refresh
  • VVC Versatile Video Coding
  • JVET Joint Video Experts Team
  • HEVC ISO/IEC 23008-2 - MPEG-H Part 2, High Efficiency Video Coding / ITU-T H.265
  • AVC ((ISO/CEI 14496-10)
  • EVC Essential Video Coding/MPEG-5
  • AVI AVI
  • AV2 AV2 and VP9.
  • Fig- 1 illustrates schematically a context in which embodiments are implemented.
  • a system 11 that could be a camera, a storage device, a computer, a server or any device capable of delivering a video stream, transmits a video stream to a system 13 using a communication channel 12.
  • the video stream is either encoded and transmitted by the system 11 or received and/or stored by the system 11 and then transmitted.
  • the communication channel 12 is a wired (for example Internet or Ethernet) or a wireless (for example WiFi, 3G, 4G or 5G) network link.
  • the system 13 that could be for example a set top box, receives and decodes the video stream to generate a sequence of decoded pictures.
  • a post processing may be applied to the decoded pictures.
  • the obtained sequence of decoded pictures is then transmitted to a display system 15 using a communication channel 14, that could be a wired or wireless network.
  • the display system 15 then displays said pictures.
  • the system 13 is comprised in the display system 15.
  • the system 13 and display 15 are comprised in a TV, a computer, a tablet, a smartphone, a head-mounted display, etc.
  • Figs. 2, 3 and 4 introduce an example of video format.
  • Fig- 2 illustrates an example of partitioning undergone by a picture of pixels 21 of an original video sequence 20. It is considered here that a pixel is composed of three components: a luminance component and two chrominance components. Other types of pixels are however possible comprising less or more components such as only a luminance component or an additional depth component or transparency component.
  • a picture is divided into a plurality of coding entities.
  • a picture is divided in a grid of blocks called coding tree units (CTU).
  • CTU coding tree units
  • a CTU consists of an N x N block of luminance samples together with two corresponding blocks of chrominance samples.
  • N is generally a power of two having a maximum value of “128” for example.
  • a picture is divided into one or more groups of CTU. For example, it can be divided into one or more tile rows and tile columns, a tile being a sequence of CTU covering a rectangular region of a picture. In some cases, a tile could be divided into one or more bricks, each of which consisting of at least one row of CTU within the tile.
  • another encoding entity, called slice exists, that can contain at least one tile of a picture or at least one brick of a tile.
  • the picture 21 is divided into three slices SI, S2 and S3 of the raster-scan slice mode, each comprising a plurality of tiles (not represented), each tile comprising only one brick.
  • a CTU may be partitioned into the form of a hierarchical tree of one or more sub-blocks called coding units (CU).
  • the CTU is the root (i.e. the parent node) of the hierarchical tree and can be partitioned in a plurality of CU (i.e. child nodes).
  • Each CU becomes a leaf of the hierarchical tree if it is not further partitioned in smaller CU or becomes a parent node of smaller CU (i.e. child nodes) if it is further partitioned.
  • the CTU 24 is first partitioned in “4” square CU using a quadtree type partitioning.
  • the upper left CU is a leaf of the hierarchical tree since it is not further partitioned, i.e. it is not a parent node of any other CU.
  • the upper right CU is further partitioned in “4” smaller square CU using again a quadtree type partitioning.
  • the bottom right CU is vertically partitioned in “2” rectangular CU using a binary tree type partitioning.
  • the bottom left CU is vertically partitioned in “3” rectangular CU using a ternary tree type partitioning.
  • the partitioning is adaptive, each CTU being partitioned so as to optimize a compression efficiency of the CTU criterion.
  • HEVC In HEVC appeared the concept of prediction unit (PU) and transform unit (TU). Indeed, in HEVC, the coding entity that is used for prediction (i.e. a PU) and transform (i.e. a TU) can be a subdivision of a CU. For example, as represented in Fig. 2, a CU of size 2N x 2N, can be divided in PU 2411 of size N x 2N or of size 2N x N. In addition, said CU can be divided in “4” TU 2412 of size N x N or in “16” TU of size $ X
  • a CU comprises generally one TU and one PU.
  • the term “block” or “picture block” can be used to refer to any one of a CTU, a CU, a PU and a TU.
  • the term “block” or “picture block” can be used to refer to a macroblock, a partition and a sub-block as specified in H.264/AVC or in other video coding standards, and more generally to refer to an array of samples of numerous sizes.
  • the terms “reconstructed” and “decoded” may be used interchangeably, the terms “pixel” and “sample” may be used interchangeably, the terms “image,” “picture”, “sub-picture”, “slice” and “frame” may be used interchangeably.
  • the term “reconstructed” is used at the encoder side while “decoded” is used at the decoder side.
  • Fig. 3 depicts schematically a method for encoding a video stream executed by an encoding module. Variations of this method for encoding are contemplated, but the method for encoding of Fig. 3 is described below for purposes of clarity without describing all expected variations.
  • a current original picture of an original video sequence may go through a pre-processing.
  • a color transform is applied to the current original picture (e.g., conversion from RGB 4:4:4 to YCbCr 4:2:0), or a remapping is applied to the current original picture components in order to get a signal distribution more resilient to compression (for instance using a histogram equalization of one of the color components).
  • Pictures obtained by pre-processing are called pre-processed pictures in the following.
  • the encoding of a pre-processed picture begins with a partitioning of the pre- processed picture during a step 302, as described in relation to Fig. 2.
  • the pre-processed picture is thus partitioned into CTU, CU, PU, TU, etc.
  • the encoding module determines a coding mode between an intra prediction and an inter prediction.
  • the intra prediction consists of predicting, in accordance with an intra prediction method, during a step 303, the pixels of a current block from a prediction block derived from pixels of reconstructed blocks situated in a causal vicinity of the current block to be coded.
  • the result of the intra prediction is a prediction direction indicating which pixels of the blocks in the vicinity to use, and a residual block resulting from a calculation of a difference between the current block and the prediction block.
  • the inter prediction consists in predicting the pixels of a current block from a block of pixels, referred to as the reference block, of a picture preceding or following the current picture, this picture being referred to as the reference picture.
  • a block of the reference picture closest, in accordance with a similarity criterion, to the current block is determined by a motion estimation step 304.
  • a motion vector indicating the position of the reference block in the reference picture is determined.
  • Said motion vector is used during a motion compensation step 305 during which a residual block is calculated in the form of a difference between the current block and the reference block.
  • the mono-directional inter prediction mode described above was the only inter mode available. As video compression standards evolve, the family of inter modes has grown significantly and comprises now many different inter modes.
  • the prediction mode optimising the compression performances in accordance with a rate/distortion optimization criterion (i.e. RDO criterion), among the prediction modes tested (Intra prediction modes, Inter prediction modes), is selected by the encoding module.
  • a rate/distortion optimization criterion i.e. RDO criterion
  • the residual block is transformed during a step 307.
  • the transformed block is then quantized during a step 309.
  • the encoding module can skip the transform and apply quantization directly to the non-transformed residual signal.
  • a prediction direction and the transformed and quantized residual block are encoded by an entropic encoder during a step 310.
  • a motion vector of the block is predicted from a prediction vector selected from a set of motion vector predictors derived from reconstructed blocks situated in a spatial and temporal vicinity of the block to be encoded.
  • the motion information is next encoded by the entropic encoder during step 310 in the form of a motion residual and an index for identifying the prediction vector.
  • the transformed and quantized residual block is encoded by the entropic encoder during step 310.
  • the encoding module can bypass both transform and quantization, i. e. , the entropic encoding is applied on the residual without the application of the transform or quantization processes.
  • the result of the entropic encoding is inserted in an encoded video stream 311.
  • Metadata such as SEI (supplemental enhancement information) messages can be attached to the encoded video stream 311.
  • a SEI message as defined for example in standards such as AVC, HEVC or VVC (or in standard Versatile supplemental enhancement information (VSEI) messages for coded video bitstreams - H.274) is a data container or a syntax structure associated to a video stream and comprising metadata providing information relative to the video stream.
  • VSEI Versatile supplemental enhancement information
  • the prediction block of the block is reconstructed.
  • the encoding module applies, when appropriate, during a step 316, a motion compensation using the motion vector of the current block in order to identify the reference block of the current block.
  • the prediction direction corresponding to the current block is used for reconstructing the prediction block of the current block. The prediction block and the reconstructed residual block are added in order to obtain the reconstructed current block.
  • In-loop filtering intended to reduce the encoding artefacts is applied, during a step 317, to the reconstructed block.
  • This filtering is called in-loop filtering since this filtering occurs in the prediction loop to obtain at the decoder the same reference pictures as the encoder and thus avoid a drift between the encoding and the decoding processes.
  • In-loop filtering tools comprises deblocking filtering, SAO (Sample adaptive Offset) and ALF (Adaptive Loop Filtering).
  • DPB Decoded Picture Buffer
  • Fig. 4 depicts schematically a method for decoding the encoded video stream 311 encoded according to method described in relation to Fig. 3 executed by a decoding module. Variations of this method for decoding are contemplated, but the method for decoding of Fig. 4 is described below for purposes of clarity without describing all expected variations.
  • the decoding is done block by block. For a current block, it starts with an entropic decoding of the current block during a step 410. Entropic decoding allows to obtain, at least, the prediction mode of the block. If the block has been encoded according to an inter prediction mode, the entropic decoding allows to obtain, when appropriate, a prediction vector index, a motion residual and a residual block. During a step 408, a motion vector is reconstructed for the current block using the prediction vector index and the motion residual.
  • Steps 412, 413, 414, 415, 416 and 417 implemented by the decoding module are in all respects identical respectively to steps 312, 313, 314, 315, 316 and 317 implemented by the encoding module.
  • Decoded blocks are saved in decoded pictures and the decoded pictures are stored in a DPB 419 in a step 418.
  • the decoding module decodes a given picture
  • the pictures stored in the DPB 419 are identical to the pictures stored in the DPB 319 by the encoding module during the encoding of said given picture.
  • the decoded picture can also be outputted by the decoding module for instance to be displayed.
  • a post-processing step 421 may be applied and comprise, for example:
  • a conformance point Cl is at the level of the decoded picture (i.e. the output of the decoder without any post-processing).
  • the conformance point C2 is at the level of the output of the postprocessing step 421 and allowing checking the conformance of the post-processed picture with an expected result.
  • dph_sei_hash_type indicates a method used to calculate the checksum as specified in Table TAB2. Values of dph sei hash type that are not listed in Table TAB2 are reserved for future use and shall not be present. Decoders shall ignore decoded picture hash SEI messages that contain reserved values of dph sei hash type, values different from “0”, “1” or “2”.
  • dph sei single component Jag indicates that the picture associated with the decoded picture hash SEI message contains a single colour component
  • dph sei single component Jlag indicates that the picture associated with the decoded picture hash SEI message contains three colour components.
  • dph_sei_reserved_zero_7bits shall be equal to “0”. Values greater than “0” for dph_sei_reserved_zero_7bits are reserved for future use and shall not be present. Decoders shall ignore the value of dph_sei_reserved_zero_7bits .
  • dph sei _picture_md5[ cldx ][ i ] is the /-th byte of a 16-byte MD5 hash of the c/rfc-th colour component of the decoded picture.
  • dph sei _picture_crc[ cldx ] is the cyclic redundancy check (CRC) of the colour component cldx of the decoded picture.
  • dph sei _picture_checksum[ cldx ] is the checksum of the colour component cldx of the decoded picture.
  • the decoded picture hash SEI message allows defining hashes of different type but doesn’t allow to specify the position of a conformance point, i.e. doesn’t allow to define at which level of the reconstruction process (or on which data ) the hash is to be computed.
  • the decoded picture hash SEI message doesn’t allow to specify that the conformance point is before or after a post-processing process.
  • the decoded picture hash SEI message doesn’t allow neither to define several conformance points.
  • Fig. 5A, 5B and 5C describes examples of device, apparatus and/or system allowing implementing the various embodiments.
  • Fig. 5A illustrates schematically an example of hardware architecture of a processing module 500 able to implement an encoding module or a decoding module capable of implementing respectively a method for encoding of Fig. 3 and a method for decoding of Fig. 4 modified according to different aspects and embodiments.
  • the encoding module is for example comprised in the system 11 when this system is in charge of encoding the video stream.
  • the decoding module is for example comprised in the system 13.
  • the processing module 500 comprises, connected by a communication bus 5005: a processor or CPU (central processing unit) 5000 encompassing one or more microprocessors, general purpose computers, special purpose computers, and processors based on a multi-core architecture, as non-limiting examples; a random access memory (RAM) 5001; a read only memory (ROM) 5002; a storage unit 5003, which can include non-volatile memory and/or volatile memory, including, but not limited to, Electrically Erasable Programmable Read-Only Memory (EEPROM), Read- Only Memory (ROM), Programmable Read-Only Memory (PROM), Random Access Memory (RAM), Dynamic Random Access Memory (DRAM), Static Random Access Memory (SRAM), flash, magnetic disk drive, and/or optical disk drive, or a storage medium reader, such as a SD (secure digital) card reader and/or a hard disc drive (HDD) and/or a network accessible storage device; at least one communication interface 5004 for exchanging data with other modules, devices or system.
  • the communication interface 5004 can include
  • the communication interface 5004 enables for instance the processing module 500 to receive encoded video streams and to provide a sequence of decoded pictures. If the processing module 500 implements an encoding module, the communication interface 5004 enables for instance the processing module 500 to receive a sequence of original picture data to encode and to provide an encoded video stream.
  • the processor 5000 is capable of executing instructions loaded into the RAM 5001 from the ROM 5002, from an external memory (not shown), from a storage medium, or from a communication network. When the processing module 500 is powered up, the processor 5000 is capable of reading instructions from the RAM 5001 and executing them.
  • These instructions form a computer program causing, for example, the implementation by the processor 5000 of a decoding method as described in relation with Fig. 4 and/or an encoding method described in relation to Fig. 3, and methods described in relation to Figs. 6A and 6B, these methods comprising various aspects and embodiments described below in this document.
  • a programmable machine such as a DSP (digital signal processor) or a microcontroller
  • a dedicated component such as a FPGA (field-programmable gate array) or an ASIC (application-specific integrated circuit).
  • microprocessors general purpose computers, special purpose computers, processors based or not on a multi-core architecture, DSP, microcontroller, FPGA and ASIC are electronic circuitry adapted or configured to implement at least partially the methods of Figs. 3, 4, 6A and 6B.
  • Fig. 5C illustrates a block diagram of an example of the system 13 in which various aspects and embodiments are implemented.
  • the system 13 can be embodied as a device including the various components described below and is configured to perform one or more of the aspects and embodiments described in this document. Examples of such devices include, but are not limited to, various electronic devices such as personal computers, laptop computers, smartphones, tablet computers, digital multimedia set top boxes, digital television receivers, personal video recording systems, connected home appliances and head mounted display.
  • Elements of system 13, singly or in combination can be embodied in a single integrated circuit (IC), multiple ICs, and/or discrete components.
  • the system 13 comprises one processing module 500 that implements a decoding module.
  • system 13 is communicatively coupled to one or more other systems, or other electronic devices, via, for example, a communications bus or through dedicated input and/or output ports. In various embodiments, the system 13 is configured to implement one or more of the aspects described in this document.
  • the input to the processing module 500 can be provided through various input modules as indicated in block 531.
  • Such input modules include, but are not limited to, (i) a radio frequency (RF) module that receives an RF signal transmitted, for example, over the air by a broadcaster, (ii) a component (COMP) input module (or a set of COMP input modules), (iii) a Universal Serial Bus (USB) input module, and/or (iv) a High Definition Multimedia Interface (HDMI) input module.
  • RF radio frequency
  • COMP component
  • USB Universal Serial Bus
  • HDMI High Definition Multimedia Interface
  • the input modules of block 531 have associated respective input processing elements as known in the art.
  • the RF module can be associated with elements suitable for (i) selecting a desired frequency (also referred to as selecting a signal, or band-limiting a signal to a band of frequencies), (ii) down-converting the selected signal, (iii) band-limiting again to a narrower band of frequencies to select (for example) a signal frequency band which can be referred to as a channel in certain embodiments, (iv) demodulating the down-converted and bandlimited signal, (v) performing error correction, and (vi) demultiplexing to select the desired stream of data packets.
  • the RF module of various embodiments includes one or more elements to perform these functions, for example, frequency selectors, signal selectors, band-limiters, channel selectors, filters, downconverters, demodulators, error correctors, and demultiplexers.
  • the RF portion can include a tuner that performs various of these functions, including, for example, down-converting the received signal to a lower frequency (for example, an intermediate frequency or a near-baseband frequency) or to baseband.
  • the RF module and its associated input processing element receives an RF signal transmitted over a wired (for example, cable) medium, and performs frequency selection by filtering, downconverting, and filtering again to a desired frequency band.
  • Adding elements can include inserting elements in between existing elements, such as, for example, inserting amplifiers and an analog-to-digital converter.
  • the RF module includes an antenna.
  • USB and/or HDMI modules can include respective interface processors for connecting system 13 to other electronic devices across USB and/or HDMI connections.
  • various aspects of input processing for example, Reed-Solomon error correction, can be implemented, for example, within a separate input processing IC or within the processing module 500 as necessary.
  • aspects of USB or HDMI interface processing can be implemented within separate interface ICs or within the processing module 500 as necessary.
  • the demodulated, error corrected, and demultiplexed stream is provided to the processing module 500.
  • Various elements of system 13 can be provided within an integrated housing. Within the integrated housing, the various elements can be interconnected and transmit data therebetween using suitable connection arrangements, for example, an internal bus as known in the art, including the Inter-IC (I2C) bus, wiring, and printed circuit boards.
  • I2C Inter-IC
  • the processing module 500 is interconnected to other elements of said system 13 by the bus 5005.
  • the communication interface 5004 of the processing module 500 allows the system 13 to communicate on the communication channel 12.
  • the communication channel 12 can be implemented, for example, within a wired and/or a wireless medium.
  • Data is streamed, or otherwise provided, to the system 13, in various embodiments, using a wireless network such as a Wi-Fi network, for example IEEE 802. 11 (IEEE refers to the Institute of Electrical and Electronics Engineers).
  • the WiFi signal of these embodiments is received over the communications channel 12 and the communications interface 5004 which are adapted for Wi-Fi communications.
  • the communications channel 12 of these embodiments is typically connected to an access point or router that provides access to external networks including the Internet for allowing streaming applications and other over-the-top communications.
  • Other embodiments provide streamed data to the system 13 using the RF connection of the input block 531.
  • various embodiments provide data in a nonstreaming manner.
  • various embodiments use wireless networks other than Wi-Fi, for example a cellular network or a Bluetooth network.
  • the system 13 can provide an output signal to various output devices, including the display system 15, speakers 535, and other peripheral devices 536.
  • the display system 15 of various embodiments includes one or more of, for example, a touchscreen display, an organic light-emitting diode (OLED) display, a curved display, and/or a foldable display.
  • the display 15 can be for a television, a tablet, a laptop, a cell phone (mobile phone), a head mounted display or other devices.
  • the display system 15 can also be integrated with other components (for example, as in a smart phone), or separate (for example, an external monitor for a laptop).
  • the other peripheral devices 536 include, in various examples of embodiments, one or more of a stand-alone digital video disc (or digital versatile disc) (DVR, for both terms), a disk player, a stereo system, and/or a lighting system.
  • DVR digital video disc
  • Various embodiments use one or more peripheral devices 536 that provide a function based on the output of the system 13. For example, a disk player performs the function of playing an output of the system 13.
  • control signals are communicated between the system 13 and the display system 15, speakers 535, or other peripheral devices 536 using signaling such as AV. Link, Consumer Electronics Control (CEC), or other communications protocols that enable device-to-device control with or without user intervention.
  • the output devices can be communicatively coupled to system 13 via dedicated connections through respective interfaces 532, 533, and 534. Alternatively, the output devices can be connected to system 13 using the communications channel 12 via the communications interface 5004 or a dedicated communication channel corresponding to the communication channel 12 in Fig. 5C via the communication interface 5004.
  • the display system 15 and speakers 535 can be integrated in a single unit with the other components of system 13 in an electronic device such as, for example, a television.
  • the display interface 532 includes a display driver, such as, for example, a timing controller (T Con) chip.
  • T Con timing controller
  • the display system 15 and speaker 535 can alternatively be separate from one or more of the other components.
  • the output signal can be provided via dedicated output connections, including, for example, HDMI ports, USB ports, or COMP outputs.
  • Fig. 5B illustrates a block diagram of an example of the system 11 in which various aspects and embodiments are implemented.
  • System 11 is very similar to system 13.
  • the system 11 can be embodied as a device including the various components described below and is configured to perform one or more of the aspects and embodiments described in this document. Examples of such devices include, but are not limited to, various electronic devices such as personal computers, laptop computers, smartphones, tablet computers, a camera and a server.
  • Elements of system 11, singly or in combination can be embodied in a single integrated circuit (IC), multiple ICs, and/or discrete components.
  • the system 11 comprises one processing module 500 that implements an encoding module.
  • system 11 is communicatively coupled to one or more other systems, or other electronic devices, via, for example, a communications bus or through dedicated input and/or output ports.
  • system 11 is configured to implement one or more of the aspects described in this document.
  • the input to the processing module 500 can be provided through various input modules as indicated in block 531 already described in relation to Fig. 5D.
  • Various elements of system 11 can be provided within an integrated housing. Within the integrated housing, the various elements can be interconnected and transmit data therebetween using suitable connection arrangements, for example, an internal bus as known in the art, including the Inter-IC (I2C) bus, wiring, and printed circuit boards.
  • I2C Inter-IC
  • the processing module 500 is interconnected to other elements of said system 11 by the bus 5005.
  • the communication interface 5004 of the processing module 500 allows the system 11 to communicate on the communication channel 12.
  • Data is streamed, or otherwise provided, to the system 11, in various embodiments, using a wireless network such as a Wi-Fi network, for example IEEE 802.11 (IEEE refers to the Institute of Electrical and Electronics Engineers).
  • the WiFi signal of these embodiments is received over the communications channel 12 and the communications interface 5004 which are adapted for Wi-Fi communications.
  • the communications channel 12 of these embodiments is typically connected to an access point or router that provides access to external networks including the Internet for allowing streaming applications and other over-the-top communications.
  • Other embodiments provide streamed data to the system 11 using the RF connection of the input block 531.
  • various embodiments provide data in a non-streaming manner. Additionally, various embodiments use wireless networks other than Wi-Fi, for example a cellular network or a Bluetooth network.
  • the data provided to the system 11 can be provided in different format.
  • these data are encoded and compliant with a known video compression format such as AVI, VP9, VVC, HEVC, AVC, etc.
  • these data are raw data provided for example by a picture and/or audio acquisition module connected to the system 11 or comprised in the system 11. In that case, the processing module 500 take in charge the encoding of these data.
  • the system 11 can provide an output signal to various output devices capable of storing and/or decoding the output signal such as the system 13.
  • Decoding can encompass all or part of the processes performed, for example, on a received encoded video stream in order to produce a final output suitable for display.
  • processes include one or more of the processes typically performed by a decoder, for example, entropy decoding, inverse quantization, inverse transformation, and prediction.
  • processes also, or alternatively, include processes performed by a decoder of various implementations described in this application, for example, for applying a post-processing or for applying conformance checking processes at conformance points specified for instance in SEI messages.
  • decoding process is intended to refer specifically to a subset of operations or generally to the broader decoding process will be clear based on the context of the specific descriptions and is believed to be well understood by those skilled in the art.
  • encoding can encompass all or part of the processes performed, for example, on an input video sequence in order to produce an encoded video stream.
  • processes include one or more of the processes typically performed by an encoder, for example, partitioning, prediction, transformation, quantization, and entropy encoding.
  • processes also, or alternatively, include processes performed by an encoder of various implementations described in this application, for example, for signaling various conformance points to a decoder.
  • syntax elements names as used herein are descriptive terms. As such, they do not preclude the use of other syntax element names.
  • Various embodiments refer to rate distortion optimization.
  • the rate distortion optimization is usually formulated as minimizing a rate distortion function, which is a weighted sum of the rate and of the distortion.
  • the approaches may be based on an extensive testing of all encoding options, including all considered modes or coding parameters values, with a complete evaluation of their coding cost and related distortion of a reconstructed signal after coding and decoding.
  • Faster approaches may also be used, to save encoding complexity, in particular with computation of an approximated distortion based on a prediction or a prediction residual signal, not the reconstructed one.
  • the implementations and aspects described herein can be implemented in, for example, a method or a process, an apparatus, a software program, a data stream, or a signal. Even if only discussed in the context of a single form of implementation (for example, discussed only as a method), the implementation of features discussed can also be implemented in other forms (for example, an apparatus or program).
  • An apparatus can be implemented in, for example, appropriate hardware, software, and firmware.
  • the methods can be implemented, for example, in a processor, which refers to processing devices in general, including, for example, a computer, a microprocessor, an integrated circuit, or a programmable logic device. Processors also include communication devices, such as, for example, computers, cell phones, portable/personal digital assistants ("PDAs”), and other devices that facilitate communication of information between end-users.
  • PDAs portable/personal digital assistants
  • references to “one embodiment” or “an embodiment” or “one implementation” or “an implementation”, as well as other variations thereof, means that a particular feature, structure, characteristic, and so forth described in connection with the embodiment is included in at least one embodiment.
  • the appearances of the phrase “in one embodiment” or “in an embodiment” or “in one implementation” or “in an implementation”, as well any other variations, appearing in various places throughout this application are not necessarily all referring to the same embodiment.
  • Determining the information can include one or more of, for example, estimating the information, calculating the information, predicting the information, retrieving the information from memory or obtaining the information for example from another device, module or from user.
  • Accessing the information can include one or more of, for example, receiving the information, retrieving the information (for example, from memory), storing the information, moving the information, copying the information, calculating the information, determining the information, predicting the information, or estimating the information.
  • this application may refer to “receiving” various pieces of information.
  • Receiving is, as with “accessing”, intended to be a broad term.
  • Receiving the information can include one or more of, for example, accessing the information, or retrieving the information (for example, from memory).
  • “receiving” is typically involved, in one way or another, during operations such as, for example, storing the information, processing the information, transmitting the information, moving the information, copying the information, erasing the information, calculating the information, determining the information, predicting the information, or estimating the information.
  • any of the following “and/or”, and “at least one of’, “one or more of’ for example, in the cases of “A/B”, “A and/or B” and “at least one of A and B”, “one or more of A and B” is intended to encompass the selection of the first listed option (A) only, or the selection of the second listed option (B) only, or the selection of both options (A and B).
  • the word “signal” refers to, among other things, indicating something to a corresponding decoder.
  • the encoder signals a use of some coding tools.
  • the same parameters can be used at both the encoder side and the decoder side.
  • an encoder can transmit (explicit signaling) a particular parameter to the decoder so that the decoder can use the same particular parameter.
  • signaling can be used without transmitting (implicit signaling) to simply allow the decoder to know and select the particular parameter. By avoiding transmission of any actual functions, a bit savings is realized in various embodiments.
  • signaling can be accomplished in a variety of ways. For example, one or more syntax elements, flags, and so forth are used to signal information to a corresponding decoder in various embodiments. While the preceding relates to the verb form of the word “signal”, the word “signal” can also be used herein as a noun.
  • implementations can produce a variety of signals formatted to carry information that can be, for example, stored or transmitted.
  • the information can include, for example, instructions for performing a method, or data produced by one of the described implementations.
  • a signal can be formatted to carry the encoded video stream and SEI messages of a described embodiment.
  • Such a signal can be formatted, for example, as an electromagnetic wave (for example, using a radio frequency portion of spectrum) or as a baseband signal.
  • the formatting can include, for example, encoding an encoded video stream and modulating a carrier with the encoded video stream.
  • the information that the signal carries can be, for example, analog or digital information.
  • the signal can be transmitted over a variety of different wired or wireless links, as is known.
  • the signal can be stored on a processor-readable medium.
  • Fig. 6A illustrates schematically an example of process for associating metadata comprising information representative of at least one conformance point among a plurality of conformance points to a video stream.
  • the process of Fig. 6A is for instance executed by the processing module 500 of the system 11.
  • the processing module 500 obtains video data and encodes these video data into a video stream.
  • the processing module 500 generates metadata comprising an information representative of at least one conformance point among a plurality of conformance points, a conformance point specifying which data reconstructed from the video stream is to be tested for conformance.
  • Step 602 comprise a generation of a hash using for instance a reference standard decoder or applying a reference post-processing process identical to the process applied in step 421.
  • the processing module 500 associates the metadata to the video stream.
  • the processing module 500 provides the video stream along with the metadata to allow checking a conformance of data decoded from the video stream specified by the at least one conformance points. For instance, the processing module 500 provides the video stream and the metadata to the system 13.
  • the data reconstructed from the video stream can be data extracted from the video stream without any decoding or interpretation of these data or data resulting from a decoding or interpretation such as a semantical interpretation of the video stream.
  • Fig. 6B illustrates schematically an example of process for obtaining a video stream along with metadata comprising information representative of at least one conformance point among a plurality of conformance points.
  • the process of Fig. 6B is for instance executed by the processing module 500 of the system 13.
  • the processing module 500 obtains a video stream. For instance, the processing module 500 obtains the video stream provided by the system 11 in step 604.
  • the processing module 500 obtains the metadata associated with the video stream in step 603 and comprising the information representative of at least one conformance point among a plurality of conformance points.
  • the processing module 500 decodes the video stream.
  • the processing module 500 checks a conformance of data of the decoded video stream specified by the at least one conformance points.
  • the conformance points of the plurality of conformance points are pre-defined, for instance, in a standard specification such as the VVC (respectively AVC, HEVC, AVI, AV2, VP9, VSEI-H.274) standard specification.
  • VVC video coding
  • HSS conformance or decoded picture conformance i.e. the criterion to fulfil to consider that a video stream or a picture decoded from a video stream is conform to pre-defined specifications and/or requirements of a standard, are indicated for each conformance point of the plurality of conformance points.
  • specifications and/or requirements of conformance are specified for various profiles, tiers or levels defined in a standard (or for each profile, tier and level defined by a standard).
  • Profiles, tiers and levels specify restrictions on bitstreams and hence limits on the capabilities needed to decode the bitstreams.
  • Profiles, tiers and levels are also used to indicate the capability of individual decoder implementations and interoperability points between encoders and decoders.
  • specifications and/or requirements of conformance are specified for each profile, tier or level defined in Annex A of the VVC specification.
  • conformance points are defined for various profiles, tiers or levels defined in a standard (or for each profile, tier and level defined by a standard)
  • the metadata are encoded in the form of a SEI message.
  • the decoded picture hash SEI message is extended (i.e. modified) to signal several conformance points.
  • the semantic of the syntax element dph sei hash type is modified. Thanks to this new semantic, a decoder not supporting the extended decoded picture hash SEI message describing several conformance points correctly decodes this SEI message if a value of the syntax element dph sei hash type is in the range of the previous semantic, i.e. 0, 1 or 2, and ignores this SEI message if the value of the syntax element dph sei hash type is not in this range indicating that the extended decoded picture hash SEI message described several conformance points.
  • a decoder not supporting extended decoded picture hash SEI messages describing several conformance points typically ignores this SEI message when dph sei hash type equals “3”.
  • the semantics of the syntax elements dphm_sei_hash_type[cf], dphm sei single component Jlag[cf], dphm sei _picture_md5[cf], dphm sei _picture_crc[cf], dphm sei _picture checksum [cf] are respectively the same as the semantics of the syntax elements dph sei hash type, dph sei single component Jlag, dph sei _picture_md5, dph sei _picture_crc, dph sei _picture checksum, the variable cf indicating that the syntax elements concerns the conformance point number c in the plurality of conformance points.
  • the syntax element dph sei nb conformance _point indicates the number of conformance points in the plurality of conformance points.
  • the number of conformance points is at least equal to 1 and dph sei nb conformance _point is replaced by dph sei nb conformance _point minus 1.
  • an additional syntax element dphm_sei_hash_conf_id[ cf] is added to the syntax and indicates for each conformance point cf the position of the conformance point in the decoding process.
  • reference Cl and C2 in Fig. 4 correspond to two different positions of conformance points.
  • Table TAB5 represents examples of values taken by the syntax element dphm sei hash conf id [ cf]'.
  • dphm sei hash conf id [ cf ] is inferred to “0”, i.e. the conformance process is applied on reconstructed samples after in-loop filtering (step 417).
  • the value of dphm_sei_hash_conf_id[ cf ] is inferred from cf the value of the c/index being implicitly associated with a pre-defined conformance point.
  • all conformance points use the same type of hash.
  • the syntax element dph_sei_hash_type[ cf ] is not present and the type of hash for the several conformance points is signaled by additional values of the syntax element dph sei hash type as indicated in table TAB6:
  • the syntax element dph_sei_reserved_zero_7bits that was previously ignored is now used.
  • the semantics of the syntax element dph_sei_reserved_zero_7bits is modified to indicate that the extended decoded picture hash SEI message may contain several hash codes corresponding to several conformance points. If the syntax element dph_sei_reserved_zero_7bits is equal to zero, the previous syntax is used, i.e. no signalling of several conformance points is possible.
  • Table TAB7 describes the syntax of the extended decoded picture hash SEI message when the syntax element dph_sei_reserved_zero_7bits is used. Again, the new syntax is in bold.
  • the syntax element dph_sei reserved_zero_7bits is replaced by a new syntax element indicating the number of conformance points (and the number of hashes).
  • dph_sei_reserved_zero_7bits The name of the syntax element dph_sei_reserved_zero_7bits is changed to dph sei _nb hashes minus 1, and semantics is changed so that it indicates the number of conformance points minus “1” if different from “0”. If the syntax element dph sei nb hashes minus 1 is equal to “0” a single conformance point is signalled by the extended decoded picture hash SEI message.
  • the generic syntax of the SEI messages allows extending a SEI message with a payload extension mechanism. That is, if a number of payload bytes in a SEI NAL (network Abstraction Layer) unit is larger than a regular payload size, then a pay load extension _present() syntax element is set to true. When the syntax element payload extension _present() is equal to true, a syntax element sei reserved _payload_extension_data is read. Decoders generally ignores the syntax element sei reserved _payload extension data. In an embodiment, the syntax element sei reserved _payload_extension_data is used to extend the existing decoded picture hash SEI message described in table TAB1.
  • Table TABU An example of extension of the decoded picture hash SEI message replacing the syntax element sei reserved _payload_extension_data is depicted in Table TAB12.
  • the extended decoded picture hash SEI message is a combination of the decoded picture hash SEI message of table TAB1 with the extension decoded jncture hash ext ension() represented in Table TAB12.
  • the syntax elements dphm sei hash type [ cf ] and dphm sei single component Jlag[ cf ] are not present and inferred equal to the values of the syntax elements dph sei hash type and dph sei single component Jlag present in the non-extended part of the extended decoded picture hash SEI message.
  • a new decoded picture hash SEI message is proposed.
  • This new decoded picture hash SEI message represented in table TAB 13 enables signalling several conformance points.
  • the syntax elements dphm sei hash type/cfl and dphm sei single component Jlag[cf] are not present but inferred equal to dph sei hash type and dph sei single component Jlag as represented in table TAB 14.
  • the decoded picture hash SEI message contains a single conformance point with a single conformance point position (dphm sei hash conf id).
  • several SEI messages with different values of conformance point position may be inserted in the bitstreams corresponding to a same decoded picture (e.g. with the same POC (picture order count)).
  • the ordering of the SEI message allows deriving implicitly the conformance position Ci.
  • a cropping window that delimits the samples of a picture to be considered for the conformance checks.
  • the hash is computed with the samples of this cropping window.
  • several cropping windows are defined and a hash is signalled for some or each of these cropping windows.
  • Each cropping window (position and size into a decoded picture) may be signalled explicitly in the SEI message or implicitly. In case of implicit signalling, it can re-use other areas definition. For instance, a cropping window can be aligned on a slice, a sub-picture or a tile and each area is associated to a hash and corresponds to a conformance point.
  • a cropping window can be implicitly defined with virtual boundaries as defined for the GDR (gradual decoding refresh) tool for example.
  • GDR is a technique to limit large bitrate variations with a regular refreshment of pictures with INTRA blocks but without inserting INTRA pictures.
  • the CTUs situated at the left of a GDR virtual boundary are considered as a "clean” area and the CTUs situated at the right of the virtual boundary are considered as “dirty” (may have been corrupted), then a hash computed on the left part only would be useful.
  • each cropping window could be associated to a single decoded picture hash SEI message
  • each cropping window could correspond to a conformance point of the plurality of conformance points indicated in an extended decoded picture SEI message.
  • several conformance points can be obtained by computing the hash on reconstructed picture data with their full bit-depth precision or computed using the most significant bits only of the reconstructed picture data.
  • This feature may be signalled in the decoded picture hash SEI message.
  • a hash may be computed with most-significant 8 -bits only or with the 10 bits.
  • a first conformance point of an extended decoded picture hash SEI message could be associated to a hash computed on picture data with a 8-bits-depth precision and a second conformance point of the extended decoded picture hash SEI message could be associated with a hash computed on picture data with a 10-bits-depth precision.
  • This disclosure has described various pieces of information representing conformance points, such as for example syntax, that can be transmitted or stored, for example.
  • This conformance points information can be packaged or arranged in a variety of manners, including for example manners common in video standards such as putting the conformance points information into a SEI message such as the SEI messages described in this disclosure.
  • Other manners are also available, including for example manners common for system level or application level standards such as putting the conformance points information into:
  • SDP session description protocol
  • RTP Real-time Transport Protocol
  • DASH MPD Media Presentation Description
  • a descriptor is associated to a representation or collection of representations to provide additional characteristics to the content representation.
  • RTP header extensions for example as used during RTP streaming.
  • HLS HTTP live Streaming
  • a manifest is associated to a version or collection of versions of a content to provide characteristics of the version or collection of versions.
  • embodiments can be provided alone or in any combination. Further, embodiments can include one or more of the following features, devices, or aspects, alone or in any combination, across various claim categories and types:
  • a TV, set-top box, cell phone, tablet, or other electronic device that performs at least one of the embodiments described.
  • a TV, set-top box, cell phone, tablet, or other electronic device that performs at least one of the embodiments described, and that displays (e.g. using a monitor, screen, or other type of display) a resulting picture.
  • a TV, set-top box, cell phone, tablet, or other electronic device that tunes (e.g. using a tuner) a channel to receive a signal including an encoded video stream, and performs at least one of the embodiments described.
  • a TV, set-top box, cell phone, tablet, or other electronic device that receives (e.g. using an antenna) a signal over the air that includes an encoded video stream, and performs at least one of the embodiments described.
  • a server camera, cell phone, tablet or other electronic device that transmits (e.g. using an antenna) a signal over the air that includes an encoded video stream, and performs at least one of the embodiments described.
  • a server camera, cell phone, tablet or other electronic device that tunes (e.g. using a tuner) a channel to transmit a signal including an encoded video stream, and performs at least one of the embodiments described.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

Procédé comprenant l'obtention (611) d'un flux vidéo ; l'obtention (612) de métadonnées associées au flux vidéo comprenant des informations représentatives d'au moins un point de conformité parmi une pluralité de points de conformité, un point de conformité spécifiant quelles données reconstruites à partir du flux vidéo doivent être testées du point de vue de la conformité ; le décodage (613) du flux vidéo, et ; la vérification (614) d'une conformité de données du flux vidéo décodé spécifié par le ou les points de conformité.
PCT/EP2023/062663 2022-05-19 2023-05-11 Sei conçues pour de multiples points de conformité WO2023222521A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
EP22305746.4 2022-05-19
EP22305746 2022-05-19

Publications (1)

Publication Number Publication Date
WO2023222521A1 true WO2023222521A1 (fr) 2023-11-23

Family

ID=81975404

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/EP2023/062663 WO2023222521A1 (fr) 2022-05-19 2023-05-11 Sei conçues pour de multiples points de conformité

Country Status (1)

Country Link
WO (1) WO2023222521A1 (fr)

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150103888A1 (en) * 2013-10-15 2015-04-16 Qualcomm Incorporated Support of multi-mode extraction for multi-layer video codecs
US20170085878A1 (en) * 2015-09-22 2017-03-23 Qualcomm Incorporated Video decoder conformance for high dynamic range (hdr) video coding using a core video standard
US20210407139A1 (en) * 2020-06-26 2021-12-30 Sony Corporation Decoded tile hash sei message for v3c/v-pcc

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150103888A1 (en) * 2013-10-15 2015-04-16 Qualcomm Incorporated Support of multi-mode extraction for multi-layer video codecs
US20170085878A1 (en) * 2015-09-22 2017-03-23 Qualcomm Incorporated Video decoder conformance for high dynamic range (hdr) video coding using a core video standard
US20210407139A1 (en) * 2020-06-26 2021-12-30 Sony Corporation Decoded tile hash sei message for v3c/v-pcc

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
"Text of ISO/IEC FDIS 23002-7:202X Versatile supplemental enhancement information messages for coded video bitstreams (2nd edition)", no. n21134, 15 April 2022 (2022-04-15), XP030302263, Retrieved from the Internet <URL:https://dms.mpeg.expert/doc_end_user/documents/137_OnLine/wg11/MDS21134_WG05_N00100.zip WG5_N0100_FDIS_23002-7_202X-2ndEd_VSEI_text.pdf> [retrieved on 20220415] *

Similar Documents

Publication Publication Date Title
US20230164360A1 (en) Method and device for image encoding and decoding
US20230188757A1 (en) Method and device to finely control an image encoding and decoding process
US20230164311A1 (en) Signaling chroma offset presence in video coding
WO2023222521A1 (fr) Sei conçues pour de multiples points de conformité
US20240080484A1 (en) Method and device for luma mapping with cross component scaling
US20230262268A1 (en) Chroma format dependent quantization matrices for video encoding and decoding
US20230379482A1 (en) Spatial resolution adaptation of in-loop and post-filtering of compressed video using metadata
US20220224902A1 (en) Quantization matrices selection for separate color plane mode
US20240187649A1 (en) High precision 4x4 dst7 and dct8 transform matrices
WO2022258356A1 (fr) Syntaxe de haut niveau pour un rééchantillonnage d&#39;image
WO2023046463A1 (fr) Procédés et appareils pour coder/décoder une vidéo
EP4356608A1 (fr) Codage de dernier coefficient significatif dans un bloc d&#39;une image
WO2023198527A1 (fr) Codage et décodage vidéo en utilisant une contrainte d&#39;opérations
WO2023041317A1 (fr) Procédé et appareil de codage et de décodage vidéo avec échantillonnage de résidus de chrominance
JP2024522138A (ja) ビデオを符号化/復号するための方法及び装置
EP4360313A1 (fr) Procédés et appareils pour coder/décoder une vidéo
WO2021197979A1 (fr) Procédé et appareil de codage et de décodage vidéo
WO2024002675A1 (fr) Simplification d&#39;une prédiction intra inter-composantes
WO2022268608A2 (fr) Procédé et appareil de codage et de décodage vidéo
EP4320868A1 (fr) Matrices 4x4 de transformées de dst7 et de dct8 de haute précision
KR20230150293A (ko) 비디오를 인코딩/디코딩하기 위한 방법들 및 장치들
WO2023099249A1 (fr) Indication de phase de sous-échantillonnage
EP4038876A1 (fr) Dérivation de matrices de quantification pour le codage conjoint de cb et cr

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 23726110

Country of ref document: EP

Kind code of ref document: A1