EP2805490A1 - Ausgabe von dekodierten referenzbildern - Google Patents

Ausgabe von dekodierten referenzbildern

Info

Publication number
EP2805490A1
EP2805490A1 EP12818654.1A EP12818654A EP2805490A1 EP 2805490 A1 EP2805490 A1 EP 2805490A1 EP 12818654 A EP12818654 A EP 12818654A EP 2805490 A1 EP2805490 A1 EP 2805490A1
Authority
EP
European Patent Office
Prior art keywords
picture
output
value
decoded
order count
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
EP12818654.1A
Other languages
English (en)
French (fr)
Inventor
Jonaton SAMUELSSON
Rickard Sjöberg
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Telefonaktiebolaget LM Ericsson AB
Original Assignee
Telefonaktiebolaget LM Ericsson AB
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Telefonaktiebolaget LM Ericsson AB filed Critical Telefonaktiebolaget LM Ericsson AB
Publication of EP2805490A1 publication Critical patent/EP2805490A1/de
Withdrawn legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/42Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by implementation details or hardware specially adapted for video compression or decompression, e.g. dedicated software implementation
    • H04N19/423Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by implementation details or hardware specially adapted for video compression or decompression, e.g. dedicated software implementation characterised by memory arrangements
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/174Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a slice, e.g. a line of blocks or a group of blocks
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/46Embedding additional information in the video signal during the compression process
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/70Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/30Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
    • H04N19/31Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability in the temporal domain

Definitions

  • the embodiments generally relate to encoding and decoding of pictures, and in particular to outputting decoded reference pictures from a decoded picture buffer.
  • H.264 Moving Picture Experts Group-4 (MPEG-4) Advanced Video Coding (AVC)
  • MPEG-4 Moving Picture Experts Group-4
  • AVC Advanced Video Coding
  • JCT-VC Joint Collaborative Team - Video Coding
  • JCT-VC is a collaborative project between MPEG and International Telecommunication Union Telecommunication standardization sector (ITU-T).
  • ITU-T International Telecommunication Union Telecommunication standardization sector
  • WD HEVC Working Draft
  • the HEVC WD specifies that each picture shall belong to a temporal layer and that a syntax element called temporaljd shall be present for each picture in the bitstream, corresponding to the temporal layer the picture belongs to.
  • the temporal layers are ordered and have the property that a lower temporal layer never references a higher temporal layer. Thus, higher temporal layers can be removed without affecting the lower temporal layers.
  • the removal of temporal layers can be referred to as temporal scaling. Removal of layers can be done in an entity that is neither an encoder nor a decoder, such as a network node. Such an entity can, but is not limited to, forward video bitstream packets from an encoder to a decoder and perform removal of temporal layers without performing full video decoding on the incoming data.
  • the resulting bitstream after one or more temporal layers have been removed is called a subsequence.
  • HEVC High Efficiency Video Coding
  • HEVC it is possible to signal that a picture is a temporal layer switching point, which indicates that at this picture it is possible for a decoder to start decoding more temporal layers than what was decoded before the switching point.
  • the switching point indication guarantees that no picture following the switching point references a picture from before the switching point that might not have been decoded because it belongs to a higher temporal layer than what was decoded before the switching point.
  • the switching points are therefore very useful for a layer removal entity in order to know when to stop removing a certain temporal layer and start forwarding it.
  • the output process is changed compared to H.264/AVC so that marking of pictures as "unused for prediction" is performed prior to decoding of the current picture.
  • the output process is also performed prior to the decoding of the current picture.
  • HEVC defines a Decoded Picture Buffer (DPB) that consists of frame buffers, also referred to as picture slots or picture buffers, in which decoded pictures are stored.
  • DPB Decoded Picture Buffer
  • the DPB size is determined from syntax elements in the bitstream.
  • the DPB fullness increases with one when a picture is inserted into the DPB and decreases with one when a picture is removed from the DPB. If the DPB fullness is equal to the DPB size there are no empty frame buffers and the DPB is said to be full.
  • the "bumping" process that is used for outputting pictures basically consists of filling up the DPB and then start outputting as few pictures as possible, in correct output order, to free up a frame buffer for the current picture.
  • DPB fullness is equal to DPB size
  • the "bumping" is invoked repeatedly until there is an empty frame buffer in which to store the current decoded picture.
  • Picture Order Count represented by the variable PicOrderCntVal is used in HEVC to define the display order of pictures. POC is also used to identify reference pictures.
  • the "bumping" process consists of the following steps:
  • the picture that is first for output is selected as the one having the smallest value of PicOrderCntVal of all pictures in the DPB marked as "needed for output”.
  • IDR Instant Decoder Refresh
  • the "bumping" process includes filling up the DPB before starting to output.
  • a subsequence consisting of a subset of the temporal layers in the original sequence uses a different DPB size.
  • An encoder that would signal different DPB sizes for different temporal layers would, however, be required to evaluate the "bumping" operations for each subsequence, i.e. for each temporal layer, to validate that DPB size requirements are fulfilled and the display order of pictures is correct.
  • the encoder uses different DPB sizes for different temporal layers and additionally signals switching points, it has to track all possible switching alternatives and keep track of the output status in each subsequence to ensure that the picture order is correct for each possible subsequence. This is also the case for a layer removing entity.
  • the encoder can control how many subsequences there are and thereby the complexity of this problem a layer removing entity can not, it has to handle every possible incoming bitstream.
  • An aspect of the embodiments relates to a method of outputting decoded pictures of a video stream from a decoded picture buffer in a decoder.
  • the method comprises calculating a picture order count limit based on at least one syntax element retrieved based on an encoded representation of a current picture of the video stream.
  • the method also outputs decoded reference pictures stored in the decoded picture buffer and having a respective picture order count value that is lower than the calculated picture order count limit.
  • a related aspect of the embodiments defines a decoder comprising a decoded picture buffer configured to store decoded pictures of a video stream.
  • a limit calculator of the decoder is configured to calculate a picture order count limit based on at least one syntax element retrieved based on an encoded representation of a current picture of the video stream.
  • the decoder also comprises a picture outputting unit configured to output decoded reference pictures stored in the decoded picture buffer and having a respective picture order count value lower than the picture order count limit calculated by the limit calculator.
  • a further related aspect of the embodiments defines a receiver comprising a decoder comprising a decoded picture buffer configured to store decoded pictures of a video stream.
  • a limit calculator of the decoder is configured to calculate a picture order count limit based on at least one syntax element retrieved based on an encoded representation of a current picture of the video stream.
  • the decoder also comprises a picture outputting unit configured to output decoded reference pictures stored in the decoded picture buffer and having a respective picture order count value lower than the picture order count limit calculated by the limit calculator.
  • Another aspect of the embodiments relates to a method of encoding a current picture of a video stream in an encoder.
  • the method comprises determining a picture order count limit to have a value enabling a target state of a decoded picture buffer in a decoder for the current picture.
  • the picture order count limit defines a number of decoded reference pictures to be output from the decoded picture buffer in a picture output process for the current picture.
  • the method also comprises determining at least one syntax element representative of the picture order count limit.
  • the current picture is encoded to get an encoded representation of the current picture.
  • the at least one syntax element is associated with this encoded representation.
  • Another related aspect of the embodiments defines an encoder comprising a limit determiner configured to determine a picture order count limit to have a value enabling a target state of a decoded picture buffer in a decoder for a current picture of a video stream.
  • the determined picture order count defines a number of decoded reference pictures to be output from the decoded picture buffer in a picture output process for the current picture.
  • a syntax element determiner is configured to determine at least one syntax element representative of the value determined for the picture order count limit.
  • An encoding unit is configured to encode the current picture to get an encoded representation of the current picture.
  • the encoder also comprises an associating unit configured to associate the at least one syntax element with the encoded representation.
  • Yet another related aspect of the embodiments defines a transmitter comprising an encoder comprising a limit determiner configured to determine a picture order count limit to have a value enabling a target state of a decoded picture buffer in a decoder for a current picture of a video stream.
  • the determined picture order count defines a number of decoded reference pictures to be output from the decoded picture buffer in a picture output process for the current picture.
  • a syntax element determiner is configured to determine at least one syntax element representative of the value determined for the picture order count limit.
  • An encoding unit is configured to encode the current picture to get an encoded representation of the current picture.
  • the encoder also comprises an associating unit configured to associate the at least one syntax element with the encoded representation.
  • the present embodiments provide a picture output process enabling a control of which decoded reference pictures to be output from the decoded picture buffer and the timing of outputting the decoded reference pictures. As a consequence, any delay in outputting decoded reference pictures can be reduced as compared to the prior art "bumping" process. I n fact, it is with the present embodiments possible to start outputting decoded reference pictures even if the decoded picture buffer fullness has not reached the defined decoded picture buffer size.
  • the embodiments are particular advantageous in connection with temporally scaled video streams and sequences where otherwise significant output delays can occur due to removal of various temporal layers.
  • Fig. 1 is a schematic illustration of a video stream of pictures comprising one or more slices
  • Fig. 2 is an illustration of a data packet comprising a NAL unit
  • Fig. 3 is an illustration of an encoded representation of a slice
  • Fig. 4 is a flow diagram of a method of outputting decoded pictures according to an embodiment
  • Fig. 5 is a flow diagram illustrating additional, optional steps of the method in Fig. 4 according to an embodiment
  • Fig. 6 is a flow diagram illustrating additional, optional steps of the method in Figs. 4 and 5 according to an embodiment
  • Fig. 7 is a flow diagram illustrating additional, optional steps of the method in Fig. 4 according to another embodiment
  • Fig. 8 is a flow diagram illustrating additional, optional steps of the method in Fig. 7 according to an embodiment
  • Fig. 9 is a flow diagram of a method of encoding a picture according to an embodiment
  • Fig. 10 is a flow diagram illustrating additional, optional steps of the method in Fig. 9 according to an embodiment
  • Fig. 11 is a schematic block diagram of a receiver according to an embodiment
  • Fig. 12 is a schematic block diagram of a decoder according to an embodiment
  • Fig. 13 is a schematic block diagram of a decoder according to another embodiment
  • Fig. 14 is a schematic block diagram of a transmitter according to an embodiment
  • Fig. 15 is a schematic block diagram of an encoder according to an embodiment
  • Fig. 16 is a schematic block diagram of an encoder according to another embodiment. DETAILED DESCRIPTION
  • the present embodiments generally relate to the field of encoding and decoding of pictures of a video stream.
  • the embodiments relate to outputting decoded reference pictures from a decoded picture buffer (DPB).
  • DPB decoded picture buffer
  • the embodiments hence, provide an outputting process that could be used instead of the current "bumping" process mentioned in the background section.
  • the outputting process provides significant advantages over the prior art bumping process in terms of reducing output delay. These advantages are in particular obtained in connection with temporarily scaled video sequences. It is proposed herein to use a limit calculated from syntax elements signaled in the bitstream, i.e. in the encoded data generated by an encoder and transmitted to a decoder.
  • This limit is then used at the decoder to identify those decoded references pictures stored in the DPB that should be output, such as for display.
  • all pictures in the DPB with a picture order count value, such as PicOrderCntVal, lower than the limit and that have not yet been output are output, such as for display.
  • a syntax element is a codeword or data element forming part of the encoded data generated by an encoder and to be decoded by a decoder.
  • a syntax element is typically a codeword or data element forming part of the control data associated with an encoded representation or such control data or header data present in an encoded representation of a picture.
  • a syntax element can, for instance, be a codeword in a slice header of the encoded representation of a picture.
  • a syntax element can, for instance, be a codeword in a parameter set or other control data associated with the encoded representation of a picture, e.g. retrievable from the bitstream based on data present in the encoded representation or sent outside of the bitstream but retrievable based on data present in the encoded representation.
  • a picture is typically output for display on a screen of or connected to the decoder.
  • a picture could also be output for other reasons including, but not limited, storage on a file; coded in another format or with other properties, such as transcoded; delivered to another unit or device for post-decoding processing; etc.
  • Output as used herein therefore typically relates to output for display but also encompasses other forms of picture output, such as any of the above mentioned examples.
  • Fig. 1 is a schematic illustration of a video stream 1 of pictures 2.
  • a picture 2 in HEVC is partioned into one or more slices 3, where each slice 3 is an independently decodable segment of a picture 2. This means that if a slice 3 is missing, for instance got lost during transmission, the other slices 3 of that picture 2 can still be decoded correctly.
  • slices 3 In order to make slices 3 independent, they should not depend on each other. Hence, in a particular embodiment, no bitstream element of a slice 3 is required for decoding any element of another slice 3.
  • NAL Network Abstraction Layer
  • a coded video stream or sequence i.e. bitstream
  • NAL unit 1 1 comprises either a slice with a corresponding slice header including control information for that slice or the NAL unit 11 comprises a parameter set.
  • the parameter set comprises control information.
  • a NAL unit 11 as output from an encoder is typically complemented with headers 12 to form a data packet 10 that can be transmitted as a part of a bistream from the encoder to the decoder.
  • headers 12 For instance, Real-time Transport Protocol (RTP), User Datagram Protocol (UDP) and Internet Protocol (IP) headers 12 could be added to the NAL unit 11.
  • RTP Real-time Transport Protocol
  • UDP User Datagram Protocol
  • IP Internet Protocol
  • This form of packetization of NAL units 11 merely constitutes an example in connection with video transport.
  • Other approaches of handling NAL units 11 such as file format, MPEG-2 transport streams, MPEG-2 program streams, etc. are possible.
  • APS comprises control information valid for more than one slice.
  • the control information may differ between the slices.
  • PPS comprises control information valid for several pictures, and may be the same for multiple pictures of the same video sequence.
  • SPS comprises control information valid for an entire video sequence.
  • an encoded representation 20 of a slice comprises a slice header 21 which independently provides all required data for the slice to be independently decodable.
  • An example of a syntax element or data element present in the slice header 21 is the slice address, which is used by the decoder to know the spatial location of the slice.
  • the encoded representation 20 also comprises, in addition to the slice header 21 , slice data 22 that comprises the encoded data of the particular slice, e.g. encoded color values of the pixels in the slice.
  • Fig. 4 is a flow diagram illustrating a method of outputting decoded pictures of a video stream from a DPB in a decoder.
  • the method generally starts in step S1.
  • a picture order count (POC) limit sometimes referred to as limit X herein, is calculated based on at least one syntax element retrieved based on an encoded representation of a current picture of the video stream.
  • POC picture order count
  • step S2 of the method decoded reference pictures stored in the DPB and having a respective POC value lower than the POC limit calculated in step S1 are output.
  • one or more syntax elements obtained based on the encoded representation of a current picture are used to calculate a POC limit, which in turn defines those decoded reference pictures in the DPB that should be output.
  • a POC limit which in turn defines those decoded reference pictures in the DPB that should be output.
  • step S1 comprises calculating the POC limit based a POC value of the current picture and the at least one syntax element retrieved based on the encoded representation of the current picture.
  • the at least one syntax element represents a so-called output distance value or OutputDistance.
  • the output distance value is calculated based on the at least one syntax element.
  • the POC limit is preferably calculated based on the POC value of the current picture and the output distance value and more preferably as a difference between the POC value and the output distance value.
  • the output distance value defines a distance or step among POC values from the position of the current picture in the output order to the position of the latest, according to the output order, picture that should be output.
  • any pictures stored in the decoded picture buffer and that precede, according to the output order as defined by the POC values, this latest picture should also be output in step S2.
  • the output distance value is thereby used, preferably together with the POC value of the current picture, to define the picture with the highest POC value, i.e. the POC limit, which should be output. Pictures stored in the decoded picture buffer and that have respective POC values lower than this highest POC value should also be output.
  • step S1 involves calculating the POC limit as PicOrderCnt ⁇ CurrPic ) - OutputDistance, wherein PicOrderCnt ⁇ CurrPic ) denotes the POC value of the current picture and OutputDistance denotes the output distance value.
  • PicOrderCnt ⁇ CurrPic denotes the POC value of the current picture
  • OutputDistance denotes the output distance value.
  • the POC limit can be determined in other ways.
  • OutputDistance as the retrieved syntax element, or syntax elements used to calculate the output distance value is (are) preferably signaled in the bitstream from the encoder to the decoder. If syntax elements are signaled, the output distance value is calculated at the decoder. The output distance value is then used for outputting "old" pictures, i.e. already decoded reference pictures, from the DPB in order to reduce the content of the DPB such that the output delay is decreased.
  • syntax elements are used to calculate OutputDistance, i.e. the output distance value, the syntax elements may be signaled in the slice header or in a parameter set.
  • the encoded representation 20 of the current picture may include the at least one syntax element that is used to calculate the POC limit.
  • the at least one syntax element is then used to calculate the output distance value, which in turn is used, preferably together with the POC value of the current picture, to calculate the POC limit.
  • the decoder can directly retrieve the at least one syntax element during the parsing and decoding of the slice header 21.
  • the at least one syntax element is not necessarily present in the encoded representation 20 of the picture.
  • the encoded representation 20 comprises data enabling retrieval of the at least one syntax element.
  • the encoded representation 20 preferably comprises a parameter set identifier, such as an APS identifier or a PPS identifier.
  • the parameter set identifier is typically present in the slice header 21 of the encoded representation 20. In such a case, the parameter set identifier is retrieved during parsing and decoding of the slice header 21 and the decoder can then locate the particular parameter set that is identified by the parameter set identifier.
  • the at least one syntax element is thereby retrieved from the identified parameter set.
  • the encoded representation 20 comprises, such as in the slice header 21 , a first parameter set identifier, such as a PPS identifier, identifying a first parameter set, such as a PPS.
  • This first parameter set could then comprise a second parameter set identifier, such as an SPS identifier, identifying a second parameter set, such an SPS.
  • the at least one syntax element could then be present in the second parameter set.
  • the at least one syntax element is thereby obtained based on the second parameter set identifier as obtained using the first parameter set identifier retrieved from the encoded representation 20.
  • any syntax element could be retrieved directly from the encoded representation 20 or all could be retrieved from a parameter set identified based on a parameter set identifier obtained based on the encoded representation 20.
  • at least one of the multiple syntax element is directly retrieved from the encoded representation 20 and at least one of the multiple syntax element is retrieved from one or more parameter sets.
  • the multiple syntax elements could be carried in different parameter sets, which are identified based on data, i.e. one or more parameter set identifiers, carried in the encoded representation 20.
  • the variable OutputDistance i.e. the output distance value
  • the syntax elements may be signaled in slice headers or in parameter sets.
  • Reference pictures in the DPB that have not yet been output when a current picture [CurrPic) is to be decoded and that have a POC value, such as represented by PicOrderCntVal, lower than PicOrderCnt ⁇ CurrPic ) - OutputDistance are output.
  • OutputDistance is a non-negative integer. In an embodiment there is no limitation of the maximum value for OutputDistance.
  • OutputDistance is, though, limited to be in the range from 0 and N, inclusive, for some positive integer N.
  • N is MaxPicOrderCntLsbA , which is the maximum possible value that can be signaled for a syntax element pic_order_cnt_lsb.
  • PicOrderCnt ⁇ is defined to be consistently increasing relative to the latest IDR picture in decoding order.
  • the variable PicOrderCntVal is derived as PicOrderCntMsb + pic_order_cnt_lsb, where pic_order_cnt_lsb is signaled for each picture and PicOrderCntMsb is calculated from a previous picture.
  • the variable MaxPicOrderCntLsb which is calculated from syntax elements in the SPS, is used to define the range of possible values for pic_order_cnt_lsb, i.e. 0 to MaxPicOrderCntLsbA , inclusive.
  • the maximum value of the variable OutputDistance is defined based on the syntax element in the SPS used to define the range of the parameter pic_order_cnt_lsb, based on which the POC value ⁇ PicOrderCntVal) of a picture is calculated.
  • the maximum value of the variable OutputDistance, i.e. N is maxPOCA if wrapped POC is used.
  • PicOrderCnt ⁇ the function PicOrderCnt ⁇ ) is defined with wrap around.
  • POC wrap around works like this. Assume a sequence of pictures that constructs a video sequence. Regardless of the length of the video sequence a display order number can be assigned to each picture that simply represents the order in which the pictures should be displayed. Denote this number the "true_POC". However, in order to save bits during encoding the true_POC is not signaled in the bitstream. Instead a range-limited syntax element is encoded referred to as "POC" and limited to be in the range from 0 to MaxPOCA , inclusive. One possible way to calculate POC from the true_POC is to use the modulo operation denoted "%", i.e.
  • POC true_POC % MaxPOC. That means, for example, that the true_POC value equal to MaxPOC+ ⁇ will be given the POC value 1.
  • a DiffPOC ⁇ ) function can be defined for wrapped POC to give correct differences, i.e. distances, between two POC values as long as -MaxPOCIl ⁇ true_POC distance ⁇ MaxPOCIl.
  • Fig. 5 illustrates additional, optional steps of the method in Fig. 4.
  • the method starts in step S10, in which a first output flag is retrieved based on the encoded representation.
  • This retrieval of the first output flag is advantageously performed by retrieving the first output flag from the encoded representation, such as from the slice header.
  • the first output flag is retrieved from a parameter set identified based on data retrieved from the encoded representation, such as a parameter set identifier present in the slice header.
  • this first output flag is investigated in the optional step S11. If the first output flag has a first predefined value, such as It , the method continues to step S12. In this step S12 an output distance syntax element is retrieved based on the encoded representation.
  • This output distance syntax element is preferably present in the encoded representation, such as in the slice header. Alternatively, it is present in a parameter set identifiable based on a parameter set identifier present in the encoded parameter set.
  • the retrieved output distance syntax element is employed in step S13 to determine an output distance value.
  • the method then continues to step S1 of Fig. 4 where the POC limit is calculated based on a POC value of the current picture and the output distance value, preferably equal to the POC value of the current picture subtracted by the output distance value determined in step S13.
  • the output distance value is determined in step S13 to be equal to half of a difference syntax element representing a largest POC difference that can be signaled from the encoder to the decoder, e.g. MaxPicOrderCntLsbl2, if the output distance syntax element has a predefined value, preferably 0 or ( n.
  • the output distance value is preferably determined in step S13 to be equal to the output distance syntax element if the output distance syntax element has a value different from the predefined value, such as 0.
  • step S14 the output distance value is preferably determined to be equal to zero.
  • step S1 of Fig. 4 the POC limit is calculated based on the output distance value.
  • the POC limit is calculated based on the POC value of the current picture and the output distance value, more preferably the POC limit is equal to the POC value of the current picture subtracted by the (zero) output distance value.
  • output_all jprecedingjpicsjiag first output flag
  • output_distance_idc output distance syntax element
  • a further syntax element could be used to signal whether syntax element(s) to be used for calculating the output distance value and therefore the POC limit is present for each picture or if the output distance value could be inferred to be a specific value.
  • Such an embodiment is shown in Fig. 6.
  • Fig. 6 starts in step S20 where a second output flag is retrieved based on the encoded representation.
  • This second output flag is preferably retrieved from a parameter set, such as PPS or SPS, using a parameter set identifier retrieved from the encoded representation or retrieved from a second parameter set identified based on a second parameter set identifier present in the encoded representation. It could alternatively be possible to include the second output flag in the encoded representation, such as in the slice header.
  • step S21 it is investigated whether the second output flag has a predefined value, preferably I t .
  • the method continues to step S22 where the output distance value is determined or inferred to be set equal to a predefined value, preferably zero.
  • the method then continues to step S1 where the POC limit is calculated based on the zero output distance value.
  • the POC limit is calculated to be based on, preferably equal to, the POC value of the current picture subtracted by the output distance value. Since the output distance value is zero in this case the POC limit is preferably equal to the POC value of the current picture.
  • the method continues to step S10.
  • the second output flag is, for instance, equal to (Li the retrieving steps S10, S12 and the determining step S13 of Fig. 5 or the retrieving step S10 and the determining step S14 of Fig. 5 are preferably performed in order to determine the output distance value, which is used in step S1 to calculate the POC limit.
  • another syntax element output_distance_always_zero (second output flag) in SPS, PPS or in another appropriate data field defines whether syntax elements used to calculate the variable OutputDistance (output distance value) shall be present for each picture or if OutputDistance shall be inferred to be a specific value. If syntax elements used to calculate OutputDistance are not present it is preferred that OutputDistance is inferred to be set equal to 0. In an embodiment the syntax element output_distance_always_zero is a one bit flag. OutputDistance can then be calculated as exemplified below.
  • step S2 of Fig. 4 comprises outputting, in increasing order of POC values starting from a lowest POC value, decoded reference pictures that i) are stored in the DPB, ii) have a respective POC value that is lower than the POC limit and iii) are marked as "needed for output".
  • decoded reference pictures stored in frame buffers of the DPB can be marked as needed for output if they need to be output, e.g. for display.
  • a decoded reference picture that is not needed for output, such as has already been output, e.g. for display, is typically marked as "not needed for output”.
  • the method also comprises an additional step S30 as shown in Fig. 7.
  • step S30 the decoded reference pictures that are output in step S2 are marked as not needed for output. This means that the decoded reference picture(s) that previously was(were) marked as needed for output and that was(were) output in step S2 of Fig. 4 is(are) remarked as not needed for output in step S30. This remarking is used to indicate that the decoded reference picture(s) has(have) already been output for display and therefore do(es) not need to be output any longer.
  • An optional but preferred additional step S31 of the method comprises emptying any frame buffer of the DPB that stores a decoded reference picture marked as "unused for reference” and marked, such as in step S30, as "not needed for output". Hence, at this step S31 one or more of the frame buffers of the DPB could become empty and available for storing a new decoded picture. If any frame buffer is emptied in step S31 the DPB fullness is preferably reduced by the corresponding number of frame buffers that have been emptied in step S31.
  • the encoded representation can be decoded in step S40 of Fig. 8 to get a current decoded picture.
  • Decoding of pictures using the slice data of the encoded representation and control information as defined by the slice header is performed according techniques well known in the art.
  • the current decoded picture obtained in step S40 can then be stored in an empty frame buffer in the DPB in step S41.
  • outputting of decoded reference pictures are preferably performed prior to decoding the slice data of the encoded representation of the current slice.
  • the removal of pictures from the DPB before decoding of the current picture, but after parsing the slice header of the first slice of the current picture proceeds as follows.
  • the decoding process for reference picture set is invoked. If the current picture is not an IDR picture, frame buffers containing a picture which is marked as "not needed for output” and "unused for reference” are emptied without output.
  • the DPB fullness is decremented by the number of frame buffers emptied.
  • the output process consists, in an embodiment, of the following ordered steps: 1.
  • the picture that is first for output is selected as the one having the smallest value of PicOrderCntVal of all pictures in the DPB marked as "needed for output”.
  • the picture is optionally cropped, using the cropping rectangle specified in the active SPS for the picture, the optionally cropped picture is output, and the picture is marked as "not needed for output".
  • the frame buffer that included the picture that was output and optionally cropped contains a picture marked as "unused for reference”
  • the frame buffer is emptied and the DPB fullness is decremented by 1.
  • wrapped POC is used to signal the POC values. For that case PicOrderCntVal and PicOrderCnt ⁇ ) are calculated relative each current picture.
  • Fig. 9 is a flow diagram of a method of encoding a current picture of a video stream in an encoder according to an embodiment.
  • the method generally starts in step S50 where a POC limit is determined to have a value enabling, determining or defining a target state of a DPB in a decoder for the current picture.
  • the POC limit determined in step S50 defines a number of decoded reference pictures to be output from the DPB in a picture output process invoked for the current picture.
  • a next step S51 at least one syntax element representative of the value of the POC limit determined in step S50 is determined.
  • the at least one syntax element determined in step S51 enables determination or calculation of the POC limit.
  • the current picture is encoded in step S52 to get an encoded representation of the current picture.
  • This encoded representation may be in the form of an encoded representation 20 of a slice comprising a slice header 21 and slice data 22, such as packed into a NAL unit 11 as shown in Fig. 2.
  • the encoded representation of the current picture could be in form of, if the current picture consists of multiple slices, multiple respective encoded representations 20 of slices, each having a respective slice header 21 and slice data 22. These multiple encoded representations 20 of slices could be packed in separate NAL units 11.
  • Encoding of a picture is performed according to techniques well known in the art of picture and video coding.
  • the at least one syntax element determined in step S51 is associated with or to the encoded representation in step S53.
  • This step S53 can be performed prior to, after or substantially in parallel with step S52.
  • Associating the at least one syntax element with the encoded representation can be performed according to various embodiments as mentioned herein.
  • the at least one syntax element could, for instance, be added to the encoded representation, such as inserted into the slice header of the encoded representation in step S53.
  • each encoded representation of a slice for the picture preferably comprises the at least one syntax element.
  • the at least one syntax element could be inserted, in step S53, into one or more parameter sets.
  • step S53 involves inserting at least one syntax element into the slice header and inserting at least one parameter set identifier into the slice header, where this at least one parameter set identifier enables identification of at least one parameter set carrying at least one syntax element as determined in step S51.
  • the encoding method as shown in Fig. 9 can therefore, by determining the POC limit to have a particular value, determine or define a target state that the DPB in the decoder will have for the current picture. This means that if a particular target state of the DPB is desired, such as a desired DPB fullness, the POC limit is determined in step S50 to have a value that will achieve the particular target state, such as DPB fullness, when processing the encoded representation of the current picture at the decoder.
  • the DPB preferably comprises a number of frame buffers in which decoded reference pictures are stored.
  • the POC limit is determined to have a value such that at least one frame buffer in the DPB is emptied from a decoded reference picture marked as unused for reference if there are no empty frame buffers in the DPB prior to emptying the at least one frame buffer.
  • the DPB fullness prior to processing the current picture at the decoder, is equal to the DPB size so there are no empty frame buffers in which the current picture can be entered once it has been decoded by the decoder.
  • the desired target state of the DPB is in this case therefore to achieve a DPB fullness that is lower than the DPB size to allow room for the current picture in the DPB.
  • the encoder has all the knowledge of the DPB status and therefore knows which decoded reference pictures that are stored in the frame buffers of the DPB in the decoder at the time of decoding of the current picture. This means that the encoder also knows the respective POC values of these stored decoded reference pictures, their markings and the POC value of the current picture. All these, i.e. the POC values and the markings, are actually determined and set by the encoder.
  • the encoder can thereby, such as based on the POC values of the decoded reference pictures stored in the DPB prior to decoding the current picture and the POC value of the current picture, determine a POC limit to have a value so that when the decoder performs the previously described picture output method or process at least one frame buffer of the DPB is emptied.
  • the POC limit is determined to have value defined based on the coding structure of the video stream and preferably based on POC values of future pictures of the video stream, i.e. pictures following the current picture in decoding order.
  • This embodiment is particular suitable if there already is at least one empty frame buffer in the DPB for the current picture.
  • the POC limit value is in this embodiment determined based on the coding structure of the video stream, i.e. the encoding and decoding relationships between pictures in the video stream. Information of pictures that are encoded and thereby decoded based on other pictures, i.e.
  • the encoder can, by determining a suitable value of the POC limit in step S50, make sure that reference pictures that have already been marked as "unused for reference", i.e. are no longer needed as reference picture for the current and/or following pictures, are emptied from the DPB so that new reference pictures can be added to the DPB.
  • the POC limit is thereby determined so that a target status of the DPB is achieved and any reference pictures that might be needed as reference for future picture decoding could be entered in frame buffers of the DPB.
  • the syntax element that is determined in step S51 of Fig. 9 depends on the particular embodiment.
  • the syntax element determined in step S51 could be at least one of output_distance_idc, MaxPicOrderCntLsb, output_all receding icsjlag, output_distance_always_zero.
  • the at least one syntax element determined in step S51 could, for instance, be the previously mentioned output distance syntax element and preferably also the first output flag and optionally the second output flag.
  • the maximum picture order count value could also be included as syntax element.
  • the output distance value [OutputDistance) could be determined based on the POC limit, such as PicOrderCnt ⁇ CurrPic ) - POC limit. In an embodiment, if this OutputDistance becomes zero, the first output flag [outpu all jirecedingjiicsjlag) could be set to zero. Alternatively, if OutputDistance should be equal to half of the largest POC difference [MaxPicOrderCntLsblT) the first output flag is preferably set to one and the output distance syntax element [output_distance_idc) is preferably set to zero. Otherwise the first output flag is preferably set to one and the output distance syntax element is preferably set to the determined value of the output distance value.
  • the POC limit such as PicOrderCnt ⁇ CurrPic ) - POC limit.
  • the first output flag [outpu all jirecedingjiicsjlag) could be set to zero.
  • the first output flag is preferably set to
  • the second output flag [output_distance_always_zero) is used. This is in particular beneficial if multiple of the pictures in the video stream should have an output distance value of zero.
  • a bitstream restriction may be imposed on the value of the output distance.
  • a reason for this is that the bitstream otherwise could become somewhat sensitive to loss of data packets carrying encoded slices (see Fig. 2) if very large values are allowed for the output distance.
  • Fig. 10 is a flow diagram illustrating additional, optional steps of the method in Fig. 9 when using such bitstream restrictions. This embodiment is particular suitable for HEVC or other video codecs in which each picture has a respective POC value and a respective temporal identifier.
  • a value X is compared to another value X'.
  • the value X represents and is preferably equal to the highest POC value of all decoded pictures of the video stream with temporal identifier lower than or equal to a temporal identifier of the current picture and that have been output prior to invoking the output method or process for the current picture.
  • the value X' correspondingly represents and is preferably equal to the highest POC value of all decoded pictures of the video stream with temporal identifier lower than or equal to the temporal identifier of the current picture and that have been output after invoking the output method or process for a previous picture in the video stream.
  • This previous picture is previous to the current picture according to the decoding order of the video stream.
  • the previous picture has a temporal identifier lower than or equal to the temporal identifier of the current picture.
  • the previous picture is preferably the closest, according to the decoding order, picture with temporal identifier equal to or lower than the temporal identifier of the current picture that precedes the current picture according to the decoding order.
  • step S60 If the value X is equal to the value X' as investigated in step S60 the method continues to step S51 of Fig. 9. Hence, in this embodiment no restriction is needed for the output distance value.
  • step S61 a new value of the syntax element is set, e.g. by setting a new value of the output distance value, which is smaller than the POC of the current value subtracted by the value X.
  • the output distance value is determined to be smaller than PicOrderCntVal ⁇ CurrPic ) - X.
  • Fig. 12 is a schematic block diagram of a decoder 40 according to an embodiment.
  • the decoder 40 comprises a decoded picture buffer (DPB) 48 configured to store decoded pictures of a video stream.
  • the decoder 40 also comprises a limit calculator 41 , also denoted limit calculating unit, means or module.
  • the limit calculator 41 is configured to calculate a POC limit based on at least one syntax element retrieved based on an encoded representation of a current picture of the video stream.
  • a picture outputting unit 42 also denoted picture output or picture outputting means or module, is implemented in the decoder 40 and configured to output decoded reference pictures stored in the DPB 48 and having a respective POC value that is lower than the POC limit.
  • the decoder 40 comprises an optional flag retriever 43, also denoted flag retrieving unit, means or module.
  • the flag retriever 43 is configured to retrieve a first output flag based on the encoded representation.
  • the flag retriever 43 could be configured to retrieve the first output flag from a slice header of the encoded representation or from a parameter set using a parameter set identifier obtained based on, such as present in, the slice header of the encoded representation as previously discussed herein.
  • the decoder 40 preferably, in this particular embodiment, also comprises an optional element retriever 44, also denoted element retrieving unit, means or module.
  • the element retriever 44 is configured to retrieve an output distance syntax element if the first output flag retrieved by the flag retriever 43 has a first predefined value, such as It .
  • the element retriever 44 is configured to retrieve this output distance element based on the encoded representation, such as from a slice header in the encoded representation.
  • An optional value determiner 45, also denoted value determining unit, means or module, of the decoder 40 is configured to determine an output distance value based on the output distance syntax element retrieved by the element retriever 44.
  • the limit calculator 41 is configured to calculate the POC limit to be based on, preferably equal to, the POC value of the current picture subtracted by the output distance value determined by the value determiner 45.
  • the value determiner 45 is, in an embodiment, configured to determine the output distance value to be based on, preferably equal to, half of a largest POC difference if the output distance syntax element retrieved by the element retriever 44 has a predefined value, such as 0.
  • the value determiner 45 is preferably configured to determine the output distance value to be based on, preferably equal to, the value defined or represented by the output distance syntax element.
  • the value determiner 45 is configured to determine the output distance value to be equal to zero if the first output flag retrieved by the flag retriever 43 has a second predefined value, such as (Li.
  • the limit calculator 41 preferably calculates the POC limit to be equal to the POC value of the current picture subtracted by the output distance value, i.e. equal to the POC value of the current picture since the output distance value is zero in this example.
  • the flag retriever 43 is configured to retrieve a second output flag based on the encoded representation.
  • the flag retriever 43 typically retrieves the second output flag from a parameter set identified based on a parameter set identifier present in a slice header of the encoded representation or present in a second parameter set identified based on a second parameter set identifier present in the slice header of the encoded representation.
  • the flag retriever 43 is then configured to retrieve the first output flag if this second output flag has a predefined value, such as (Li. If the second output flag instead has a second predefined value, such as It , the output distance value could have a predefined value, such as zero, so that no retrieval of any first output flag or retrieval of any output distance syntax element is needed to calculate the POC limit.
  • the picture outputting unit 42 is, in an embodiment, configured to output, in increasing order of POC values starting with a lower POC value, decoded reference pictures that are stored in the DPB and have a respective POC value lower than the POC limit calculated by the limit calculator 41 and are marked as needed for output.
  • the decoder 40 comprises, in an optional embodiment, a picture marking unit 46, also denoted picture marker or picture marking means or module.
  • the picture marking unit 46 is configured to mark the decoded reference pictures that are output by the picture outputting unit 42 as "not needed for output" to indicate that the decoded reference picture(s) already has(have) been output.
  • the picture outputting unit 42 is preferably configured to empty any frame buffer 49 of the DPB 48 storing a decoded reference picture that is marked as "not needed for output” and marked as "unused for reference”.
  • the decoder 40 preferably also comprises a decoding unit 47, also denoted picture decoder or decoding means or module.
  • the decoding unit 47 is configured to decode the encoded representation to get a current decoded picture. This current decoded picture can then be stored in the DPB 48 in an empty frame buffer 49.
  • the decoder decodes 40 the information needed to calculate the POC limit (limit X), e.g. PicOrderCntVal of the current picture and OutputDistance.
  • the syntax elements that may be used for calculating OutputDistance may be specified by standard specifications.
  • the decoder 40 calculates OutputDistance based on the received syntax elements.
  • the decoder 40 displays pictured in the DPB 48 marked as "needed for output” with PicOrderCntVal lower than the POC limit (limit X), e.g. defined as PicOrderCnt ⁇ CurrPic ) - OutputDistance, in increasing order of PicOrderCntVal starting with the one with lowest PicOrderCntVal. Pictures that have been displayed are marked as "not needed for output”.
  • limit X e.g. defined as PicOrderCnt ⁇ CurrPic
  • the current picture is decoded and marked according to its OutputFlag, e.g. "needed for output” or “not needed for output”.
  • the decoder 40 of Fig. 12 with its including units 41 -47 could be implemented in hardware.
  • circuitry elements that can be used and combined to achieve the functions of the units 41-47 of the decoder 40. Such variants are encompassed by the embodiments.
  • Particular examples of hardware implementation of the decoder 40 is implementation in digital signal processor (DSP) hardware and integrated circuit technology, including both general-purpose electronic circuitry and application- specific circuitry.
  • DSP digital signal processor
  • the decoder 50 can also be implemented by means of a processor 52 and a memory 54 as illustrated 5 in Fig. 13.
  • the decoder 50 is implemented e.g. by one or more of a processor 52 and adequate software with suitable storage or memory 54 therefore, a programmable logic device (PLD) or other electronic component(s).
  • the decoder 50 preferably comprises an input or input unit 51 configured to receive the encoded representations of the video stream, such as in the form of NAL units.
  • a corresponding output or output unit 53 is configured to output the decoded 10 pictures.
  • the decoder can be implemented in a device, such as a mobile device exemplified as mobile phones, tablets, video camera, set-top-box, etc.
  • Fig. 1 1 illustrates such an example where the decoder 32 is located in a receiver 30, such as in a video camera or a display, e.g. in a mobile device.
  • the receiver 30 such as in a video camera or a display, e.g. in a mobile device.
  • the 15 30 then comprises an input or input unit 31 configured to receive a coded bitstream, such as data packets of NAL units as shown in Fig. 2.
  • the encoded representations of the NAL units are decoded by the decoder 32 as disclosed herein.
  • the decoder 32 preferably comprises or is connected to a reference picture buffer 34 that temporarily stores already decoded reference pictures 35 that are to be used as reference pictures for other pictures in the video stream. Decoded pictures are output from the
  • These output pictures are sent to be displayed to a user on a screen or display of or connected, including wirelessly connected, to the receiver 30.
  • Fig. 15 is a schematic block diagram of an encoder 70 according to an embodiment.
  • the encoder 70 25 comprises a limit determiner 71 , also denoted limit determining unit, means or module.
  • the limit determiner 71 is configured to determine a POC limit to have a value enabling, defining or determining a target state of a DPB in a decoder for a current picture of a video stream. This determined POC limit defines a number of decoded reference pictures to be output from the DPB in a picture output process invoked by the decoder for the current picture.
  • a syntax element determiner 72 also denoted syntax element determining unit, means or module, is configured to determine at least one syntax element representative of the value of the POC limit determined by the limit determiner 71.
  • An encoding unit 73 also denoted picture encoder or encoding means or module, of the encoder 70 is configured to encode the current picture to get an encoded representation of the current picture.
  • the encoder 70 also comprises an associating unit 74, also denoted associator or associating means or module.
  • the associating unit 74 is configured to associate the at least one syntax element determined by the syntax element determiner 72 with or to the encoded representation.
  • the associating unit 74 could be configured to include the a syntax element into the slice header of the encoded representation and/or include a parameter set identifier in the slice header where this parameter set identifier enables identification of a parameter set comprising a syntax element.
  • the limit determiner 71 is configured to determine the POC limit to have a value selected so that at least one frame buffer in the DPB is emptied from a decoded reference picture marked as unused for reference if there are no such empty frame buffers in the DPB prior to emptying the at least one frame buffer.
  • the particular value determined for the POC limit frees a frame buffer and thereby makes room in the DPB for the current picture during decoding.
  • the limit determiner 71 is configured to determine the POC limit to have a value defined based on the coding structure of at least a portion of the video stream and preferably of POC values of future pictures of the video stream. This embodiment is particular suitable if there is at least one empty frame buffer in the DPB for the current picture.
  • the limit determiner 71 preferably and at least partly determines the POC limit based on coding structure, i.e. how pictures of the video stream are encoded and decoded relative to each other, i.e. used as reference pictures, for the current picture but preferably also for future pictures of the video stream that follow the current picture according to the decoding order.
  • the encoder 70 may impose bitstream restriction to syntax element determined by the syntax element determiner 72.
  • the encoder 70 preferably comprises a comparator 75, also denoted comparing unit, means or module.
  • the comparator 75 is configured to compare value X with a value X'.
  • the value X is preferably equal to the highest POC of all decoded pictures of the video stream with temporal identifier lower than or equal to a temporal identifier of the current picture and that have been output prior to invoking the picture output process for the current picture.
  • the value X' is preferably equal to the higher POC value of all decoded pictures with temporal identifier lower than or equal to the temporal identifier of the current picture and that have been output after invoking a picture output process for a previous picture.
  • the previous picture is previous to the current picture according to the decoding order of the video stream and has a temporal identifier lower than or equal to the temporal identifier of the current picture.
  • the syntax element determiner 72 is configured to set a new value of at least one syntax element of the at least one syntax element, which value is smaller than the POC value of the current picture subtracted by the value X if the value X is different from the value X' as determined by the comparator 75. If the value X is equal to the value X' no new value of the at least one syntax element needs to be determined.
  • the encoder 70 preferably ensures that there is an empty frame buffer in the DPB that can be used by the new picture. If the DPB is full, i.e.
  • a frame buffer is emptied by the encoder by selecting a value of OutputDistance such that at least one picture marked as "unused for reference" is output. Otherwise the encoder 70 preferably selects any value for OutputDistance within the specified allowed range according to what is needed for the coding structure and POC values of future pictures. If the selected value for OutputDistance is larger than what is allowed by an optional bitstream restriction the encoder 70 preferably selects a new value for OutputDistance that is not larger than what is required by the bitstream restriction. The encoder 70 encodes the value of OutputDistance using the syntax elements form which the OutDistance is calculated at the decoder. The picture is encoded.
  • the encoder 70 of Fig. 15 with its including units 71-75 could be implemented in hardware.
  • circuitry elements that can be used and combined to achieve the functions of the units 71-75 of the encoder 70. Such variants are encompassed by the embodiments.
  • Particular examples of hardware implementation of the encoder 70 is implementation in digital signal processor (DSP) hardware and integrated circuit technology, including both general-purpose electronic circuitry and application- specific circuitry.
  • DSP digital signal processor
  • the encoder 80 can also be implemented by means of a processor 82 and a memory 84 as illustrated in Fig. 16.
  • the encoder 80 is implemented e.g.
  • the encoder 80 preferably comprises an input or input unit 81 configured to receive the pictures of the video stream.
  • a corresponding output or output unit 83 is configured to output the encoded representations of the pictures, such as in the form of NAL units.
  • the encoder can be implemented in a device, such as a mobile device exemplified as mobile phones, tablets, video camera, etc.
  • Fig. 14 illustrates an example of such a device in the form of a transmitter 60, e.g. implemented in a video camera e.g. in a mobile device.
  • the transmitter 60 then comprises an input or input unit 61 configured to receive pictures of a video stream to be encoded.
  • the pictures are encoded by the encoder 62 as disclosed herein.
  • Encoded pictures are output from the transmitter 60 by an output or output unit 63 in the form of a coded bitstream, such as of NAL units or data packets carrying such NAL units as shown in Fig. 2.
  • a network node may use the embodiments. For instance, pictures are forwarded by the network node and temporal layer switches are performed at temporal layer switching points. According to the embodiments this picture forwarding can be performed by the network node without having to care about the DPB status in the decoder, i.e. without regard to what pictures that have been output, for different temporal layers.
  • the present embodiments can be applied to different video codecs and different types of extensions, including, but not limited to, multi-view video codecs and scalable video codecs.
  • Temporal identifiers as discussed herein could, in alternative embodiments, be replaced by general layer identifiers that do not necessarily have to relate to different temporal layers. Such layer identifier could, for instance, define various camera view, different scalability layers, spatial layers, etc.
  • picture order count or POC is used herein as identifier of the pictures in the video stream, either in consistently increasing order relative to a latest IDR picture or by using POC wrap around. The embodiments are, however, not limited to using picture order count values as picture identifiers. In alternative embodiment other types of picture identifiers could be used instead of POC values.
  • the functional blocks may include or encompass, without limitation, digital signal processor (DSP) hardware, reduced instruction set processor, hardware (e.g., digital or analog) circuitry including but not limited to application specific integrated circuit(s) (ASIC), and (where appropriate) state machines capable of performing such functions.
  • DSP digital signal processor
  • ASIC application specific integrated circuit

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)
EP12818654.1A 2012-01-20 2012-12-11 Ausgabe von dekodierten referenzbildern Withdrawn EP2805490A1 (de)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US201261588764P 2012-01-20 2012-01-20
PCT/SE2012/051372 WO2013109179A1 (en) 2012-01-20 2012-12-11 Output of decoded reference pictures

Publications (1)

Publication Number Publication Date
EP2805490A1 true EP2805490A1 (de) 2014-11-26

Family

ID=47604013

Family Applications (1)

Application Number Title Priority Date Filing Date
EP12818654.1A Withdrawn EP2805490A1 (de) 2012-01-20 2012-12-11 Ausgabe von dekodierten referenzbildern

Country Status (2)

Country Link
EP (1) EP2805490A1 (de)
WO (1) WO2013109179A1 (de)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2015056182A2 (en) 2013-10-15 2015-04-23 Nokia Technologies Oy Video encoding and decoding
US9648348B2 (en) * 2013-10-23 2017-05-09 Qualcomm Incorporated Multi-layer video file format designs
KR20210118183A (ko) * 2019-02-01 2021-09-29 프라운호퍼 게젤샤프트 쭈르 푀르데룽 데어 안겐반텐 포르슝 에. 베. 서브픽처 또는 영역 단위 랜덤 액세스가 가능한 비디오 코덱 및 이를 이용한 비디오 합성을 위한 개념

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See references of WO2013109179A1 *

Also Published As

Publication number Publication date
WO2013109179A1 (en) 2013-07-25

Similar Documents

Publication Publication Date Title
US11849144B2 (en) Signaling of state information for a decoded picture buffer and reference picture lists
US10057571B2 (en) Reference picture list handling
US10951899B2 (en) Extension data handling
RU2581566C2 (ru) Сигнализация опорного изображения
US9774927B2 (en) Multi-layer video stream decoding
US20200413042A1 (en) Multi-Layer Video Stream Encoding and Decoding
EP2805490A1 (de) Ausgabe von dekodierten referenzbildern
EP2936809B1 (de) Mehrschicht videostrom-dekodierung
US20240357117A1 (en) Extension data handling
US20140233653A1 (en) Decoder and encoder for picture outputting and methods thereof

Legal Events

Date Code Title Description
PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

17P Request for examination filed

Effective date: 20140801

AK Designated contracting states

Kind code of ref document: A1

Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR

DAX Request for extension of the european patent (deleted)
STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE APPLICATION HAS BEEN WITHDRAWN

18W Application withdrawn

Effective date: 20170407