WO2004098196A1 - Picture coding method - Google Patents

Picture coding method Download PDF

Info

Publication number
WO2004098196A1
WO2004098196A1 PCT/FI2004/050053 FI2004050053W WO2004098196A1 WO 2004098196 A1 WO2004098196 A1 WO 2004098196A1 FI 2004050053 W FI2004050053 W FI 2004050053W WO 2004098196 A1 WO2004098196 A1 WO 2004098196A1
Authority
WO
WIPO (PCT)
Prior art keywords
pictures
coded
picture
redundant
picture information
Prior art date
Application number
PCT/FI2004/050053
Other languages
French (fr)
Inventor
Miska Hannuksela
Original Assignee
Nokia Corporation
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nokia Corporation filed Critical Nokia Corporation
Priority to JP2006505656A priority Critical patent/JP2006526908A/en
Priority to MXPA05011533A priority patent/MXPA05011533A/en
Priority to EP04730282A priority patent/EP1618747A1/en
Priority to BRPI0409491-3A priority patent/BRPI0409491A/en
Publication of WO2004098196A1 publication Critical patent/WO2004098196A1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/157Assigned coding mode, i.e. the coding mode being predefined or preselected to be further used for selection of another element or parameter
    • H04N19/159Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/30Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
    • H04N19/31Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability in the temporal domain
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • H04N19/107Selection of coding mode or of prediction mode between spatial and temporal predictive coding, e.g. picture refresh
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/172Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a picture, frame or field
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/20Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using video object coding
    • H04N19/29Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using video object coding involving scalability at the object level, e.g. video object layer [VOL]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/85Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression
    • H04N19/89Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression involving methods or arrangements for detection of transmission errors at the decoder

Definitions

  • the invention relates to method for encoding pictures, in which primary coded pictures and redundant coded pictures of primary coded pictures are formed.
  • the invention also relates to a system, encoder, decoder, transmitting device, receiving device, software program, a storage medium and a bitstream.
  • Published video coding standards include ITU-T H.261, ITU-T H.263, ISO/IEC MPEG-1 , ISO/IEC MPEG-2, and ISO/IEC MPEG-4 Part 2. These standards are herein referred to as conventional video coding standards.
  • Video communication systems can be divided into conversational and non-conversational systems.
  • Conversational systems include video conferencing and video telephony. Examples of such systems include ITU-T Recommendations H.320, H.323, and H.324 that specify a video conferencing/telephony system operating in ISDN, IP, and PSTN networks respectively.
  • Conversational systems are characterized by the intent to minimize the end-to-end delay (from audio-video capture to the far-end audio-video presentation) in order to improve the user experience.
  • Non-conversational systems include playback of stored content, such as Digital Versatile Disks (DVDs) or video files stored in a mass memory of a playback device, digital TV, and streaming.
  • DVDs Digital Versatile Disks
  • video files stored in a mass memory of a playback device, digital TV, and streaming.
  • JVT Joint Video Team
  • H.26L An earlier standardization project in ITU-T
  • H.26L An earlier standardization project in ITU-T
  • the goal of the JVT standardization is to release the same standard text as ITU-T Recommendation H.264 and ISO/IEC International Standard 14496-10 (MPEG-4 Part 10).
  • the draft standard is referred to as the JVT coding standard in this paper, and the codec according to the draft standard is referred to as the JVT codec.
  • the codec specification itself distinguishes conceptually between a video coding layer (VCL), and a network abstraction layer (NAL).
  • VCL contains the signal processing functionality of the codec, things such as transform, quantization, motion search/compensation, and the loop filter. It follows the general concept of most of today's video codecs, a macroblock-based coder that utilizes inter picture prediction with motion compensation, and transform coding of the residual signal.
  • the output of the VCL encoder are slices: a bit string that contains the macroblock data of an integer number of macroblocks, and the information of the slice header (containing the spatial address of the first macroblock in the slice, the initial quantization parameter, and similar).
  • Macroblocks in slices are ordered consecutively in scan order unless a different macroblock allocation is specified, using the so-called Flexible Macroblock Ordering syntax. In-picture prediction is used only within a slice.
  • the NAL encapsulates the slice output of the VCL into Network Abstraction Layer Units (NALUs), which are suitable for the transmission over packet networks or the use in packet oriented multiplex environments.
  • NALUs Network Abstraction Layer Units
  • JVT's Annex B defines an encapsulation process to transmit such NALUs over byte-stream oriented networks.
  • the optional reference picture selection mode of H.263 and the NEWPRED coding tool of MPEG-4 Part 2 enable selection of the reference frame for motion compensation per each picture segment, e.g., per each slice in H.263. Furthermore, the optional Enhanced Reference Picture Selection mode of H.263 and the JVT coding standard enable selection of the reference frame for each macroblock separately.
  • FIG. 1 shows an example of a temporal scalability scheme, which is herein referred to as recursive temporal scalability.
  • the example scheme can be decoded with three constant frame rates.
  • Figure 2 depicts a scheme referred to as Video Redundancy Coding, where a sequence of pictures is divided into two or more independently coded threads in an interleaved manner.
  • the arrows in these and all the subsequent figures indicate the direction of motion compensation and the values under the frames correspond to the relative capturing and displaying times of the frames.
  • Figure 8 shows a block diagram of a general video communications system 800. Due to the fact that uncompressed video requires a huge bandwidth, the input video 801 is compressed in a transmitting device 802 by a source coder 803 to a desired bit rate.
  • the source coder 803 can be divided into two components, namely waveform coder 803.1 and entropy coder 803.2.
  • the waveform coder 803.1 performs lossy video signal compression whereas the entropy coder 803.2 losslessly converts the output of waveform coder 803.1 into a binary sequence.
  • the transport coder 804 encapsulates the compressed video according to the transport protocols in use. It may manipulate the compressed video in other ways too. For example, it may interleave and modulate the data.
  • the receiver 807 performs inverse operations to obtain reconstructed video signal for display.
  • the receiver 807 comprises a transport decoder 808 and a source decoder 809.
  • the transport decoder 808 de-capsulates the compressed video input from the transmission channel 805 according to the transport protocols in use.
  • the source decoder 809 can also be divided into two components, namely entropy decoder 809.1 and waveform decoder 809.2.
  • the entropy decoder 809.1 converts the binary sequence from the transport decoder 808 into a waveform to be input to the waveform decoder 809.1.
  • the waveform decoder 809.1 performs video signal decompression and outputs the video signal 810.
  • the receiver 807 may also give feedback to the transmitter. For instance, the receiver may signal the rate of successfully received transmission units. Parameter Set Concept
  • One very fundamental design concept of the JVT codec is to generate self-contained packets, to make mechanisms such as the header duplication unnecessary.
  • the way how this was achieved is to decouple information that is relevant to more than one slice from the media stream.
  • This higher layer meta information should be sent reliably, asynchronously and in advance from the RTP packet stream that contains the slice packets.
  • This information can also be sent in-band in such applications that do not have an out-of-band transport channel appropriate for the purpose.
  • the combination of the higher level parameters is called a Parameter Set.
  • the Parameter Set contains information such as picture size, display window, optional coding modes employed, macroblock allocation map, and others.
  • the encoder and decoder can maintain a list of more than one Parameter Set.
  • Each slice header contains a codeword that indicates the Parameter Set to be used.
  • This mechanism allows to decouple the transmission of the Parameter Sets from the packet stream, and transmit them by external means, e.g. as a side effect of the capability exchange, or through a (reliable or unreliable) control protocol. It may even be possible that they get never transmitted but are fixed by an application design specification.
  • the decoding order of pictures is the same as the display order except for B pictures.
  • a block in a conventional B picture can be bi-directionally temporally predicted from two reference pictures, where one reference picture is temporally preceding and the other reference picture is temporally succeeding in display order. Only the latest reference picture in decoding order can succeed the B picture in display order (exception: interlaced coding in H.263 where both field pictures of a temporally subsequent reference frame can precede a B picture in decoding order).
  • a conventional B picture cannot be used as a reference picture for temporal prediction, and therefore a conventional B picture can be disposed without affecting the decoding of any other pictures.
  • the JVT coding standard includes the following novel technical features compared to earlier standards:
  • the decoding order of pictures is decoupled from the display order.
  • the picture number indicates decoding order and the picture order count indicates the display order.
  • Reference pictures for a block in a B picture can either be before or after the B picture in display order. Consequently, a B picture stands for a bi-predictive picture instead of a bi-directional picture.
  • a picture of any type can either be a reference picture or a non-reference picture. (Thus, a B picture can be used as a reference picture for temporal prediction of other pictures.)
  • a picture can contain slices that are coded with a different coding type.
  • a coded picture may consist of an intra- coded slice and a B-coded slice, for example.
  • Decoupling of display order from decoding order can be beneficial from compression efficiency and error resiliency point of view.
  • FIG. 3 An example of a prediction structure potentially improving compression efficiency is presented in Figure 3. Boxes indicate pictures, capital letters within boxes indicate coding types, numbers within boxes are picture numbers according to the JVT coding standard, and arrows indicate prediction dependencies. Note that picture B17 is a reference picture for pictures B18. Compression efficiency is potentially improved compared to conventional coding, because the reference pictures for pictures B18 are temporally closer compared to conventional coding with PBBP or PBBBP coded picture patterns. Compression efficiency is potentially improved compared to conventional PBP coded picture pattern, because part of reference pictures are bi-directionally predicted.
  • Figure 4 presents an example of the intra picture postponement method that can be used to improve error resiliency.
  • an intra picture is coded immediately after a scene cut or as a response to an expired intra picture refresh period, for example.
  • an intra picture is not coded immediately after a need to code an intra picture arises, but rather a temporally subsequent picture is selected as an intra picture.
  • Each picture between the coded intra picture and the conventional location of an intra picture is predicted from the next temporally subsequent picture.
  • the intra picture postponement method generates two independent inter picture prediction chains, whereas conventional coding algorithms produce a single inter picture chain. It is intuitively clear that the two- chain approach is more robust against erasure errors than the one- chain conventional approach. If one chain suffers from a packet loss, the other chain may still be correctly received. In conventional coding, a packet loss always causes error propagation to the rest of the inter picture prediction chain.
  • a multimedia streaming system consists of a streaming server and a number of players, which access the server via a network.
  • the network is typically packet-oriented and provides little or no means to guaranteed quality of service.
  • the players fetch either pre-stored or live multimedia content from the server and play it back in real-time while the content is being downloaded.
  • the type of communication can be either point-to-point or multicast.
  • point-to-point streaming the server provides a separate connection for each player.
  • multicast streaming the server transmits a single data stream to a number of players, and network elements duplicate the stream only if it is necessary.
  • Initial buffering helps to maintain pauseless playback, because, in case of occasional increased transmission delays or network throughput drops, the player can decode and play buffered data.
  • RTP and RTCP protocols can be used on top of UDP to control realtime communications.
  • RTP provides means to detect losses of transmission packets, to reassemble the correct order of packets in the receiving end, and to associate a sampling time-stamp with each packet.
  • RTCP conveys information about how large a portion of packets were correctly received, and, therefore, it can be used for flow control purposes.
  • Bit errors are typically associated with a circuit-switched channel, such as a radio access network connection in mobile communications, and they are caused by imperfections of physical channels, such as radio interference. Such imperfections may result into bit inversions, bit insertions and bit deletions in transmitted data.
  • Packet errors are typically caused by elements in packet-switched networks. For example, a packet router may become congested; i.e. it may get too many packets as input and cannot output them at the same rate. In this situation, its buffers overflow, and some packets get lost. Packet duplication and packet delivery in different order than transmitted are also possible but they are typically considered to be less common than packet losses.
  • Packet errors may also be caused by the implementation of the used transport protocol stack. For example, some protocols use checksums that are calculated in the transmitter and encapsulated with source-coded data. If there is a bit inversion error in the data, the receiver cannot end up into the same checksum, and it may have to discard the received packet.
  • Second (2G) and third generation (3G) mobile networks including GPRS, UMTS, and CDMA-2000, provide two basic types of radio link connections, acknowledged and non-acknowledged. An acknowledged connection is such that the integrity of a radio link frame is checked by the recipient (either the Mobile Station, MS, or the Base Station Subsystem, BSS), and, in case of a transmission error, a retransmission request is given to the other end of the radio link.
  • the originator Due to link layer retransmission, the originator has to buffer a radio link frame until a positive acknowledgement for the frame is received. In harsh radio conditions, this buffer may overflow and cause data loss. Nevertheless, it has been shown that it is beneficial to use the acknowledged radio link protocol mode for streaming services. A non-acknowledged connection is such that erroneous radio link frames are typically discarded.
  • Loss correction refers to the capability to restore lost data perfectly as if no losses had ever been introduced.
  • Loss concealment refers to the capability to conceal the effects of transmission losses so that they should not be visible in the reconstructed video sequence.
  • a player When a player detects a packet loss, it may request for a packet retransmission. Because of the initial buffering, the retransmitted packet may be received before its scheduled playback time.
  • Some commercial Internet streaming systems implement retransmission requests using proprietary protocols. Work is going on in IETF to standardize a selective retransmission request mechanism as a part of RTCP.
  • Point-to-point streaming systems may also benefit from non-interactive error control techniques.
  • some systems may not contain any interactive error control mechanism or they prefer not to have any feedback from players in order to simplify the system.
  • Streaming servers have to ensure that interactive error control methods do not reserve a major portion of the available network throughput. In practice, the servers may have to limit the amount of interactive error control operations.
  • transmission delay may limit the number of interactions between the server and the player, as all interactive error control operations for a specific data sample should preferably be done before the data sample is played back.
  • Non-interactive packet loss control mechanisms can be categorized to forward error control and loss concealment by post-processing.
  • Forward error control refers to techniques in which a transmitter adds such redundancy to transmitted data that receivers can recover at least part of the transmitted data even if there are transmission losses.
  • Signal-dependent methods require interpretation of the bitstream. An example of such a method is a repetition of sequence or picture header.
  • Signal-independent methods can be used to recover whatever bitstream regardless of the interpreted content of the bitstream. Examples of such methods are error correction codes (e.g. parity codes and Reed-Solomon codes).
  • Error concealment by postprocessing is totally receiver-oriented. These methods try to estimate the correct representation of erroneously received data.
  • Video communication systems can either conceal the loss in displayed images or freeze the latest correct picture onto the screen until a frame which is independent from the corrupted frame is received.
  • a primary coded picture is a primary coded representation of a picture.
  • the decoded primary coded picture covers the entire picture area, i.e., the primary coded picture contains all slices and macroblocks of the picture.
  • a redundant coded picture is a redundant coded representation of a picture that is not used for decoding unless the primary coded picture is missing or corrupted.
  • a decoded redundant coded picture contains essentially the same picture information as the respective decoded primary coded picture. However, the sample values in a decoded redundant coded picture are not required to be exactly equal to the co-located sample values in the corresponding decoded primary coded picture.
  • the number of redundant coded pictures per a primary coded picture may range from 0 to a limit specified in a coding standard (e.g. to 127 according to the JVT coding standard).
  • a redundant coded picture may use different reference pictures compared to the respective primary coded picture. Thus, if one of the reference pictures of the primary coded picture is missing or corrupted and all the reference pictures of a corresponding redundant coded picture are correctly decoded, it is advantageous from the picture quality point of view to decode the redundant coded picture instead of the primary coded picture.
  • Most conventional video coding standards include a concept of "not coded” or "skipped" macroblocks.
  • the decoding process of such a macroblock consists of a copy of the spatially corresponding macroblock in the reference picture.
  • MPEG-4 Visual includes optional object-based coding tools.
  • MPEG-4 video objects may be of any shape, and furthermore the shape, size, and position of the object may vary from one frame to the next.
  • a video object is composed of three color components (YUV) and an alpha component.
  • the alpha component defines the object's shape on an image-by-image basis.
  • Binary objects form the simplest class of object. They are represented by a sequence of binary alpha maps, i.e., 2-D images where each pixel is either black or white.
  • MPEG-4 provides a binary shape only mode for compressing these objects. The compression process is defined exclusively by a binary shape encoder for coding the sequence of alpha maps.
  • the representation comprises the colors of all the pixels within the interior of the object shape.
  • MPEG-4 encodes these objects using a binary shape encoder and then a motion compensated discrete cosine transform (DCT)-based algorithm for the interior texture coding.
  • DCT discrete cosine transform
  • the alpha map is a gray-level image with 256 possible levels. This gray-level alpha information is used to specify an object's transparency characteristics during the video composition process.
  • MPEG-4 encodes these objects using a binary shape encoder for the support of the alpha map and a motion compensated DCT-based algorithm for the coding of the alpha map and the interior texture.
  • Streaming clients typically have a receiver buffer that is capable of storing a relatively large amount of data. Initially, when a streaming session is established, a client does not start playing the stream back immediately, but rather it typically buffers the incoming data for a few seconds. This buffering helps to maintain continuous playback, because, in case of occasional increased transmission delays or network throughput drops, the client can decode and play buffered data. Otherwise, without initial buffering, the client has to freeze the display, stop decoding, and wait for incoming data. The buffering is also necessary for either automatic or selective retransmission in any protocol level. If any part of a picture is lost, a retransmission mechanism may be used to resend the lost data. If the retransmitted data is received before its scheduled decoding or playback time, the loss is perfectly recovered.
  • Coded pictures can be ranked according to their importance in the subjective quality of the decoded sequence. For example, non- reference pictures, such as conventional B pictures, are subjectively least important, because their absence does not affect decoding of any other pictures. Subjective ranking can also be made on data partition or slice group basis. Coded slices and data partitions that are subjectively the most important can be sent earlier than their decoding order indicates, whereas coded slices and data partitions that are subjectively the least important can be sent later than their natural coding order indicates. Consequently, any retransmitted parts of the most important slice and data partitions are more likely to be received before their scheduled decoding or playback time compared to the least important slices and data partitions.
  • the slice header syntax has to provide means to detect picture boundaries to let decoders operate on picture basis. If a decoder conformant to the JVT coding standard receives an error-free bitstream that includes both primary and redundant coded pictures, the decoder must detect the boundaries of primary and redundant coded pictures and decode only the primary coded pictures in order to reconstruct the sample values exactly as required in the standard. Moreover, if redundant pictures are conveyed over a connectionless channel such as RTP/UDP/IP, each one of them may be encapsulated to more than one IP packet. Because of the connectionless nature of UDP, the packets may be received in different order from the one they were transmitted.
  • a connectionless channel such as RTP/UDP/IP
  • the receiver has to deduce which coded slices belong to redundant coded pictures and which ones belong to primary coded pictures, and which redundant coded pictures correspond to a particular primary coded picture. If the receiver did not do this, slices overlapping with each other could be decoded unnecessarily.
  • a redundant coded representation of a picture can be used to provide unequal error protection in error-prone video transmission. If a primary coded representation of a picture is not received, a redundant representation can be used. If one of the reference pictures of the primary coded picture is missing or corrupted and all the reference pictures of a corresponding redundant coded picture are correctly decoded, the redundant coded picture can be decoded. Many times the subjective importance of different spatial parts of a picture can vary. The invention enables transmission of incomplete redundant pictures that do not cover the entire picture area. Consequently, the invention enables protection of only the subjectively most important parts of selected pictures. This improves compression efficiency compared to earlier standards and allows spatial focusing of unequal error protection.
  • the stored video signals can be either uncoded signals stored before encoding, as encoded signals stored after encoding, or as decoded signals stored after encoding and decoding process.
  • an encoder produces bitstreams in decoding order.
  • a file system receives audio and/or video bitstreams which are encapsulated e.g. in decoding order and stored as a file.
  • the encoder and the file system can produce metadata which informs subjective importance of the pictures and NAL units, contains information on sub-sequences, inter alia.
  • the file can be stored into a database from which a direct playback server can read the NAL units and encapsulate them into RTP packets.
  • the direct playback server can modify the transmission order of the packets different from the decoding order, remove sub-sequences, decide what SEI messages will be transmitted, if any, etc.
  • the RTP packets are received and buffered.
  • the NAL units are first reordered into correct order and after that the NAL units are delivered to the decoder.
  • Some networks or inter-networks and/or the communication protocols used in these networks for video communication may be constructed such that a sub-network is error-prone whereas another sub-network provides essentially error-free link.
  • a mobile terminal connects to a streaming server that resides on a public IP-based network
  • reliable link layer protocols may be used in the radio link and the private mobile operator core network may be over-provisioned such that the sub-network controlled by the mobile operator is essentially error-free.
  • the public IP-based network e.g. the Internet
  • the gateway preferably analyzes the bitstream transmitted from a terminal connected to the error-prone sub-network to a terminal connected to the error-free sub-network. If no errors have hit a particular part of the bitstream, the gateway preferably removes the application-level redundancy for forward error control that corresponds to that part of the bitstream. This operation reduces the amount of traffic in the error-free network, and the saved amount of traffic can then be used for other purposes.
  • each primary coded picture comprising essentially the same picture information as the respective redundant coded picture
  • at least one of the redundant coded pictures comprises picture information corresponding to only a part of the picture information of the respective primary coded picture.
  • the decoding method according to the present invention is primarily characterized in that the primary coded pictures having been formed using essentially the same picture information as what has been used to form the respective redundant coded pictures, and at least one of the redundant coded pictures comprises picture information corresponding to only a part of the picture information of the respective primary coded picture; detecting in the bitstream a parameter indicating that the coded picture information belongs to a redundant coded picture; using the parameter to control the decoding of the coded picture information belonging to a redundant coded picture wherein the redundant coded picture information corresponds to only a part of the picture information used to form the respective primary coded picture.
  • the system according to the present invention is primarily characterized in that the encoder comprises encoding means for forming primary coded pictures and redundant coded pictures of primary coded pictures, each primary coded picture comprising essentially the same picture information as the respective redundant coded picture, and at least one of the redundant coded pictures comprises picture information corresponding to only a part of the picture information of the respective primary coded picture; and the decoder comprises detecting means for detecting in the bitstream a parameter indicating that the coded picture information belongs to a redundant coded picture; and controlling means using the parameter to control the decoding of the coded picture information belonging to a redundant coded picture wherein the redundant coded picture information corresponds to only a part of the picture information used to form the respective primary coded picture.
  • the encoder according to the present invention is primarily characterized in that the encoder comprises encoding means for forming primary coded pictures and redundant coded pictures of primary coded pictures, each primary coded picture comprising essentially the same picture information as the respective redundant coded picture, and at least one of the redundant coded pictures comprises picture information corresponding to only a part of the picture information of the respective primary coded picture.
  • the decoder according to the present invention is primarily characterized in that the decoder comprises detecting means for detecting in the bitstream a parameter indicating that the coded picture information belongs to a redundant coded picture; and controlling means using the parameter to control the decoding of the coded picture information belonging to a redundant coded picture wherein the redundant coded picture information corresponds to only a part of the picture information used to form the respective primary coded picture.
  • the software program for encoding according to the present invention is primarily characterized in that it comprises machine executable steps for encoding pictures, comprising machine executable steps for forming primary coded pictures and redundant coded pictures of primary coded pictures, each primary coded picture comprising essentially the same picture information as the respective redundant coded picture, and at least one of the redundant coded pictures comprises picture information corresponding to only a part of the picture information of the respective primary coded picture.
  • the software program for decoding according to the present invention is primarily characterized in that it comprises machine executable steps for detecting in the bitstream a parameter indicating that the coded picture information belongs to a redundant coded picture; and using the parameter to control the decoding of the coded picture information belonging to a redundant coded picture wherein the redundant coded picture information corresponds to only a part of the picture information used to form the respective primary coded picture.
  • the storage medium for storing a software program comprising machine executable steps for encoding pictures according to the present invention is primarily characterized in that primary coded pictures and redundant coded pictures of primary coded pictures, each primary coded picture comprising essentially the same picture information as the respective redundant coded picture, and at least one of the redundant coded pictures comprises picture information corresponding to only a part of the picture information of the respective primary coded picture.
  • the transmitting device is primarily characterized in that it comprises an encoder for encoding pictures, comprising encoding means for forming primary coded pictures and redundant coded pictures of primary coded pictures, each primary coded picture comprising essentially the same picture information as the respective redundant coded picture, and at least one of the redundant coded pictures comprises picture information corresponding to only a part of the picture information of the respective primary coded picture.
  • the receiving device is primarily characterized in that it comprises a decoder comprising detecting means for detecting in the bitstream a parameter indicating that the coded picture information belongs to a redundant coded picture; and controlling means using the parameter to control the decoding of the coded picture information belonging to a redundant coded picture wherein the redundant coded picture information corresponds to only a part of the picture information used to form the respective primary coded picture.
  • the bitstream according to the present invention is primarily characterized in that it comprises primary coded pictures and redundant coded pictures of primary coded pictures, each primary coded picture comprising essentially the same picture information as the respective redundant coded picture, and at least one of the redundant coded pictures comprises picture information corresponding to only a part of the picture information of the respective primary coded picture.
  • the present invention enables decoders to detect boundaries between primary and redundant coded pictures and to avoid unnecessary decoding of redundant coded pictures if the primary coded picture is correctly decoded.
  • the present invention improves the reliability of the coding systems.
  • the correct decoding order of the pictures can be more reliably determined than in prior art systems even if some packets of a video stream are not available in the decoder.
  • Fig. 1 shows an example of a recursive temporal scalability scheme
  • Fig. 2 depicts a scheme referred to as Video Redundancy Coding, where a sequence of pictures is divided into two or more independently coded threads in an interleaved manner
  • Fig. 3 presents an example of a prediction structure potentially improving compression efficiency
  • Fig. 4 presents an example of the intra picture postponement method that can be used to improve error resiliency
  • Fig. 5 depicts an advantageous embodiment of the system according to the present invention
  • Fig. 6 depicts an advantageous embodiment of the encoder according to the present invention
  • Fig. 7 depicts an advantageous embodiment of the decoder according to the present invention
  • Fig. 8 shows a block diagram of a general video communications system.
  • Slice data partitioning is a method of partitioning the syntax elements of the slice syntax structure into slice data partition syntax structures based on the type of each syntax element.
  • slice data partition A contains all syntax elements in the slice header and slice data syntax structures other than the syntax elements for coding the difference between the predicted sample values and the decoded sample values.
  • slice data partition B contains the syntax elements for coding the difference between the predicted sample values and the decoded sample values in intra macroblock types (I and SI macroblocks).
  • Slice data partition C contains the syntax elements for coding the difference between the predicted sample values and the decoded sample values in inter-predicted macroblock types (P, SP, and B macroblocks).
  • Primary coded data partition is a data partition which belongs to a primary coded picture.
  • Primary coded picture is a primary coded representation of a picture.
  • Primary coded slice is a slice which belongs to a primary coded picture.
  • Redundant coded data partition is a data partition which belongs to a redundant coded picture.
  • Redundant coded picture is a redundant coded representation of a picture that should only be used if the primary coded or decoded picture is corrupted.
  • the decoded redundant picture may not cover the entire picture area. There should be no noticeable difference between the co- located areas of the decoded primary picture and any decoded redundant slices.
  • the redundant coded picture is not required to contain all macroblocks in the primary coded picture.
  • Redundant coded slice is a slice belonging to a redundant coded picture.
  • not coded macroblocks There are two major differences between "not coded" macroblocks and the macroblocks that are not included in a redundant picture: First, macroblocks that are not included in a redundant coded picture are not signaled, whereas "not coded" macroblocks are coded in the bitstream (typically by one bit per "not coded” macroblock). Second, decoders must not decode the areas not included in a redundant picture. If any macroblock is not included in the received primary coded picture or any corresponding redundant coded picture, decoders should conceal these missing macroblocks using any proprietary error concealment algorithm. In contrast to this, there is a specific normative decoding process for "not coded" macroblocks.
  • the pictures to be encoded can be, for example, pictures of a video stream from a video source 3, e.g. a camera, a video recorder, etc.
  • the pictures (frames) of the video stream can be divided into smaller portions such as slices.
  • the slices can further be divided into blocks.
  • the video stream is encoded to reduce the information to be transmitted via a transmission channel 4, or to a storage media (not shown).
  • Pictures of the video stream are input to the encoder 1.
  • the encoder has an encoding buffer 1.1 (Fig.
  • the encoder 1 also includes a memory 1.3 and a processor 1.2 in which the encoding tasks according to the invention can be applied.
  • the memory 1.3 and the processor 1.2 can be common with the transmitting device 6 or the transmitting device 6 can have another processor and/or memory (not shown) for other functions of the transmitting device 6.
  • the encoder 1 performs motion estimation and/or some other tasks to compress the video stream. In motion estimation similarities between the picture to be encoded (the current picture) and a previous and/or latter picture are searched. If similarities are found the compared picture or part of it can be used as a reference picture for the picture to be encoded.
  • the display order and the decoding order of the pictures are not necessarily the same, wherein the reference picture has to be stored in a buffer (e.g. in the encoding buffer 1.1 ) as long as it is used as a reference picture.
  • the encoder 1 also inserts information on display order of the pictures into the transmission stream. In practice, either the timing information SEI message or timestamps external to the JVT syntax (such as RTP timestamps) can be used.
  • the encoded pictures are moved to an encoded picture buffer 1.2, if necessary.
  • the encoded pictures are transmitted from the encoder 1 to the decoder 2 via the transmission channel 4.
  • the encoded pictures are decoded to form uncompressed pictures corresponding as much as possible to the encoded pictures.
  • Each decoded picture is buffered in the DPB 2.1 of the decoder 2 unless it is displayed substantially immediately after the decoding and is not used as a reference picture.
  • both the reference picture buffering and the display picture buffering are combined and they use the same decoded picture buffer 2.1. This eliminates the need for storing the same pictures in two different places thus reducing the memory requirements of the decoder 2.
  • the decoder 1 also includes a memory 2.3 and a processor 2.2 in which the decoding tasks according to the invention can be applied.
  • the memory 2.3 and the processor 2.2 can be common with the receiving device 8 or the receiving device 8 can have another processor and/or memory (not shown) for other functions of the receiving device 8.
  • Pictures from the video source 3 are entered to the encoder 1 and advantageously stored in the pre-encoding buffer 1.1.
  • pictures arriving after a picture to be encoded are analyzed in a bit rate control algorithm so that there would be no remarkable variations in the quality of the pictures.
  • the coding order of the pictures (and the decoding order) is different from the capture order of the pictures. This kind of arrangement can be effective on the compression efficiency point of view (for example, a PBBBP frame sequence in which the B frame in between two other B frames is a reference frame for the other two B frames) and/or on the error resilience point of view (intra picture postponement).
  • the encoding process is not necessarily started immediately after the first picture is entered to the encoder, but after a certain amount of pictures are available in the encoding buffer 1.1. Then the encoder 1 tries to find suitable candidates from the pictures to be used as the reference frames. The encoder 1 then performs the encoding to form encoded pictures.
  • the encoded pictures can be, for example, predicted pictures (P), bi-predictive pictures (B), and/or intra-coded pictures (I).
  • the intra-coded pictures can be decoded without using any other pictures, but other type of pictures need at least one reference picture before they can be decoded. Pictures of any of the above mentioned picture types can be used as a reference picture.
  • the encoder advantageously attaches two time stamps to the pictures: a decoding time stamp (DTS) and output time stamp (OTS).
  • DTS decoding time stamp
  • OTS output time stamp
  • the decoder can use the time stamps to determine the correct decoding time and time to output (display) the pictures.
  • those time stamps are not necessarily transmitted to the decoder or it does not use them.
  • the encoder 1 can form redundant coded pictures or redundant coded data partitions of the pictures to increase error resiliency. According to the present invention the encoder can form redundant pictures which do not contain all the necessary information to decode the picture but only some portion(s) of it. The encoder 1 can also form more than one different redundant coded data partitions for the same picture wherein the different redundant coded data partitions contain information from at least partly different areas of the picture.
  • the smallest redundant coded picture consists preferably of a slice. The slice contains one or more macroblocks.
  • the encoder 1 decides what pictures contain areas which should be redundantly coded.
  • the criteria for the selection may vary in different embodiments and in different situations.
  • the encoder 1 may examine if there is a possible scene change between successive pictures or if there are, for some other reason, lot of changes between successive pictures.
  • the encoder 1 can examine whether there are changes in some portions of pictures to determine which parts of the pictures should be redundantly encoded.
  • the encoder 1 can, for example, examine the motion vectors to find out important regions and/or regions which are especially sensitive to transmission/decoding errors and form redundant coded data partitions of such regions. There should be some indication in the transmission stream to indicate if there exist redundant slices in the stream.
  • the indication is preferably inserted in the slice header of each slice and/or in the picture parameter set.
  • One advantageous embodiment of the indication uses two syntax elements for redundant slices: the first syntax element is "redundant_slice_flag" which resides in the picture parameter set, and the other syntax element is “redundant_pic_cnt” and it resides in slice header.
  • the "redundant_pic_cnt” is optional and it is included in the slice header only when "redundant_slice_flag" in the referred picture parameter set is set to 1.
  • redundant_slice_flag indicates the presence of the redundant_pic_cnt parameter in all slice headers referencing the picture parameter set.
  • the picture parameter set can be common for more than one slice if all the parameters are equal for the slices. If the value of the redundant_slice_flag is true, then the slice headers of those slices which refer to this parameter set contain the second syntax element (redundant_pic_cnt).
  • the value of the redundant_pic_cnt is 0 for coded slices and data partitions belonging to the primary representation of the picture contents.
  • the redundant_pic_cnt is greater than 0 for coded slices and data partitions that contain redundant coded representation of the picture contents. There should be no noticeable difference between the co-located areas of the decoded primary representation of the picture and any decoded redundant slices. Redundant slices and data partitions having the same value of redundant_pic_cnt belong to the same redundant picture. Decoded slices having the same redundant_pic_cnt shall not overlap. Decoded slices having a redundant_pic_cnt greater than 0 may not cover the entire picture area.
  • the pictures may have a parameter called as nal_storage_idc.
  • the value of the nal_storage_idc in a primary picture is 0, the value of the nal_storage_idc in corresponding redundant pictures shall be 0. If the value of the nal_storage_idc in a primary picture is non-zero, the value of the nal_storage_idc in corresponding redundant pictures shall be non-zero.
  • the above described syntax design works well when data partitioning is not applied for redundant slices. However, when data partitioning is used, i.e. each redundant slice has three data partitions DPA, DPB, and DPC, a further mechanism is needed to inform the decoder which redundant slice is in question.
  • the redundant_pic_cnt is included, not only the slice header in DPA but also in slice headers of both DPB and DPC. If slice data partitioning is in use, slice data partition B and C have to be associated to the respective slice data partition A in order to enable decoding of the slice.
  • Slice data partition A includes a slicejd syntax element whose value uniquely identifies a slice within a coded picture.
  • Slice data partitions B and C include the redundant_pic_cnt syntax element if it is also present in the slice header included in the slice data partition A (which is conditional to the value of "redundant_slice_flag" in the referred picture parameter set).
  • redundant_pic_cnt syntax element is used to associate slice data partitions B and C to a particular primary or redundant coded picture.
  • slice data partitions B and C include the slicejd syntax element, which is used to associate the data partition with the respective data partition A of the same coded picture.
  • the transmission and/or storing of the encoded pictures (and the optional virtual decoding) can be started immediately after the first encoded picture is ready.
  • This picture is not necessarily the first one in decoder output order because the decoding order and the output order may not be the same.
  • the transmission can be started.
  • the encoded pictures are optionally stored to the encoded picture buffer 1.2.
  • the transmission can also start at a later stage, for example, after a certain part of the video stream is encoded.
  • the number of transmitted redundant pictures depends inter alia on network conditions such as amount of traffic, bit error rate in a radio link, etc. In other words, all redundant pictures are not necessarily transmitted.
  • the receiver 8 collects all packets belonging to a picture, bringing them into a reasonable order. The strictness of the order depends on the profile employed.
  • the received packets are advantageously stored into the receiving buffer 9.1 (pre-decoding buffer).
  • the receiver 8 discards anything that is unusable, and passes the rest to the decoder 2.
  • the decoder can use some of the redundant coded slices to decode the picture.
  • the decoder 2 can send the slice id:s or some other information identifying the picture in question to the encoder 1.
  • the decoder 2 can try e.g. some error recovery methods to diminish the effects of the error on picture quality, or the decoder 2 can discard the erroneous picture and use some previous picture instead.
  • the present invention can be applied in many kind of systems and devices.
  • the transmitting device 6 including the encoder 1 and optionally the HRD 5 advantageously include also a transmitter 7 to transmit the encoded pictures to the transmission channel 4.
  • the receiving device 8 include the receiver 9 to receive the encoded pictures, the decoder 2, and a display 10 on which the decoded pictures can be displayed.
  • the transmission channel can be, for example, a landline communication channel and/or a wireless communication channel.
  • the transmitting device and the receiving device include also one or more processors 1.2, 2.2 which can perform the necessary steps for controlling the encoding/decoding process of video stream according to the invention. Therefore, the method according to the present invention can mainly be implemented as machine executable steps of the processors.
  • the buffering of the pictures can be implemented in the memory 1.3, 2.3 of the devices.
  • the program code 1.4 of the encoder can be stored into the memory 1.3.
  • the program code 2.4 of the decoder can be stored into the memory 2.3.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)
  • Compression, Expansion, Code Conversion, And Decoders (AREA)

Abstract

The invention relates to method for encoding pictures, method for encoding pictures, wherein primary coded pictures and redundant coded pictures of primary coded pictures are formed, each primary coded picture comprising essentially the same picture information as the respective redundant coded picture. At least one of the redundant coded pictures comprises picture information corresponding to only a part of the picture information of the respective primary coded picture. The invention also relates to a system, encoder, decoder, a transmitting device, a receiving device, a software program, a storage medium and a bitstream.

Description

PICTURE CODING METHOD
Field of the Invention
The invention relates to method for encoding pictures, in which primary coded pictures and redundant coded pictures of primary coded pictures are formed. The invention also relates to a system, encoder, decoder, transmitting device, receiving device, software program, a storage medium and a bitstream.
Background of the Invention
Published video coding standards include ITU-T H.261, ITU-T H.263, ISO/IEC MPEG-1 , ISO/IEC MPEG-2, and ISO/IEC MPEG-4 Part 2. These standards are herein referred to as conventional video coding standards.
Video communication systems
Video communication systems can be divided into conversational and non-conversational systems. Conversational systems include video conferencing and video telephony. Examples of such systems include ITU-T Recommendations H.320, H.323, and H.324 that specify a video conferencing/telephony system operating in ISDN, IP, and PSTN networks respectively. Conversational systems are characterized by the intent to minimize the end-to-end delay (from audio-video capture to the far-end audio-video presentation) in order to improve the user experience.
Non-conversational systems include playback of stored content, such as Digital Versatile Disks (DVDs) or video files stored in a mass memory of a playback device, digital TV, and streaming.
There is a standardization effort going on in a Joint Video Team (JVT) of ITU-T and ISO/IEC. The work of JVT is based on an earlier standardization project in ITU-T called H.26L. The goal of the JVT standardization is to release the same standard text as ITU-T Recommendation H.264 and ISO/IEC International Standard 14496-10 (MPEG-4 Part 10). The draft standard is referred to as the JVT coding standard in this paper, and the codec according to the draft standard is referred to as the JVT codec.
The codec specification itself distinguishes conceptually between a video coding layer (VCL), and a network abstraction layer (NAL). The VCL contains the signal processing functionality of the codec, things such as transform, quantization, motion search/compensation, and the loop filter. It follows the general concept of most of today's video codecs, a macroblock-based coder that utilizes inter picture prediction with motion compensation, and transform coding of the residual signal. The output of the VCL encoder are slices: a bit string that contains the macroblock data of an integer number of macroblocks, and the information of the slice header (containing the spatial address of the first macroblock in the slice, the initial quantization parameter, and similar). Macroblocks in slices are ordered consecutively in scan order unless a different macroblock allocation is specified, using the so-called Flexible Macroblock Ordering syntax. In-picture prediction is used only within a slice.
The NAL encapsulates the slice output of the VCL into Network Abstraction Layer Units (NALUs), which are suitable for the transmission over packet networks or the use in packet oriented multiplex environments. JVT's Annex B defines an encapsulation process to transmit such NALUs over byte-stream oriented networks.
The optional reference picture selection mode of H.263 and the NEWPRED coding tool of MPEG-4 Part 2 enable selection of the reference frame for motion compensation per each picture segment, e.g., per each slice in H.263. Furthermore, the optional Enhanced Reference Picture Selection mode of H.263 and the JVT coding standard enable selection of the reference frame for each macroblock separately.
Reference picture selection enables many types of temporal scalability schemes. Figure 1 shows an example of a temporal scalability scheme, which is herein referred to as recursive temporal scalability. The example scheme can be decoded with three constant frame rates. Figure 2 depicts a scheme referred to as Video Redundancy Coding, where a sequence of pictures is divided into two or more independently coded threads in an interleaved manner. The arrows in these and all the subsequent figures indicate the direction of motion compensation and the values under the frames correspond to the relative capturing and displaying times of the frames.
Figure 8 shows a block diagram of a general video communications system 800. Due to the fact that uncompressed video requires a huge bandwidth, the input video 801 is compressed in a transmitting device 802 by a source coder 803 to a desired bit rate. The source coder 803 can be divided into two components, namely waveform coder 803.1 and entropy coder 803.2. The waveform coder 803.1 performs lossy video signal compression whereas the entropy coder 803.2 losslessly converts the output of waveform coder 803.1 into a binary sequence. The transport coder 804 encapsulates the compressed video according to the transport protocols in use. It may manipulate the compressed video in other ways too. For example, it may interleave and modulate the data. Then, the data is transmitted to the receiver side via a transmission channel 805 which may comprise server device(s) 806, gateways (not shown) etc. The receiver 807 performs inverse operations to obtain reconstructed video signal for display. The receiver 807 comprises a transport decoder 808 and a source decoder 809. The transport decoder 808 de-capsulates the compressed video input from the transmission channel 805 according to the transport protocols in use. The source decoder 809 can also be divided into two components, namely entropy decoder 809.1 and waveform decoder 809.2. The entropy decoder 809.1 converts the binary sequence from the transport decoder 808 into a waveform to be input to the waveform decoder 809.1. The waveform decoder 809.1 performs video signal decompression and outputs the video signal 810. The receiver 807 may also give feedback to the transmitter. For instance, the receiver may signal the rate of successfully received transmission units. Parameter Set Concept
One very fundamental design concept of the JVT codec is to generate self-contained packets, to make mechanisms such as the header duplication unnecessary. The way how this was achieved is to decouple information that is relevant to more than one slice from the media stream. This higher layer meta information should be sent reliably, asynchronously and in advance from the RTP packet stream that contains the slice packets. This information can also be sent in-band in such applications that do not have an out-of-band transport channel appropriate for the purpose. The combination of the higher level parameters is called a Parameter Set. The Parameter Set contains information such as picture size, display window, optional coding modes employed, macroblock allocation map, and others.
In order to be able to change picture parameters (such as the picture size), without having the need to transmit Parameter Set updates synchronously to the slice packet stream, the encoder and decoder can maintain a list of more than one Parameter Set. Each slice header contains a codeword that indicates the Parameter Set to be used.
This mechanism allows to decouple the transmission of the Parameter Sets from the packet stream, and transmit them by external means, e.g. as a side effect of the capability exchange, or through a (reliable or unreliable) control protocol. It may even be possible that they get never transmitted but are fixed by an application design specification.
Transmission order
In conventional video coding standards, the decoding order of pictures is the same as the display order except for B pictures. A block in a conventional B picture can be bi-directionally temporally predicted from two reference pictures, where one reference picture is temporally preceding and the other reference picture is temporally succeeding in display order. Only the latest reference picture in decoding order can succeed the B picture in display order (exception: interlaced coding in H.263 where both field pictures of a temporally subsequent reference frame can precede a B picture in decoding order). A conventional B picture cannot be used as a reference picture for temporal prediction, and therefore a conventional B picture can be disposed without affecting the decoding of any other pictures.
The JVT coding standard includes the following novel technical features compared to earlier standards:
The decoding order of pictures is decoupled from the display order. The picture number indicates decoding order and the picture order count indicates the display order.
Reference pictures for a block in a B picture can either be before or after the B picture in display order. Consequently, a B picture stands for a bi-predictive picture instead of a bi-directional picture.
Pictures that are not used as reference pictures are marked explicitly. A picture of any type (intra, inter, B, etc.) can either be a reference picture or a non-reference picture. (Thus, a B picture can be used as a reference picture for temporal prediction of other pictures.)
A picture can contain slices that are coded with a different coding type. In other words, a coded picture may consist of an intra- coded slice and a B-coded slice, for example.
Decoupling of display order from decoding order can be beneficial from compression efficiency and error resiliency point of view.
An example of a prediction structure potentially improving compression efficiency is presented in Figure 3. Boxes indicate pictures, capital letters within boxes indicate coding types, numbers within boxes are picture numbers according to the JVT coding standard, and arrows indicate prediction dependencies. Note that picture B17 is a reference picture for pictures B18. Compression efficiency is potentially improved compared to conventional coding, because the reference pictures for pictures B18 are temporally closer compared to conventional coding with PBBP or PBBBP coded picture patterns. Compression efficiency is potentially improved compared to conventional PBP coded picture pattern, because part of reference pictures are bi-directionally predicted. Figure 4 presents an example of the intra picture postponement method that can be used to improve error resiliency. Conventionally, an intra picture is coded immediately after a scene cut or as a response to an expired intra picture refresh period, for example. In the intra picture postponement method, an intra picture is not coded immediately after a need to code an intra picture arises, but rather a temporally subsequent picture is selected as an intra picture. Each picture between the coded intra picture and the conventional location of an intra picture is predicted from the next temporally subsequent picture. As Figure 4 shows, the intra picture postponement method generates two independent inter picture prediction chains, whereas conventional coding algorithms produce a single inter picture chain. It is intuitively clear that the two- chain approach is more robust against erasure errors than the one- chain conventional approach. If one chain suffers from a packet loss, the other chain may still be correctly received. In conventional coding, a packet loss always causes error propagation to the rest of the inter picture prediction chain.
Transmission of multimedia streams
A multimedia streaming system consists of a streaming server and a number of players, which access the server via a network. The network is typically packet-oriented and provides little or no means to guaranteed quality of service. The players fetch either pre-stored or live multimedia content from the server and play it back in real-time while the content is being downloaded. The type of communication can be either point-to-point or multicast. In point-to-point streaming, the server provides a separate connection for each player. In multicast streaming, the server transmits a single data stream to a number of players, and network elements duplicate the stream only if it is necessary.
When a player has established a connection to a server and requested for a multimedia stream, the server begins to transmit the desired stream. The player does not start playing the stream back immediately, but rather it typically buffers the incoming data for a few seconds. Herein, this buffering is referred to as initial buffering. Initial buffering helps to maintain pauseless playback, because, in case of occasional increased transmission delays or network throughput drops, the player can decode and play buffered data.
In order to avoid unlimited transmission delay, it is uncommon to favor reliable transport protocols in streaming systems. Instead, the systems prefer unreliable transport protocols, such as UDP, which, on one hand, inherit a more stable transmission delay, but, on the other hand, also suffer from data corruption or loss.
RTP and RTCP protocols can be used on top of UDP to control realtime communications. RTP provides means to detect losses of transmission packets, to reassemble the correct order of packets in the receiving end, and to associate a sampling time-stamp with each packet. RTCP conveys information about how large a portion of packets were correctly received, and, therefore, it can be used for flow control purposes.
Transmission errors
There are two main types of transmission errors, namely bit errors and packet errors. Bit errors are typically associated with a circuit-switched channel, such as a radio access network connection in mobile communications, and they are caused by imperfections of physical channels, such as radio interference. Such imperfections may result into bit inversions, bit insertions and bit deletions in transmitted data. Packet errors are typically caused by elements in packet-switched networks. For example, a packet router may become congested; i.e. it may get too many packets as input and cannot output them at the same rate. In this situation, its buffers overflow, and some packets get lost. Packet duplication and packet delivery in different order than transmitted are also possible but they are typically considered to be less common than packet losses. Packet errors may also be caused by the implementation of the used transport protocol stack. For example, some protocols use checksums that are calculated in the transmitter and encapsulated with source-coded data. If there is a bit inversion error in the data, the receiver cannot end up into the same checksum, and it may have to discard the received packet. Second (2G) and third generation (3G) mobile networks, including GPRS, UMTS, and CDMA-2000, provide two basic types of radio link connections, acknowledged and non-acknowledged. An acknowledged connection is such that the integrity of a radio link frame is checked by the recipient (either the Mobile Station, MS, or the Base Station Subsystem, BSS), and, in case of a transmission error, a retransmission request is given to the other end of the radio link. Due to link layer retransmission, the originator has to buffer a radio link frame until a positive acknowledgement for the frame is received. In harsh radio conditions, this buffer may overflow and cause data loss. Nevertheless, it has been shown that it is beneficial to use the acknowledged radio link protocol mode for streaming services. A non-acknowledged connection is such that erroneous radio link frames are typically discarded.
Packet losses can either be corrected or concealed. Loss correction refers to the capability to restore lost data perfectly as if no losses had ever been introduced. Loss concealment refers to the capability to conceal the effects of transmission losses so that they should not be visible in the reconstructed video sequence.
When a player detects a packet loss, it may request for a packet retransmission. Because of the initial buffering, the retransmitted packet may be received before its scheduled playback time. Some commercial Internet streaming systems implement retransmission requests using proprietary protocols. Work is going on in IETF to standardize a selective retransmission request mechanism as a part of RTCP.
A common feature for all of these retransmission request protocols is that they are not suitable for multicasting to a large number of players, as the network traffic may increase drastically. Consequently, multicast streaming applications have to rely on non-interactive packet loss control.
Point-to-point streaming systems may also benefit from non-interactive error control techniques. First, some systems may not contain any interactive error control mechanism or they prefer not to have any feedback from players in order to simplify the system. Second, retransmission of lost packets and other forms of interactive error control typically take a larger portion of the transmitted data rate than non-interactive error control methods. Streaming servers have to ensure that interactive error control methods do not reserve a major portion of the available network throughput. In practice, the servers may have to limit the amount of interactive error control operations. Third, transmission delay may limit the number of interactions between the server and the player, as all interactive error control operations for a specific data sample should preferably be done before the data sample is played back.
Non-interactive packet loss control mechanisms can be categorized to forward error control and loss concealment by post-processing. Forward error control refers to techniques in which a transmitter adds such redundancy to transmitted data that receivers can recover at least part of the transmitted data even if there are transmission losses. There are two categories of forward error control methods: signal-dependent and signal-independent. Signal-dependent methods require interpretation of the bitstream. An example of such a method is a repetition of sequence or picture header. Signal-independent methods can be used to recover whatever bitstream regardless of the interpreted content of the bitstream. Examples of such methods are error correction codes (e.g. parity codes and Reed-Solomon codes). Error concealment by postprocessing is totally receiver-oriented. These methods try to estimate the correct representation of erroneously received data.
Most video compression algorithms generate temporally predicted INTER or P pictures. As a result, a data loss in one picture causes visible degradation in the consequent pictures that are temporally predicted from the corrupted one. Video communication systems can either conceal the loss in displayed images or freeze the latest correct picture onto the screen until a frame which is independent from the corrupted frame is received.
Primary and redundant pictures
A primary coded picture is a primary coded representation of a picture. The decoded primary coded picture covers the entire picture area, i.e., the primary coded picture contains all slices and macroblocks of the picture. A redundant coded picture is a redundant coded representation of a picture that is not used for decoding unless the primary coded picture is missing or corrupted. A decoded redundant coded picture contains essentially the same picture information as the respective decoded primary coded picture. However, the sample values in a decoded redundant coded picture are not required to be exactly equal to the co-located sample values in the corresponding decoded primary coded picture. The number of redundant coded pictures per a primary coded picture may range from 0 to a limit specified in a coding standard (e.g. to 127 according to the JVT coding standard). A redundant coded picture may use different reference pictures compared to the respective primary coded picture. Thus, if one of the reference pictures of the primary coded picture is missing or corrupted and all the reference pictures of a corresponding redundant coded picture are correctly decoded, it is advantageous from the picture quality point of view to decode the redundant coded picture instead of the primary coded picture.
Most conventional video coding standards include a concept of "not coded" or "skipped" macroblocks. The decoding process of such a macroblock consists of a copy of the spatially corresponding macroblock in the reference picture.
Object-based coding according to MPEG-4 Visual
MPEG-4 Visual includes optional object-based coding tools. MPEG-4 video objects may be of any shape, and furthermore the shape, size, and position of the object may vary from one frame to the next. In terms of its general representation, a video object is composed of three color components (YUV) and an alpha component. The alpha component defines the object's shape on an image-by-image basis. Binary objects form the simplest class of object. They are represented by a sequence of binary alpha maps, i.e., 2-D images where each pixel is either black or white. MPEG-4 provides a binary shape only mode for compressing these objects. The compression process is defined exclusively by a binary shape encoder for coding the sequence of alpha maps. In addition to the sequence of binary alpha maps representing the object shape, the representation comprises the colors of all the pixels within the interior of the object shape. MPEG-4 encodes these objects using a binary shape encoder and then a motion compensated discrete cosine transform (DCT)-based algorithm for the interior texture coding. Finally, it is possible to represent a textured object with gray-level shape. For this object, the alpha map is a gray-level image with 256 possible levels. This gray-level alpha information is used to specify an object's transparency characteristics during the video composition process. MPEG-4 encodes these objects using a binary shape encoder for the support of the alpha map and a motion compensated DCT-based algorithm for the coding of the alpha map and the interior texture.
Buffering
Streaming clients typically have a receiver buffer that is capable of storing a relatively large amount of data. Initially, when a streaming session is established, a client does not start playing the stream back immediately, but rather it typically buffers the incoming data for a few seconds. This buffering helps to maintain continuous playback, because, in case of occasional increased transmission delays or network throughput drops, the client can decode and play buffered data. Otherwise, without initial buffering, the client has to freeze the display, stop decoding, and wait for incoming data. The buffering is also necessary for either automatic or selective retransmission in any protocol level. If any part of a picture is lost, a retransmission mechanism may be used to resend the lost data. If the retransmitted data is received before its scheduled decoding or playback time, the loss is perfectly recovered.
Coded pictures can be ranked according to their importance in the subjective quality of the decoded sequence. For example, non- reference pictures, such as conventional B pictures, are subjectively least important, because their absence does not affect decoding of any other pictures. Subjective ranking can also be made on data partition or slice group basis. Coded slices and data partitions that are subjectively the most important can be sent earlier than their decoding order indicates, whereas coded slices and data partitions that are subjectively the least important can be sent later than their natural coding order indicates. Consequently, any retransmitted parts of the most important slice and data partitions are more likely to be received before their scheduled decoding or playback time compared to the least important slices and data partitions.
Identification of Redundant Pictures
Due to the fact that there are no picture headers in the JVT coding syntax, the slice header syntax has to provide means to detect picture boundaries to let decoders operate on picture basis. If a decoder conformant to the JVT coding standard receives an error-free bitstream that includes both primary and redundant coded pictures, the decoder must detect the boundaries of primary and redundant coded pictures and decode only the primary coded pictures in order to reconstruct the sample values exactly as required in the standard. Moreover, if redundant pictures are conveyed over a connectionless channel such as RTP/UDP/IP, each one of them may be encapsulated to more than one IP packet. Because of the connectionless nature of UDP, the packets may be received in different order from the one they were transmitted. Thus, the receiver has to deduce which coded slices belong to redundant coded pictures and which ones belong to primary coded pictures, and which redundant coded pictures correspond to a particular primary coded picture. If the receiver did not do this, slices overlapping with each other could be decoded unnecessarily.
Summary of the Invention
A redundant coded representation of a picture can be used to provide unequal error protection in error-prone video transmission. If a primary coded representation of a picture is not received, a redundant representation can be used. If one of the reference pictures of the primary coded picture is missing or corrupted and all the reference pictures of a corresponding redundant coded picture are correctly decoded, the redundant coded picture can be decoded. Many times the subjective importance of different spatial parts of a picture can vary. The invention enables transmission of incomplete redundant pictures that do not cover the entire picture area. Consequently, the invention enables protection of only the subjectively most important parts of selected pictures. This improves compression efficiency compared to earlier standards and allows spatial focusing of unequal error protection.
In the following description the invention is described by using encoder- decoder based system, but it is obvious that the invention can also be implemented in systems in which the video signals are stored. The stored video signals can be either uncoded signals stored before encoding, as encoded signals stored after encoding, or as decoded signals stored after encoding and decoding process. For example, an encoder produces bitstreams in decoding order. A file system receives audio and/or video bitstreams which are encapsulated e.g. in decoding order and stored as a file. In addition, the encoder and the file system can produce metadata which informs subjective importance of the pictures and NAL units, contains information on sub-sequences, inter alia. The file can be stored into a database from which a direct playback server can read the NAL units and encapsulate them into RTP packets. According to the optional metadata and the data connection in use, the direct playback server can modify the transmission order of the packets different from the decoding order, remove sub-sequences, decide what SEI messages will be transmitted, if any, etc. In the receiving end the RTP packets are received and buffered. Typically, the NAL units are first reordered into correct order and after that the NAL units are delivered to the decoder.
Some networks or inter-networks and/or the communication protocols used in these networks for video communication may be constructed such that a sub-network is error-prone whereas another sub-network provides essentially error-free link. For example, if a mobile terminal connects to a streaming server that resides on a public IP-based network, reliable link layer protocols may be used in the radio link and the private mobile operator core network may be over-provisioned such that the sub-network controlled by the mobile operator is essentially error-free. However, the public IP-based network (e.g. the Internet) provides an error-prone best-effort service. Consequently, protection against transmission errors should be used in the error-prone subnetwork, whereas application-level error protection is not useful in a sub-network providing essentially error-free connection. In such a situation, it is beneficial to have a gateway component connecting the error-prone sub-network to the error-free sub-network. The gateway preferably analyzes the bitstream transmitted from a terminal connected to the error-prone sub-network to a terminal connected to the error-free sub-network. If no errors have hit a particular part of the bitstream, the gateway preferably removes the application-level redundancy for forward error control that corresponds to that part of the bitstream. This operation reduces the amount of traffic in the error-free network, and the saved amount of traffic can then be used for other purposes.
The encoding method according to the present invention is primarily characterized in that each primary coded picture comprising essentially the same picture information as the respective redundant coded picture, and at least one of the redundant coded pictures comprises picture information corresponding to only a part of the picture information of the respective primary coded picture. The decoding method according to the present invention is primarily characterized in that the primary coded pictures having been formed using essentially the same picture information as what has been used to form the respective redundant coded pictures, and at least one of the redundant coded pictures comprises picture information corresponding to only a part of the picture information of the respective primary coded picture; detecting in the bitstream a parameter indicating that the coded picture information belongs to a redundant coded picture; using the parameter to control the decoding of the coded picture information belonging to a redundant coded picture wherein the redundant coded picture information corresponds to only a part of the picture information used to form the respective primary coded picture. The system according to the present invention is primarily characterized in that the encoder comprises encoding means for forming primary coded pictures and redundant coded pictures of primary coded pictures, each primary coded picture comprising essentially the same picture information as the respective redundant coded picture, and at least one of the redundant coded pictures comprises picture information corresponding to only a part of the picture information of the respective primary coded picture; and the decoder comprises detecting means for detecting in the bitstream a parameter indicating that the coded picture information belongs to a redundant coded picture; and controlling means using the parameter to control the decoding of the coded picture information belonging to a redundant coded picture wherein the redundant coded picture information corresponds to only a part of the picture information used to form the respective primary coded picture. The encoder according to the present invention is primarily characterized in that the encoder comprises encoding means for forming primary coded pictures and redundant coded pictures of primary coded pictures, each primary coded picture comprising essentially the same picture information as the respective redundant coded picture, and at least one of the redundant coded pictures comprises picture information corresponding to only a part of the picture information of the respective primary coded picture. The decoder according to the present invention is primarily characterized in that the decoder comprises detecting means for detecting in the bitstream a parameter indicating that the coded picture information belongs to a redundant coded picture; and controlling means using the parameter to control the decoding of the coded picture information belonging to a redundant coded picture wherein the redundant coded picture information corresponds to only a part of the picture information used to form the respective primary coded picture. The software program for encoding according to the present invention is primarily characterized in that it comprises machine executable steps for encoding pictures, comprising machine executable steps for forming primary coded pictures and redundant coded pictures of primary coded pictures, each primary coded picture comprising essentially the same picture information as the respective redundant coded picture, and at least one of the redundant coded pictures comprises picture information corresponding to only a part of the picture information of the respective primary coded picture. The software program for decoding according to the present invention is primarily characterized in that it comprises machine executable steps for detecting in the bitstream a parameter indicating that the coded picture information belongs to a redundant coded picture; and using the parameter to control the decoding of the coded picture information belonging to a redundant coded picture wherein the redundant coded picture information corresponds to only a part of the picture information used to form the respective primary coded picture. The storage medium for storing a software program comprising machine executable steps for encoding pictures according to the present invention is primarily characterized in that primary coded pictures and redundant coded pictures of primary coded pictures, each primary coded picture comprising essentially the same picture information as the respective redundant coded picture, and at least one of the redundant coded pictures comprises picture information corresponding to only a part of the picture information of the respective primary coded picture. The transmitting device according to the present invention is primarily characterized in that it comprises an encoder for encoding pictures, comprising encoding means for forming primary coded pictures and redundant coded pictures of primary coded pictures, each primary coded picture comprising essentially the same picture information as the respective redundant coded picture, and at least one of the redundant coded pictures comprises picture information corresponding to only a part of the picture information of the respective primary coded picture. The receiving device according to the present invention is primarily characterized in that it comprises a decoder comprising detecting means for detecting in the bitstream a parameter indicating that the coded picture information belongs to a redundant coded picture; and controlling means using the parameter to control the decoding of the coded picture information belonging to a redundant coded picture wherein the redundant coded picture information corresponds to only a part of the picture information used to form the respective primary coded picture. The bitstream according to the present invention is primarily characterized in that it comprises primary coded pictures and redundant coded pictures of primary coded pictures, each primary coded picture comprising essentially the same picture information as the respective redundant coded picture, and at least one of the redundant coded pictures comprises picture information corresponding to only a part of the picture information of the respective primary coded picture.
The present invention enables decoders to detect boundaries between primary and redundant coded pictures and to avoid unnecessary decoding of redundant coded pictures if the primary coded picture is correctly decoded.
The present invention improves the reliability of the coding systems. By using the present invention the correct decoding order of the pictures can be more reliably determined than in prior art systems even if some packets of a video stream are not available in the decoder. Description of the Drawings
Fig. 1 shows an example of a recursive temporal scalability scheme,
Fig. 2 depicts a scheme referred to as Video Redundancy Coding, where a sequence of pictures is divided into two or more independently coded threads in an interleaved manner,
Fig. 3 presents an example of a prediction structure potentially improving compression efficiency,
Fig. 4 presents an example of the intra picture postponement method that can be used to improve error resiliency,
Fig. 5 depicts an advantageous embodiment of the system according to the present invention,
Fig. 6 depicts an advantageous embodiment of the encoder according to the present invention,
Fig. 7 depicts an advantageous embodiment of the decoder according to the present invention,
Fig. 8 shows a block diagram of a general video communications system.
Detailed Description of the Invention
For consistency and clarity, the following definitions related to primary coded and redundant coded slices are defined for use of the description of the invention:
Slice data partitioning is a method of partitioning the syntax elements of the slice syntax structure into slice data partition syntax structures based on the type of each syntax element. In the JVT coding standard, there are three slice data partition syntax structures: slice data partition A, B, and C. Slice data partition A contains all syntax elements in the slice header and slice data syntax structures other than the syntax elements for coding the difference between the predicted sample values and the decoded sample values. Slice data partition B contains the syntax elements for coding the difference between the predicted sample values and the decoded sample values in intra macroblock types (I and SI macroblocks). Slice data partition C contains the syntax elements for coding the difference between the predicted sample values and the decoded sample values in inter-predicted macroblock types (P, SP, and B macroblocks).
Primary coded data partition is a data partition which belongs to a primary coded picture.
Primary coded picture is a primary coded representation of a picture.
Primary coded slice is a slice which belongs to a primary coded picture.
Redundant coded data partition is a data partition which belongs to a redundant coded picture.
Redundant coded picture is a redundant coded representation of a picture that should only be used if the primary coded or decoded picture is corrupted. The decoded redundant picture may not cover the entire picture area. There should be no noticeable difference between the co- located areas of the decoded primary picture and any decoded redundant slices. The redundant coded picture is not required to contain all macroblocks in the primary coded picture.
Redundant coded slice is a slice belonging to a redundant coded picture.
There are two major differences between "not coded" macroblocks and the macroblocks that are not included in a redundant picture: First, macroblocks that are not included in a redundant coded picture are not signaled, whereas "not coded" macroblocks are coded in the bitstream (typically by one bit per "not coded" macroblock). Second, decoders must not decode the areas not included in a redundant picture. If any macroblock is not included in the received primary coded picture or any corresponding redundant coded picture, decoders should conceal these missing macroblocks using any proprietary error concealment algorithm. In contrast to this, there is a specific normative decoding process for "not coded" macroblocks.
In the following the invention will be described in more detail with reference to the system of Fig. 5, the encoder 1 and optional hypothetical reference decoder (HRD) 5 of Fig. 6 and decoder 2 of Fig. 7. The pictures to be encoded can be, for example, pictures of a video stream from a video source 3, e.g. a camera, a video recorder, etc. The pictures (frames) of the video stream can be divided into smaller portions such as slices. The slices can further be divided into blocks. In the encoder 1 the video stream is encoded to reduce the information to be transmitted via a transmission channel 4, or to a storage media (not shown). Pictures of the video stream are input to the encoder 1. The encoder has an encoding buffer 1.1 (Fig. 6) for temporarily storing some of the pictures to be encoded. The encoder 1 also includes a memory 1.3 and a processor 1.2 in which the encoding tasks according to the invention can be applied. The memory 1.3 and the processor 1.2 can be common with the transmitting device 6 or the transmitting device 6 can have another processor and/or memory (not shown) for other functions of the transmitting device 6. The encoder 1 performs motion estimation and/or some other tasks to compress the video stream. In motion estimation similarities between the picture to be encoded (the current picture) and a previous and/or latter picture are searched. If similarities are found the compared picture or part of it can be used as a reference picture for the picture to be encoded. In JVT the display order and the decoding order of the pictures are not necessarily the same, wherein the reference picture has to be stored in a buffer (e.g. in the encoding buffer 1.1 ) as long as it is used as a reference picture. The encoder 1 also inserts information on display order of the pictures into the transmission stream. In practice, either the timing information SEI message or timestamps external to the JVT syntax (such as RTP timestamps) can be used.
From the encoding process the encoded pictures are moved to an encoded picture buffer 1.2, if necessary. The encoded pictures are transmitted from the encoder 1 to the decoder 2 via the transmission channel 4. In the decoder 2 the encoded pictures are decoded to form uncompressed pictures corresponding as much as possible to the encoded pictures. Each decoded picture is buffered in the DPB 2.1 of the decoder 2 unless it is displayed substantially immediately after the decoding and is not used as a reference picture. Advantageously both the reference picture buffering and the display picture buffering are combined and they use the same decoded picture buffer 2.1. This eliminates the need for storing the same pictures in two different places thus reducing the memory requirements of the decoder 2.
The decoder 1 also includes a memory 2.3 and a processor 2.2 in which the decoding tasks according to the invention can be applied. The memory 2.3 and the processor 2.2 can be common with the receiving device 8 or the receiving device 8 can have another processor and/or memory (not shown) for other functions of the receiving device 8.
Encoding
Let us now consider the encoding-decoding process in more detail. Pictures from the video source 3 are entered to the encoder 1 and advantageously stored in the pre-encoding buffer 1.1. There are two main reasons for storing the pictures. First, pictures arriving after a picture to be encoded are analyzed in a bit rate control algorithm so that there would be no remarkable variations in the quality of the pictures. Second, the coding order of the pictures (and the decoding order) is different from the capture order of the pictures. This kind of arrangement can be effective on the compression efficiency point of view (for example, a PBBBP frame sequence in which the B frame in between two other B frames is a reference frame for the other two B frames) and/or on the error resilience point of view (intra picture postponement).
The encoding process is not necessarily started immediately after the first picture is entered to the encoder, but after a certain amount of pictures are available in the encoding buffer 1.1. Then the encoder 1 tries to find suitable candidates from the pictures to be used as the reference frames. The encoder 1 then performs the encoding to form encoded pictures. The encoded pictures can be, for example, predicted pictures (P), bi-predictive pictures (B), and/or intra-coded pictures (I). The intra-coded pictures can be decoded without using any other pictures, but other type of pictures need at least one reference picture before they can be decoded. Pictures of any of the above mentioned picture types can be used as a reference picture.
The encoder advantageously attaches two time stamps to the pictures: a decoding time stamp (DTS) and output time stamp (OTS). The decoder can use the time stamps to determine the correct decoding time and time to output (display) the pictures. However, those time stamps are not necessarily transmitted to the decoder or it does not use them.
The encoder 1 can form redundant coded pictures or redundant coded data partitions of the pictures to increase error resiliency. According to the present invention the encoder can form redundant pictures which do not contain all the necessary information to decode the picture but only some portion(s) of it. The encoder 1 can also form more than one different redundant coded data partitions for the same picture wherein the different redundant coded data partitions contain information from at least partly different areas of the picture. The smallest redundant coded picture consists preferably of a slice. The slice contains one or more macroblocks.
Preferably, the encoder 1 decides what pictures contain areas which should be redundantly coded. The criteria for the selection may vary in different embodiments and in different situations. For example, the encoder 1 may examine if there is a possible scene change between successive pictures or if there are, for some other reason, lot of changes between successive pictures. Respectively, the encoder 1 can examine whether there are changes in some portions of pictures to determine which parts of the pictures should be redundantly encoded. To decide this, the encoder 1 can, for example, examine the motion vectors to find out important regions and/or regions which are especially sensitive to transmission/decoding errors and form redundant coded data partitions of such regions. There should be some indication in the transmission stream to indicate if there exist redundant slices in the stream. The indication is preferably inserted in the slice header of each slice and/or in the picture parameter set. One advantageous embodiment of the indication uses two syntax elements for redundant slices: the first syntax element is "redundant_slice_flag" which resides in the picture parameter set, and the other syntax element is "redundant_pic_cnt" and it resides in slice header. The "redundant_pic_cnt" is optional and it is included in the slice header only when "redundant_slice_flag" in the referred picture parameter set is set to 1.
The semantics of the two syntax elements are as follows: redundant_slice_flag indicates the presence of the redundant_pic_cnt parameter in all slice headers referencing the picture parameter set. The picture parameter set can be common for more than one slice if all the parameters are equal for the slices. If the value of the redundant_slice_flag is true, then the slice headers of those slices which refer to this parameter set contain the second syntax element (redundant_pic_cnt).
The value of the redundant_pic_cnt is 0 for coded slices and data partitions belonging to the primary representation of the picture contents. The redundant_pic_cnt is greater than 0 for coded slices and data partitions that contain redundant coded representation of the picture contents. There should be no noticeable difference between the co-located areas of the decoded primary representation of the picture and any decoded redundant slices. Redundant slices and data partitions having the same value of redundant_pic_cnt belong to the same redundant picture. Decoded slices having the same redundant_pic_cnt shall not overlap. Decoded slices having a redundant_pic_cnt greater than 0 may not cover the entire picture area. The pictures may have a parameter called as nal_storage_idc. If the value of the nal_storage_idc in a primary picture is 0, the value of the nal_storage_idc in corresponding redundant pictures shall be 0. If the value of the nal_storage_idc in a primary picture is non-zero, the value of the nal_storage_idc in corresponding redundant pictures shall be non-zero. The above described syntax design works well when data partitioning is not applied for redundant slices. However, when data partitioning is used, i.e. each redundant slice has three data partitions DPA, DPB, and DPC, a further mechanism is needed to inform the decoder which redundant slice is in question. To achieve this, the redundant_pic_cnt is included, not only the slice header in DPA but also in slice headers of both DPB and DPC. If slice data partitioning is in use, slice data partition B and C have to be associated to the respective slice data partition A in order to enable decoding of the slice. Slice data partition A includes a slicejd syntax element whose value uniquely identifies a slice within a coded picture. Slice data partitions B and C include the redundant_pic_cnt syntax element if it is also present in the slice header included in the slice data partition A (which is conditional to the value of "redundant_slice_flag" in the referred picture parameter set). The value of the redundant_pic_cnt syntax element is used to associate slice data partitions B and C to a particular primary or redundant coded picture. In addition to redundant_pic_cnt, slice data partitions B and C include the slicejd syntax element, which is used to associate the data partition with the respective data partition A of the same coded picture.
Transmission
The transmission and/or storing of the encoded pictures (and the optional virtual decoding) can be started immediately after the first encoded picture is ready. This picture is not necessarily the first one in decoder output order because the decoding order and the output order may not be the same.
When the first picture of the video stream is encoded the transmission can be started. The encoded pictures are optionally stored to the encoded picture buffer 1.2. The transmission can also start at a later stage, for example, after a certain part of the video stream is encoded.
In some transmission systems the number of transmitted redundant pictures depends inter alia on network conditions such as amount of traffic, bit error rate in a radio link, etc. In other words, all redundant pictures are not necessarily transmitted. Decoding
Next, the operation of the receiver 8 will be described. The receiver 8 collects all packets belonging to a picture, bringing them into a reasonable order. The strictness of the order depends on the profile employed. The received packets are advantageously stored into the receiving buffer 9.1 (pre-decoding buffer). The receiver 8 discards anything that is unusable, and passes the rest to the decoder 2.
If the primary representation of the picture or part of it is lost or decoding error(s) exist, the decoder can use some of the redundant coded slices to decode the picture. The decoder 2 can send the slice id:s or some other information identifying the picture in question to the encoder 1. When the decoder 2 has all the necessary slices available it can start to decode the picture. It may happen that, despite of the usage of redundant coded data partitions, some slices may not be available at the decoder 2. In that case the decoder 2 can try e.g. some error recovery methods to diminish the effects of the error on picture quality, or the decoder 2 can discard the erroneous picture and use some previous picture instead.
The present invention can be applied in many kind of systems and devices. The transmitting device 6 including the encoder 1 and optionally the HRD 5 advantageously include also a transmitter 7 to transmit the encoded pictures to the transmission channel 4. The receiving device 8 include the receiver 9 to receive the encoded pictures, the decoder 2, and a display 10 on which the decoded pictures can be displayed. The transmission channel can be, for example, a landline communication channel and/or a wireless communication channel. The transmitting device and the receiving device include also one or more processors 1.2, 2.2 which can perform the necessary steps for controlling the encoding/decoding process of video stream according to the invention. Therefore, the method according to the present invention can mainly be implemented as machine executable steps of the processors. The buffering of the pictures can be implemented in the memory 1.3, 2.3 of the devices. The program code 1.4 of the encoder can be stored into the memory 1.3. Respectively, the program code 2.4 of the decoder can be stored into the memory 2.3.

Claims

Claims:
1. A method for encoding pictures, wherein
- primary coded pictures and redundant coded pictures of primary coded pictures are formed, each primary coded picture comprising essentially the same picture information as the respective redundant coded picture, and
- at least one of the redundant coded pictures comprises picture information corresponding to only a part of the picture information of the respective primary coded picture.
2. The method according to claim 1 further comprising a transmission step for transmitting at least said primary coded pictures to a decoder.
3. The method according to claim 1 , wherein the pictures to be encoded comprise slices, wherein said redundant coded pictures contain part of the slices of the primary coded picture.
4. The method according to claim 1 , wherein the redundant coded pictures containing only a part of the respective primary coded picture are formed as redundant coded data portions.
5. The method according to claim 4, in which at least one parameter set is formed for the pictures, and a slice header is formed for each slice, wherein an indication whether a transmission stream contains slices of redundant coded data partitions is inserted into the parameter set, and a redundant_pic_cnt parameter is inserted into each slice header of the redundant coded data partitions.
6. A method for decoding pictures from a bitstream, wherein primary coded pictures and redundant coded pictures of primary coded pictures are contained in the bitstream, the primary coded pictures having been formed using essentially the same picture information as what has been used to form the respective redundant coded pictures, and at least one of the redundant coded pictures comprises picture information corresponding to only a part of the picture information of the respective primary coded picture; detecting in the bitstream a parameter indicating that the coded picture information belongs to a redundant coded picture; using the parameter to control the decoding of the coded picture information belonging to a redundant coded picture wherein the redundant coded picture information corresponds to only a part of the picture information used to form the respective primary coded picture.
7. The method according to claim 6 further comprising a receiving step for receiving at least said primary coded pictures.
8. The method according to claim 7 further comprising receiving said redundant coded pictures.
9. The method according to claim 8 comprising determining, if a primary coded pictures contain areas which can not be decoded, wherein the method comprising examining if the redundant coded pictures contain decodable information on the areas of the primary coded pictures which can not be decoded, and decoding redundant coded pictures found on the basis of the examination.
10. The method according to claim 9, in which at least one parameter set is formed for the pictures, and a slice header is formed for each slice, wherein an indication whether a transmission stream contains slices of redundant coded data partitions is inserted into the parameter set, and a redundant_pic 3nt parameter is inserted into each slice header of the redundant coded data partitions, wherein the indication and the redundantjDic_cnt parameter are used to distinguish between primary coded pictures and redundant coded pictures.
11. An encoder for encoding pictures, comprising encoding means for forming primary coded pictures and redundant coded pictures of primary coded pictures, each primary coded picture comprising essentially the same picture information as the respective redundant coded picture, and at least one of the redundant coded pictures comprises picture information corresponding to only a part of the picture information of the respective primary coded picture.
12. A decoder for decoding pictures from a bitstream, the bitstream comprising:
- primary coded pictures and redundant coded pictures of primary coded pictures, the primary coded pictures having been formed using essentially the same picture information as what has been used to form the respective redundant coded pictures,
- and at least one of the redundant coded pictures comprises picture information corresponding to only a part of the picture information of the respective primary coded picture; wherein the decoder comprising:
- detecting means for detecting in the bitstream a parameter indicating that the coded picture information belongs to a redundant coded picture; and
- controlling means using the parameter to control the decoding of the coded picture information belonging to a redundant coded picture wherein the redundant coded picture information corresponds to only a part of the picture information used to form the respective primary coded picture.
13. A transmitting device comprising an encoder for encoding pictures, comprising encoding means for forming primary coded pictures and redundant coded pictures of primary coded pictures, each primary coded picture comprising essentially the same picture information as the respective redundant coded picture, and at least one of the redundant coded pictures comprises picture information corresponding to only a part of the picture information of the respective primary coded picture.
14. The transmitting device according to claim 13 further comprising a transmitter for transmitting at least said primary coded pictures to a decoder.
15. The transmitting device according to claim 13, wherein the pictures to be encoded encoded comprise slices, wherein said redundant coded pictures contain part of the slices of the primary coded picture.
16. The transmitting device according to claim 13, comprising means for forming at least one parameter set for the pictures, and a slice header for each slice, means for inserting an indication whether a transmission stream contains slices of redundant coded data partitions into the parameter set, and a redundantj ic_cnt parameter into each slice header of the redundant coded data partitions.
17. A receiving device comprising a decoder for decoding pictures from a bitstream, the bitstream comprising:
- primary coded pictures and redundant coded pictures of primary coded pictures, the primary coded pictures having been formed using essentially the same picture information as what has been used to form the respective redundant coded pictures,
- and at least one of the redundant coded pictures comprises picture information corresponding to only a part of the picture information of the respective primary coded picture; wherein the decoder comprising:
- detecting means for detecting in the bitstream a parameter indicating that the coded picture information belongs to a redundant coded picture; and
- controlling means using the parameter to control the decoding of the coded picture information belonging to a redundant coded picture wherein the redundant coded picture information corresponds to only a part of the picture information used to form the respective primary coded picture.
18. A system comprising:
- an encoder for encoding pictures, comprising encoding means for forming primary coded pictures and redundant coded pictures of primary coded pictures, each primary coded picture comprising essentially the same picture information as the respective redundant coded picture, and at least one of the redundant coded pictures comprises picture information corresponding to only a part of the picture information of the respective primary coded picture;
- a transmitter for transmitting at least said primary coded pictures to a decoder; the decoder comprising:
- detecting means for detecting in the bitstream a parameter indicating that the coded picture information belongs to a redundant coded picture; and
- controlling means using the parameter to control the decoding of the coded picture information belonging to a redundant coded picture wherein the redundant coded picture information corresponds to only a part of the picture information used to form the respective primary coded picture.
19. A software program comprising machine executable steps for encoding pictures, comprising machine executable steps for forming: primary coded pictures and redundant coded pictures of primary coded pictures, each primary coded picture comprising essentially the same picture information as the respective redundant coded picture, and at least one of the redundant coded pictures comprises picture information corresponding to only a part of the picture information of the respective primary coded picture.
20. A software program comprising machine executable steps for decoding pictures from a bitstream, comprising:
- primary coded pictures and redundant coded pictures of primary coded pictures are contained in the bitstream, the primary coded pictures having been formed using essentially the same picture information as what has been used to form the respective redundant coded pictures,
- and at least one of the redundant coded pictures comprises picture information corresponding to only a part of the picture information of the respective primary coded picture; wherein the software program comprising machine executable steps for:
- detecting in the bitstream a parameter indicating that the coded picture information belongs to a redundant coded picture; and
- using the parameter to control the decoding of the coded picture information belonging to a redundant coded picture wherein the redundant coded picture information corresponds to only a part of the picture information used to form the respective primary coded picture.
21. A storage medium for storing a software program comprising machine executable steps for encoding pictures, comprising machine executable steps for forming: primary coded pictures and redundant coded pictures of primary coded pictures, each primary coded picture comprising essentially the same picture information as the respective redundant coded picture, and at least one of the redundant coded pictures comprises picture information corresponding to only a part of the picture information of the respective primary coded picture.
22. A storage medium for storing a software program comprising machine executable steps for decoding pictures from a bitstream, comprising:
- primary coded pictures and redundant coded pictures of primary coded pictures are contained in the bitstream, the primary coded pictures having been formed using essentially the same picture information as what has been used to form the respective redundant coded pictures,
- and at least one of the redundant coded pictures comprises picture information corresponding to only a part of the picture information of the respective primary coded picture; wherein the software program comprising machine executable steps for:
- detecting in the bitstream a parameter indicating that the coded picture information belongs to a redundant coded picture; and
- using the parameter to control the decoding of the coded picture information belonging to a redundant coded picture wherein the redundant coded picture information corresponds to only a part of the picture information used to form the respective primary coded picture.
23. A bitstream comprising primary coded pictures and redundant coded pictures of primary coded pictures, each primary coded picture comprising essentially the same picture information as the respective redundant coded picture, and at least one of the redundant coded pictures comprises picture information corresponding to only a part of the picture information of the respective primary coded picture.
PCT/FI2004/050053 2003-04-30 2004-04-29 Picture coding method WO2004098196A1 (en)

Priority Applications (4)

Application Number Priority Date Filing Date Title
JP2006505656A JP2006526908A (en) 2003-04-30 2004-04-29 A method for encoding a picture, a method for decoding a picture from a bitstream, an encoder for encoding a picture, a transmission device including an encoder for encoding a picture, a decoder for decoding a picture from a bitstream, and a picture from a bitstream Receiving apparatus and system including a decoder for decoding
MXPA05011533A MXPA05011533A (en) 2003-04-30 2004-04-29 Picture coding method.
EP04730282A EP1618747A1 (en) 2003-04-30 2004-04-29 Picture coding method
BRPI0409491-3A BRPI0409491A (en) 2003-04-30 2004-04-29 method for encoding images, method for decoding images, encoder, decoder, transmitting device, receiving device, system, storage device, computer program, and bit stream

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US10/427,737 2003-04-30
US10/427,737 US20040218669A1 (en) 2003-04-30 2003-04-30 Picture coding method

Publications (1)

Publication Number Publication Date
WO2004098196A1 true WO2004098196A1 (en) 2004-11-11

Family

ID=33310239

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/FI2004/050053 WO2004098196A1 (en) 2003-04-30 2004-04-29 Picture coding method

Country Status (11)

Country Link
US (1) US20040218669A1 (en)
EP (1) EP1618747A1 (en)
JP (1) JP2006526908A (en)
KR (1) KR20050122281A (en)
CN (1) CN1781314A (en)
AR (1) AR044118A1 (en)
BR (1) BRPI0409491A (en)
MX (1) MXPA05011533A (en)
MY (1) MY137090A (en)
TW (1) TWI253868B (en)
WO (1) WO2004098196A1 (en)

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2007166618A (en) * 2005-12-12 2007-06-28 Thomson Licensing Video data encoder, video data decoder, video data encoding method, and video data decoding method
JP2009508454A (en) * 2005-09-07 2009-02-26 ヴィドヨ,インコーポレーテッド Scalable low-latency video conferencing system and method using scalable video coding
FR2932036A1 (en) * 2008-06-03 2009-12-04 Thales Sa METHOD AND SYSTEM FOR PROTECTING A COMPRESSED VIDEO STREAM AGAINST ERRORS ARISING DURING TRANSMISSION
JP2010505333A (en) * 2006-09-28 2010-02-18 トムソン ライセンシング Flexible redundant coding
US8699522B2 (en) 2005-07-20 2014-04-15 Vidyo, Inc. System and method for low delay, interactive communication using multiple TCP connections and scalable coding
US8938004B2 (en) 2011-03-10 2015-01-20 Vidyo, Inc. Dependency parameter set for scalable video coding
US9313486B2 (en) 2012-06-20 2016-04-12 Vidyo, Inc. Hybrid video coding techniques
US9426499B2 (en) 2005-07-20 2016-08-23 Vidyo, Inc. System and method for scalable and low-delay videoconferencing using scalable video coding

Families Citing this family (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP4712704B2 (en) * 2004-02-09 2011-06-29 パナソニック株式会社 Broadcast receiving apparatus, broadcast receiving method, broadcast receiving program, and broadcast receiving circuit
JP4564350B2 (en) * 2004-12-27 2010-10-20 株式会社東芝 Signal processing device
BRPI0610404A2 (en) * 2005-04-07 2012-01-10 Nokia Corp method, system and device for storing media stream packets, wireless communication device, server, multipoint media broadcasting server / service, device for receiving packets having a transmission schedule and at least one transmission frame, signal associated with a media stream, computer program product
US7827467B2 (en) * 2006-01-04 2010-11-02 Nokia Corporation Method for checking of video encoder and decoder state integrity
WO2007081150A1 (en) * 2006-01-09 2007-07-19 Electronics And Telecommunications Research Institute Method defining nal unit type and system of transmission bitstream and redundant slice coding
KR101378079B1 (en) * 2006-04-20 2014-03-28 톰슨 라이센싱 Method and apparatus for redundant video coding
US7584495B2 (en) * 2006-06-30 2009-09-01 Nokia Corporation Redundant stream alignment in IP datacasting over DVB-H
US8520733B2 (en) * 2006-06-30 2013-08-27 Core Wireless Licensing S.A.R.L Video coding
US9872045B2 (en) * 2006-10-16 2018-01-16 Conversant Wireless Licensing S.A R.L. Method, electronic device, system, computer program product and circuit assembly for reducing error in video coding
KR101087379B1 (en) * 2007-02-23 2011-11-25 노키아 코포레이션 Backward-compatible characterization of aggregated media data units
US8660175B2 (en) * 2007-12-10 2014-02-25 Qualcomm Incorporated Selective display of interpolated or extrapolated video units
US20100027614A1 (en) * 2008-08-04 2010-02-04 Legend Silicon Corp. Error awareness and means for remedying same in video decoding
JP5100572B2 (en) * 2008-08-25 2012-12-19 キヤノン株式会社 Encoder
US8934530B2 (en) * 2011-02-01 2015-01-13 Vidyo, Inc. Spatial scalability using redundant pictures and slice groups
US9826238B2 (en) 2011-06-30 2017-11-21 Qualcomm Incorporated Signaling syntax elements for transform coefficients for sub-sets of a leaf-level coding unit
CN102438150B (en) * 2011-09-29 2013-10-30 华中科技大学 Anti-channel-error-code lossless compression error-detection/correction coding/decoding method for image
MX341068B (en) * 2012-04-23 2016-08-05 Panasonic Ip Corp America Image encoding method, image decoding method, image encoding device, image decoding device, and image encoding/decoding device.
CN105103554A (en) * 2013-03-28 2015-11-25 华为技术有限公司 Method for protecting video frame sequence against packet loss
CN104754341B (en) * 2013-12-31 2019-02-26 华为技术有限公司 A kind of video data encoding, decoded method and apparatus
WO2017143988A1 (en) * 2016-02-26 2017-08-31 Versitech Limited Shape-adaptive model-based codec for lossy and lossless compression of images
CN113766319A (en) * 2018-06-01 2021-12-07 北京市商汤科技开发有限公司 Image information processing method and device, and storage medium
CN108668170B (en) * 2018-06-01 2021-07-02 北京市商汤科技开发有限公司 Image information processing method and device, and storage medium

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5734443A (en) * 1995-12-28 1998-03-31 Philips Electronics North America Corporation Method and device for performing source transitions in a video system which performs entropy encoding
US5757421A (en) * 1994-08-31 1998-05-26 Sony Corporation Method and apparatus for encoding moving picture signals and recording moving picture signals
US20020176025A1 (en) * 2001-03-05 2002-11-28 Chang-Su Kim Systems and methods for encoding redundant motion vectors in compressed video bitstreams

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5289276A (en) * 1992-06-19 1994-02-22 General Electric Company Method and apparatus for conveying compressed video data over a noisy communication channel
US5557331A (en) * 1993-03-11 1996-09-17 Matsushita Electric Industrial Co., Ltd. Image encoding method, an image encoding circuit, an image encoding apparatus, and an optical disk
GB2352350B (en) * 1999-07-19 2003-11-05 Nokia Mobile Phones Ltd Video coding

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5757421A (en) * 1994-08-31 1998-05-26 Sony Corporation Method and apparatus for encoding moving picture signals and recording moving picture signals
US5734443A (en) * 1995-12-28 1998-03-31 Philips Electronics North America Corporation Method and device for performing source transitions in a video system which performs entropy encoding
US20020176025A1 (en) * 2001-03-05 2002-11-28 Chang-Su Kim Systems and methods for encoding redundant motion vectors in compressed video bitstreams

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8699522B2 (en) 2005-07-20 2014-04-15 Vidyo, Inc. System and method for low delay, interactive communication using multiple TCP connections and scalable coding
US9426499B2 (en) 2005-07-20 2016-08-23 Vidyo, Inc. System and method for scalable and low-delay videoconferencing using scalable video coding
JP2009508454A (en) * 2005-09-07 2009-02-26 ヴィドヨ,インコーポレーテッド Scalable low-latency video conferencing system and method using scalable video coding
JP2007166618A (en) * 2005-12-12 2007-06-28 Thomson Licensing Video data encoder, video data decoder, video data encoding method, and video data decoding method
JP2010505333A (en) * 2006-09-28 2010-02-18 トムソン ライセンシング Flexible redundant coding
FR2932036A1 (en) * 2008-06-03 2009-12-04 Thales Sa METHOD AND SYSTEM FOR PROTECTING A COMPRESSED VIDEO STREAM AGAINST ERRORS ARISING DURING TRANSMISSION
WO2009147182A1 (en) * 2008-06-03 2009-12-10 Thales Method and system making it possible to protect a compressed video stream against errors arising during a transmission
US8938004B2 (en) 2011-03-10 2015-01-20 Vidyo, Inc. Dependency parameter set for scalable video coding
US9313486B2 (en) 2012-06-20 2016-04-12 Vidyo, Inc. Hybrid video coding techniques

Also Published As

Publication number Publication date
TW200427335A (en) 2004-12-01
AR044118A1 (en) 2005-08-24
TWI253868B (en) 2006-04-21
EP1618747A1 (en) 2006-01-25
MY137090A (en) 2008-12-31
US20040218669A1 (en) 2004-11-04
MXPA05011533A (en) 2005-12-12
JP2006526908A (en) 2006-11-24
KR20050122281A (en) 2005-12-28
BRPI0409491A (en) 2006-05-02
CN1781314A (en) 2006-05-31

Similar Documents

Publication Publication Date Title
US20040218669A1 (en) Picture coding method
US8532194B2 (en) Picture decoding method
US7403660B2 (en) Encoding picture arrangement parameter in picture bitstream
KR100837322B1 (en) Resizing of buffer in encoder and decoder
US7116714B2 (en) Video coding
US8144764B2 (en) Video coding

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A1

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BW BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE EG ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NA NI NO NZ OM PG PH PL PT RO RU SC SD SE SG SK SL SY TJ TM TN TR TT TZ UA UG US UZ VC VN YU ZA ZM ZW

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): BW GH GM KE LS MW MZ NA SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IT LU MC NL PL PT RO SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
WWE Wipo information: entry into national phase

Ref document number: PA/a/2005/011533

Country of ref document: MX

WWE Wipo information: entry into national phase

Ref document number: 2004730282

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 2006505656

Country of ref document: JP

Ref document number: 1020057020571

Country of ref document: KR

WWE Wipo information: entry into national phase

Ref document number: 20048116252

Country of ref document: CN

WWE Wipo information: entry into national phase

Ref document number: 1200501766

Country of ref document: VN

WWP Wipo information: published in national office

Ref document number: 1020057020571

Country of ref document: KR

WWP Wipo information: published in national office

Ref document number: 2004730282

Country of ref document: EP

ENP Entry into the national phase

Ref document number: PI0409491

Country of ref document: BR