US20140092966A1 - Video encoding apparatus, video decoding apparatus, video encoding method, and video decoding method - Google Patents

Video encoding apparatus, video decoding apparatus, video encoding method, and video decoding method Download PDF

Info

Publication number
US20140092966A1
US20140092966A1 US14/040,847 US201314040847A US2014092966A1 US 20140092966 A1 US20140092966 A1 US 20140092966A1 US 201314040847 A US201314040847 A US 201314040847A US 2014092966 A1 US2014092966 A1 US 2014092966A1
Authority
US
United States
Prior art keywords
picture
video data
decoding
encoded
starting
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US14/040,847
Other languages
English (en)
Inventor
Kimihiko Kazui
Junpei KOYAMA
Satoshi Shimada
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Fujitsu Ltd
Original Assignee
Fujitsu Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Fujitsu Ltd filed Critical Fujitsu Ltd
Assigned to FUJITSU LIMITED reassignment FUJITSU LIMITED ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: SHIMADA, SATOSHI, KAZUI, KIMIHIKO, KOYAMA, JUNPEI
Publication of US20140092966A1 publication Critical patent/US20140092966A1/en
Priority to US15/015,247 priority Critical patent/US20160156922A1/en
Priority to US15/015,233 priority patent/US10582208B2/en
Priority to US15/015,221 priority patent/US20160156920A1/en
Priority to US15/015,204 priority patent/US20160219296A1/en
Priority to US16/785,706 priority patent/US20200177907A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • H04N19/00533
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/157Assigned coding mode, i.e. the coding mode being predefined or preselected to be further used for selection of another element or parameter
    • H04N19/159Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/42Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by implementation details or hardware specially adapted for video compression or decompression, e.g. dedicated software implementation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/44Decoders specially adapted therefor, e.g. video decoders which are asymmetric with respect to the encoder
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/46Embedding additional information in the video signal during the compression process
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/70Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • H04N21/23424Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving splicing one content stream with another content stream, e.g. for inserting or substituting an advertisement

Definitions

  • the embodiments discussed herein are related to a video encoding apparatus and video encoding method that can edit encoded video data without decoding the video data, and also relates to a video decoding apparatus and video decoding method for decoding video data encoded by such a video encoding apparatus.
  • an apparatus handling such video data compresses the video data by encoding before transmitting the video data to another apparatus or before storing the video data in a storage device.
  • Coding standards such as MPEG-2 (Moving Picture Experts Group Phase 2), MPEG-4, and H.264 MPEG-4 Advanced Video Coding (MPEG-4 AVC/H.264), devised by the International Standardization Organization/International Electrotechnical Commission (ISO/IEC), are typical video coding standards widely used today.
  • Such coding standards employ inter-coding, a coding method that encodes a picture by using not only information from itself but also information from pictures before and after it, and intra-coding, a coding method that encodes a picture by using only information contained in the picture to be encoded.
  • the inter-coding method uses three types of picture, referred to as the intra-coded picture (I picture), the forward predicted picture (P picture) which is usually predicted from a past picture, and the bidirectional predicted picture (B picture) which is usually predicted from both past and future pictures.
  • the amount of code of a picture or block encoded by inter-coding is smaller than the amount of code of a picture or block encoded by intra-coding.
  • the amount of code varies from picture to picture within the same video sequence, depending on the coding mode selected.
  • the amount of code varies from block to block within the same picture, depending on the coding mode selected. Therefore, in order to enable a data stream containing encoded video to be transmitted at a constant transmission rate even if the amount of code temporally varies, a transmit buffer for buffering the data stream is provided at the transmitting end, and a receive buffer for buffering the data stream is provided at the receiving end.
  • MPEG-2 and MPEG-4 AVC/H.264 each define the behavior of a receive buffer in an idealized video decoding apparatus called the video buffering verifier (VBV) or the coded picture buffer (CPB), respectively.
  • VBV video buffering verifier
  • CPB coded picture buffer
  • the idealized video decoding apparatus will hereinafter be referred to simply as the idealized decoder. It is specified that the idealized decoder performs instantaneous decoding that takes zero time to decode.
  • Japanese Laid-open Patent Publication No. 2003-179938 discloses a video encoder control method concerning the VBV.
  • the video encoder controls the amount of code to guarantee that all the data needed to decode a given picture are available in the receive buffer when the idealized decoder decodes that given picture.
  • the receive buffer may underflow if the transmission of the data needed to decode the picture has not been completed by the time the picture is to be decoded and displayed by the video decoder. That is, the receive buffer underflow refers to a situation in which the data needed to decode the picture are not available in the receive buffer of the video decoder. If this happens, the video decoder is unable to perform decoding, and frame skipping occurs.
  • the video decoder displays the picture after delaying the stream by a prescribed time from its receive time so that the decoding can be done without causing the receive buffer to underflow.
  • the idealized decoder accomplishes decoding in zero time.
  • Solid line graph 100 depicts the buffer occupancy as a function of time.
  • the buffer occupancy of the receive buffer is restored at a rate synchronized to a prescribed transmission rate, and the data used for decoding each picture is retrieved from the buffer at the decode time of the picture.
  • the data of the i-th picture starts to be input to the receive buffer at time at(i), and the final data of the i-th picture is input at time ft(i).
  • the idealized decoder completes the decoding of the i-th picture at time tr(i), and thus the i-th picture becomes ready for display at time tr(i).
  • the data stream contains a B picture, the actual display time of the i-th picture may become later than tr(i) due to the occurrence of picture reordering (changing the encoding order).
  • SEI Supplemental Enhancement Information
  • BPSEI Buffering Period SEI
  • a parameter InitialCpbRemovalDelay is described in the BPSEI.
  • the InitialCpbRemovalDelay parameter indicates the difference between the time of arrival in the receive buffer of the first bit of the BPSEI-appended picture and the decode time of the BPSEI-appended picture. The resolution of the difference is 90 kHz.
  • the decode time tr(0) of the first picture is the time of arrival in the video decoder of the first bit of the encoded video data (the time is designated as 0); i.e., the decode time is delayed from time at(0) by an amount of time equal to InitialCpbRemovalDelay ⁇ 90,000 [sec].
  • PTSEI Picture Timing SEI
  • CpbRemovalDelay indicates the difference between the decode time of the immediately preceding BPSEI-appended picture and the decode time of the PTSEI-appended picture.
  • the DpbOutputDelay parameter indicates the difference between the decode time of the PTSEI-appended picture and the display time of that picture. The resolution of these differences is one field picture interval. Accordingly, when the picture is a frame, the value of each of the parameters CpbRemovalDelay and DpbOutputDelay is a multiple of 2.
  • the decode time tr(i) of each of the second and subsequent pictures is delayed from the decode time tr(0) of the first picture by an amount of time equal to tc*CpbRemovalDelay(i) [sec].
  • CpbRemovalDelay(i) is the CpbRemovalDelay appended to the i-th picture.
  • tc is the inter-picture time interval [sec]; for example, in the case of 29.97-Hz progressive video, tc is 1001/60000.
  • the display time of each of the pictures, including the BPSEI-appended picture, is delayed from tr(i) by an amount of time equal to tc*DpbOutputDelay(i).
  • DpbOutputDelay(i) is the DpbOutputDelay appended to the i-th picture. That is, after time tr(0), each picture is decoded and displayed at time equal to an integral multiple of tc.
  • the encoded video may be edited. Editing the encoded video involves dividing the encoded video data into smaller portions and splicing them to generate a new encoded video data stream. For example, insertion of another video stream (for example, an advertisement) into the currently broadcast video stream (i.e., splicing) is one example of edit operation.
  • an encoded video data editing machine When editing inter-frame predictive coded video, particularly in the case of an inter-coded picture, the encoded picture cannot be decoded correctly by itself. Accordingly, when splicing two encoded video data streams at a desired picture position, an encoded video data editing machine first decodes the two encoded video data streams to be spliced and then splice them on a decoded picture-by-picture basis, and thereafter re-encodes the spliced video data.
  • the task of re-encoding can be very laborious, in particular, in the case of real time processing such as splicing, it is common to restrict the splicing point and edit the encoded video data directly by eliminating the need for re-encoding.
  • the first picture of the encoded video data stream to be spliced on the temporally downstream side has to be an I picture.
  • the GOP structure of the encoded video data stream to be spliced on the temporally downstream side is limited to the so-called closed GOP structure in which all the pictures that follow the starting I picture are decodable without referring to any pictures temporally preceding the starting I picture. With this arrangement, it is possible to correctly decode all the pictures that follow the starting I picture of the encoded video data stream spliced on the downstream side by editing at the desired splicing point.
  • the non-closed GOP structure may be employed. In that case, some of the pictures immediately following the starting I picture after the splicing point are not correctly decoded, but since these pictures are pictures preceding the starting I picture in display order, there will be no problem if they are not displayed. Therefore, as a general practice, after displaying the last picture of the temporally preceding encoded video data stream, the video decoder performs processing such as freezing the display, thereby masking the display of the pictures that failed to be decoded correctly.
  • the header information is also edited so that a discrepancy does not occur between the two encoded video data stream spliced together.
  • POC Picture Order Count
  • FrameNum is appended to the slice header in order to maintain the inter-picture temporal relationship and identify the reference picture.
  • POC indicates the relative display order of the picture.
  • FrameNum is a value that increments by 1 each time the reference picture appears in the encoded video.
  • a CRA Create Random Access
  • BLA Broken Link Access
  • TFD Tagged For Discard
  • DLP Decodable Leading Picture
  • TP Trailing Picture
  • the CRA picture and the BLA picture are both self-contained pictures, i.e., pictures that do not refer to any other pictures, so that pictures that follow the CRA picture or the BLA picture can be decoded correctly.
  • the video decoder starts decoding starting with a CRA picture, for example, any subsequent pictures other than the TFD picture that immediately follows the CRA picture can be decoded correctly.
  • the TFD picture is a picture that appears immediately following the CRA picture or the BLA picture, and that refers to a picture appearing earlier than the CRA picture or the BLA picture in time order and in decoding order.
  • the plurality of B pictures immediately following the I picture at the head of the GOP each correspond to the TFD picture.
  • the BLA picture occurs as a result of editing of the encoded video data.
  • the encoded video data stream spliced on the downstream side generally begins with a CRA picture, but if this CRA picture appears partway through the spliced encoded video data, its picture type is changed from the CRA picture to the BLA picture.
  • the POC values are permitted to become discontinuous.
  • the TFD picture that immediately follows this BLA picture is unable to be decoded correctly from any point in the spliced encoded video data because the picture to be referred to by it is lost from the spliced encoded video data. Therefore, the video encoder may delete from the encoded video data any TFD picture that follows the BLA picture at the head of the encoded video data stream to be spliced on the downstream side.
  • a DLP picture like the TFD picture, is a picture that appears immediately following a CRA picture or BLA picture. Unlike the TFD picture, the DLP picture does not refer to a picture that precedes the CRA picture or BLA picture in both time order and decoding order. As a result, even if the decoding is started from the CRA picture or BLA picture, the DLP picture can be decoded correctly.
  • a TP picture is a picture that appears later than the CRA picture or BLA picture and the TFD and DLP pictures in decoding order, and that follows the CRA picture or BLA picture in time order. As a result, even if the decoding is started from the CRA picture or BLA picture, the TP picture can be decoded correctly.
  • the decode time and display time of each encoded picture are determined by using the parameters InitialCpbRemovalDelay, CpbRemovalDelay, and DpbOutputDelay.
  • the parameters CpbRemovalDelay and DpbOutputDelay of the picture after the splicing point need to be corrected to appropriate values in order to ensure continuous video decoding and display across the splicing point.
  • the video encoder or the video decoder needs to correct the value of CpbRemovalDelay of the CRA picture at the head of the encoded video data stream spliced on the downstream side, based on the number of pictures that follow the last BPSEI-appended picture in the temporally preceding encoded video data stream. Furthermore, the video encoder or the video decoder increments the value of CpbRemovalDelay in order to ensure CPB buffer continuity.
  • the video encoder or the video decoder needs to correct the value of CpbRemovalDelay of the picture to be decoded after the discarded TFD picture as well as the value of DpbOutputDelay of the first CRA picture after the splicing point.
  • a video encoding apparatus for generating spliced encoded video data by splicing together first video data and second video data both encoded using inter-frame predictive coding.
  • the video encoding apparatus includes: a splicing point identification information processing unit which obtains decoding delay and display delay correction information for ensuring that even when one or more pictures that are later in encoding order than a starting encoded picture in the second video data to be spliced to a trailing end of the first encoded video data have been discarded from among pictures contained in the second video data, the starting encoded picture and subsequent pictures contained in the second encoded video data can be continuously decoded and displayed by a video decoding apparatus, and which appends the correction information to the spliced video data; and a data splicing unit which discards from the spliced video data any picture that is later in encoding order than the starting encoded picture and that is not guaranteed to be decoded correctly if decoding is started from the starting encoded picture.
  • a video decoding apparatus for decoding video data encoded by inter-frame predictive coding.
  • the video decoding apparatus includes: a picture decoding/display time determining unit which uses information indicating that the encoded video data has been generated by splicing second encoded video data to a trailing end of first encoded video data and information indicating that one or more encoded pictures that are later in encoding order than a starting encoded picture in the second encoded video data spliced downstream of a splicing point at which the first encoded video data and the second encoded video data are spliced have been discarded, and which corrects, based on correction information, a decoding delay for a picture that follows the starting encoded picture, while also correcting, based on the correction information, display delays for the starting encoded picture and the picture that follows the starting encoded picture.
  • the correction information carries a value calculated based on a decoding interval between each picture that is later in decoding order than the starting encoded picture, and that has been discarded from the second encoded video data, and a picture that immediately precedes the discarded picture in decoding order.
  • FIG. 1 is a diagram illustrating the relationship between the buffer occupancy of a receive buffer and the display time.
  • FIG. 2 is a diagram illustrating the relationship between the display order and decoding order of various pictures contained in video data and the decoding delay and display delay values of the respective pictures.
  • FIG. 3 is a diagram for explaining the decoding delay and display delay values of the pictures after the splicing point when two encoded video data streams are spliced.
  • FIG. 4 is a diagram for explaining the data structure of one picture in encoded video according to a first embodiment.
  • FIG. 5 is a simplified block diagram of a video encoding apparatus according to the first embodiment.
  • FIG. 6 is an operation flowchart of a video encoding process according to the first embodiment.
  • FIG. 7 is an operation flowchart of a video editing process according to the first embodiment.
  • FIG. 8 is a simplified block diagram of a video decoding apparatus according to the first embodiment.
  • FIG. 9 is an operation flowchart of a video decoding process according to the first embodiment.
  • FIG. 10 is a diagram for explaining the decoding delay and display delay values of the pictures after the splicing point when two encoded video data streams are spliced according to a second embodiment.
  • FIG. 11 is a diagram for explaining the data structure of one picture in encoded video according to the second embodiment.
  • FIG. 12 is a diagram illustrating the configuration of a computer that operates as the video encoding apparatus or video decoding apparatus by executing a computer program for implementing the functions of the various units in the video encoding apparatus or video decoding apparatus according to any one of the embodiments or their modified examples.
  • Video encoding apparatus and video decoding apparatus will be described below with reference to the drawings.
  • the video encoding apparatus computes the values to be used for correcting the parameters indicating the decode time and display time of each picture appearing after the splicing point, and adds the values in the header information of each picture appearing after the splicing point. In this way, the video encoding apparatus eliminates the need to edit the parameters in the header of the original encoded video data when splicing two encoded video data streams.
  • the picture is a frame.
  • the picture may not be limited to a frame, but may be a field.
  • a frame refers to one complete still image in the video data
  • a field refers to a still image obtained by extracting data only in the odd-numbered lines or even-numbered lines from one frame.
  • the encoded video data may be color video data or monochrome video data.
  • the picture coding structure 201 as one example of the picture coding structure contains a plurality of pictures.
  • Each block in the picture coding structure 201 represents one picture.
  • the alphabetic character at the left indicates the coding mode applied to that picture.
  • the characters I, P, and B mean I picture, P picture, and B picture, respectively.
  • the number at the right indicates the order of input to the video encoding apparatus. The input order coincides with the order in which the picture is output from the video decoding apparatus.
  • Each arrow depicted above the picture coding structure 201 indicates the reference picture to which a picture to be encoded by forward frame prediction refers.
  • a picture P4 refers to a picture I0 that appears earlier than the picture P4.
  • each arrow depicted below the picture coding structure 201 indicates the reference picture to which a picture to be encoded by backward frame prediction refers.
  • a picture B2 refers to the picture P4 that appears later than the picture B2.
  • the picture sequence presented below the picture coding structure 201 is the decoding order 202 of the pictures contained in the picture coding structure 201 .
  • Each block in the decoding order 202 represents one picture and, as in the case of the picture coding structure 201 , the characters carried in the block indicate the coding mode and the order of input to the video encoding apparatus.
  • the decoding order 202 coincides with the order of encoding performed by the video encoding apparatus.
  • Each arrow depicted above the picture decoding order 202 and each arrow depicted below the picture decoding order 202 indicate, respectively, the reference picture to which a picture encoded by forward frame prediction refers and the reference picture to which a picture encoded by backward frame prediction refers.
  • BPSEI is appended to each picture that carries the designation “BPSEI” thereunder.
  • BPSEI is appended to every I picture. That is, for every I picture, the InitialCpbRemovalDelay parameter that indicates the difference between the time of arrival in the receive buffer of the first bit of the I picture and the decode time of the I picture is defined.
  • a block array 203 presented below the decoding order 202 indicates the values of CpbRemovalDelay and DpbOutputDelay carried in PTSEI appended to each picture.
  • Each block in the upper row of the block array 203 carries the value of CpbRemovalDelay for the corresponding picture in the decoding order 202 located directly above that block.
  • each block in the lower row of the block array 203 carries the value of DpbOutputDelay for the corresponding picture in the decoding order 202 located directly above that block.
  • CpbRemovalDelay corresponds to the order of encoding as counted from the most recent picture in the encoding order among the pictures to which BPSEI is appended.
  • a picture P8 is the fifth picture in the encoding order as counted from the picture I0.
  • the inter-picture time interval tc is a value expressed in units of fields
  • DpbOutputDelay specifies the display delay that is needed in the video decoding apparatus to sequentially output the pictures in the correct order.
  • the DpbOutputDelay value of the picture P4 is 10. This is the delay needed to correctly display the picture B1 whose difference between the input order and the encoding order in the video encoding apparatus is the largest. That is, since the picture B1 is decoded with a delay of two pictures after the picture P4 is decoded, the display time of the picture P4 has to be further delayed by three picture times from the earliest time at which the picture B1 becomes ready for display, that is, the time at which the picture B1 is decoded. Since the difference between the decode time and display time of the picture P4 is equal to five picture times, and since tc is expressed in units of fields, the value of DpbOutputDelay is given as 10.
  • the decoding delay CpbRemovalDelay and display delay DpbOutputDelay of each picture in the encoded video data stream spliced downstream of the splicing point may take in order to eliminate any inconsistencies in decoding delay and display delay before and after the splicing point of the two encoded video data streams.
  • Each block in the first encoded video data stream 301 spliced upstream of the splicing point represents one picture, and the characters carried in the block indicate the coding mode and the order of input to the video encoding apparatus, as in the case of FIG. 2 .
  • the coding structure of the first encoded video data stream 301 is identical to the coding structure 201 depicted in FIG. 2 .
  • the second encoded video data stream 302 is spliced immediately following the last picture B15 of the first encoded video data stream.
  • each block represents one picture, and the characters carried in each block indicate the coding mode and the order of input to the video encoding apparatus.
  • Arrows depicted above the second encoded video data stream 302 indicate the reference pictures to which pictures B70, B69, and B71, respectively, refer when encoded by forward frame prediction.
  • arrows depicted below the second encoded video data stream 302 indicate the reference pictures to which the pictures B70, B69, and B71, respectively, refer when encoded by backward frame prediction.
  • the coding structure of the second encoded video data stream 302 is identical to the coding structure 201 depicted in FIG. 2 , except for the pictures B70, B69, and B71.
  • the encoding order of the pictures B70, B69, and B71 is the same as the encoding order of the bidirectional predicted pictures contained in the coding structure 201 depicted in FIG. 2 .
  • the reference pictures for the pictures B70, B69, and B71 are different from the reference pictures for the bidirectional predicted pictures contained in the coding structure 201 .
  • the pictures B70 and B71 each refer only to a picture later in display time, i.e., the picture I72.
  • the picture B69 refers only to a picture earlier in display time, i.e., the picture I68.
  • the picture I68 is a TFD picture
  • B70 and B71 are DLP pictures.
  • the picture I72 and the subsequent pictures in the second encoded video data stream 302 are spliced so as to follow the picture B15 in the first encoded video data stream.
  • a block array 303 presented below the second encoded video data stream 302 indicates the values of the decoding delay CpbRemovalDelay and display delay DpbOutputDelay carried in the PTSEI appended to each picture of the second encoded video data stream 302 .
  • Each block in the upper row of the block array 303 carries the value of the decoding delay CpbRemovalDelay for the corresponding picture in the second encoded video data stream 302 located directly above that block.
  • each block in the lower row of the block array 303 carries the value of the display delay DpbOutputDelay for the corresponding picture in the second encoded video data stream 302 located directly above that block.
  • Spliced encoded video data 304 generated by splicing the first and second encoded video data streams 301 and 302 is depicted below the block array 303 .
  • the picture B67 in the second encoded video data stream 302 and the pictures preceding the picture B67 in the encoding order are not contained in the spliced encoded video data 304 .
  • the picture B69 is a TFD picture that refers to the encoded picture I68 that precedes the picture I72 in the encoding order.
  • the picture B69 may not be discarded and may be retained in the spliced encoded video data.
  • the pictures B70 and B71 are each a DLP picture that does not refer to a picture preceding the picture I72 in the encoding order, and can therefore be correctly reproduced.
  • the pictures B70 and B71 are pictures neither of which is referred to from the picture P76 or its subsequent pictures, if the pictures B70 and B71 were discarded simultaneously with the TFD picture 69 , that would not affect the reproduction of the picture P76 and its subsequent pictures.
  • a block array 305 indicates the values of the decoding delay CpbRemovalDelay and display delay DpbOutputDelay that the pictures I72, B70, B71, P76, B74, B73, and B75 in the spliced encoded video data 304 are supposed to have.
  • Each block in the upper row of the block array 305 carries the value of the decoding delay CpbRemovalDelay for the corresponding picture in the spliced encoded video data 304 located directly above that block.
  • each block in the lower row of the block array 305 carries the value of the display delay DpbOutputDelay for the corresponding picture in the spliced encoded video data 304 located directly above that block.
  • the decoding delay CpbRemovalDelay of the picture I72 needs to be made to match the encoded picture interval relative to the picture I12 which is the picture having the immediately preceding BPSEI.
  • the display delay DpbOutputDelay of the picture I72 needs also to be corrected so that the picture B73 to be decoded later than the picture I72 can be correctly displayed.
  • the value of the display delay DpbOutputDelay of the picture I72 is different before and after the discarding of the picture B69.
  • the value of the display delay DpbOutputDelay after the discarding of the picture B69 decreases by a value equal to the decoding interval which is defined by the difference between the decoding time of the discarded picture that is later in decoding order than I72 and the decoding time of the picture that immediately precedes the discarded picture in decoding order.
  • the picture B69 is the discarded picture
  • the decoding interval of B69 i.e., the difference between the decoding time of B69 and the decoding time of the picture B70 immediately preceding it in decoding order
  • the value of the display delay DpbOutputDelay of the picture I72 is given as 2.
  • the display delay DpbOutputDelay of the picture B70 also decreases by a value equal to the decoding interval of the discarded picture that follows B70 in decoding order, i.e., by 2, and is thus given as 2.
  • the value of the decoding delay CpbRemovalDelay of each of the pictures B71, P76, B74, B73, and B75 is also different before and after the discarding of the picture B69.
  • the value of the decoding delay CpbRemovalDelay of each of the pictures B71, P76, B74, B73, and B75 after the discarding of the picture B69 decreases from the original value of the decoding delay CpbRemovalDelay by a value equal to the decoding interval of the discarded picture that precedes the picture I72 in decoding order.
  • the values of the decoding delay CpbRemovalDelay of the pictures B71, P76, B74, B73, and B75 are given as 4, 6, 8, 10, and 12, respectively, by subtracting the decoding interval, 2, of the TFD picture B69 from the original values of the decoding delay CpbRemovalDelay of the respective pictures.
  • the value of CpbRemovalDelay remains unchanged after the discarding of the picture B69, since there is no discarded picture that precedes B70 in decoding order.
  • the values of the display delay DpbOutputDelay of the pictures P76, B74, B73, and B75 also remain unchanged. Further, for any picture that was input later than the picture that would become the first CRA picture in the spliced encoded video data, neither the decoding delay CpbRemovalDelay nor the display delay DpbOutputDelay need be corrected.
  • the video encoding apparatus instead of correcting the values of the decoding delay CpbRemovalDelay and display delay DpbOutputDelay of the affected pictures contained in the original encoded video data before splicing, the video encoding apparatus adds in the header of the encoded video data the parameters that can be used to change the values of the decoding delay CpbRemovalDelay and display delay DpbOutputDelay to appropriate values when the video decoding apparatus decodes the spliced encoded video data.
  • the structure of the encoded video data according to the first embodiment will be described that contains the parameters that can be used to change the values of the decoding delay CpbRemovalDelay and display delay DpbOutputDelay to appropriate values.
  • the data structure 400 of one picture contains six kinds of network abstraction layer (NAL) units 410 to 415 .
  • NAL network abstraction layer
  • These NAL units 410 to 415 conform to the NAL units defined in MPEG-4 AVC/H.264 and the method disclosed in the non-patent document (JCTVC-J1003, “High-Efficiency Video Coding (HEVC) text specification Draft 8”, Joint Collaborative Team on Video Coding of ITU-T SG16 WP3 and ISO/IEC JTC1/SC29/WG11, July 2012).
  • a header NUH 420 is appended to each NAL unit.
  • the header NUH 420 contains a NalUnitType field that indicates the type of the NAL unit.
  • NalUnitType When NalUnitType is 1 or 2, it indicates that the picture is a TP picture. When NalUnitType is 7, it indicates that the picture is a self-contained BLA picture immediately after which a TFD picture and a DLP picture may appear. When NalUnitType is 8, it indicates that the picture is a self-contained BLA picture immediately after which a DLP picture may appear. When NalUnitType is 9, it indicates that the picture is a self-contained BLA picture immediately after which neither a TFD picture nor a DLP picture appears. When NalUnitType is 12, it indicates that the picture is a self-contained CRA picture. When NalUnitType is 13, it indicates that the picture is a DLP picture. When NalUnitType is 14, it indicates that the picture is a TFD picture.
  • the NalUnitType value of each picture need not be limited to the above specific value, but may be set to some other suitable value.
  • the NAL units will be explained below.
  • the NAL unit 410 is a delimiter (DELIM) NAL unit and indicates a picture boundary.
  • DELIM delimiter
  • the NAL unit 411 is a sequence parameter set (SPS) NAL unit which carries a set of parameters common to the entire sequence of the encoded video.
  • SPS sequence parameter set
  • the NAL unit 411 is appended to a self-contained picture.
  • the NAL unit 412 is a picture parameter set (PPS) NAL unit which carries a set of parameters common to a plurality of encoded pictures.
  • PPS picture parameter set
  • the PPS NAL unit 412 is appended to a self-contained picture, and may sometimes be appended to other types of picture.
  • the NAL unit 413 is a BPSEI NAL unit which is appended only to a self-contained picture.
  • the parameters to be used by the video decoding apparatus to correct the decoding delay and display delay of each picture appearing after the splicing point are added in this NAL unit 413 .
  • the NAL unit 414 is a PTSEI NAL unit which is appended to every picture.
  • the NAL unit 415 is a slice (SLICE) NAL unit which is the substance of the encoded picture.
  • the BPSEI NAL unit 413 contains a number, (N+1), of InitialCpbRemovalDelay/InitialCpbRemovalDelayOffset field pairs (where N is an integer not smaller than 0).
  • N is an integer not smaller than 0.
  • the definition of these field may be the same as that given in the method disclosed in the earlier cited non-patent document (JCTVC-J1003, “High-Efficiency Video Coding (HEVC) text specification Draft 8”, Joint Collaborative Team on Video Coding of ITU-T SG16 WP3 and ISO/IEC JTC1/SC29/WG11, July 2012) or in MPEG-4 AVC/H.264.
  • the reason that there is more than one InitialCpbRemovalDelay/InitialCpbRemovalDelayOffset field pair is to describe the InitialCpbRemovalDelay and InitialCpbRemovalDelayOffset parameters suitable for the case where the encoded bit stream is transmitted at (N+1) different bit rates.
  • the InitialCpbRemovalDelayOffset parameter defines the difference between the encoding completion time of the first picture in the video encoding apparatus and the start time of the transmission of the encoded picture data to the video decoding apparatus.
  • the PTSEI NAL unit 414 contains a decoding delay CpbRemovalDelay field, a display delay DpbOutputDelay field, and a NumRemovedTfds field.
  • the NumRemovedTfds field is one example of correction information used for the correction of the decoding delay and display delay.
  • the NumRemovedTfds field carries the sum of the decoding intervals of the pictures discarded during the interval between the PTSEI-appended picture and the next BPSEI-appended picture in decoding order.
  • the decoding interval of a picture is defined as the value obtained by subtracting, from the CpbRemovalDelay field value carried in the PTSEI appended to the picture, the CpbRemovalDelay field value carried in the PTSEI appended to the picture immediately preceding in decoding order.
  • the picture immediately preceding in decoding order is a BLA picture
  • the CpbRemovalDelay field value carried in the PTSEI appended to the BLA picture is treated as 0.
  • the NumRemovedTfds field value is set to 0.
  • FIG. 5 is a simplified block diagram of the video encoding apparatus according to the first embodiment.
  • the video encoding apparatus 1 includes a control unit 11 , an encoding control unit 12 , a picture encoding unit 13 , a splicing point identification information processing unit 14 , and a data splicing unit 15 .
  • These units constituting the video encoding apparatus 1 are each implemented as a separate circuit on the video encoding apparatus 1 .
  • these units constituting the video encoding apparatus 1 may be implemented on the video encoding apparatus 1 in the form of a single integrated circuit on which the circuits implementing the functions of the respective units are integrated.
  • these units constituting the video encoding apparatus 1 may be functional modules implemented by executing a computer program on a processor incorporated in the video encoding apparatus 1 .
  • the control unit 11 controls the operation of each designated unit of the video encoding apparatus 1 when encoding the video data or when editing the encoded video data. For example, based on the nature of the video data such as the scene change position, etc., and the reproduced image quality, compression ratio, etc. needed of the encoded video data, the control unit 11 determines the GOP structure, etc. that are applied to the video data to be encoded. Then, the control unit 11 signals the GOP structure, etc. to the encoding control unit 12 .
  • the video encoding process is carried out using the encoding control unit 12 and the picture encoding unit 13 .
  • the encoding control unit 12 determines the encoding order, the coding mode (for example, intra-coding, forward prediction, or bidirectional prediction), etc. for each picture in accordance with the GOP structure signaled from the control unit 11 . Then, based on the coding mode, the position within the GOP structure, etc. of each picture, the encoding control unit 12 determines the CRA picture insertion interval, the number of pictures to be reordered at the time of encoding, and the maximum display delay. In the example illustrated in FIG. 2 , the CRA picture insertion interval is 12, the number of pictures to be reordered is 2, and the maximum display delay is 5. The encoding control unit 12 generates the header information of each picture based on these values.
  • the coding mode for example, intra-coding, forward prediction, or bidirectional prediction
  • the encoding control unit 12 sets NalUnitType in the NUH 420 of each slice of the picture to 12.
  • NalUnitType in the NUH 420 of each slice of the starting picture of the encoded video data is set to 10 (IDR picture).
  • the encoding control unit 12 sets NalUnitType to (TFD picture) for a picture that immediately follows the CRA picture and that refers to a picture that is earlier in both decoding order and display order than the CRA picture.
  • the encoding control unit 12 sets NalUnitType to 13 (DLP picture). For the other pictures, the encoding control unit 12 sets NalUnitType to 1 or 2 (TP picture).
  • the encoding control unit 12 notifies the picture encoding unit 13 of the value of NalUnitType that has been set in the header NUH 420 of each slice of the picture to be encoded. Further, the encoding control unit 12 obtains the values of the decoding delay CpbRemovalDelay and display delay DpbOutputDelay in the PTSEI of each picture from the picture prediction structure, as shown in FIG. 2 , and notifies the picture encoding unit 13 accordingly.
  • the encoding control unit 12 appends BPSEI to that picture.
  • the encoding control unit 12 For each picture, the encoding control unit 12 signals the coding mode and the header information of the picture to the picture encoding unit 13 , and issues an instruction to encode the picture.
  • the picture encoding unit 13 in response to the instruction from the encoding control unit 12 , encodes the picture with the specified coding mode by employing one of video coding methods capable of inter-frame predictive coding.
  • the video coding method that the picture encoding unit 13 employs may be, for example, MPEG-4 AVC/H.264 or MPEG-2.
  • the picture encoding unit 13 stores the encoded video data, containing each encoded picture, in a storage unit (not depicted).
  • the editing process is carried out using the splicing point identification information processing unit 14 and the data splicing unit 15 .
  • the splicing point identification information processing unit 14 reads out from a storage unit (not depicted) the two encoded video data streams selected, for example, via a user interface unit (not depicted). Then, in accordance with an externally applied control signal (not depicted), the splicing point identification information processing unit 14 identifies the splicing point starting picture in the second encoded video data stream to be spliced temporally downstream of the other one of the encoded video data streams.
  • the external control signal specifies, for example, the number of encoded pictures as counted from the beginning of the second encoded video data stream, and the splicing point identification information processing unit 14 identifies, for example, the latest CRA picture within this number of encoded pictures as being the splicing point picture.
  • the splicing point identification information processing unit 14 changes the value of the NalUnitType of each slice of that picture from 12 to 7 which indicates that the picture is a BLA picture that may potentially be followed by a TFD picture.
  • This NalUnitType value indicates that the two encoded video data streams has been spliced together at that splicing point, and also that one or more encoded pictures later than the splicing point BLA picture in both encoding order and decoding order have been discarded.
  • the splicing point identification information processing unit 14 supplies the splicing point CRA picture and its subsequent pictures in the second encoded video data stream to the data splicing unit 15 , and issues an instruction to discard the TFD picture immediately following the splicing point CRA picture.
  • the splicing point identification information processing unit 14 changes the value of the NalUnitType of each slice of the splicing point CRA picture from 12 to 9 which indicates that the picture is a BLA picture immediately after which neither a TFD picture nor a DLP picture appears.
  • the splicing point identification information processing unit 14 calculates the decoding interval of the TFD picture to be discarded, and the value of the NumRemovedTfds field of a non-TFD picture immediately preceding the TFD picture to be discarded is incremented by a value equal to the decoding interval of the discarded TFD picture that follows the non-TFD picture.
  • the value of the NumRemovedTfds field of the non-TFD picture will eventually represent the number of discarded pictures in units of fields that follow the non-TFD picture in decoding order.
  • the splicing point identification information processing unit 14 corrects the value of the NumRemovedTfds field of the PTSEI appended to any picture that precedes in decoding order the TFD picture to be discarded in the second encoded video data stream.
  • the data splicing unit 15 receives the second encoded video data stream from the splicing point identification information processing unit 14 , and splices it to the trailing end of the first encoded video data stream that is spliced temporally upstream of the splicing point. At this time, the data splicing unit 15 discards any TFD picture that immediately follows the starting picture in the second encoded video data stream and that is not guaranteed to be decoded correctly. In this case, the data splicing unit 15 may also discard a DLP picture by regarding it as a TFD picture. Then, the data splicing unit 15 stores in the storage unit (not depicted) the spliced encoded video data created by splicing together the first and second encoded video data streams.
  • FIG. 6 is an operation flowchart illustrating the video encoding process performed by the video encoding apparatus according to the first embodiment.
  • the video encoding apparatus 1 encodes the entire video sequence in accordance with the operation flowchart of FIG. 6 .
  • the picture prediction structure such as the GOP structure is determined, for example, by the control unit 11 (step S 101 ).
  • the picture prediction structure determined is signaled to the encoding control unit 12 .
  • the encoding control unit 12 determines the coding mode to be applied to encode the target picture, and generates the header information of the target picture to be encoded (step S 102 ).
  • the encoding control unit 12 supplies the data of the target picture to the picture encoding unit 13 along with the type of the coding mode and the header information of that picture. Then, the picture encoding unit 13 encodes the target picture in accordance with the coding mode and the header information, and appends the header information to the data of the encoded picture (step S 103 ).
  • control unit 11 determines whether there is any picture remaining to be encoded in the video sequence (step S 104 ). If there is any picture remaining to be encoded (Yes in step S 104 ), the control unit 11 carries out the process from step S 102 onward to encode the next target picture. On the other hand, if there is no longer any picture remaining to be encoded (No in step S 104 ), the control unit 11 terminates the encoding process.
  • FIG. 7 is an operation flowchart illustrating the video editing process performed by the video encoding apparatus according to the first embodiment.
  • DLP pictures are not discarded, but only TFD pictures are discarded.
  • the splicing point identification information processing unit 14 initializes a list L[ ] of pictures that are not discarded among the TFD and DLP pictures, and initializes to 2 a variable m that represents the value obtained by adding 2 to the number of pictures not discarded (step S 201 ). If there is no TFD picture that follows the last DLP picture in decoding order, the variable m may be set to represent the number of pictures not discarded among the TFD and DLP pictures.
  • the splicing point identification information processing unit 14 sequentially reads out from the storage unit (not depicted) the encoded pictures up to the splicing point out of the first encoded video data stream to be spliced upstream of the splicing point (step S 202 ).
  • the splicing point identification information processing unit 14 sequentially reads out from the storage unit (not depicted) the encoded pictures after the splicing point out of the second encoded video data stream to be spliced downstream of the splicing point (step S 203 ).
  • the splicing point identification information processing unit 14 changes the value of NalUnitType in each slice NUH to the value that indicates a BLA picture (step S 204 ).
  • the splicing point identification information processing unit 14 determines whether the value of NalUnitType of the next picture in decoding order is 14 or not, that is, whether the next picture is a TFD picture or not (step S 205 ). If the picture is a TFD picture (Yes in step S 205 ), the splicing point identification information processing unit 14 issues an instruction to the splicing unit 15 to discard the TFD picture, and adds the decoding interval of that TFD picture, i.e., the difference in PTSEI CpbRemovalDelay value between that TFD picture and the picture immediately preceding it in decoding order, to each of the 0th to mth entries in the list [ ] (step S 206 ). After that, the splicing point identification information processing unit 14 returns to step S 205 to evaluate the NalUnitType of the next picture.
  • the splicing point identification information processing unit 14 determines whether the value of NalUnitType of the next picture in decoding order is 13 or not, i.e., whether the next picture is a DLP picture or not (step S 207 ). If the next picture is a DLP picture (Yes in step S 207 ), the splicing point identification information processing unit 14 increments the variable m by 1 (step S 208 ). After that, the splicing point identification information processing unit 14 returns to step S 205 to repeat the above process.
  • next picture in decoding order is not a DLP picture (No in step S 207 )
  • the next picture is neither a TFD picture nor a DLP picture, but is a TP picture.
  • the splicing point identification information processing unit 14 updates the NumRemovedTfds field of the PTSEI appended to each of the BLA and DLP pictures (step S 209 ).
  • the splicing point identification information processing unit 14 updates the value of the NumRemovedTfds field of the PTSEI appended to the kth picture to L[k]. After that, the splicing point identification information processing unit 14 supplies the BLA picture and its subsequent pictures to the data splicing unit 15 .
  • the splicing unit 15 splices the BLA picture and its subsequent pictures in the second encoded video data stream so as to follow the last picture of the first encoded video data stream upstream of the splicing point. At this time, the splicing unit 15 discards the TFD pictures specified by the splicing point identification information processing unit 14 to be removed.
  • FIG. 8 is a simplified block diagram of the video decoding apparatus according to the first embodiment.
  • the video decoding apparatus 2 includes a control unit 21 , a header information analyzing unit 22 , a picture decoding/display time determining unit 23 , a picture decoding unit 24 , and a frame memory 25 .
  • These units constituting the video decoding apparatus 2 are each implemented as a separate circuit on the video decoding apparatus 2 .
  • these units constituting the video decoding apparatus 2 may be implemented on the video decoding apparatus 2 in the form of a single integrated circuit on which the circuits implementing the functions of the respective units are integrated.
  • these units constituting the video decoding apparatus 2 may be functional modules implemented by executing a computer program on a processor incorporated in the video decoding apparatus 2 .
  • the control unit 21 controls the operation of each designated unit of the video decoding apparatus 2 when decoding the encoded video data.
  • the header information analyzing unit 22 analyzes the header information of the encoded video data, and passes the parameters necessary for the determination of the picture decoding and display times, for example, NalUnitType of each picture and CpbRemovalDelay, DpbOutputDelay, and NumRemovedTfds carried in PTSEI, to the picture decoding/display time determining unit 23 .
  • the picture decoding/display time determining unit 23 that received the parameters from the header information analyzing unit 22 checks the slice NUH of the picture to be decoded. When the value of NalUnitType carried in the NUH is 7 or 8 or 9, the picture decoding/display time determining unit 23 determines that the picture to be decoded is a BLA picture.
  • the picture decoding/display time determining unit 23 uses, as the decoding delay CpbRemovalDelay of the BLA picture, the value calculated in the following manner, not the value of the CpbRemovalDelay carried in the PTSEI appended to the BLA picture.
  • the picture decoding/display time determining unit 23 calculates the sum A of picture decoding intervals counted from the picture that immediately follows the most recent BPSEI-appended picture before the BLA picture up to the BLA picture. Then, the picture decoding/display time determining unit 23 sets the decoding delay CpbRemovalDelay of the BLA picture equal to A. When the decoding interval of each picture is equal, the picture decoding/display time determining unit 23 may set the decoding delay CpbRemovalDelay of the BLA picture equal to the number of pictures counted in units of fields from the picture immediately following the most recent BPSEI-appended picture before the BLA picture up to the BLA picture.
  • the picture decoding/display time determining unit 23 checks the NumRemovedTfds field of the PTSEI appended to the BLA picture. If the value of NumRemovedTfds is not zero, the picture decoding/display time determining unit 23 determines that the TFD picture immediately following the BLA picture has been discarded, and corrects the display delay CpbRemovalDelay of the BLA picture by subtracting the value of NumRemovedTfds from the value of the display delay CpbRemovalDelay of the BLA picture.
  • the picture decoding/display time determining unit 23 further performs the following processing on every picture that follows the BLA picture in decoding order until the next BPSEI-appended picture appears.
  • the picture decoding/display time determining unit 23 corrects the decoding delay CpbRemovalDelay by subtracting from the original value of CpbRemovalDelay the difference between the value of NumRemovedTfds carried in the PTSEI appended to the BLA picture and the value of NumRemovedTfds carried in the PTSEI appended to the designated picture (i.e., the sum of the decoding intervals of the discarded pictures that occurred after the designated picture).
  • the picture decoding/display time determining unit 23 corrects the display delay DpbOutputDelay by subtracting from the original value of DpbOutputDelay the value of NumRemovedTfds carried in the PTSEI appended to the designated picture.
  • the picture decoding/display time determining unit 23 corrects the decoding delay CpbRemovalDelay by subtracting from the original value of the decoding delay CpbRemovalDelay of that picture the value of NumRemovedTfds carried in the PTSEI appended to the BLA picture.
  • the picture decoding/display time determining unit 23 determines the decoding delay CpbRemovalDelay and display delay DpbOutputDelay of the picture by directly taking the values of CpbRemovalDelay and DpbOutputDelay carried in the PTSEI appended to that picture.
  • the picture decoding/display time determining unit 23 determines the decoding time of each picture based on the above decoding delay CpbRemovalDelay, and issues a decoding instruction to the picture decoding unit 24 at the determined decoding time. Further, the picture decoding/display time determining unit 23 determines the display time of each picture based on the above display delay DpbOutputDelay, and issues a display instruction to the frame memory 25 at the determined display time.
  • the picture decoding unit 24 Upon receiving the decoding instruction for the picture to be decoded, the picture decoding unit 24 decodes the picture by using a reference picture stored in the frame memory 25 . Then, the picture decoding unit 24 stores the decoded picture in the frame memory 25 . The picture decoding unit 24 performs the decoding by employing the same coding method as that employed by the picture encoding unit in the video encoding apparatus 1 .
  • the frame memory 25 stores the decoded picture. Further, the frame memory 25 outputs the decoded picture to the picture decoding unit 24 to present a reference picture to a picture to be decoded subsequently. The frame memory 25 also outputs the decoded picture to a display unit (not depicted) in accordance with the display instruction received from the picture decoding/display time determining unit 23 .
  • FIG. 9 is an operation flowchart illustrating the video decoding process performed by the video decoding apparatus according to the first embodiment.
  • the video decoding apparatus 2 decodes the entire video sequence in accordance with the operation flowchart of FIG. 9 .
  • the control unit 21 initializes a variable “flag” to 0 (step S 301 ).
  • the variable “flag” is a variable that indicates whether the picture is a non-BLA picture whose CpbRemovalDelay and DpbOutputDelay need correcting. If the flag is 1, CpbRemovalDelay and DpbOutputDelay need correcting, and if the flag is 0, neither CpbRemovalDelay nor DpbOutputDelay need correcting.
  • the header information analyzing unit 22 analyzes the header information of the picture to be decoded, and passes the parameters necessary for determining the decoding time and display time of the picture to the picture decoding/display time determining unit 23 (step S 302 ).
  • the picture decoding/display time determining unit 23 determines whether the variable “flag” is 1 or not (step S 303 ).
  • the picture decoding/display time determining unit 23 corrects the decoding delay CpbRemovalDelay of the picture to be decoded, which is, in this case, a non-BLA picture, by using the NumRemovedTfds of the picture to be decoded and the NumRemovedTfds of the most recent BLA picture (step S 304 ).
  • the picture decoding/display time determining unit 23 also corrects the display delay DpbOutputDelay of the picture to be decoded, by using the NumRemovedTfds of the picture to be decoded.
  • step S 304 or after it is determined in step S 303 that the variable “flag” is 0 (No in step S 303 ), the picture decoding/display time determining unit 23 proceeds to determine whether the picture to be decoded is a BPSEI-appended picture or not (step S 305 ).
  • the picture decoding/display time determining unit 23 determines whether the picture to be decoded is a BLA picture or not (step S 306 ). If the picture to be decoded is not a BLA picture (No step S 306 ), the picture decoding/display time determining unit 23 resets the variable “flag” to 0 (step S 307 ).
  • the picture decoding/display time determining unit 23 corrects the decoding delay CpbRemovalDelay and display delay DpbOutputDelay of the picture, and sets the variable “flag” to 1 (step S 308 ).
  • the picture decoding/display time determining unit 23 determines the decoding delay CpbRemovalDelay of the BLA picture by taking the sum of picture decoding intervals counted from the picture that immediately follows the most recent BPSEI-appended picture up to the BLA picture. Further, the picture decoding/display time determining unit 23 determines the display delay DpbOutputDelay of the picture by subtracting the value of NumRemovedTfds from the original value of DpbOutputDelay.
  • step S 307 or S 308 or after it is determined in step S 305 that the picture to be decoded is not a BPSEI-appended picture (No in step S 305 ), the control unit 21 proceeds to determine whether there is any picture remaining to be decoded in the encoded video data (step S 309 ). If there is any picture remaining to be decoded (Yes in step S 309 ), the control unit 21 returns the process to step S 302 . Then, the picture to be decoded next in decoding order is taken as the target picture, and the process starting from step S 302 is repeated. On the other hand, if there is no longer any picture remaining to be decoded (No in step S 309 ), the control unit 21 terminates the video decoding process.
  • Each block in the first encoded video data stream 1001 to be spliced upstream of the splicing point represents one picture, and the characters carried in each block indicate the coding mode and the order of input to the video encoding apparatus, as in FIG. 2 .
  • the second encoded video data stream 1002 is spliced immediately following the last picture B11 of the first encoded video data stream.
  • each block represents one picture, and the characters carried in each block indicate the coding mode and the order of input to the video encoding apparatus.
  • Arrows depicted above the second encoded video data stream 1002 indicate the reference pictures to which pictures B4 to B7, respectively, refer when encoded by forward frame prediction.
  • arrows depicted below the second encoded video data stream 1002 indicate the reference pictures to which the pictures B4 to B7, respectively, refer when encoded by backward frame prediction.
  • the pictures B4, B2, B1, B3, and B5 are TFD pictures, as designated below the second encoded video data stream 1002 .
  • the pictures B6 and B7 are DLP pictures.
  • a block array 1003 presented below the second encoded video data stream 1002 indicates the values of the decoding delay CpbRemovalDelay and display delay DpbOutputDelay carried in PTSEI appended to each picture of the second encoded video data stream 1002 .
  • Each block in the upper row of the block array 1003 carries the value of the decoding delay CpbRemovalDelay for the corresponding picture in the second encoded video data stream 1002 located directly above that block.
  • each block in the lower row of the block array 1003 carries the value of the display delay DpbOutputDelay for the corresponding picture in the second encoded video data stream 1002 located directly above that block.
  • Spliced encoded video data 1004 generated by splicing the first and second encoded video data streams 1001 and 1002 is depicted below the block array 1003 .
  • the TFD pictures B4, B2, B1, B3, and B5 in the second encoded video data stream 1002 are discarded and are therefore not contained in the spliced encoded video data 1004 .
  • NumRemovedTfds 1005 of the spliced encoded video data 1004 is depicted below the block array 1004 .
  • the NumRemovedTfds field of the BLA picture I8 carries the sum of the decoding intervals of the discarded TFD pictures (B4, B2, B1, B3, and B5) that follow I8 in decoding order, i.e., in the illustrated example, the value “10” which represents the number of pictures in units of fields that have been discarded after I8.
  • the NumRemovedTfds field of the DLP picture B6 carries the sum of the decoding intervals of the discarded TFD pictures (B5) that follow B6 in decoding order, i.e., in the illustrated example, the value “2” which represents the number of pictures in units of fields that have been discarded after B6.
  • the value of NumRemovedTfds remains 0 because there are no discarded TFD pictures that follow in decoding order.
  • a block array 1006 presented below the NumRemovedTfds 1005 of the spliced encoded video data 1004 indicates the values of the decoding delay CpbRemovalDelay and display delay DpbOutputDelay of the spliced encoded video data 1004 that have been corrected based on the NumRemovedTfds values.
  • Each block in the upper row of the block array 1006 carries the corrected value of the decoding delay CpbRemovalDelay for the picture located directly above that block, and each block in the lower row of the block array 1006 carries the corrected value of the display delay DpbOutputDelay for the picture located directly above that block.
  • the corrected value of the display delay DpbOutputDelay is given as “10” by subtracting the NumRemovedTfds value “10” from the original value “20” of the display delay DpbOutputDelay.
  • the corrected value of the display delay DpbOutputDelay of the picture I8 can be expressed in terms of the difference between the decode time and the display time of the picture I8 taken by reference to the display time of the picture B9 for which the number of pictures to be reordered is the largest among the pictures that follow the picture I8.
  • the corrected value of the decoding delay CpbRemovalDelay is obtained by subtracting the NumRemovedTfds value of the picture I8 from the original value of the decoding delay CpbRemovalDelay.
  • the display delay DpbOutputDelay remains unchanged.
  • the video encoding apparatus need only store within the encoded video data the decoding delay and display delay correction parameters determined based on the number of pictures discarded at the time of splicing, and need not correct the decoding delay and display delay parameters determined at the time of encoding. Then, since the decoding delay and display delay of each picture can be corrected using the decoding delay and display delay correction parameters added at the time of splicing the encoded video data streams, the video decoding apparatus according to the present embodiment can decode and display each picture at correct timing.
  • the second embodiment differs from the first embodiment in the structure of the encoded video data.
  • the data structure 1100 of one picture contains six kinds of NAL units 1110 to 1115 .
  • the BPSEI 1113 and PTSEI 1114 are different from the BPSEI 413 and PTSEI 414 depicted in FIG. 4 .
  • the DELIM 1110 , SPS 1111 , PPS 1112 , SLICE 1115 , and NUH 1120 are identical to the DELIM 410 , SPS 411 , PPS 412 , SLICE 415 , and NUH 420 , respectively, depicted in FIG. 4 .
  • the BPSEI 1113 contains a NumEntries field which carries a number calculated by adding 1 to the variable m that represents the value obtained by adding 2 to the number of pictures not discarded at the time of splicing among the TFD and DLP pictures located between a BLA picture and the next CRA picture.
  • the BPSEI 1113 further contains as many AltCpbRemovalDelayOffset fields and AltDpbOutputDelayOffset fields as there are NumEntries.
  • the NumEntries field, the AltCpbRemovalDelayOffset field, and the AltDpbOutputDelayOffset field together constitute another example of the correction information used for the correction of the decoding delay and display delay.
  • the PTSEI 1140 unlike the PTSEI 440 , does not contain the NumRemovedTfds field.
  • the video decoding apparatus When the value of the NumEntries field is 0, the video decoding apparatus need not correct the values of CpbRemovalDelay and DpbOutputDelay for any of the BPSEI-appended picture and its subsequent pictures (preceding the next BPSEI-appended picture).
  • the video decoding apparatus corrects the decoding delay CpbRemovalDelay of the kth picture in decoding order as counted from the BPSEI-appended picture by subtracting the value of AltCpbRemovalDelayOffset [k] from the original value of the decoding delay CpbRemovalDelay.
  • the video decoding apparatus corrects the display delay DpbOutputDelay by subtracting the value of AltDpbOutputDelayOffset [k] from the original value of the display delay DpbOutputDelay.
  • the difference from the first embodiment lies in the type of SEI that carries the corrected values of the CpbRemovalDelay and DpbOutputDelay fields. Therefore, the video encoding apparatus of the second embodiment differs from the video encoding apparatus of the first embodiment in the operation of the splicing point identification information processing unit 14 . The following therefore describes the operation of the splicing point identification information processing unit 14 .
  • the configuration of the video decoding apparatus according to the second embodiment is essentially the same as that of the video decoding apparatus according to the first embodiment.
  • the video decoding apparatus of the second embodiment differs from that of the first embodiment in the operation of the picture decoding/display time determining unit 23 .
  • the following therefore describes the operation of the picture decoding/display time determining unit 23 .
  • the picture decoding/display time determining unit 23 correct the values of the decoding delay CpbRemovalDelay and display delay DpbOutputDelay in the PTSEI of the picture in the following manner.
  • k is equal to or larger than NumEntries
  • the picture decoding/display time determining unit 23 corrects the value of the decoding delay CpbRemovalDelay by subtracting the value of AltCpbRemovalDelayOffset [NumEntries ⁇ 1] from the original value of the decoding delay CpbRemovalDelay of the kth picture.
  • the picture decoding/display time determining unit 23 corrects the value of CpbRemovalDelay for the kth picture by subtracting the value of AltCpbRemovalDelayOffset [k] from the original value of the decoding delay CpbRemovalDelay, and corrects the value of DpbOutputDelay by subtracting the value of AltDpbOutputDelayOffset from the original value of the display delay DpbOutputDelay.
  • FIG. 12 is a diagram illustrating the configuration of a computer that operates as the video encoding apparatus or video decoding apparatus by executing a computer program for implementing the functions of the various units in the video encoding apparatus or video decoding apparatus according to any one of the above embodiments or their modified examples.
  • the computer 100 includes a user interface unit 101 , a communication interface unit 102 , a storage unit 103 , a storage media access device 104 , and a processor 105 .
  • the processor 105 is connected to the user interface unit 101 , communication interface unit 102 , storage unit 103 , and storage media access device 104 , for example, via a bus.
  • the user interface unit 101 includes, for example, an input device such as a keyboard and a mouse, and a display device such as a liquid crystal display.
  • the user interface unit 101 may include a device, such as a touch panel display, into which an input device and a display device are integrated.
  • the user interface unit 101 generates, for example, in response to a user operation, an operation signal for selecting video data to be encoded, encoded video data to be edited, or encoded video data to be decoded, and supplies the operation signal to the processor 105 .
  • the interface unit 101 may also display the decoded video data received from the processor 105 .
  • the communication interface unit 102 may include a communication interface for connecting the computer 100 to a video data generating apparatus, for example, a video camera, and a control circuit for the communication interface.
  • a communication interface may be, for example, a Universal Serial Bus (USB) interface.
  • the communication interface unit 102 may include a communication interface for connecting to a communication network conforming to a communication standard such as the Ethernet (registered trademark), and a control circuit for the communication interface.
  • a communication standard such as the Ethernet (registered trademark)
  • the communication interface 102 receives video data to be encoded, encoded video data to be edited, or encoded video data to be decoded, from another apparatus connected to the communication network, and passes the received data to the processor 105 . Further, the communication interface 102 may receive encoded video data, spliced encoded video data, or decoded video data from the processor 105 and may transmit the data to another apparatus over the communication network.
  • the storage unit 103 includes, for example, a readable/writable semiconductor memory and a read-only semiconductor memory.
  • the storage unit 103 stores a computer program for video encoding or video decoding to be executed on the processor 105 , and also stores the data generated as a result of or during the execution of the program.
  • the storage media access device 104 is a device that accesses a storage medium 106 such as a magnetic disk, a semiconductor memory card, or an optical storage medium.
  • the storage media access device 104 accesses the storage medium 106 to read out, for example, the computer program for video encoding or video decoding to be executed on the processor 105 , and passes the readout computer program to the processor 105 .
  • the processor 105 generates encoded video data by executing the video encoding computer program according to any one of the above embodiments or their modified examples.
  • the processor 105 then stores the encoded video data thus generated in the storage unit 103 , or transmits the generated data to another apparatus via the communication interface unit 102 .
  • the processor 105 generates spliced encoded video data by splicing together two encoded video data streams.
  • the processor 105 then stores the spliced encoded video data generated in the storage unit 103 , or transmits the generated data to another apparatus via the communication interface unit 102 .
  • the processor 105 decodes the encoded video data by executing the video decoding computer program according to any one of the above embodiments or their modified examples.
  • the processor 105 then stores the decoded video data in the storage unit 103 , presents the data to the user interface unit 101 for display, or transmits the data to another apparatus via the communication interface unit 102 .
  • a computer program which is executed on a computer to implement the functions of the various units constituting the video encoding apparatus or video decoding apparatus according to each of the above embodiments or their modified examples may be distributed in the form stored in a semiconductor memory or in the form recorded on a recording medium such as an optical recording medium.
  • the term “recording medium” used here does not a carrier wave.
  • the video encoding apparatus and video decoding apparatus are used in various applications.
  • the video encoding apparatus and video decoding apparatus are incorporated in a video camera, a video transmitting apparatus, a video receiving apparatus, a video telephone system, a computer, or a mobile telephone.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Business, Economics & Management (AREA)
  • Marketing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)
US14/040,847 2012-10-01 2013-09-30 Video encoding apparatus, video decoding apparatus, video encoding method, and video decoding method Abandoned US20140092966A1 (en)

Priority Applications (5)

Application Number Priority Date Filing Date Title
US15/015,247 US20160156922A1 (en) 2012-10-01 2016-02-04 Video encoding apparatus, video decoding apparatus, video encoding method, and video decoding method
US15/015,233 US10582208B2 (en) 2012-10-01 2016-02-04 Video encoding apparatus, video decoding apparatus, video encoding method, and video decoding method
US15/015,221 US20160156920A1 (en) 2012-10-01 2016-02-04 Video encoding apparatus, video decoding apparatus, video encoding method, and video decoding method
US15/015,204 US20160219296A1 (en) 2012-10-01 2016-02-04 Video encoding apparatus, video decoding apparatus, video encoding method, and video decoding method
US16/785,706 US20200177907A1 (en) 2012-10-01 2020-02-10 Video encoding apparatus, video decoding apparatus, video encoding method, and video decoding method

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2012219663A JP6094126B2 (ja) 2012-10-01 2012-10-01 動画像復号装置
JP2012-219663 2012-10-01

Related Child Applications (4)

Application Number Title Priority Date Filing Date
US15/015,233 Division US10582208B2 (en) 2012-10-01 2016-02-04 Video encoding apparatus, video decoding apparatus, video encoding method, and video decoding method
US15/015,204 Division US20160219296A1 (en) 2012-10-01 2016-02-04 Video encoding apparatus, video decoding apparatus, video encoding method, and video decoding method
US15/015,247 Division US20160156922A1 (en) 2012-10-01 2016-02-04 Video encoding apparatus, video decoding apparatus, video encoding method, and video decoding method
US15/015,221 Division US20160156920A1 (en) 2012-10-01 2016-02-04 Video encoding apparatus, video decoding apparatus, video encoding method, and video decoding method

Publications (1)

Publication Number Publication Date
US20140092966A1 true US20140092966A1 (en) 2014-04-03

Family

ID=49212655

Family Applications (6)

Application Number Title Priority Date Filing Date
US14/040,847 Abandoned US20140092966A1 (en) 2012-10-01 2013-09-30 Video encoding apparatus, video decoding apparatus, video encoding method, and video decoding method
US15/015,204 Abandoned US20160219296A1 (en) 2012-10-01 2016-02-04 Video encoding apparatus, video decoding apparatus, video encoding method, and video decoding method
US15/015,233 Active US10582208B2 (en) 2012-10-01 2016-02-04 Video encoding apparatus, video decoding apparatus, video encoding method, and video decoding method
US15/015,247 Abandoned US20160156922A1 (en) 2012-10-01 2016-02-04 Video encoding apparatus, video decoding apparatus, video encoding method, and video decoding method
US15/015,221 Abandoned US20160156920A1 (en) 2012-10-01 2016-02-04 Video encoding apparatus, video decoding apparatus, video encoding method, and video decoding method
US16/785,706 Abandoned US20200177907A1 (en) 2012-10-01 2020-02-10 Video encoding apparatus, video decoding apparatus, video encoding method, and video decoding method

Family Applications After (5)

Application Number Title Priority Date Filing Date
US15/015,204 Abandoned US20160219296A1 (en) 2012-10-01 2016-02-04 Video encoding apparatus, video decoding apparatus, video encoding method, and video decoding method
US15/015,233 Active US10582208B2 (en) 2012-10-01 2016-02-04 Video encoding apparatus, video decoding apparatus, video encoding method, and video decoding method
US15/015,247 Abandoned US20160156922A1 (en) 2012-10-01 2016-02-04 Video encoding apparatus, video decoding apparatus, video encoding method, and video decoding method
US15/015,221 Abandoned US20160156920A1 (en) 2012-10-01 2016-02-04 Video encoding apparatus, video decoding apparatus, video encoding method, and video decoding method
US16/785,706 Abandoned US20200177907A1 (en) 2012-10-01 2020-02-10 Video encoding apparatus, video decoding apparatus, video encoding method, and video decoding method

Country Status (9)

Country Link
US (6) US20140092966A1 (pt)
EP (1) EP2713620B1 (pt)
JP (1) JP6094126B2 (pt)
KR (5) KR101612162B1 (pt)
CN (5) CN106878708B (pt)
BR (4) BR102013025344B1 (pt)
CA (5) CA2828843C (pt)
MX (5) MX344755B (pt)
TW (1) TWI488471B (pt)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150269967A1 (en) * 2014-03-24 2015-09-24 Microsoft Corporation Fast and smart video trimming at frame accuracy on generic platform
US10897623B2 (en) * 2018-09-05 2021-01-19 Fujitsu Limited Video coding device and video coding method

Families Citing this family (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113038188B (zh) * 2015-03-31 2023-06-06 松下电器(美国)知识产权公司 发送方法、接收方法、发送装置以及接收装置
CN106331835B (zh) * 2015-06-26 2019-06-07 成都鼎桥通信技术有限公司 一种动态调整数据接收缓存的方法及视频解码设备
KR102477964B1 (ko) * 2015-10-12 2022-12-16 삼성전자주식회사 미디어 전송 시스템에서 비디오 비트스트림의 임의 접근 및 재생을 가능하게 하는 기법
CN110636306B (zh) * 2018-06-22 2021-07-20 杭州海康威视数字技术股份有限公司 一种图像显示方法、解码设备及电视墙系统
CN111479111B (zh) * 2019-01-23 2024-04-09 华为技术有限公司 图像显示顺序的确定方法、装置和视频编解码设备
CN114615506B (zh) * 2019-06-13 2023-07-04 北京达佳互联信息技术有限公司 视频解码方法、计算设备、存储介质
CN114302173B (zh) 2021-12-31 2022-07-15 广东工业大学 一种平面编码靶标的二维图像拼接系统及其拼接方法

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6154496A (en) * 1997-11-25 2000-11-28 Philips Electronics N.A. Corp. Video buffer for seamless splicing of MPEG streams
US6380991B1 (en) * 1998-11-19 2002-04-30 Tektronix, Inc. Method for splicing MPEG-2 transport streams
US20060107187A1 (en) * 2004-11-16 2006-05-18 Nokia Corporation Buffering packets of a media stream
US20090328096A1 (en) * 2008-06-30 2009-12-31 Rgb Networks, Inc. Preconditioning ad content for digital program insertion
US20100246662A1 (en) * 2009-03-25 2010-09-30 Kabushiki Kaisha Toshiba Image encoding method and image decoding method
US20110292995A1 (en) * 2009-02-27 2011-12-01 Fujitsu Limited Moving image encoding apparatus, moving image encoding method, and moving image encoding computer program
US20120016965A1 (en) * 2010-07-13 2012-01-19 Qualcomm Incorporated Video switching for streaming video data
US20130170561A1 (en) * 2011-07-05 2013-07-04 Nokia Corporation Method and apparatus for video coding and decoding
US20130272430A1 (en) * 2012-04-16 2013-10-17 Microsoft Corporation Constraints and unit types to simplify video random access
US20130294499A1 (en) * 2012-04-27 2013-11-07 Qualcomm Incorporated Parameter set updates in video coding

Family Cites Families (24)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5917830A (en) * 1996-10-18 1999-06-29 General Instrument Corporation Splicing compressed packetized digital video streams
JPH11261958A (ja) * 1998-03-09 1999-09-24 Sony Corp 映像編集装置及び映像編集方法
US7031348B1 (en) * 1998-04-04 2006-04-18 Optibase, Ltd. Apparatus and method of splicing digital video streams
US20060093045A1 (en) 1999-06-29 2006-05-04 Roger Anderson Method and apparatus for splicing
EP1310092A1 (en) 2000-07-25 2003-05-14 Agilevision, L.L.C. Splicing compressed, local video segments into fixed time slots in a network feed
FI114527B (fi) * 2002-01-23 2004-10-29 Nokia Corp Kuvakehysten ryhmittely videokoodauksessa
JP4390112B2 (ja) * 2002-09-05 2009-12-24 エージェンシー フォー サイエンス,テクノロジー アンド リサーチ ビデオシーケンスのレートを制御する方法及び装置並びにビデオ符号化装置
JP4276420B2 (ja) 2002-11-05 2009-06-10 三菱電機株式会社 動き補償方法
US7266147B2 (en) * 2003-03-31 2007-09-04 Sharp Laboratories Of America, Inc. Hypothetical reference decoder
KR101148701B1 (ko) * 2004-08-31 2012-05-23 파나소닉 주식회사 동화상 부호화 방법 및 장치
TWI285055B (en) * 2004-09-30 2007-08-01 Nippon Telegraph & Telephone Method for progressive lossless video coding, method for progressive lossless video decoding, progressive lossless video coding apparatus, progressive lossless video decoding apparatus, and recording media of progressive lossless video coding program...
US20060215755A1 (en) * 2005-03-24 2006-09-28 Mediatek Incorporation Video encoding methods and systems for battery-powered apparatus
US7724305B2 (en) * 2006-03-21 2010-05-25 Mediatek Inc. Video data conversion method and system for multiple receivers
JP2008109637A (ja) * 2006-09-25 2008-05-08 Toshiba Corp 動画像符号化装置及びその方法
TWI339073B (en) * 2006-11-13 2011-03-11 Univ Nat Chiao Tung Video coding method using image data skipping
AR064274A1 (es) * 2006-12-14 2009-03-25 Panasonic Corp Metodo de codificacion de imagenes en movimiento, dispositivo de codificacion de imagenes en movimiento, metodo de grabacion de imagenes en movimiento, medio de grabacion, metodo de reproduccion de imagenes en movimiento, dispositivo de reproduccion de imagenes en movimiento, y sistema de reproducci
EP2123044A1 (en) * 2007-01-08 2009-11-25 Thomson Licensing Methods and apparatus for video stream splicing
WO2009052262A2 (en) * 2007-10-16 2009-04-23 Cisco Technology, Inc. Conveyance of concatenation properties and picture orderness in a video stream
US8781003B2 (en) * 2008-07-17 2014-07-15 Cisco Technology, Inc. Splicing of encrypted video/audio content
US8170401B2 (en) * 2008-11-25 2012-05-01 Cisco Technology, Inc. Optimizing ad insertion by removing low information frames
JP5227875B2 (ja) * 2009-04-06 2013-07-03 株式会社日立製作所 動画像符号化装置
CN101742321B (zh) * 2010-01-12 2011-07-27 浙江大学 基于图层分解的视频编、解码方法及装置
US9979958B2 (en) * 2012-04-20 2018-05-22 Qualcomm Incorporated Decoded picture buffer processing for random access point pictures in video sequences
JP5891975B2 (ja) * 2012-07-02 2016-03-23 富士通株式会社 動画像符号化装置、動画像復号装置、動画像符号化方法および動画像復号方法

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6154496A (en) * 1997-11-25 2000-11-28 Philips Electronics N.A. Corp. Video buffer for seamless splicing of MPEG streams
US6380991B1 (en) * 1998-11-19 2002-04-30 Tektronix, Inc. Method for splicing MPEG-2 transport streams
US20060107187A1 (en) * 2004-11-16 2006-05-18 Nokia Corporation Buffering packets of a media stream
US20090328096A1 (en) * 2008-06-30 2009-12-31 Rgb Networks, Inc. Preconditioning ad content for digital program insertion
US20110292995A1 (en) * 2009-02-27 2011-12-01 Fujitsu Limited Moving image encoding apparatus, moving image encoding method, and moving image encoding computer program
US20100246662A1 (en) * 2009-03-25 2010-09-30 Kabushiki Kaisha Toshiba Image encoding method and image decoding method
US20120016965A1 (en) * 2010-07-13 2012-01-19 Qualcomm Incorporated Video switching for streaming video data
US20130170561A1 (en) * 2011-07-05 2013-07-04 Nokia Corporation Method and apparatus for video coding and decoding
US20130272430A1 (en) * 2012-04-16 2013-10-17 Microsoft Corporation Constraints and unit types to simplify video random access
US20130294499A1 (en) * 2012-04-27 2013-11-07 Qualcomm Incorporated Parameter set updates in video coding

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
Kazui K., J. Koyama, S. Shimada, and A. Nakagawa, "AHG9: New high-level syntax for simple HEVC stream editing", Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG 16 WP3 and ISO/IEC JTC 1/SC 29/WG11, 10th Meeting: Stockholm, SE, 11-20 July 2012 *
The Society of Motion Picture and Television Engineers (SMPTE Standard for Television - Splice Points for MPEG-2 Transport Streams, SMPTE 312M-2001, Revision of SMPTE 312M-1999, 2001 *

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150269967A1 (en) * 2014-03-24 2015-09-24 Microsoft Corporation Fast and smart video trimming at frame accuracy on generic platform
US9911460B2 (en) * 2014-03-24 2018-03-06 Microsoft Technology Licensing, Llc Fast and smart video trimming at frame accuracy on generic platform
US10897623B2 (en) * 2018-09-05 2021-01-19 Fujitsu Limited Video coding device and video coding method

Also Published As

Publication number Publication date
TWI488471B (zh) 2015-06-11
CN106851266B (zh) 2019-06-28
KR101612162B1 (ko) 2016-04-12
CA2910306A1 (en) 2014-04-01
TW201415900A (zh) 2014-04-16
EP2713620A2 (en) 2014-04-02
MX344755B (es) 2017-01-05
KR101612235B1 (ko) 2016-04-12
JP2014072854A (ja) 2014-04-21
MX344754B (es) 2017-01-05
US20160156922A1 (en) 2016-06-02
CA2910432A1 (en) 2014-04-01
CA2910488C (en) 2018-05-01
US20160156920A1 (en) 2016-06-02
CN106878708A (zh) 2017-06-20
KR20150124438A (ko) 2015-11-05
BR122020018231B1 (pt) 2023-01-10
CA2828843A1 (en) 2014-04-01
MX344753B (es) 2017-01-05
CA2910306C (en) 2018-01-16
BR122020018234B1 (pt) 2023-01-17
KR101612236B1 (ko) 2016-04-26
KR101612237B1 (ko) 2016-04-12
CN106851266A (zh) 2017-06-13
CN107105277B (zh) 2020-03-06
MX344446B (es) 2016-12-15
CA2910309A1 (en) 2014-04-01
US20160219296A1 (en) 2016-07-28
US20160156921A1 (en) 2016-06-02
CN107105277A (zh) 2017-08-29
CN103716621B (zh) 2017-11-21
MX344749B (es) 2017-01-05
KR20150124439A (ko) 2015-11-05
KR20140043681A (ko) 2014-04-10
US20200177907A1 (en) 2020-06-04
US10582208B2 (en) 2020-03-03
MX2013011346A (es) 2014-04-28
CN106878708B (zh) 2019-10-18
CN103716621A (zh) 2014-04-09
CA2828843C (en) 2017-10-03
EP2713620B1 (en) 2019-07-03
BR102013025344B1 (pt) 2023-01-10
CA2910488A1 (en) 2014-04-01
BR122020018232B1 (pt) 2023-01-10
KR101612234B1 (ko) 2016-04-26
CN107071402A (zh) 2017-08-18
BR102013025344A2 (pt) 2014-10-14
EP2713620A3 (en) 2014-09-03
KR20150126337A (ko) 2015-11-11
JP6094126B2 (ja) 2017-03-15
KR20150124440A (ko) 2015-11-05

Similar Documents

Publication Publication Date Title
US20200177907A1 (en) Video encoding apparatus, video decoding apparatus, video encoding method, and video decoding method
US10070144B2 (en) Video encoding apparatus, video decoding apparatus, video encoding method, and video decoding method
US20140003519A1 (en) Video encoding apparatus, video decoding apparatus, video encoding method, and video decoding method

Legal Events

Date Code Title Description
AS Assignment

Owner name: FUJITSU LIMITED, JAPAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:KAZUI, KIMIHIKO;KOYAMA, JUNPEI;SHIMADA, SATOSHI;SIGNING DATES FROM 20130829 TO 20130906;REEL/FRAME:031471/0681

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION