WO2013061839A1 - 映像信号の符号化システム及び符号化方法 - Google Patents
映像信号の符号化システム及び符号化方法 Download PDFInfo
- Publication number
- WO2013061839A1 WO2013061839A1 PCT/JP2012/076813 JP2012076813W WO2013061839A1 WO 2013061839 A1 WO2013061839 A1 WO 2013061839A1 JP 2012076813 W JP2012076813 W JP 2012076813W WO 2013061839 A1 WO2013061839 A1 WO 2013061839A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- encoding
- stream
- video
- frame
- prediction
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Ceased
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L65/00—Network arrangements, protocols or services for supporting real-time applications in data packet communication
- H04L65/60—Network streaming of media packets
- H04L65/70—Media network packetisation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/103—Selection of coding mode or of prediction mode
- H04N19/105—Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/103—Selection of coding mode or of prediction mode
- H04N19/11—Selection of coding mode or of prediction mode among a plurality of spatial predictive coding modes
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/132—Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/17—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
- H04N19/176—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/18—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a set of transform coefficients
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/503—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
- H04N19/51—Motion estimation or motion compensation
- H04N19/513—Processing of motion vectors
- H04N19/517—Processing of motion vectors by encoding
- H04N19/52—Processing of motion vectors by encoding by predictive encoding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/503—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
- H04N19/51—Motion estimation or motion compensation
- H04N19/55—Motion estimation with spatial constraints, e.g. at image or region borders
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/593—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial prediction techniques
Definitions
- the present invention relates to a video signal encoding system and encoding method.
- the present invention relates to an encoding technique suitable for composing a single combined bit stream by arbitrarily connecting each MB line in a plurality of tile streams in units of each MB line.
- Non-Patent Document 1 proposes a system that divides and encodes videos acquired from a plurality of video cameras and omnidirectional cameras on a tile, and decodes and displays only tile videos at the viewpoint position requested by the user. ing.
- Non-Patent Document 2 proposes a system that realizes access to high-resolution panoramic images obtained from a plurality of video cameras based on Multi-View Coding, which is an extension standard of H.264 / AVC. .
- the input video is divided and encoded, and a plurality of encoded streams are transmitted according to the viewing area requested by the user (client terminal).
- this encoded stream can be decoded and panoramic video can be displayed.
- the client terminal may be simply referred to as a client.
- Non-Patent Documents 1 and 2 require simultaneous decoding and synchronous display of a plurality of streams at the client.
- Non-Patent Document 1 does not mention the transmission method
- Non-Patent Document 2 also requires multiple session control for simultaneous acquisition of multiple streams. Since these increase the complexity of processing in the client, it is considered that it is difficult to use the multi-vision service particularly in an environment where the calculation resources such as a smartphone are limited.
- non-patent document 3 and patent document 1 below a system (non-patent document 3 and patent document 1 below) is proposed in which a single stream is generated by combining a plurality of streams on the server side without transmitting a plurality of streams, and then transmitting this single stream. ing.
- a plurality of streams before combining are referred to as tile streams, and a single stream after combining is referred to as a combined stream.
- Non-Patent Document 3 and Patent Document 1 only the combined stream acquired from the distribution server is decoded and displayed by the client. For this reason, with this technique, complicated processing such as simultaneous decoding of a plurality of streams and synchronous display of decoded video signals can be avoided on the client side. As a result, the client system can simultaneously reproduce a plurality of tile images using a conventional image reproduction system.
- intra prediction encoding As intra prediction encoding, “4 ⁇ 4 intra prediction encoding that refers to adjacent pixels in 4 ⁇ 4 pixel block units” and “16 ⁇ 16 intra frame that refers to adjacent pixels in 16 ⁇ 16 pixel block units”.
- One of “predictive coding” can be selected. For example, in “4 ⁇ 4 intra prediction encoding”, there is a mode for referring to an adjacent 4 ⁇ 4 pixel block for encoding the 4 ⁇ 4 pixel block.
- tile streams are encoded using such a mode
- image quality due to pixel reference information mismatch Deterioration will occur.
- inconsistency can also occur in other scenes in encoding (for example, variable length encoding for the number of non-zero coefficients after DCT).
- Non-Patent Document 3 proposes a method of correcting the prediction difference information in order to avoid this problem. Specifically, a part of the MB in which the mismatch occurs is decoded to the pixel region, and the pixel signal is corrected (variable length decoding of the MB, coefficient inverse quantization, inverse DCT, residual by re-prediction from adjacent pixel values) Signal regeneration, DCT, quantization) and correction of prediction information from adjacent MB.
- Non-Patent Document 4 is a technique related to Video-in-Video, in which one different video is superimposed on one video screen.
- a method for saving various information related to encoding mode control and encoding in a separate file Is adopted.
- a recalculation process and a re-encoding process of motion vectors and non-zero coefficients are assumed in the superimposition process, there is a problem that the process at the server increases.
- the present invention has been made in view of the above situation.
- One of the objects of the present invention is to provide a technique capable of generating a combined stream while suppressing a load on a server by devising an encoding method of a video tile stream.
- Another object of the present invention is to provide a technique for composing a single bit stream by arbitrarily connecting MB lines in a video tile stream.
- the video signal receiving unit receives a video signal to be encoded
- the encoding processing unit is configured to generate a video tile stream by encoding the video signal using appropriate prediction reference information
- the encoding processing unit is a prediction reference information restriction method so that an error caused by a mismatch in signal prediction relationship does not occur even if each MB line in the video tile stream is arbitrarily connected in the encoding. Or, it is configured to use the prediction reference information fixing method,
- the encoding system, wherein the stream output unit is configured to output the video tile stream obtained by encoding in the encoding processing unit.
- the prediction reference information restriction method is a prediction method in which encoding information is restricted so as not to depend on a combination of encoding information held by mutually adjacent MBs between MB lines in different video tile streams.
- the encoding system according to item 1, wherein the prediction reference information restriction method includes the following processes: (1) A process of encoding a frame constituting the video signal in one of two types of encoding modes of intra-frame prediction encoding and inter-frame prediction encoding; (2) In a plurality of MBs in a frame to be subjected to intraframe prediction encoding, encoding is performed using a prediction mode that refers to pixel values that do not depend on the contents of MBs adjacent to each other between MB lines in different video tile streams. Process.
- the encoding system according to item 1, wherein the prediction reference information fixing method includes the following processes: (1) At least some of the luminance coefficient sequences and color differences in the MB in at least some of the MBs that constitute the video tile stream and are located in the peripheral portion of the frame of the video tile stream. A process of encoding the number of non-zero coefficients in the coefficient sequence as a preset fixed value; (2) In the case of MB referring to the number of non-zero coefficients of MB to be adjacent to the peripheral part of the frame of the video tile stream, there is an adjacent MB having the number of non-zero coefficients of the fixed value Assuming the encoding process.
- the encoding system according to item 1, wherein the prediction reference information fixing method includes the following processes: (1) A process of performing inter-frame predictive encoding by fixing a motion vector held by the MB to a predetermined motion vector in at least a part of MBs located in a peripheral portion of the frame of the video tile stream; (2) In the case of an MB that refers to an MB motion vector to be adjacent to a peripheral portion of a frame of the video tile stream, it is assumed that there is an adjacent MB having the predetermined motion vector, and inter-frame predictive coding Processing to do.
- the encoding processing unit includes an MB line code amount inserting unit, and the MB line code amount inserting unit encodes additional information for specifying a position of the MB line in the video tile stream.
- the encoding system according to any one of items 1 to 6, wherein the encoding system is configured to be generated at a time.
- Additional information for specifying the position of the MB line in the video tile stream can be used when MB lines are combined.
- the video tile stream reception unit is configured to receive the video tile stream,
- the combination processing unit is configured to generate a combined stream by performing the following processing: (1) Processing for detecting an end of the MB line in the video tile stream and acquiring a stream corresponding to the MB line; (2) A process of inserting a peripheral adjustment MB at the end of the MB line so as to be adjacent to the position of the peripheral edge of the frame in the combined stream in which the video tile streams are combined.
- a part of the peripheral adjustment MBs is encoded by the encoding system according to any one of items 1 to 7;
- the combined stream output unit is configured to output the combined stream generated by the combination processing unit.
- the detection of the end of the MB line includes a process of detecting the end of the MB line by reading the code amount of the MB line generated and embedded by the MB line code amount insertion unit described in Item 7.
- receiving a video signal to be encoded There, (1) receiving a video signal to be encoded; (2) generating a tile stream by encoding the video signal using appropriate prediction reference information; (3) outputting the video tile stream obtained by encoding,
- the prediction reference information does not cause an error caused by a mismatch in the prediction relationship of signals even when a stream composed of MB lines of frames in the video tile stream is arbitrarily connected.
- the encoding method is characterized in that a restriction method or the prediction reference information fixing method is used.
- (Item 11) A data structure generated by combining streams corresponding to MB lines constituting a tile stream encoded by the system according to any one of items 1 to 7, wherein the video tile stream is combined
- a peripheral adjustment MB is inserted at the end of the MB line so as to be adjacent to the position of the peripheral edge of the frame in the combined stream.
- 8. A data structure in which at least a part of the peripheral adjustment MB is encoded by the encoding system according to any one of items 1 to 7.
- the computer program and / or data structure described above can be stored in an appropriate recording medium, for example, an electrical, magnetic, or optical medium, and used by a computer.
- the recording medium may be a recording medium that can be used via a network, for example, on cloud computing.
- a single bit stream can be configured by arbitrarily connecting MB lines in a video tile stream.
- This system includes a video input unit 1, a server 2, a client terminal 3, and a network 4.
- the video input unit 1 includes a camera 11 or an external video distribution server 12.
- the camera 11 is preferably capable of acquiring a high-definition moving image.
- the external video distribution server 12 has already stored the encoded video bitstream, and the server 2 can acquire the video bitstream from the server 12 upon request. Since an existing camera or video distribution server can be used as the video input unit 1, further detailed description is omitted.
- the server 2 includes a tile stream encoding unit 21, a bit stream group storage unit 22, a combined stream generation unit 23, a client status management server 24, a combined stream transmission unit 25, and a video stream decoding unit 26. Yes.
- the video stream decoding unit 26 decodes the video bit stream sent from the external video distribution server 12 to generate a video signal, and sends this video signal to the tile stream encoding unit 21.
- the video signal means an uncompressed signal.
- the tile stream encoding unit 21 is a functional element corresponding to an example of the encoding system of the present invention.
- the tile stream encoding unit 21 receives a video signal to be encoded from the camera 11 or the video stream decoding unit 26.
- the tile stream encoding unit 21 according to the present embodiment can arbitrarily connect MB lines in a plurality of video tile streams in units of MB lines so that a single combined stream can be configured.
- the video tile stream is encoded.
- MB means a macroblock.
- the tile stream encoding unit 21 includes a video signal receiving unit 211, an encoding processing unit 212, and a video tile stream output unit 213.
- the video signal reception unit 211 receives a video signal to be encoded, sent from the camera of the video input unit 1 or the video stream decoding unit 26.
- the encoding processing unit 212 is configured to generate a video tile stream by encoding a video signal using appropriate prediction reference information. Further, the encoding processing unit 212 may prevent the prediction reference information restriction method or the prediction so that an error caused by a mismatch in the prediction relationship of signals does not occur even if each MB line in the video tile stream is arbitrarily connected in the encoding.
- the reference information fixing method is used. The prediction reference information restriction method or the prediction reference information fixing method will be described later.
- the encoding processing unit 212 is configured to use an MB line code amount insertion method in encoding.
- the MB line code amount insertion method is a bit amount of each MB line code string in all frames (in this specification, referred to as an MB line code amount) in order to execute the combination processing of each video tile stream at high speed. Is stored in the stream. However, the MB line code amount can be stored as a separate file or information without being stored in the tile stream.
- the prediction reference information restriction method in the present embodiment is that the encoding information is limited so that it does not depend on the combination of encoding information held by mutually adjacent MBs between MB lines in different video tile streams. This is the prediction method.
- the prediction reference information restriction method in the present embodiment includes the following processing: (1) A video signal is encoded for each frame in two types of encoding modes of intra-frame prediction encoding and inter-frame prediction encoding, and intra-frame prediction frames are inserted periodically or aperiodically; (2) In addition, in a plurality of MBs in an intra-frame prediction frame, encoding is performed using a prediction mode that refers to pixel values that do not depend on the contents of MBs adjacent to each other between MB lines in different video tile streams. .
- the prediction reference information fixing method in the present embodiment is a method using prediction information fixed to a preset value.
- the prediction reference information fixing method includes the following processes: (1) MB constituting a video tile stream and located at the peripheral part of the frame of the video tile stream, and at least a part of the luminance coefficient sequence and the color difference coefficient sequence in the MB in at least a part of the MB.
- the prediction reference information fixing method of the present embodiment includes the following processing: (1) A process of performing inter-frame predictive encoding by fixing a motion vector held by the MB to a predetermined motion vector in at least a part of MBs located in a peripheral portion of the frame of the video tile stream; (2) In the case of an MB that refers to an MB motion vector that should be adjacent to the peripheral portion of the frame of the video tile stream, inter-frame prediction encoding is performed on the assumption that an adjacent MB having a predetermined motion vector exists. processing. A specific example of the prediction reference information fixing method will be described later.
- the encoding processing unit 212 includes an orthogonal transform unit 2121a, a quantization unit 2121b, a coefficient adjustment unit 2122, a variable length coding unit 2123, an inverse quantization unit 2124a, an inverse orthogonal transform unit 2124b, a frame A memory 2125, a frame position and MB position management unit 2126, an encoding mode determination unit 2127, a motion search / compensation unit 2128, an intra prediction mode determination unit 2129, and an MB line code amount insertion unit 21291 are provided.
- the configurations and operations of the orthogonal transform unit 2121a, the quantization unit 2121b, the inverse quantization unit 2124a, the inverse orthogonal transform unit 2124b, and the frame memory 2125 may be the same as those of the conventional (for example, in H.264). Detailed explanation is omitted. The operations of the remaining functional elements will be described in detail in the description of the encoding processing method described later.
- the tile stream output unit 213 is configured to output a video tile stream obtained by encoding in the encoding processing unit 212 to the bit stream group storage unit 22.
- the bit stream group storage unit 22 is a part that stores the video tile stream generated by the tile stream encoding unit 21.
- the bit stream group accumulation unit 22 sends a predetermined MB bit stream sequence (video tile stream) that is a part of the video tile stream to the combined stream generation unit 23. It can be done.
- the combined stream generation unit 23 is an example of a combined system for combining the MB lines constituting the video tile stream encoded by the tile stream encoding unit 21. As illustrated in FIG. 4, the combined stream generation unit 23 includes a video tile stream reception unit 231, a combination processing unit 232, and a combined stream output unit 233.
- the video tile stream reception unit 231 is configured to receive a video tile stream from the bit stream group storage unit 22.
- the combination processing unit 232 includes a peripheral edge adjusting MB information insertion unit 2321, an MB line code amount reading unit 2322, an MB line extraction unit 2323, and a combined stream header information generation / insertion unit 2324.
- the peripheral edge adjustment MB information insertion unit 2321 performs the following processing to generate a combined stream: A process of inserting a peripheral adjustment MB at least at the end of the MB line so as to be adjacent to the position of the peripheral edge of the frame in the combined stream in which the video tile streams are combined.
- the peripheral adjustment MB is encoded by the encoding system described above.
- the MB line code amount reading unit 2322 is a part that reads the MB line code amount inserted by the MB line code amount insertion unit 21291 of the encoding processing unit 212. By reading the MB line code amount, the end of the MB line can be detected at high speed.
- the MB line extraction unit 2323 performs a process of extracting a code string from the tile stream by the bit amount of the MB line code string acquired by the MB line code quantity reading unit 2322. As a result, it is possible to avoid the variable length decoding process that is originally necessary to obtain the MB line code string bit amount. Of course, it is also possible to extract a code string without using the bit amount of the MB line code string by performing variable length decoding processing.
- the combined stream header information generation / insertion unit 2324 generates and inserts header information for the combined stream.
- the generation and insertion of the combined stream header may be the same as in the conventional process, and detailed description thereof is omitted.
- the combined stream output unit 233 is configured to output the combined stream generated by the combining processing unit 232.
- An example of the generated combined stream will be described later.
- the client status management server 24 receives a request sent from the client terminal 3, for example, information on a video area that the user wants to view (a specific example will be described later).
- the combined stream transmission unit 25 sends the combined stream generated by the combined stream generation unit 23 to the client terminal 3 via the network 4.
- the client terminal 3 is a terminal for a user to send a necessary command to the server 2 or to receive information sent from the server 2.
- the client terminal 3 is normally operated by a user, but may operate automatically without requiring a user operation.
- a mobile phone including a so-called smartphone
- a mobile computer including a so-called desktop computer, or the like can be used.
- the network 4 is for transmitting and receiving information between the server 2 and the client terminal 3.
- the network 4 is usually the Internet, but may be a network such as a LAN or WAN.
- the network only needs to be able to transmit and receive necessary information, and the protocol and physical medium used are not particularly limited.
- Steps SA-1 and 2 in FIG. 5 First, a video signal is captured from the video input unit 1 to the encoding processing unit 21 of the server 2. Details of the encoding process in the encoding processing unit 21 will be described with reference to FIG.
- the subsequent encoding processing is basically processing in MB units.
- an MB line is composed of MBs
- a tile stream frame is composed of MB lines
- a combined stream frame is composed of tile stream frames.
- the encoding processing unit 21 first determines an encoding mode for each MB.
- the coding mode is either intra-frame predictive coding (so-called intra coding) or inter-frame predictive coding (so-called inter coding).
- FIG. 1 An example of the encoding mode determination processing algorithm is shown in FIG.
- Step SC-1 in FIG. 7 it is determined whether or not the frame to which the processing target MB belongs is a refresh frame.
- This determination uses the number of processing frames obtained from the frame position and MB position management unit 2126. That is, the frame position and MB position management unit 2126 holds therein variables for counting the number of frames and the number of MBs for each process, and by referring to these variables, the number of frames to be processed and the number of MBs Can be obtained. Since the encoding processing unit 21 knows in advance which timing frame should be a refresh frame, the refresh frame is determined using the number of frames to be processed and predetermined timing information. Can do. In addition, refresh frames are usually inserted periodically (that is, at predetermined time intervals), but periodicity is not essential.
- Step SC-2 in FIG. 7 When the determination in step SC-1 is Yes (that is, in the case of a refresh frame), the MB is determined to be intra-coded.
- Step SC-3 in FIG. 7 When the determination in step SC-1 is No, it is determined that the MB should be subjected to interframe predictive coding.
- the encoding mode of each MB can be determined by the above algorithm.
- Step SB-2 in FIG. 6 a motion search / compensation technique by the motion search / compensation unit 2128 will be described with reference mainly to FIG.
- H.264 motion search / compensation is performed in units of pixel division called “partition” in MB.
- partition pixel sizes 16 ⁇ 16, 8 ⁇ 16, 16 ⁇ 8, 8 ⁇ 8, 4 ⁇ 8, 8 ⁇ 4, and 4 ⁇ 4 (see FIG. 9).
- the motion vector information held by the partition E shown in FIG. 10A is encoded as a difference value from the median value of the motion vectors held by the adjacent partitions A, B, and C.
- FIG. 10A shows a case where the sizes of the partitions are the same. However, as shown in FIG. 10B, adjacent partitions may have different sizes, and the encoding method in this case is the same as described above.
- Step SD-1 in FIG. 8 The flag is set to 0 as initialization processing. In subsequent processing, it is determined to which position of the frame the processing MB belongs based on the MB position obtained from the frame position and MB position management unit 2126.
- Steps SD-1-1 to SD-1-3 in FIG. 8 Next, it is determined whether the MB to which the partition to be processed belongs is at the left end of the frame.
- Steps SD-2 to 4 in FIG. 8 If the determination result in step SD-1-1 is No, it is determined whether the MB to which the partition to be processed belongs belongs to the right end of the frame.
- Steps SD-5 to 7 in Fig. 8 If the determination result in step SD-2 is No, it is determined whether the MB to which the partition to be processed belongs belongs to the lower end of the frame.
- Steps SD-8 to 9 in Fig. 8 When the flag attached to the MB is not 1 (that is, remains 0), the prediction reference information is limited to refer to the block information in the frame, and the pixel value of the previous frame obtained from the frame memory is set. Based on the motion search.
- This method is an example of a prediction reference information restriction method.
- “restricting prediction information to refer to block information in a frame” is realized by providing a restriction that a motion vector search range is within a frame.
- the limitation on the search range of the motion vector is pointed out in the literature (paragraphs 0074 to 0084 of JP 2011-55219 A).
- the motion vector search restriction range is not within the target MB line but within the frame.
- Step SD-10 in FIG. 8 If the determination result in step SD-8 is Yes, a fixed motion vector value is set. That is, a fixed value stored on the system side is taken out.
- the setting of the fixed motion vector value corresponds to an example of a prediction reference information fixing method. Specifically, the same part of the previous frame is referred to (when the motion vector is fixed as (0, 0)).
- Step SD-11 in FIG. 8 the motion search / compensation unit 2128 performs a motion compensation process using the searched motion vector value or the fixed motion vector value. Since the motion compensation process itself may be the same as the normal process in H.264, detailed description is omitted.
- a “motion vector value of a partition that may be referred to from an adjacent partition because it is at the right end, left end, or lower end of a tile stream frame” can be a fixed value. In this way, even when adjacent MBs are different between encoding and combining, correct decoding can be performed without being affected by the contents of adjacent MBs.
- Step SB-3 in FIG. 6 a processing algorithm in the intra prediction mode determination unit 2129 will be described with reference to FIG.
- the intra prediction mode determination unit 2129 sets the prediction mode shown in FIG. 12 according to the MB position. As shown in FIG. 12, in this mode, in the plurality of MBs at the left end in the video tile stream, a prediction mode that refers to the pixel value of the MB in contact with each MB is used, and in each of the plurality of MBs at the top, A prediction mode that refers to the pixel value of the MB that touches the left of the MB is used. In the rightmost MB, the “prediction mode other than the two modes for prediction from the upper right MB (see FIG. 12)” is used. Further, in the upper left MB in the same frame, no other MB is referred to.
- IPCM mode Use prediction mode
- Such a prediction mode restriction corresponds to an example of a prediction reference information restriction method. By setting in this way, it is possible to encode without referring to the MB value of the adjacent frame, so even if the prediction information referred to at the time of encoding and combining is different in each frame of the tile stream, correct decoding is performed. Is possible. In other words, in this figure The prediction mode restriction is performed.
- Step SE-2 in FIG. 11 According to the prediction mode set in step SE-1, a prediction reference pixel value is generated from one of “an adjacent pixel signal that has already been encoded and decoded” and “a pixel signal of a previous frame acquired from the frame memory”. The predicted reference pixel value is output. Since this process may be the same as the normal H.264 process, detailed description thereof is omitted.
- Steps SB-4 and SB-5 in FIG. 6 a prediction difference signal with respect to the input signal is generated using the result of the processing in steps SB-2 and SB-3. Further, orthogonal transformation and quantization are performed. Since the prediction difference signal generation, orthogonal transform, and quantization methods may be the same as those in normal H.264 processing, detailed description thereof is omitted.
- variable length coding is performed by the coefficient adjustment unit 2122 and the variable length coding unit 2123 (see FIG. 3).
- a coefficient adjustment process is performed before a normal variable length coding process. Therefore, in the following description, first, the coefficient adjustment processing in the coefficient adjustment unit 2122 will be described based on FIG. 13, and then, the variable length coding processing in the variable length coding unit 2123 based on FIG. Will be explained.
- Step SF-1 in FIG. 13 In order to determine the coefficient adjustment target block based on the MB position and the block position therein, the flag is set to 0.
- the MB position information is obtained from the frame position and MB position management unit 2126.
- the coefficient adjustment and variable length coding processing is performed in units of blocks, which are a set of transform coefficients in the MB.
- the point of processing in block units is the same as the normal processing in H.264, so a detailed description is omitted.
- Steps SF-2 to SF-4 in FIG. 13 When the MB to be processed is at the right end of the frame, it is determined whether or not the processing block is at the right end of the block (that is, the right end of the frame).
- Step SF-5 to SF-7 in FIG. 13 If the determination in step SF-5 is No, the process proceeds to step SF-5.
- the processing block is at the lower end of the block (that is, the lower end of the frame).
- Step SF-8 in FIG. 13 Thereafter, it is determined whether or not the flag in the MB is 1, and if No, the process proceeds to variable length coding processing.
- Step SF-9 to 10 in FIG. 13 If the determination result in step SF-8 is Yes, the number of non-zero coefficients in the block is compared with the number of non-zero coefficients set in advance (ie, held on the system side).
- the preset number of non-zero coefficients may be different between the luminance space (Y) and the color difference space (UV) in the YUV signal.
- the number of non-zero coefficients in the block is smaller than the preset number of non-zero coefficients, coefficients having values other than 0 are inserted from the high-frequency component side of the number of non-zero coefficients. Thereby, the number of non-zero coefficients can be adjusted to a predetermined value. Even if a coefficient having a value other than 0 is inserted on the high frequency component side, the influence on the image quality is small.
- Steps SF-11 to 12 in FIG. 13 When the number of non-zero coefficients in the block is larger than the preset number of non-zero coefficients, a coefficient having a value of 0 is inserted from the high frequency component side of the number of non-zero coefficients instead of a coefficient having a value other than 0. To do. Thereby, the number of non-zero coefficients can be adjusted to a predetermined value. Even if a coefficient having a value of 0 is inserted instead of a coefficient having a value other than 0 on the high frequency component side, the influence on the image quality is small. Note that using a fixed number of non-zero coefficients corresponds to an example of a prediction reference information fixing method.
- Step SG-1 in FIG. 14 a specific example of the variable length encoding process will be described with reference to FIG.
- the coefficient-adjusted MB is subjected to variable length coding according to a command from the frame position and MB position management unit 2126.
- initialization is performed by setting both the values of the flag 1 and the flag 2 used for determining the processing of the target MB to 0.
- Steps SG-1-1 to SG-1-3 in FIG. 14 If the MB to be processed is the right end of the frame and the partition to be processed in the MB is the right end of the MB, flag 1 is set to 1.
- Steps SG-2 to 6 in FIG. 14 If the MB to be processed is the left end of the frame and the block to be processed in the MB is the left end of the MB, flag 1 is set to 1. Further, flag 2 is set to 1 when the partition to be processed is at the left end.
- Steps SG-7 to 11 in FIG. 14 The flag 1 is set to 1 when the MB to be processed is the upper end of the frame and the block to be processed in the MB is the upper end of the MB. Further, when the partition to be processed is at the upper end, the flag 2 of the MB is set to 1.
- Step SG-7 when the determination result in Step SG-7 is No, normal variable-length encoding processing is performed, and thus illustration is omitted. If the determination of SG-10 is No, the process proceeds to step SG-12.
- Step SG-12 in FIG. 14 Next, encoding such as skip information and MB encoding mode is performed. Since this processing may be the same as the processing in the conventional H.264, detailed description is omitted.
- Steps SG-13 to 15 in FIG. 14 Next, when the flag 2 is not 1 and the MB is for inter-frame prediction encoding, the motion vector held by the partition to be processed is encoded by a normal method. If the MB is an intra-frame encoding, the process proceeds to SG-17.
- Step SG-16 in FIG. 14 If the determination result in Step SG-13 is Yes, an adjacent partition on the left, upper or upper right of the partition to be processed is assumed. Then, the motion vector of the partition to be processed is encoded on the assumption that the motion vector held by the partition is a predetermined fixed value.
- prediction reference information is generated from the left, upper, and upper right adjacent partitions, and a difference value from the prediction reference information is encoded, as shown in FIG. Therefore, in order to suppress prediction reference information mismatch at the time of combining, motion vectors are encoded on the assumption that these partitions exist.
- Step SG-17 in FIG. 14 Next, other MB information is encoded.
- Steps SG-18 to 19 in FIG. 14 Next, if the flag 1 of the MB to be processed is not 1, the variable length table is selected based on the average value of the number of non-zero coefficients in the block adjacent to the left or above. Since this process is the same as the normal H.264 process, detailed description is omitted.
- Step SG-20 in FIG. 14 If the flag 1 of the MB to be processed is 1, a left or upper adjacent block that does not exist is assumed. Then, the variable length table is selected on the assumption that the number of non-zero coefficients of these left or upper adjacent blocks is a fixed value. As a result, even if the tile stream frame differs between the time of encoding and the time of combination, the correct variable length table can be selected, and variable length decoding can be normally performed.
- variable length coding processing is performed.
- Other variable-length encoding processing may be the same as normal processing in H.264, and thus detailed description thereof is omitted. In this way, a variable-length encoded bitstream can be generated.
- Step SB-6-1 in FIG. 6 Next, the MB line code amount insertion procedure by the MB line code amount insertion unit 21291 will be described with further reference to FIG.
- Step SJ-1 in FIG. 20 First, the amount of MB processed by the variable length encoding unit 2123 (hereinafter referred to as CurrentMBBit) is acquired.
- Steps SJ-2 to 4 in FIG. 20 Next, if the MB position is at the left end of the frame, the bit amount (MBLinebit) of all MBs included in the MB line to be processed is set to 0. Otherwise, CurrentMBBit is added to the previous MBLinebit to make a new MBLinebit.
- Steps SJ-5 to 6 in FIG. 20 When the MB position to be processed reaches the right end of the frame, the MBLinebit obtained by adding up to that point is inserted into the header of the MB line code string to form a bit stream. As long as the right end is not reached, the process from step SJ-1 is repeated each time a new MB is acquired.
- Steps SB-7 to 9 in FIG. 6 The encoded bitstream is then inverse transformed for prediction and stored in the frame memory. Since these processes may be the same as the normal H.264 process, detailed description thereof is omitted. Next, the processing procedure returns to Step SB-1. Thereafter, when there is no more MB to be processed, the process is terminated.
- Step SA-3 in FIG. 5 the tile stream encoding unit 21 stores the bit stream generated by the above procedure in the bit stream group storage unit 22.
- each of the frames constituting the video is composed of tile stream frames (sometimes referred to as divided areas) Ap00 to Apmn.
- the entire video frame composed of the tile stream frames Ap00 to Apmn is referred to as a combined stream frame or an entire area Aw.
- the frames Ap00 to Apmn of each tile stream are composed of a set of MBs represented by MB00 to MBpq. Since these configurations may be the same as those described in Non-Patent Document 3 and Patent Document 1 by the present inventors, detailed description thereof will be omitted.
- the user designates an area desired to be viewed on the client terminal 3.
- the video area indicated by the frame Ap00 and the frame Ap01 of the tile stream is designated.
- the combination is performed in units of MB lines of the frames of the tile stream.
- the designation from the user is sent to the combined stream generation unit 23 via the client status management server 24.
- the method for designating the video area by the user may be the same as that of Non-Patent Document 3 and Patent Document 1 by the present inventors, and therefore will not be described in further detail.
- combining is performed in units of MB lines of the frames of the tile stream, but the viewing area may be specified in a narrower range.
- Step SA-5 in FIG. 5 the combined stream generation unit 23 combines the MB lines to generate a combined stream. This generation procedure will be described with reference mainly to FIGS.
- Step SH-1 in FIG. 16 The tile stream reception unit 231 of the combined stream generation unit 23 receives a tile stream to be transmitted to the user from the bit stream group storage unit 22 that stores the bit stream group encoded by the procedure described above (in this example, (Ap00 and Ap01 stream).
- the margin adjustment MB information insertion unit 2321 of the combination processing unit 232 inserts the margin adjustment MB information around the frame of the tile streams to be combined.
- FIG. A specific example is shown in FIG. In this example, it is assumed that frames of four tile streams are combined.
- MB information for edge adjustment is inserted into the three sides excluding the lower side.
- the MB information for edge adjustment is an MB for maintaining the consistency of encoding, and the data content and the encoding method are known in the combination processing unit 232. That is, as described above, the encoding of each tile stream frame employs an algorithm that can appropriately decode each tile stream frame even if the prediction information referred to at the time of encoding and when combined is different. .
- a peripheral adjustment MB is inserted around the frame of the tile stream so as to match the encoding condition.
- the pixel values of the MB for edge adjustment are all black. However, other pixel values can be used.
- FIG. 18 shows specific encoding conditions in the peripheral adjustment MB of the present embodiment.
- the encoding conditions for the peripheral adjustment MB are as follows.
- ⁇ Intra_16x16 MB mode and intra-frame coding (in case of refresh frame) so that the lower end block has a fixed number of non-zero coefficients; ⁇ : Interframe coding so that the bottom block has a fixed number of non-zero coefficients and a fixed motion vector (except for refresh frames); ⁇ : No encoding limit; ⁇ : Intra_16x16 MB mode, intra-frame coding (in case of refresh frame) so that the rightmost block has a fixed number of non-zero coefficients; ⁇ : Interframe coding so that the rightmost block has a fixed number of non-zero coefficients and a fixed motion vector (except for refresh frames); ⁇ : Intraframe coding (in the case of a refresh frame) assuming that the number of non-zero coefficients of the boundary block adjacent to the left side of the MB is a fixed value; ⁇ : Assuming that the number of non-zero coefficients of the boundary block adjacent to the left side of the MB is a fixed value and the
- Steps SH-3 to 4 in FIG. 16 Subsequently, the MB line code amount written in the header of the bit stream is read, and the MB line is extracted based on the MB line code amount.
- the end of the MB line can be detected without performing variable length decoding. This is important for implementation in order to reduce the load on the system.
- Step SH-5 in FIG. 16 the combined stream header information generation / insertion unit 2324 generates header information for the combined stream.
- the generated header information is inserted into the extracted MB line code string.
- a conceptual diagram of the combined stream with the header inserted is shown in FIG.
- MB for edge adjustment at the left end of the m-th line Code string MB line code string (m-th line) of tile stream Ap00 to be combined, MB line code string (m-th line) of tile stream Ap01 to be combined, MB code for edge adjustment at the right end of the m-th line It becomes the structure of a row.
- Step SH-6 in FIG. 16 Next, the generated combined stream is sent from the combined stream output unit 233 to the combined stream transmission unit 25.
- the encoding method of the present embodiment enables video to be configured so that a single combined stream can be configured by arbitrarily connecting each MB line in a plurality of video tile streams in units of each MB line.
- the tile stream is encoded. And this method (1) receiving a video signal to be encoded; (2) generating a tile stream by encoding the video signal using appropriate prediction reference information; (3) a step of outputting the video tile stream obtained by encoding.
- the prediction reference information fixing method is used.
- the combining method of the present embodiment is a combining method for combining the MB lines constituting the video tile stream encoded by the above-described encoding system of the present embodiment. And this method (1) detecting an end of an MB line in a video tile stream and obtaining a stream corresponding to the MB line; (2) inserting a peripheral adjustment MB at the end of the MB line so as to be adjacent to the position of the peripheral edge of the frame in the combined video stream in which the video tile streams are combined.
- peripheral adjustment MBs are encoded by the above-described encoding method, and the combined video stream output unit 25 is configured to output the combined stream generated by the combination processing unit 232. .
- the data structure shown in FIG. 19 is an example of a data structure generated by combining streams corresponding to MB lines constituting a tile stream encoded by the encoding system described above.
- the MB for edge adjustment is inserted at the end of the MB line so as to be adjacent to the position of the frame periphery in the combined stream in which the video tile streams are combined.
- at least a part of the peripheral adjustment MB is encoded by the encoding system described above.
- Step SA-6 in FIG. 5 The combined stream transmission unit 25 transmits the combined stream to the client terminal 3 via the network 4.
- the client terminal 3 can display an image by combining the combined streams. Since this decoding process may be the same as in the case of normal H.264, a detailed description is omitted.
- the stream combined by the method of this embodiment can be correctly decoded by a decoder implemented for normal H.264.
- the decoded image data can be presented to the user by displaying it on the client terminal 3. That is, according to the method of the present embodiment, it is possible to prevent deterioration in image quality displayed on the client terminal even when tile streams are arbitrarily combined. Moreover, in the method of the present embodiment, it is not necessary to decode up to the pixel level in order to correct the mismatch of the prediction reference information, so that the processing load on the server side can be reduced.
- the prediction mode is limited for the MB to be intra-frame encoded
- the prediction information referred to at the time of encoding and combining is the same in each frame of the tile stream. Decoding is possible.
- each of the above-described components only needs to exist as a functional block, and does not have to exist as independent hardware.
- a mounting method hardware or computer software may be used.
- one functional element in the present invention may be realized by a set of a plurality of functional elements, and a plurality of functional elements in the present invention may be realized by one functional element.
- each functional element constituting the present invention may exist in a discrete manner. If they exist in a discrete manner, necessary data can be transferred via a network, for example.
- each function in each part can exist discretely.
- each functional element in the present embodiment or a part thereof can be realized by using grid computing or cloud computing.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Computer Networks & Wireless Communication (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
Priority Applications (6)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| SG11201401713WA SG11201401713WA (en) | 2011-10-24 | 2012-10-17 | Encoding system and encoding method for video signals |
| IN3191DEN2014 IN2014DN03191A (enExample) | 2011-10-24 | 2012-10-17 | |
| US14/354,129 US20150127846A1 (en) | 2011-10-24 | 2012-10-17 | Encoding System and Encoding Method for Video Signals |
| CN201280057252.7A CN103947212A (zh) | 2011-10-24 | 2012-10-17 | 视频信号的编码系统以及编码方法 |
| EP12844021.1A EP2773113A4 (en) | 2011-10-24 | 2012-10-17 | SYSTEM AND METHOD FOR CODING VIDEO SIGNALS |
| KR20147010893A KR20140085462A (ko) | 2011-10-24 | 2012-10-17 | 영상 신호의 부호화 시스템 및 부호화 방법 |
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| JP2011-232863 | 2011-10-24 | ||
| JP2011232863A JP5685682B2 (ja) | 2011-10-24 | 2011-10-24 | 映像信号の符号化システム及び符号化方法 |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| WO2013061839A1 true WO2013061839A1 (ja) | 2013-05-02 |
Family
ID=48167672
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| PCT/JP2012/076813 Ceased WO2013061839A1 (ja) | 2011-10-24 | 2012-10-17 | 映像信号の符号化システム及び符号化方法 |
Country Status (8)
| Country | Link |
|---|---|
| US (1) | US20150127846A1 (enExample) |
| EP (1) | EP2773113A4 (enExample) |
| JP (1) | JP5685682B2 (enExample) |
| KR (1) | KR20140085462A (enExample) |
| CN (1) | CN103947212A (enExample) |
| IN (1) | IN2014DN03191A (enExample) |
| SG (1) | SG11201401713WA (enExample) |
| WO (1) | WO2013061839A1 (enExample) |
Cited By (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JP2016019240A (ja) * | 2014-07-10 | 2016-02-01 | 株式会社ドワンゴ | 端末装置、動画配信装置、プログラム |
| CN105612756A (zh) * | 2013-08-08 | 2016-05-25 | 国立大学法人电气通信大学 | 数据处理装置、数据处理方法、程序、记录介质以及数据处理系统 |
Families Citing this family (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| GB2557544B (en) * | 2015-09-11 | 2021-12-01 | Kt Corp | Method and device for processing video signal |
| CN105554513A (zh) * | 2015-12-10 | 2016-05-04 | Tcl集团股份有限公司 | 一种基于h.264的全景视频传输方法及系统 |
| WO2018074813A1 (ko) * | 2016-10-17 | 2018-04-26 | 에스케이텔레콤 주식회사 | 영상 부호화 또는 복호화하기 위한 장치 및 방법 |
Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JP2001285876A (ja) * | 2000-03-30 | 2001-10-12 | Sony Corp | 画像符号化装置とその方法、ビデオカメラ、画像記録装置、画像伝送装置 |
| WO2006006564A1 (ja) * | 2004-07-12 | 2006-01-19 | Sony Corporation | 符号化方法、符号化装置、復号方法、復号装置およびそれらのプログラム |
| JP2011024018A (ja) | 2009-07-16 | 2011-02-03 | Gunzoo Inc | 送信装置、受信装置、送信方法、受信方法及び伝送システム |
| JP2011055219A (ja) | 2009-09-01 | 2011-03-17 | Sony Corp | 画像送信装置、画像受信装置、画像送信方法及び画像受信方法 |
Family Cites Families (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US7599438B2 (en) * | 2003-09-07 | 2009-10-06 | Microsoft Corporation | Motion vector block pattern coding and decoding |
| CN101895760B (zh) * | 2010-07-29 | 2013-05-01 | 西安空间无线电技术研究所 | 基于jpeg-ls算法的码流拼接实现系统及方法 |
| CN102036073B (zh) * | 2010-12-21 | 2012-11-28 | 西安交通大学 | 基于视觉潜在注意力目标区域的jpeg2000图像编解码方法 |
-
2011
- 2011-10-24 JP JP2011232863A patent/JP5685682B2/ja active Active
-
2012
- 2012-10-17 IN IN3191DEN2014 patent/IN2014DN03191A/en unknown
- 2012-10-17 CN CN201280057252.7A patent/CN103947212A/zh active Pending
- 2012-10-17 KR KR20147010893A patent/KR20140085462A/ko not_active Withdrawn
- 2012-10-17 SG SG11201401713WA patent/SG11201401713WA/en unknown
- 2012-10-17 WO PCT/JP2012/076813 patent/WO2013061839A1/ja not_active Ceased
- 2012-10-17 EP EP12844021.1A patent/EP2773113A4/en not_active Withdrawn
- 2012-10-17 US US14/354,129 patent/US20150127846A1/en not_active Abandoned
Patent Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JP2001285876A (ja) * | 2000-03-30 | 2001-10-12 | Sony Corp | 画像符号化装置とその方法、ビデオカメラ、画像記録装置、画像伝送装置 |
| WO2006006564A1 (ja) * | 2004-07-12 | 2006-01-19 | Sony Corporation | 符号化方法、符号化装置、復号方法、復号装置およびそれらのプログラム |
| JP2011024018A (ja) | 2009-07-16 | 2011-02-03 | Gunzoo Inc | 送信装置、受信装置、送信方法、受信方法及び伝送システム |
| JP2011055219A (ja) | 2009-09-01 | 2011-03-17 | Sony Corp | 画像送信装置、画像受信装置、画像送信方法及び画像受信方法 |
Non-Patent Citations (8)
| Title |
|---|
| E.KAMINSKY; D.GROIS; O.HADAR: "Efficient real-time video-in-video insertion into a pre-encoded video stream for the H.264/AVC", IEEE INTERNATIONAL CONFERENCE ON IMAGING SYSTEMS AND TECHNIQUES (IST, 1 July 2010 (2010-07-01), pages 436 - 441, XP055148682, DOI: doi:10.1109/IST.2010.5548511 |
| EVGENY KAMINSKY ET AL.: "Efficient real-time Video-in-Video insertion into a pre-encoded video stream for the H.264/AVC", INTERNATIONAL CONFERENCE ON IMAGING SYSTEMS AND TECHNIQUES (IST), July 2010 (2010-07-01), pages 436 - 441, XP055148682 * |
| H. KIMATA; S. SHIMIZU; Y. KUNITA; M. ISOGAI; Y. OHTANI: "Panorama video coding for user-driven interactive video application", IEEE INTERNATIONAL SYMPOSIUM ON CONSUMER ELECTRONICS (ISCE2009), KYOTO, 2009 |
| N. UCHIHARA; H. KASAI: "Fast H.264/AVC stream joiner for interactive free view-area multivision video", IEEE TRANSACTIONS ON CONSUMER ELECTRONICS, vol. 57, no. 3, August 2011 (2011-08-01), pages 1311 - 1319, XP011386575, DOI: doi:10.1109/TCE.2011.6018889 |
| NAOFUMI UCHIHARA ET AL.: "A Study on Video Delivery System with Free View-Area and Multiple Resolution Access for Multiple Users", IEICE TECHNICAL REPORT, vol. 109, no. 398, N, 2010, pages 89 - 94, XP008169220 * |
| NAOFUMI UCHIHARA ET AL.: "Fast H.264/AVC stream joiner for interactive free view-area multivision video", IEEE TRANSACTIONS ON CONSUMER ELECTRONICS, vol. 57, no. 3, August 2011 (2011-08-01), pages 1311 - 1319, XP011386575 * |
| S. HEYMANN; A. SMOLIC; K. MULLER; Y. GUO; J. RURAINSKI; P. EISERT; T. WIEGAND: "Representation, Coding and Interactive Rendering or High-Resolution Panoramic Images and Video using MPEG-4", PROC. PANORAMIC PHOTOGRAMMETRY WORKSHOP, BERLIN, GERMANY, February 2005 (2005-02-01) |
| See also references of EP2773113A4 |
Cited By (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN105612756A (zh) * | 2013-08-08 | 2016-05-25 | 国立大学法人电气通信大学 | 数据处理装置、数据处理方法、程序、记录介质以及数据处理系统 |
| JP2016019240A (ja) * | 2014-07-10 | 2016-02-01 | 株式会社ドワンゴ | 端末装置、動画配信装置、プログラム |
Also Published As
| Publication number | Publication date |
|---|---|
| KR20140085462A (ko) | 2014-07-07 |
| SG11201401713WA (en) | 2014-09-26 |
| US20150127846A1 (en) | 2015-05-07 |
| EP2773113A4 (en) | 2015-06-03 |
| CN103947212A (zh) | 2014-07-23 |
| JP5685682B2 (ja) | 2015-03-18 |
| IN2014DN03191A (enExample) | 2015-05-22 |
| JP2013093656A (ja) | 2013-05-16 |
| EP2773113A1 (en) | 2014-09-03 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US20200252650A1 (en) | Video processing method for blocking in-loop filtering from being applied to at least one boundary in reconstructed frame and associated video processing apparatus | |
| US8462196B2 (en) | Method and apparatus for generating block-based stereoscopic image format and method and apparatus for reconstructing stereoscopic images from block-based stereoscopic image format | |
| CN115443660B (zh) | 用于解码与符号数据隐藏相关的成像的方法和设备 | |
| US20220007028A1 (en) | Method and device for signaling information on chroma format | |
| CN114762336B (zh) | 基于变换跳过和调色板编码相关信息的信令的图像或视频编码 | |
| CN102984543A (zh) | 图像处理装置、图像处理方法和图像处理系统 | |
| KR20250142470A (ko) | 변환에 기반한 영상 코딩 방법 및 그 장치 | |
| US8704932B2 (en) | Method and system for noise reduction for 3D video content | |
| US11902530B2 (en) | Image coding method based on transform and apparatus therefor | |
| JP5685682B2 (ja) | 映像信号の符号化システム及び符号化方法 | |
| CN114556933A (zh) | 基于调色板转义编码的图像或视频编码 | |
| JP7596572B2 (ja) | サブピクチャに基づく映像コーディング装置及び方法 | |
| CN117676169A (zh) | 使用去块滤波的图像编译方法和装置 | |
| CN115299060A (zh) | 基于act残差的图像或视频编码 | |
| US20220174270A1 (en) | Method and device for configuring mpm list | |
| US9554131B1 (en) | Multi-slice/tile encoder with overlapping spatial sections | |
| CN116134821A (zh) | 用于在图像/视频编码系统中处理高级语法的方法和设备 | |
| CN116134816A (zh) | 在图像/视频编码系统中处理一般约束信息的方法和设备 | |
| CN115104313A (zh) | 发信号通知在图片或切片级适用的视频信息的方法和装置 | |
| CN118354065A (zh) | 用于编译关于合并数据的信息的方法和装置 | |
| KR20250090386A (ko) | 픽처 분할 정보를 시그널링 하는 방법 및 장치 | |
| CN114902667A (zh) | 基于色度量化参数偏移信息的图像或视频编码 | |
| CN121193920A (zh) | 与切片相关的信令信息的方法和设备 | |
| CN118158444A (zh) | 用于图像/视频编译的高级别语法信令方法和装置 | |
| CN114208182A (zh) | 用于基于去块滤波对图像进行编码的方法及其设备 |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| 121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 12844021 Country of ref document: EP Kind code of ref document: A1 |
|
| WWE | Wipo information: entry into national phase |
Ref document number: 2012844021 Country of ref document: EP |
|
| ENP | Entry into the national phase |
Ref document number: 20147010893 Country of ref document: KR Kind code of ref document: A |
|
| NENP | Non-entry into the national phase |
Ref country code: DE |
|
| WWE | Wipo information: entry into national phase |
Ref document number: 14354129 Country of ref document: US |