WO2014166096A1 - Reference view derivation for inter-view motion prediction and inter-view residual prediction - Google Patents

Reference view derivation for inter-view motion prediction and inter-view residual prediction Download PDF

Info

Publication number
WO2014166096A1
WO2014166096A1 PCT/CN2013/074102 CN2013074102W WO2014166096A1 WO 2014166096 A1 WO2014166096 A1 WO 2014166096A1 CN 2013074102 W CN2013074102 W CN 2013074102W WO 2014166096 A1 WO2014166096 A1 WO 2014166096A1
Authority
WO
WIPO (PCT)
Prior art keywords
view
inter
reference view
picture
ivrp
Prior art date
Application number
PCT/CN2013/074102
Other languages
French (fr)
Inventor
Jicheng An
Jian-Liang Lin
Kai Zhang
Original Assignee
Mediatek Singapore Pte. Ltd.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Mediatek Singapore Pte. Ltd. filed Critical Mediatek Singapore Pte. Ltd.
Priority to PCT/CN2013/074102 priority Critical patent/WO2014166096A1/en
Publication of WO2014166096A1 publication Critical patent/WO2014166096A1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/513Processing of motion vectors
    • H04N19/517Processing of motion vectors by encoding
    • H04N19/52Processing of motion vectors by encoding by predictive encoding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/174Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a slice, e.g. a line of blocks or a group of blocks
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/597Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/70Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards

Definitions

  • the invention relates generally to Three-Dimensional (3D) video processing.
  • the present invention relates to methods for reference view derivation for inter-view motion prediction (IVMP) and inter-view residual prediction (IVRP).
  • IVMP inter-view motion prediction
  • IVRP inter-view residual prediction
  • 3D video coding is developed for encoding or decoding video data of multiple views simultaneously captured by several cameras. Since all cameras capture the same scene from different viewpoints, multi-view video data contains a large amount of inter-view redundancy.
  • additional tools which employ DV such as inter-view motion prediction (IVMP), and inter-view residual prediction (IVRP), have been integrated to conventional 3D-HEVC (High Efficiency Video Coding) or 3D-AVC (Advanced Video Coding) codec.
  • IVMP inter-view motion prediction
  • IVRP inter-view residual prediction
  • a disparity vector is firstly derived for current PU and then the motion information of the reference block in the reference view that is referenced by the DV is used for predicting the motion information of the current PU.
  • IVRP is applied for a current PU
  • a DV is firstly derived for a current PU and the residual block in the reference view that is referenced by the DV is used for predicting the residual of the current PU.
  • the IVMP needs to access the motion data in the reference view, and the IVRP needs to access the residual data in the reference view.
  • the reference view for IVMP is the view that is included in one of the reference lists with the smallest view index.
  • the reference view for IVRP is the view with index equal to 0.
  • the IVRP is currently combined with IVMP in merge mode, therefore the reference view for IVMP and IVRP should be the same.
  • This invention proposes to unify the reference view for IVMP and IVRP by providing the syntax to identify the picture of reference view from the reference lists and prohibiting the reference view to change on every slice.
  • the proposed method has benefits for the management of decoded motion and residual data, and provides more flexibility for reference view selection, with little impact on coding efficiency.
  • This invention also proposes to unify the reference view for IVMP and IVRP by the fixed reference view selection algorithm instead of explicit signaling, such as either one of the current reference view selection algorithm for IVMP or IVRP.
  • Fig.l is a diagram illustrating the proposed new syntax according to an embodiment of the invention.
  • the situation of the reference view selection here is similar to that of the collocated reference picture selection in HEVC. Therefore, imitating the collocated reference picture selection, it is proposed to signal the reference view in slice header from the reference lists since the reference view choice must be available to every slice to preserve error resiliency, and semantically restrict the selection to always point to the same inter-view reference picture for every slice in a picture since the frequent reference view changes will create a performance burden for decoder.
  • the proposed syntax table is shown in Fig. 1.
  • the new syntax element simply gives the reference index of the reference view picture, from list 0 in P slices and from either list 0 or list 1 (according to refoiew Jrom O Jlag) in B slices. This value is semantically restricted to always point to the same inter-view reference picture for every slice in a picture.
  • This invention also proposes to unify the reference view for IVMP and IVRP by the fixed reference view selection algorithm instead of explicit signaling, such as either one of the current reference view selection algorithm for IVMP or IVRP.
  • Embodiments of disparity vector derivation methods according to the present invention as described above may be implemented in various hardware, software codes, or a combination of both.
  • an embodiment of the present invention can be a circuit integrated into a video compression chip or program codes integrated into video compression software to perform the processing described herein.
  • An embodiment of the present invention may also be program codes to be executed on a Digital Signal Processor (DSP) to perform the processing described herein.
  • DSP Digital Signal Processor
  • the invention may also involve a number of functions to be performed by a computer processor, a digital signal processor, a microprocessor, or field programmable gate array (FPGA).
  • processors can be configured to perform particular tasks according to the invention, by executing machine-readable software code or firmware code that defines the particular methods embodied by the invention.
  • the software code or firmware codes may be developed in different programming languages and different format or style.
  • the software code may also be compiled for different target platform.
  • different code formats, styles and languages of software codes and other means of configuring code to perform the tasks in accordance with the invention will not depart from the spirit and scope of the invention.

Abstract

Methods of reference view selection for inter-view motion prediction and inter-view residual prediction for multi-view video coding and 3D video coding are disclosed. It is proposed to unify the reference view for inter-view motion prediction and inter-view residual prediction by providing the syntax to identify the picture of reference view from the reference lists and prohibiting the reference view to change on every slice. The proposed method has benefits for the management of decoded motion and residual data, and provides more flexibility for reference view selection, with little impact on coding efficiency.

Description

REFERENCE VIEW DERIVATION FOR INTER- VIEW MOTION PREDICTION AND INTER- VIEW RESIDUAL
PREDICTION
TECHNICAL FIELD The invention relates generally to Three-Dimensional (3D) video processing. In particular, the present invention relates to methods for reference view derivation for inter-view motion prediction (IVMP) and inter-view residual prediction (IVRP).
BACKGROUND
3D video coding is developed for encoding or decoding video data of multiple views simultaneously captured by several cameras. Since all cameras capture the same scene from different viewpoints, multi-view video data contains a large amount of inter-view redundancy. To exploit the inter-view redundancy, additional tools which employ DV such as inter-view motion prediction (IVMP), and inter-view residual prediction (IVRP), have been integrated to conventional 3D-HEVC (High Efficiency Video Coding) or 3D-AVC (Advanced Video Coding) codec.
When IVMP is applied for a current prediction unit (PU), a disparity vector (DV) is firstly derived for current PU and then the motion information of the reference block in the reference view that is referenced by the DV is used for predicting the motion information of the current PU. When IVRP is applied for a current PU, a DV is firstly derived for a current PU and the residual block in the reference view that is referenced by the DV is used for predicting the residual of the current PU.
Therefore, the IVMP needs to access the motion data in the reference view, and the IVRP needs to access the residual data in the reference view.
In current HTM6.0, the reference view for IVMP is the view that is included in one of the reference lists with the smallest view index. On the other hand, the reference view for IVRP is the view with index equal to 0. The IVRP is currently combined with IVMP in merge mode, therefore the reference view for IVMP and IVRP should be the same.
In order to unify these two kinds of reference view, facilitate the management of decoded motion and residual data, and provide more flexibility for reference view selection, a more general reference view selection approach is proposed.
SUMMARY
This invention proposes to unify the reference view for IVMP and IVRP by providing the syntax to identify the picture of reference view from the reference lists and prohibiting the reference view to change on every slice. The proposed method has benefits for the management of decoded motion and residual data, and provides more flexibility for reference view selection, with little impact on coding efficiency.
This invention also proposes to unify the reference view for IVMP and IVRP by the fixed reference view selection algorithm instead of explicit signaling, such as either one of the current reference view selection algorithm for IVMP or IVRP.
Other aspects and features of the invention will become apparent to those with ordinary skill in the art upon review of the following descriptions of specific embodiments. BRIEF DESCRIPTION OF DRAWINGS
The invention can be more fully understood by reading the subsequent detailed description and examples with references made to the accompanying drawings, wherein:
Fig.l is a diagram illustrating the proposed new syntax according to an embodiment of the invention.
DETAILED DESCRIPTION
The following description is of the best-contemplated mode of carrying out the invention. This description is made for the purpose of illustrating the general principles of the invention and should not be taken in a limiting sense. The scope of the invention is best determined by reference to the appended claims.
The situation of the reference view selection here is similar to that of the collocated reference picture selection in HEVC. Therefore, imitating the collocated reference picture selection, it is proposed to signal the reference view in slice header from the reference lists since the reference view choice must be available to every slice to preserve error resiliency, and semantically restrict the selection to always point to the same inter-view reference picture for every slice in a picture since the frequent reference view changes will create a performance burden for decoder. The proposed syntax table is shown in Fig. 1.
The new syntax element simply gives the reference index of the reference view picture, from list 0 in P slices and from either list 0 or list 1 (according to refoiew Jrom O Jlag) in B slices. This value is semantically restricted to always point to the same inter-view reference picture for every slice in a picture.
This invention also proposes to unify the reference view for IVMP and IVRP by the fixed reference view selection algorithm instead of explicit signaling, such as either one of the current reference view selection algorithm for IVMP or IVRP.
The reference view selection methods described above can be used in a video encoder as well as in a video decoder. Embodiments of disparity vector derivation methods according to the present invention as described above may be implemented in various hardware, software codes, or a combination of both. For example, an embodiment of the present invention can be a circuit integrated into a video compression chip or program codes integrated into video compression software to perform the processing described herein. An embodiment of the present invention may also be program codes to be executed on a Digital Signal Processor (DSP) to perform the processing described herein. The invention may also involve a number of functions to be performed by a computer processor, a digital signal processor, a microprocessor, or field programmable gate array (FPGA). These processors can be configured to perform particular tasks according to the invention, by executing machine-readable software code or firmware code that defines the particular methods embodied by the invention. The software code or firmware codes may be developed in different programming languages and different format or style. The software code may also be compiled for different target platform. However, different code formats, styles and languages of software codes and other means of configuring code to perform the tasks in accordance with the invention will not depart from the spirit and scope of the invention.
The invention may be embodied in other specific forms without departing from its spirit or essential characteristics. The described examples are to be considered in all respects only as illustrative and not restrictive. To the contrary, it is intended to cover various modifications and similar arrangements (as would be apparent to those skilled in the art). Therefore, the scope of the appended claims should be accorded the broadest interpretation so as to encompass all such modifications and similar arrangements.

Claims

I . A method of reference view selection for multi-view video coding or 3D video coding, used for inter-view motion prediction (IVMP) and inter-view residual prediction (IVRP), comprising determining a reference view for performing IVMP and IVRP.
2. The method as claimed in claim 1, the reference view for IVMP and IVRP is unified.
3. The method as claimed in claim 1, the reference view is a reference picture in current reference picture lists.
4. The method as claimed in claim 1, the reference view is signaled in a bitstream.
5. The method as claimed in claim 4, first information is signaled to indicate which reference picture list is selected, and second information is signaled to indicate which reference picture in the reference picture list indicated by the first information is used as the reference view for IVMP and IVRP.
6. The method as claimed as claim 5, the first information is transmitted only when the slice type is B slice.
7. The method as claimed in claim 4, the reference view is signaled in slice level.
8. The method as claimed in claim 4, the reference view is signaled in video, view, picture, sequence level, or LCU, CU, PU level, including VPS, SPS, PPS or PS.
9. The method as claimed in claim 3, the reference view is the same inter-view reference picture for every slice in one picture.
10. The method as claimed in claim 1, the reference view selection is according to a fixed algorithm.
I I . The method as claimed in claim 10, the reference view is the view with index equal to 0.
12. The method as claimed in claim 10, the reference view is the view that is included in one of the reference lists with the smallest view index.
13. The method as claimed in claim 10, the reference view is the view that closest to a current view (in terms of view index) that is included in one of the reference lists.
14. The method as claimed in claim 10, the reference view is implicitly derived by both encoder and decoder.
15. The method as claimed in claim 14, wherein the reference view is implicitly derived by the encoder and decoder according to view ID.
16. The method as claimed in claim 4, wherein the reference view is not signaled if there are no more than one inter- view reference in reference lists.
PCT/CN2013/074102 2013-04-11 2013-04-11 Reference view derivation for inter-view motion prediction and inter-view residual prediction WO2014166096A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
PCT/CN2013/074102 WO2014166096A1 (en) 2013-04-11 2013-04-11 Reference view derivation for inter-view motion prediction and inter-view residual prediction

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2013/074102 WO2014166096A1 (en) 2013-04-11 2013-04-11 Reference view derivation for inter-view motion prediction and inter-view residual prediction

Publications (1)

Publication Number Publication Date
WO2014166096A1 true WO2014166096A1 (en) 2014-10-16

Family

ID=51688868

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2013/074102 WO2014166096A1 (en) 2013-04-11 2013-04-11 Reference view derivation for inter-view motion prediction and inter-view residual prediction

Country Status (1)

Country Link
WO (1) WO2014166096A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017060587A1 (en) * 2015-10-08 2017-04-13 Orange Multi-view coding and decoding

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20060108475A (en) * 2005-04-14 2006-10-18 삼성전자주식회사 Syntax optimization method for scalable video coding
WO2007110757A2 (en) * 2006-03-28 2007-10-04 Nokia Corporation Slice groups and data partitioning in scalable video coding
CN101292538A (en) * 2005-10-19 2008-10-22 汤姆森特许公司 Multi-view video coding using scalable video coding
CN102257818A (en) * 2008-10-17 2011-11-23 诺基亚公司 Sharing of motion vector in 3d video coding
US20120075436A1 (en) * 2010-09-24 2012-03-29 Qualcomm Incorporated Coding stereo video data

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20060108475A (en) * 2005-04-14 2006-10-18 삼성전자주식회사 Syntax optimization method for scalable video coding
CN101292538A (en) * 2005-10-19 2008-10-22 汤姆森特许公司 Multi-view video coding using scalable video coding
WO2007110757A2 (en) * 2006-03-28 2007-10-04 Nokia Corporation Slice groups and data partitioning in scalable video coding
CN102257818A (en) * 2008-10-17 2011-11-23 诺基亚公司 Sharing of motion vector in 3d video coding
US20120075436A1 (en) * 2010-09-24 2012-03-29 Qualcomm Incorporated Coding stereo video data

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017060587A1 (en) * 2015-10-08 2017-04-13 Orange Multi-view coding and decoding
FR3042368A1 (en) * 2015-10-08 2017-04-14 Orange MULTI-VIEW ENCODING AND DECODING METHOD, MULTI-VIEW ENCODING AND DECODING DEVICE AND CORRESPONDING COMPUTER PROGRAMS
US10893295B2 (en) 2015-10-08 2021-01-12 Orange Multi-view coding and decoding

Similar Documents

Publication Publication Date Title
AU2011354441B2 (en) Method and apparatus of improved intra luma prediction mode coding
WO2015003383A1 (en) Methods for inter-view motion prediction
WO2015109598A1 (en) Methods for motion parameter hole filling
US9800896B2 (en) Method for depth lookup table signaling
US10142610B2 (en) Method for sub-range based coding a depth lookup table
WO2014166068A1 (en) Refinement of view synthesis prediction for 3-d video coding
WO2016008161A1 (en) Temporal derived bi-directional motion vector predictor
WO2015006920A1 (en) An adaptive disparity vector derivation method
WO2015062002A1 (en) Methods for sub-pu level prediction
WO2016115708A1 (en) Methods for chroma component coding with separate intra prediction mode
US10499075B2 (en) Method for coding a depth lookup table
WO2015100710A1 (en) Existence of inter-view reference picture and availability of 3dvc coding tools
CA2896132C (en) Method and apparatus of compatible depth dependent coding
WO2015100731A1 (en) Methods for determining the prediction partitions
WO2014166109A1 (en) Methods for disparity vector derivation
WO2015135175A1 (en) Simplified depth based block partitioning method
WO2015006922A1 (en) Methods for residual prediction
WO2014029086A1 (en) Methods to improve motion vector inheritance and inter-view motion prediction for depth map
WO2014166096A1 (en) Reference view derivation for inter-view motion prediction and inter-view residual prediction
WO2015131404A1 (en) Methods for depth map coding
WO2013159326A1 (en) Inter-view motion prediction in 3d video coding
WO2014106346A1 (en) Method of signalling additional collocated picture for 3dvc
WO2014106327A1 (en) Method and apparatus for inter-view residual prediction in multiview video coding
WO2015143603A1 (en) An improved method for temporal motion vector prediction in video coding
WO2015103747A1 (en) Motion parameter hole filling

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 13881887

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 13881887

Country of ref document: EP

Kind code of ref document: A1