WO2020070409A1 - Codage et décodage d'une vidéo omnidirectionnelle - Google Patents
Codage et décodage d'une vidéo omnidirectionnelleInfo
- Publication number
- WO2020070409A1 WO2020070409A1 PCT/FR2019/052254 FR2019052254W WO2020070409A1 WO 2020070409 A1 WO2020070409 A1 WO 2020070409A1 FR 2019052254 W FR2019052254 W FR 2019052254W WO 2020070409 A1 WO2020070409 A1 WO 2020070409A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- view
- data
- coded
- processed
- Prior art date
Links
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/597—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/161—Encoding, multiplexing or demultiplexing different image signal components
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/132—Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/167—Position within a video image, e.g. region of interest [ROI]
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/17—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
- H04N19/172—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a picture, frame or field
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/59—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial sub-sampling or interpolation, e.g. alteration of picture size or resolution
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/80—Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/85—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/111—Transformation of image signals corresponding to virtual viewpoints, e.g. spatial image interpolation
Definitions
- the present invention relates generally to the field of omnidirectional videos, such as in particular 360 °, 180 ° videos, etc. More particularly, the invention relates to the encoding and decoding of 360 °, 180 ° views, etc. which are captured to generate such videos, as well as the synthesis of non-captured intermediate viewpoints.
- the invention can in particular, but not exclusively, be applied to the video coding implemented in current AVC and HEVC video coders and their extensions (MVC, 3D-AVC, MV-HEVC, 3D-HEVC, etc.), and corresponding video decoding.
- Ui® such a 360 ° camera consists of several 2D cameras (two dimensions) installed on a spherical platform. Each 2D camera captures a particular angle of a 3D scene (three dimensions), the set of views captured by the cameras making it possible to generate a video representing the 3D scene according to a 360 ° x180 ° field of view. It is also possible to use a single 360 ° camera to capture the 3D scene with a 360 ° x180 ° field of view. Such a field of vision may of course have been smaller, for example at 270 ° x 135 °.
- Such 360 ° videos then allow the user to look at the scene as if it were placed in the center of it and to look around, 360 °, thus providing a new way of watching videos.
- Such videos are generally reproduced on virtual reality headsets, also known by the English name HMD for "Head Mounted Devices". However, they can also be displayed on 2D screens equipped with suitable user interaction means. The number of 2D cameras to capture a 360 ° scene varies depending on the platforms used.
- the diverging views captured by the different 2D cameras are placed end to end taking into account the overlaps between views, to create a panoramic 2D image.
- This step is also known as name for "stitching" in English.
- ERP EquiRectangular projection
- the views captured by each of the 2D cameras are projected onto a spherical surface.
- Other types of projections are also possible, such as a Cube Mapping type projection (projection on the faces of a cube).
- the views projected onto a surface are then projected onto a 2D plane to obtain a 2D panoramic image comprising at a given instant all the views of the scene that have been captured.
- a 360 ° camera can be a real camera, that is to say a physical object, or else a virtual camera, in which case the view is obtained by software for generating views.
- a virtual camera makes it possible to generate views representative of points of view of the 3D scene which have not been captured by real cameras.
- the image of the 360 ° view obtained using a single 360 ° camera or the images of 360 ° views obtained using several 360 ° cameras (real and virtual) are then coded using for example :
- a conventional 2D video coder for example a coder conforming to the HEVC standard (English abbreviation for “High Efficiency Video Coding”),
- a conventional 3D video encoder for example an encoder conforming to the MV-HEVC and 3D-HEVC standards.
- Such coders are not efficient enough in terms of compression, given the very large amount of data of the image of a 360 ° view to be coded, and moreover, images of several 360 ° views to be coded, and of the particular geometry of the 360 ° representation of the 3D scene using such 360 ° views.
- the views captured by the 2D cameras of a 360 ° camera being divergent, the abovementioned coders are not sufficiently suitable for coding the various 360 ° view images, since the inter-image prediction will be little, if not used by these coders. Indeed, between two views captured respectively by two 2D cameras, there is little similar content that can be predicted. As a result, all 360 ° view images are compressed in the same way.
- One of the aims of the invention is to remedy the drawbacks of the aforementioned state of the art.
- an object of the present invention relates to a method of coding an image of a view forming part of a plurality of views, the plurality of views simultaneously representing a 3D scene at different positions or different viewing angles, implemented by a coding device, comprising the following:
- the present invention also relates to a method of decoding a data signal representative of an image of a view forming part of a plurality of views, the plurality of views simultaneously representing a 3D scene at different positions or different viewing angles. , implemented by a decoding device, comprising the following:
- the processed image of the view image is data of the image of the view which has not been deleted following the application of a cropping of the image of the view
- the image processing description information is location information, in the view image, of one or more cropped areas.
- Such a cropping treatment applied to the image of said view makes it possible not to code part of the original data of the latter, for the benefit of a significant reduction in the transmission rate of the coded data associated with the image of said view, since the data belonging to the area (s) which have been cropped are neither coded nor signaled to the decoder.
- the reduction in flow will depend on the size of the cropped area (s).
- the image of the view which will be reconstructed after decoding, then possibly processing of its processed data, using the corresponding image processing description information, will therefore not contain all of its original data or at least will be different by compared to the original view image.
- the processed data of the view image are the data of at least one area of the view image which has been sampled, according to a given sampling factor and according to at least one given direction,
- the image processing description information includes at least one location information, in the image of the view, of the at least one sampled area.
- Such processing favors a uniform degradation of the image of said view, always with the aim of optimizing the reduction in data throughput resulting from the sampling applied, then coded.
- the subsequent reconstruction of such an image of the view thus sampled even if it provides a reconstructed image of the view which is degraded / different compared to the original image of the view, the original data of which have been sampled, then coded, does not call into question the efficiency of the synthesis of an intermediate image which would use such an image of said reconstructed sampled view.
- the data processed from the view image are the data from at least one area of the view image which has been filtered
- the image processing description information includes at least one location information, in the view image, of the at least one filtered area.
- Such processing favors the deletion of the image data from said view which are considered not to be useful for coding, with a view to optimizing the reduction in bit rate of the coded data which advantageously consists only of the filtered image data.
- the data processed from the image of the view are pixels from the image of the view, corresponding to an occlusion detected using an image from another view of the plurality,
- the image processing description information includes an indicator of the pixels of the image of the view which are found in the image of another view.
- such processing favors the deletion of the image data of said view which are considered not useful to code, in order to optimize the reduction in bit rate of the coded data which are advantageously constituted only by the pixels of the image of said view, the absence of which has been detected in another image of a current view of said plurality.
- the data processed from an image of a first view and the data processed from an image of at least a second view are combined into a single image.
- the processed image image data which is obtained according to the second decoding method includes the processed image data from a first view and the processed data from a image of at least a second view.
- the processed coded / decoded data of the view image is image type data
- the coded / decoded information describing the image processing is image type and / or text type data.
- the invention also relates to a device for coding a picture of a view forming part of a plurality of views, the plurality of views simultaneously representing a 3D scene at different positions or different viewing angles, the coding device comprising a processor that is configured to implement the following, at a current time:
- Such a coding device is in particular capable of implementing the aforementioned coding method.
- the invention also relates to a device for decoding a data signal representative of an image of a view forming part of a plurality of views, the plurality of views simultaneously representing a 3D scene at different positions or different viewing angles.
- the decoding device comprising a processor which is configured to implement the following, at a current instant:
- Such a decoding device is in particular capable of implementing the aforementioned decoding method.
- the invention also relates to a data signal containing data coded according to the aforementioned coding method.
- the invention also relates to a computer program comprising instructions for implementing the coding method or the decoding method according to the invention, according to any one of the particular embodiments described above, when said program is executed. by a processor.
- This program can use any programming language, and be in the form of source code, object code, or intermediate code between source code and object code, such as in a partially compiled form, or in any other desirable form.
- the invention also relates to a recording medium or information medium readable by a computer, and comprising instructions of a computer program as mentioned above.
- the recording medium can be any entity or device capable of storing the program.
- the support may include a storage means, such as a ROM, for example a CD ROM or a microelectronic circuit ROM, or else a magnetic recording means, for example a USB key or a hard disk.
- the recording medium can be a transmissible medium such as an electrical or optical signal, which can be routed via an electric cable or optical, radio or other means.
- the program according to the invention can in particular be downloaded from a network of the Internet type.
- the recording medium can be an integrated circuit in which the program is incorporated, the circuit being adapted to execute or to be used in the execution of the above-mentioned coding or decoding method.
- FIG. 1 shows the main actions performed by the coding method according to an embodiment of the invention
- FIG. 2A represents a first type of data signal capable of being generated following the implementation of the coding method of FIG. 1,
- FIG. 2B represents a second type of data signal capable of being generated following the implementation of the coding method of FIG. 1,
- FIG. 2C represents a third type of data signal capable of being generated following the implementation of the coding method of FIG. 1,
- FIG. 3A represents a first embodiment of a method for coding all the view images available at a current instant
- FIG. 3B represents a second embodiment of a method for coding all the view images available at a current time
- FIGS. 4A to 4E each represent an example of a processing applied to the image of a view, according to a first embodiment
- FIGS. 5A to 5D each represent an example of a processing applied to the image of a view, according to a second embodiment
- FIG. 6 represents an example of a treatment applied to the image of a view, according to a third embodiment
- FIG. 7 represents an example of a processing applied to the image of a view, according to a fourth embodiment
- FIG. 8 represents an example of a processing applied to the image of a view, according to a fifth embodiment
- FIG. 9 represents an example of a treatment applied to the image of a view, according to a sixth embodiment.
- FIG. 10 represents a coding device implementing the coding method of FIG. 1,
- FIG. 12A represents a first embodiment of a method for decoding all the view images available at a current instant
- FIG. 12B represents a second embodiment of a method for decoding all the view images available at a current instant
- FIG. 13 represents a decoding device implementing the decoding method of FIG. 11,
- FIG. 14 represents an embodiment of a synthesis of view images, in which are used view images reconstructed according to the decoding method of FIG. 11,
- FIGS. 15A to 15D each represent an example of a processing applied to the image of a view after reconstruction of the latter, according to a first embodiment
- FIG. 16 represents an example of a processing applied to the image of a view after reconstruction of the latter, according to a second embodiment
- FIG. 17 represents an example of a processing applied to the image of a view after reconstruction of the latter, according to a third embodiment
- FIG. 18 shows an example of a processing applied to the image of a view after reconstruction of the latter, according to a fourth embodiment.
- the invention mainly provides a coding scheme for a plurality of current images of respectively a plurality of views, the plurality of views representing at the current time a 3D scene according to a given position or a given angle of view, in which two coding techniques are available:
- a first coding technique according to which at least one current image of a view is coded using a conventional coding mode, such as for example HEVC, MV-HEVC, 3D-HEVC,
- the invention proposes a decoding scheme which makes it possible to combine two decoding techniques:
- a first decoding technique according to which at least one current image of a coded view is reconstructed using a conventional decoding mode, such as for example HEVC, MV-HEVC, 3D-HEVC, and corresponding to a conventional coding mode used for coding and having been signaled to the decoder, so as to obtain at least one reconstructed image of a view which is of very good quality,
- a conventional decoding mode such as for example HEVC, MV-HEVC, 3D-HEVC
- the coded processing data of at least one image of a view are decoded using a decoding mode corresponding to the coding mode signaled to the decoder, namely either the mode conventional coding and / or the other suitable coding mode, so as to obtain processed image data and description information of the image processing from which the obtained processed data originate.
- the processed data obtained during decoding for this image therefore does not correspond to the original data thereof, unlike the image data decoded according to the first decoding technique.
- a method of encoding omnidirectional 360 °, 180 °, or other videos which can use any type of multi-view video codajrs, for example conforming to the 3D-HEVC or MV-HEVC standard, or the like, is described below.
- such a coding method is applied to a current image of a view which is part of a plurality of views Vi, ..., V N , the plurality of views representing a 3D scene according to respectively a plurality of viewing angles or a plurality of positions / orientations.
- a first omnidirectional camera can for example be placed in the center of the 3D scene, according to a 360 ° x180 ° angle of view,
- a second omnidirectional camera can for example be placed on the left in the 3D scene, according to a 360 ° x180 ° angle of view,
- a third omnidirectional camera can for example be placed on the right in the 3D scene, according to a 360 ° x180 ° angle of view.
- a first omnidirectional camera can for example be placed in the center of the 3D scene, according to a 360 ° x180 ° angle of view,
- a second omnidirectional camera can for example be placed on the left in the 3D scene, at a viewing angle of 270 ° x 135 °,
- a third omnidirectional camera can for example be placed on the right in the 3D scene, according to a viewing angle, 180 ° x90 °.
- At least two views of said plurality of views can represent the 3D scene at the same viewing angle or not.
- the coding method according to the invention consists in coding at a current time:
- An image of a view considered can be both a texture image and a depth image.
- the image of a view considered for example the image IV k , contains a number Q (Q> 1) of original data (d1 k , ..., dQ k ), such as for example Q pixels.
- the coding method then comprises the following, for at least one IV k image of a V k view, to be coded:
- a first coding method MC1 or a second coding method MC2 of the image IV k is selected.
- flag_proc information is coded, for example on a bit positioned at 0, to indicate that the coding method MC1 is selected.
- the original Q data (pixels) d1 k , ..., dQ k of the image IV k are coded using a conventional coder, such as for example conforming to the HEVC standard, MV- HEVC, 3D-HEVC, etc.
- a coded image IVC k of the view V k is obtained.
- the coded image IVC k then contains Q original coded data dc1 k , dc2 k , ..., dcQ k .
- a data signal F1 k is generated.
- flag_proc information is coded, for example on a bit positioned at 1, to indicate that the coding method MC2 is selected.
- the coding method MC2 is applied to data DT k resulting from a processing of the image IV k , carried out before the coding step.
- Such data DT k include:
- processed data coded DTC k are obtained. They are representative of a processed image coded IVTC k .
- the processed data DT k do not correspond to the original data of the image IV k .
- these processed data DT k correspond to an image whose resolution is greater or less than that of the image IV k before processing.
- the image IV k processed may for example be larger, since it is obtained from images of other views, or on the contrary be smaller, since it results from a deletion of one or more original pixels from the image IV k .
- these processed data DT k correspond to an image whose representation format (YUV, RGB, etc.) of the image IV k processed is different from the original format of the image IV k before processing, as well as the number of bits to represent a pixel (16 bit, 10 bit, 8 bit, etc.).
- these processed data DT k correspond to a color or texture component which is degraded with respect to the original texture or color component of the image IV k before processing.
- these processed data DT k correspond to a particular representation of the original content of the image IV k before processing, for example a representation of the original filtered content of the image IV k .
- the coding method MC2 can be implemented by a coder which is similar to the coder implementing the first coding method MC1. It could be a lossy or lossless encoder.
- the MC2 coding method can be implemented:
- a lossless coder to specifically code textual data
- a lossy or lossless coder to specifically code the image data, such a coder being able to be identical to the coder implementing the first coding method MC1 or else different.
- a data signal F2 k is generated.
- the data signal F'3 k contains the processed data coded DTC k of textual type.
- the first n views V 1 to V n are called master views because once reconstructed, the images of n master views will contain all of their original data and will be relevant for use with one or more of the Nn other views in order to synthesize the images of arbitrary views required by a user,
- the first n images IV-i, ..., IV n are coded using a conventional HEVC type coder, independently of each other.
- a conventional HEVC type coder At the end of the coding C1 1 a, there are obtained respectively n coded images IVC-i, ..., IVC n .
- n data signals F1 -i, ..., F1 n are generated.
- these n data signals F1 1; ..., F1 n are concatenated, generating an F1 data signal.
- the images IV-i, ..., IV n can be used when processing the Nn other images IV n + 1 , ..., IV N.
- the Mn data processed DT n + 1 , ... , DT M are coded using a conventional HEVC type coder, independently of each other, if these Mn data processed are all of image type or, if they are both of image type and of text type , are coded using a conventional HEVC type coder, independently of each other, for the processed image type data, and are coded by a lossless coder for the processed text type data.
- Nn processed data coded DTC n + i , ..., DTC N are obtained at the end of the coding C1 1 b, as respectively Nn coded data associated respectively with the Nn images IV n + 1 , ..., IV N .
- Nn data signals F2 n + i , ..., F2 N are respectively generated containing the Nn processed data coded DTC n + i , ... , DTC N.
- Nn data signals F3 n + i , ..., F3 N respectively containing the Nn processed coded data of image type
- the data signals F3 n + i , ..., F3 N and F'3 n + i , ..., F'3 N are concatenated, generating a data signal F3.
- the first n images IV-i, ..., IV n are coded simultaneously using a conventional coder of the MV-FIEVC or 3D-FIEVC type.
- a conventional coder of the MV-FIEVC or 3D-FIEVC type At the end of the coding C1 1 a, there are obtained respectively n coded images IVC-i, ..., IVC n .
- a single signal F1 is generated which contains the original coded data associated with each of these n coded images.
- the first images IV-i, ..., IV n can be used when processing the Nn other images IV n + 1 , ..., IV N.
- the Mn data processed DT n + i , ..., DT M of the image type are coded simultaneously using a conventional encoder of the MV-FIEVC or 3D type -FIEVC if these Mn data processed are all of the image type or, if they are both of the image type and of the text type, are coded simultaneously using a conventional encoder of the MV-FIEVC or 3D-FIEVC type for processed image data, and are encoded by a lossless encoder for processed text data.
- Nn coded processed data DTC n + i , ..., DTC N as respectively Nn coded processed data associated respectively with the Nn images IV n + i , ..., IV N that have been processed.
- a single signal F2 is generated which contains the N- n processed data coded DTC n + i, ..., DTC N of image type.
- C12c in the case where the Mn data processed are both image and text type:
- a signal F'3 is generated which contains the processed data coded as text.
- the signals F1 and F2 are concatenated, or the signals F1, F3 and F'3 are concatenated, providing a data signal F capable of being decoded by a decoding method which will be described later in the description.
- the coder implementing the MC1 coding method could be a coder of the FIEVC type and the coder implementing the MC2 coding method could be a coder of the MV-FIEVC or 3D-HEVC type. or else include an MV-HEVC or 3D-HEVC type encoder and a lossy encoder.
- the coder implementing the coding method MC1 could be a coder of the MV-FIEVC or 3D-FIEVC type and the coder implementing the coding method MC2 could be a coder of the FIEVC type or else include a FIEVC type encoder and a lossless encoder.
- FIGS. 4A to 4E a first embodiment of a processing applied to the original data of an image IV k , before the coding step C1 1 b (FIG. 1) according to the second method MC2 coding.
- the processing applied to the original data of the image IV k is a cropping of one or more zones of this image, in a horizontal or vertical direction or else in both directions at the same time .
- the left edge B1 and the right edge B2 of the image IV k are cropped, which means the removal of the pixels from the rectangular area of the image IV k formed by each of the edges B1 and B2.
- the upper edge B3 and the lower edge B4 of the image IV k are cropped, which means the removal of the pixels from the rectangular area of the image IV k formed by each of the edges B3 and B4.
- the cropping is applied in a vertical direction to a rectangular zone Z1 situated in the image IV k .
- the cropping is applied in a horizontal direction to a rectangular zone Z2 located in the image IV k .
- the cropping is applied both in a horizontal and vertical direction to an area Z3 located in the image IV k .
- the processed data DT k to be coded then includes: - the pixels of the remaining area Z R of the image IV k which have not been deleted following the cropping (figs 4A, 4B, 4E), or else the pixels of the remaining areas Z1 R and Z2 R (figs 4C, 4D) of image IV k which have not been deleted following the cropping,
- the information describing the trimming applied is of textual type and contains:
- the information describing the trimming applied contains:
- the original data (pixels) of the rectangular area defined by the coordinates of the pixel located at the top, leftmost, in the remaining area Z R (resp. Z1 R and Z2 r ) and the coordinates of the pixel located below, the further to the right, in the remaining zone Z R (resp. Z1 R and Z2 R ) are then coded in C1 1 b (FIG. 1) by an encoder of the HEVC, 3D-HEVC, MV-HEVC type, etc.
- the description information of the applied trimming is coded in C1 1 b (FIG. 1) by a lossless coder.
- the information describing the cropping applied contains the number of rows and / or columns of pixels to be deleted, as well as the position of these rows and / or these columns in the image IV k .
- the number of data to be deleted by trimming is fixed. It can for example be decided to systematically delete X rows and / or Y columns from the image of a view considered. In this case, the description information contains only the cropping information or not for each view. According to another embodiment, the number of data to be deleted by trimming is variable between the image IV k of the view V k and an image of another available view.
- the number of data to be deleted by trimming can also depend, for example, on the position in the 3D scene of the camera which captured the image IV k .
- a quantity of data to be deleted different from the amount of data deleted for the image IV k will for example be used.
- the application of a cropping may also depend on the instant at which the image IV k is coded. At the current instant, it can for example be decided to apply a cropping to the image IV k , while at the instants preceding or following the current instant, it can be decided not to apply such a cropping, nor by the way no processing, as in image IV k .
- cropping can be applied to images from one or more views at the current time.
- the cropped area in the image IV k of the view V k may or may not be the same as the cropped area of an image of another view to be coded at the current time.
- FIGS. 5A to 5D a second embodiment of a processing applied to the original data of an image IV k , before the coding step C1 1 b (FIG. 1) according to the second coding method. MC2.
- the processing applied to the original data of the image IV k is a subsampling of one or more zones of this image, in a horizontal or vertical direction.
- a subsampling is applied to an area Z4 of the image IV k , in a vertical direction.
- a subsampling is applied to an area Z5 of the image IV k , in a horizontal direction.
- a subsampling is applied to the whole image IV k .
- a subsampling is applied to an area Z6 of the image IV k , in both a horizontal direction and a vertical direction.
- the processed data DT k to be coded then includes:
- the sub-sampled image data are then coded in C11 b (FIG. 1) by an encoder of the HEVC, 3D-HEVC, MV-HEVC type, etc.
- the description information of the applied sub-sampling is coded in C1 1 b (FIG. 1) by a lossless coder.
- the value of the sub-sampling factor can be fixed or else depend on the position in the 3D scene of the camera which captured the image IV k .
- another sub-sampling factor will be used for example.
- the application of a sub-sampling can moreover depend on the instant at which the image of the view V k is coded. At the current instant, it can for example be decided to apply a sub-sampling to the image IV k , while at the instants preceding or following the current instant, it can be decided not to apply such a sub -sampling, or moreover no processing, like the view V k .
- subsampling can be applied to images from one or more views at the current time.
- the sub-sampled area in the image IV k can be the same or not as the sub-sampled area of an image from another view to be coded at the current instant.
- FIG. 6 a third embodiment of a processing applied to the original data of an image IV k , before the coding step C1 1 b (FIG. 1) according to the second coding method MC2 .
- the processing applied to the original data of the image IV k is a detection of contours by filtering this image.
- Two contours ED1 and ED2 are for example present in an IV k image.
- filtering comprises for example the following:
- the processed data DT k to be coded in the case of such filtering, then includes:
- the image data (pixels) corresponding to the hatched areas are then coded in C1 1 b (FIG. 1) by an encoder of HEVC, 3D-HEVC, MV-HEVC type, etc.
- the information describing the filtering applied is coded in C11 b (FIG. 1) by a lossless coder.
- the filtering which has just been described with respect to the image IV k can be applied to one or more images of other views among said N views, on areas of this or these images which may be different from an image d 'a view to another image of a view.
- FIG. 7 a fourth embodiment of a processing applied to the original data of an image IV k , before the coding step C1 1 b (FIG. 1) according to the second coding method MC2 .
- the processing applied to the original data of the image IV k is a detection of occlusion of at least one zone Z Q c of the image IV k using at least one IV P image from another view V p among N views (1 ⁇ p £ N).
- such occlusion detection consists in searching for the zone Z 0 c of the image IV k from the image IV P , using for example an estimate of disparity.
- the occluded zone Z 0 c is then expanded, using for example a mathematical morphology algorithm.
- the zone Zoc thus dilated is represented by hatching in FIG. 7. All the original data of the image IV k which are not part of the zone Z 0 c hatched and which are therefore considered useless to code are deleted.
- the processed data DT k to be coded in the case of such occlusion detection, then includes:
- the image data (pixels) corresponding to the hatched area are then coded in C1 1 b (FIG. 1) by an encoder of HEVC, 3D-HEVC, MV-HEVC type, etc.
- the description information of the occlusion detection applied is coded in C1 1 b (FIG. 1) by a lossless coder.
- FIG. 8 a fifth embodiment of a processing applied to the original data of an image IV k , before the coding step C1 1 b (FIG. 1) according to the second coding method MC2 .
- the processing applied to the original data of the image IV k consists in calculating pixels:
- the processed data DT k to be coded in the case of such a calculation, then includes: an indicator of the pixels of the image IV k of said view which have been calculated,
- the abovementioned calculation consists for example in subtracting the original pixels from the image IV k of a view and possibly the original pixels from the image IV
- FIG. 9 a sixth embodiment of a processing applied to the original data of an image IV k , before the coding step C1 1 b (FIG. 1) according to the second coding method MC2 .
- the processing applied to the original data of the image IV k consists of:
- FIG. 10 shows the simplified structure of a COD coding device suitable for implementing the coding method according to any one of the particular embodiments of the invention.
- the actions executed by the coding method are implemented by computer program instructions.
- the coding device COD has the conventional architecture of a computer and comprises in particular a memory MEM_C, a processing unit UT_C, equipped for example with a processor PROC_C, and controlled by the computer program PG_C stored in MEM_C memory.
- the computer program PG_C includes instructions for implementing the actions of the coding method as described above, when the program is executed by the processor PROC_C.
- the code instructions of the computer program PG_C are for example loaded into a RAM memory (not shown) before being executed by the processor PROC_C.
- the processor PROC_C of the processing unit UT_C implements in particular the actions of the coding method described above, according to the instructions of the computer program PG_C.
- such a decoding method applies to a data signal representative of a current image of a view which is part of said plurality of views Vi, ..., V N.
- the decoding method according to the invention consists in decoding:
- An image of a view considered to be reconstructed using the aforementioned decoding method can be both a texture image and a depth image.
- the decoding method comprises the following, for a data signal F1 k , F2 k or F3 k and F'3 k representative of at least one image IV k of a view V k , to be reconstructed:
- the flag_proc information indicating whether the image IV k has been coded using the first coding method MC1 or the second coding method MC2
- the data signal F1 k , F2 k or F3 k and F'3 k is read from the data signal F1 k , F2 k or F3 k and F'3 k , as shown respectively in FIGS. 2A, 2B and 2C.
- D1 1 a the coded data dc1 k , dc2 k , dcQ k associated with the coded image IVC k are read in the data signal F1 k .
- an IVD k image is reconstructed from the coded data dc1 k , dc2 k , ..., dcQ k read in D1 1 a, using a decoding method MD1 corresponding to the method of coding MC1 applied to the coding, in C1 1 a in FIG. 1.
- the image IV k is reconstructed using a conventional decoder, such as for example in accordance with the FIEVC, MVC-FIEVC, 3D-FIEVC standard, etc.
- the image IVD k thus reconstructed contains the original data d1 k , d2 k , ..., dQ k of the image IV k which has been coded in C1 1 a in FIG. 1.
- the IVD k image being consistent with the original IV k image thus constitutes a master image which is relevant to use, for example in the context of a synthesis of intermediate views.
- the processed data coded DTC k associated with the processed image coded IVTC k are read in the signal of data F2 k .
- a processed IVTD k image is reconstructed from the DTC k coded data read in D1 1 b, using a decoding method MD2 corresponding to the coding method MC2 applied to coding, in C1 1 b in Figure 1.
- coded data DTC k are decoded using a conventional decoder, such as for example conforming to the standard FIEVC, MVC-FIEVC, 3D-HEVC, etc.
- the processed image thus reconstructed IVTD k corresponding to the decoded DTC k data, contains the processed data DT k of the image IV k before their coding in C1 1 b in FIG. 1.
- the reconstructed IVTD k processed image contains image data (pixels) corresponding to all or part of the original data of the IV k image which has been processed using a specific image processing, including various examples detailed have been described with reference to Figures 4 to 9.
- the processed data coded DTC k associated with the processed image coded IVTC k are read in D1 1 c.
- processed data coded DTC k different from the image data, such as for example data of textual type, or else comprising both image data and data of a type other than image data , are read in the signal F'3 k .
- the processed data coded DTC k are decoded in D12b, using the decoding method MD2, which can be implemented:
- a lossy or lossless decoder to specifically decode the image data, such a decoder being able to be identical to the decoder implementing the first MD1 decoding method or else different,
- Such a reconstructed processed image IVTD k according to the second decoding method MD2 does not contain all the original data of the image IV k before processing then coding in C1 1 b.
- Such a reconstructed image of a view according to the second MD2 decoding method may however be used in addition to an image of a master view which will have been reconstructed using the first MD1 decoding method, in the context for example of a synthesis of intermediate images, in order to obtain images of synthesized views which are of good quality.
- the first n coded images IVCi, ..., IVC n are reconstructed using the first decoding technique MD1 to obtain respectively the image of each of the n master views,
- the data signal F1 as generated in C13a in FIG. 3A
- the data signals F3 and F’3 as generated in C13c in FIG. 3A.
- the data signal F1 is in turn separated into n data signals F1, ..., F1 n representative respectively of the n coded view images IVCi ,. .., IVC n .
- D1 1 a in each of the n data signals F1 -i, ..., F1 n are determined respectively the original coded data dc1, ..., dcQ-i, ..., dc1 n , ..., dcQ n which are associated with each of these n coded images.
- D12a the IVD-i, ..., IVD n images are reconstructed from their respective original coded data, read in D11 a, using a conventional HEVC type decoder, independently of one another.
- the data signal F2 is in turn separated into Nn data signals F2 n + 1 , ..., F2 N representative respectively of the Nn processed data coded DTC n + i ,. .., DTC N -
- D1 1 b in each of the Nn data signals F2 n + i, ..., F2 N are read respectively Nn processed data coded DTC n + i , ..., DTC N which correspond respectively to each of the Nn images IV n + 1 , ..., IV N to be reconstructed.
- the processed images are reconstructed respectively from Nn processed data coded DTC n + i, ..., DTC N read in D11 b, using a conventional FIEVC type decoder, independently of each other .
- the processed images reconstructed IVTD n + 1 , ..., IVTD N are then obtained.
- the data signal F1 is in turn separated into n data signals F1 -i, ..., F1 n respectively representative of the n images coded IVC-i, ..., IVC n .
- D1 1 a in each of the n data signals F1 1; ..., F1 n are read respectively the original coded data dd -i, ..., dcQi, ..., dc1 n , ..., dcQ n which are associated with each of these n coded images.
- the IVD-i, ..., IVD n images are reconstructed from their respective original coded data, read in D11 a, using a conventional FIEVC type decoder, independently of each other.
- the data signal F3 is in turn separated into Nn data signals F3 n + i , ..., F3 N respectively representative of the Nn processed data coded DTC n + i , ..., DTC N of image type and,
- the data signal F'3 is in turn separated into Nn data signals F'3 n + i , ..., F'3 N respectively representative of the Nn processed data coded DTC n + i , ..., DTC No text type or another.
- each of the Nn data signals F3 n + i, ..., F3N are respectively read Nn processed data coded DTC n + i , ..., DTC N of image type which correspond respectively to each of the Nn images IV n + 1 , ..., IV N to be reconstructed
- - in each of the Nn data signals F'3 n + i, ..., F'3 N are respectively read Nn processed data coded DTC n + i , ..., DTC N of text type or others which correspond respectively to processing description information for each of the Nn images IV n + 1 , ..., IV N to be reconstructed.
- the Nn processed images are reconstructed respectively from Nn processed data coded DTC n + i ..., DTC N read in D1 1b, using a conventional FIEVC type decoder, independently of each of the other.
- the processed images reconstructed IVTD n + 1 , ..., IVTD N are then obtained.
- D12b are reconstructed information describing the processing applied to each of the images IV n + i ..., IV N , before their coding C1 1 b (FIG. 3A), from Nn processed data coded DTC n + i , ..., DTCN of text type or other read in D1 1 c, using a decoder corresponding to the lossless coder used for coding.
- the data signal F1 as generated in C12a in FIG. 3B
- the data signals F3 and F’3 as generated in C12c in FIG. 3B.
- signals F1 and F2 in D11 a, in the data signal F1 are read the original coded data dc1 1; ..., dcQ-i, ..., dc1 n , ⁇ , dcQ n which are respectively associated with each of the IVC-g images, ..., IVC n of n coded views.
- the IVD- I , ..., IVD n images are reconstructed from their respective original coded data, read in D1 1 a, simultaneously, using a conventional MV-FIEVC decoder. or 3D-FIEVC.
- D11 b in the data signal F2 are read the Nn coded processed data DTC n + i , ..., DTC N which are respectively associated with each of the images IV n + i , ..., IV N of the Nn views to be reconstructed.
- the Nn processed images are reconstructed respectively from the Nn processed data coded DTC n + i , ..., DTC N , read in D1 1b, so simultaneous, using a conventional MV-HEVC or 3D-HEVC decoder.
- the reconstructed processed IVTD n + i , ..., IVTD N images are then obtained.
- the IVD-i, ..., IVD n images are reconstructed from their respective original coded data, read in D1 1 a, simultaneously, using a conventional MV-FIEVC decoder. or 3D-FIEVC.
- D1 1 c in the data signal F3 are read the Nn processed data coded DTC n + i , ..., DTC N of image type which are associated respectively with each of the Nn images IV n + i , ..., IV N to be reconstructed.
- the processed images are reconstructed respectively from Nn processed data coded DTC n + i , ..., DTC N , read in D1 1 c, simultaneously, using a conventional MV type decoder -FIEVC or 3D-FIEVC.
- the reconstructed processed view images IVTD n + i , ..., IVTD N are then obtained.
- D12b in FIG. 12B, are reconstructed information describing the processing applied to each of the images IV n + 1 ..., IV N , before their coding C1 1 b (FIG. 3B), from the Nn data processed.
- the decoder implementing the MD1 decoding method could be a HEVC type decoder and the decoder implementing the MD2 decoding method could be a MV-HEVC or 3D-HEVC decoder .
- the decoder implementing the MD1 decoding method could be a MV-HEVC or 3D-HEVC type decoder and the decoder implementing the MD2 decoding method could be a HEVC coder .
- FIG. 13 shows the simplified structure of a DEC decoding device suitable for implementing the decoding method according to any one of the particular embodiments of the invention.
- the actions executed by the decoding method are implemented by computer program instructions.
- the decoding device DEC has the conventional architecture of a computer and comprises in particular a memory MEM_D, a processing unit UT_D, equipped for example with a processor PROC_D, and controlled by the computer program PG_D stored in MEM_D memory.
- the computer program PG_D includes instructions for implementing the actions of the decoding method as described above, when the program is executed by the processor PROC_D.
- the code instructions of the computer program PG_D are for example loaded into a RAM memory (not shown) before being executed by the processor PROC_D.
- the processor PROC_D of the processing unit UT_D implements in particular the actions of the decoding method described above, according to the instructions of the computer program PG_D.
- the DEC decoding device is for example included in a terminal.
- the N reconstructed images IVD-i , ..., IVD n and IVTD n + i , ..., IVTD N can be used to synthesize an image of an intermediate view required by a user.
- the first n reconstructed view images IVD-i , ..., IVD n which are considered as master views are transmitted in S1 to an image synthesis module.
- the Nn reconstructed treated view images IVTD n + 1 , ..., IVTD N may need to be processed in S2, using the decoded image processing description information which is associated therewith.
- Nn images of reconstructed views IVD n + 1 , .., are obtained.
- IVD n is obtained.
- the Nn reconstructed images IVD n + i , ..., IVD N are then transmitted in S3 to the image synthesis module.
- an image of a view is synthesized using at least one of the IVD-i, IVD n images of the first n reconstructed views and possibly at least one of the Nn IVD n + i , ..., IVD N images of the Nn reconstructed views.
- the n reconstructed images IVD 1; IVD n can also be subject to S2 processing.
- Such processing S2 may prove to be necessary, in the case where the user UT requests an image of a view whose viewing angle represented does not correspond to the viewing angle or to the viewing angles of the n reconstructed images.
- IVD-i, ..., IVD n The user UT could for example request an image of a view representing a field of vision of 120 ⁇ 90, while the n reconstructed images IVD-i, ..., IVD n each represent a viewing angle of 360 ⁇ 180.
- Such a processing possibility for the reconstructed images IVD-i , ..., IVD n is symbolized by dotted arrows in FIG. 14.
- the reconstructed images IVD-i, ..., IVD n can be used when processing Nn other images IV n + 1 , ..., IV N.
- FIGS. 15A to 15C a first embodiment of a processing applied to the data of a reconstructed processed image IVTD k .
- Such processing consists in obtaining the initial resolution of the image IV k of the corresponding view which has been sampled before being coded in C1 1 b in FIG. 1.
- the processing applied before coding is a subsampling of an area Z4 of the image IV k , in a vertical direction, as shown in FIG. 5A.
- the processing applied to the reconstructed processed image IVTD k consists in applying an oversampling to the zone Z4, corresponding to the subsampling applied in FIG. 5A, so as to find the initial resolution of the image IV k , at l using information describing the subsampling applied, such as in particular:
- the processing applied before coding is a subsampling of an area Z5 of the image IV k , in a horizontal direction, as shown in FIG. 5B.
- the processing applied to the reconstructed processed image IVTD k consists in applying an oversampling to the zone Z5, corresponding to the undersampling applied in FIG. 5B, so as to find the initial resolution of the image IV k , at l using information describing the subsampling applied, such as in particular:
- the processing applied to the reconstructed processed image IVTD k consists in applying an oversampling to all the image data of the image IVTD k , corresponding to the subsampling applied in FIG. 5C, so as to find the initial resolution of the image IV k , using information describing the subsampling applied, such as in particular:
- the processing applied before coding is a subsampling of an area Z6 of the image IV k , in both a horizontal direction and a vertical direction, as shown in Figure 5D.
- the processing applied to the reconstructed processed image IVTD k consists in applying an oversampling to the zone Z6, corresponding to the subsampling applied in FIG. 5D, so as to find the initial resolution of the image IV k , at using information describing the subsampling applied, such as in particular:
- Such processing consists in restoring one or more contours of the image IV k of the view which has been filtered before coding of the latter.
- the processing applied before coding is a filtering of the contours ED1 and ED2 of the image IV k , as shown in FIG. 6.
- FIG. 17 a third embodiment of a processing applied to the data of a reconstructed processed image IVTD k .
- processing consists in reconstructing pixels of the image IV k of the view which, before coding, have been calculated in accordance with the embodiment of the processing in FIG. 8.
- the processing applied to the reconstructed processed image IVTD k then consists of:
- the decoding of the processed data DT k then consists in calculating the pixels of the image IV k :
- the abovementioned calculation consists for example of combining the pixels of the image IV k , with the pixels of the reconstructed image IV j and possibly with the pixels of the reconstructed image IVi.
- FIG. 18 a fourth embodiment of a processing applied to the data of a reconstructed processed image IVTD k .
- processing consists in reconstructing pixels of the image IV k of the view which, before coding, have been calculated in accordance with the embodiment of the processing in FIG. 9.
- Processing is first applied to the reconstructed IVD one image according to the second MD2 decoding method. It then consists in reconstructing, from the image IVD one , the pixels of the image IV k :
- Image IV0, from view V0, is conventionally coded using the first coding method MC1, while the other five images IV1, IV2, IV3, IV4, IV5 from views V1, V2, V3, V4, V 5 are cut before coding.
- the processing applied to each of the images IV1, IV2, IV3, IV4, IV5 consists in deleting a fixed number of columns, for example 200, to the right and to the left of each of these images. The number of columns to be deleted has been selected such that the viewing angle is reduced from 360 ° to 120 °.
- the processing applied to each of the images IV1, IV2, IV3, IV4, IV5 consists in deleting a fixed number of lines, for example 100, in the upper and lower parts respectively of each of these images. The number of lines to be deleted has been selected such that the viewing angle is reduced from 180 ° to 120 °.
- Flag_proc information is set to 0, in association with image IV0, and flag_proc information is set to 1, in association with images IV1, IV2, IV3, IV4, IV5.
- the image of view IV0 is coded using a HEVC coder.
- the data of the remaining areas after trimming of each of the images IV1, IV2, IV3, IV4, IV5 are coded using a HEVC coder.
- the data signals F1 o, F2-i, F2 2 , F2 3J F2 4 , F2 5 are concatenated, then transmitted to a decoder.
- the five data signals F2 1; F2 2 , F2 3 , F2 4 , F2 5 can also include the coordinates of the cropped areas as follows:
- the flag_proc information is read.
- the images IV1, IV2, IV3, IV4, IV5, corresponding to the coded processed data are reconstructed using a HEVC decoder. No processing is applied to the images IV1, IV2, IV3, IV4, IV5 which have been reconstructed since it is not possible to reconstruct the data of these images which have been deleted by cropping.
- a synthesis algorithm uses the six images IV0, IV1, IV2, IV3, IV4, IV5 which have been reconstructed in order to generate an image of an arbitrary view required by a user.
- the five data signals F2 1; F2 2 , F2 3 , F2 4 , F2 5 also include the coordinates of the cropped areas, these coordinates are used by the synthesis algorithm to generate an image of a view required by a user.
- the processing data such as the image data relating to the images IV1 to IV8 contains the 8 subsampled texture components of resolution 2048x1024 of the images IV1 to IV8 and the 8 subsampled depth components of resolution 1025x512 of the images IV1 to IV8.
- processing data contain textual data which indicate the sub-sampling factors for each image IV1 to IV8 of views 1 to 8. They are written as follows:
- Flag_proc information is set to 0, in association with images IV0 and IV9, and flag_proc information is set to 1, in association with images IV1 to IV8.
- the textual data which indicates the sub-sampling factors for each image IV1 to IV8 of views 1 to 8, are in turn lossless coded.
- the data signals F1 and F2 are concatenated, then transmitted to a decoder.
- the flag_proc information is read.
- flag_proc 1
- the images IV1 to IV8, corresponding to their respective coded subsampled texture and depth data are reconstructed simultaneously using an MV-HEVC decoder.
- Sub-sampled images reconstructed IVDT1 to IVDT8 are then obtained.
- the textual data corresponding to each of the 8 images IV1 to IV8 are also decoded, providing the sub-sampling factors which were used for each image IV1 to IV8.
- the reconstructed subsampled images IVDT1 to IVDT8 are then processed using their corresponding subsampling factors. At the end of the processing, reconstructed images IVD1 to IVD8 are obtained, whose 8 respective texture components are at their initial resolution 4096x2048 and whose 8 respective depth components are at their initial resolution 4096x2048.
- a synthesis algorithm uses the images of the 10 views thus reconstructed at their initial resolution, in order to generate an image of a view required by a user.
- the processing data such as the image data relating to the images IV1 and IV2 contain the 2 texture components of the occlusion masks of the images IV1 and IV2 and the 2 depth components of the occlusion masks of the images IV1 and IV2 .
- Flag_proc information is set to 0, in association with the IVO image, and flag_proc information is set to 1, in association with the IV1 and IV2 images.
- the IVO image is coded using a HEVC coder.
- the image data (texture and depth) of the occlusion masks of each of the images IV1, IV2 are coded using a HEVC coder.
- the data signals F1 0 , F2 1; F2 2 are concatenated, then transmitted to a decoder.
- the flag_proc information is read.
- the images IV1, IV2, corresponding to the image data (texture and depth) coded by the occlusion masks of each of the images IV1, IV2, are reconstructed using a HEVC decoder. No processing is applied to the images IV1, IV2 which have been reconstructed because it is not possible to reconstruct the data of these images which have been deleted after the detection of occlusion.
- a synthesis algorithm can use the IVO, IV1, IV2 images which have been reconstructed in order to generate an image of a view required by a user.
- a fourth example it is considered that two images IVO and IV1 of resolution 4096x2048 pixels are captured respectively by two cameras of the 360 ° type.
- the IVO image of the first view is conventionally coded using the first coding method MC1, while the IV1 image of the second view is processed before being coded using the second MC2 coding method.
- Such treatment includes the following:
- a filter such as for example a Sobel filter
- contours for example using a mathematical morphology operator, in order to increase the area around the contours.
- the processing data such as the image data relating to the image IV1, includes the pixels inside the area around the contours, as well as the pixels set to 0 corresponding respectively to the pixels located outside the area around the contours.
- the image data in the area around the edges of the IV1 image is encoded using a HEVC encoder, while the tagging information is encoded using a lossless encoder.
- the flag_proc information is read.
- flag_proc 1
- the image IV1 corresponding to the image data of the area around the contours of the image IV1
- a synthesis algorithm can use the two images IV0 and IV1 which have been reconstructed, in order to generate an image of a view required by a user.
- a fifth example it is considered that four images IVO to IV3, of resolution 4096x2048 pixels, are captured respectively by four cameras of the 360 ° type.
- the IVO image is conventionally coded using the first MC1 coding method, while the IV1 to IV3 images are processed before being coded according to the second MC2 coding method.
- One such processing is a filtering of the images IV1 to IV3, during which a region of interest ROI is calculated.
- a region of interest contains the zone (s) of each image IV1 to IV3 which are considered the most relevant, for example because they contain numerous details.
- Such filtering is carried out according to, for example, one of the two methods below:
- the depth map is characterized, for each texture pixel, by a value of near or far depth in the 3D scene.
- a threshold is defined, such that each pixel of an image IV1, IV2, IV3 which is located below this threshold is associated with an object in the scene which is close to the camera. All the pixels located below this threshold are then considered to be the region of interest.
- the processing data such as the image data relating to the images IV1 to IV3, include the pixels inside their respective regions of interest, as well as the pixels set to 0 corresponding respectively to the pixels located outside these regions of interest.
- textual data is generated, for example in the form of marking information which indicates the setting to 0 of the pixels located outside the region of interest. Pixels set to 0 will neither be coded nor reported to the decoder.
- the image data of the region of interest of each image IV1, IV2, IV3 is coded using a HEVC coder, while the marking information is coded using a lossless coder .
- the data signals F1 0 , F2-i, F2 2 , F2 3 are concatenated, then transmitted to a decoder.
- the flag_proc information is read.
- each image IV1 to IV3, corresponding to the image data of its respective region of interest, is reconstructed by means of a FIEVC decoder, using the marking information, which makes it possible to restore the value set to 0 of the pixels surrounding said region.
- a synthesis algorithm can directly use the four IV0 images
- IV1, IV2, IV3 which have been reconstructed, in order to generate an image of a view required by a user.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
- Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
Abstract
Description
Claims
Priority Applications (8)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US17/281,463 US11653025B2 (en) | 2018-10-01 | 2019-09-25 | Coding and decoding of an omnidirectional video |
BR112021005927-0A BR112021005927A2 (pt) | 2018-10-01 | 2019-09-25 | codificação e descodificação de um vídeo omnidirecional |
KR1020217009472A KR20210066825A (ko) | 2018-10-01 | 2019-09-25 | 전방향성 영상의 코딩 및 디코딩 |
EP19795265.8A EP3861751A1 (fr) | 2018-10-01 | 2019-09-25 | Codage et décodage d'une vidéo omnidirectionnelle |
CN201980064753.XA CN112806015B (zh) | 2018-10-01 | 2019-09-25 | 全向视频的编码和解码 |
CN202410246938.5A CN118317113A (zh) | 2018-10-01 | 2019-09-25 | 全向视频的编码和解码 |
JP2021517806A JP7520823B2 (ja) | 2018-10-01 | 2019-09-25 | 全方位映像の符号化及び復号 |
US18/296,073 US11973981B2 (en) | 2018-10-01 | 2023-04-05 | Coding and decoding of an omnidirectional video |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
FR1859067 | 2018-10-01 | ||
FR1859067A FR3086831A1 (fr) | 2018-10-01 | 2018-10-01 | Codage et decodage d'une video omnidirectionnelle |
Related Child Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US17/281,463 A-371-Of-International US11653025B2 (en) | 2018-10-01 | 2019-09-25 | Coding and decoding of an omnidirectional video |
US18/296,073 Continuation US11973981B2 (en) | 2018-10-01 | 2023-04-05 | Coding and decoding of an omnidirectional video |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2020070409A1 true WO2020070409A1 (fr) | 2020-04-09 |
Family
ID=65244050
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/FR2019/052254 WO2020070409A1 (fr) | 2018-10-01 | 2019-09-25 | Codage et décodage d'une vidéo omnidirectionnelle |
Country Status (8)
Country | Link |
---|---|
US (2) | US11653025B2 (fr) |
EP (1) | EP3861751A1 (fr) |
JP (1) | JP7520823B2 (fr) |
KR (1) | KR20210066825A (fr) |
CN (2) | CN118317113A (fr) |
BR (1) | BR112021005927A2 (fr) |
FR (1) | FR3086831A1 (fr) |
WO (1) | WO2020070409A1 (fr) |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR102065632B1 (ko) * | 2018-10-22 | 2020-02-11 | 전자부품연구원 | 복수의 가상 카메라를 이용한 게임 내 360 vr 영상 획득 장치 및 방법 |
Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20100259595A1 (en) * | 2009-04-10 | 2010-10-14 | Nokia Corporation | Methods and Apparatuses for Efficient Streaming of Free View Point Video |
Family Cites Families (23)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR100739686B1 (ko) * | 2004-08-13 | 2007-07-13 | 경희대학교 산학협력단 | 영상 코딩 방법, 코딩 장치, 영상 디코딩 방법 및 디코딩장치 |
KR101385884B1 (ko) * | 2008-01-30 | 2014-04-16 | 고려대학교 산학협력단 | 다시점 동영상의 부호화 및 복호화 방법 및 장치 |
JP4588120B2 (ja) * | 2009-02-19 | 2010-11-24 | パナソニック株式会社 | 再生装置、記録方法、記録媒体再生システム |
WO2010095440A1 (fr) * | 2009-02-20 | 2010-08-26 | パナソニック株式会社 | Support d'enregistrement, dispositif de reproduction et circuit intégré |
CN105791861B (zh) * | 2009-04-20 | 2018-12-04 | 杜比实验室特许公司 | 定向内插和数据后处理 |
US8743178B2 (en) * | 2010-01-05 | 2014-06-03 | Dolby Laboratories Licensing Corporation | Multi-view video format control |
CN102783151A (zh) * | 2010-01-08 | 2012-11-14 | 诺基亚公司 | 用于视频编码的装置、方法和计算机程序 |
US9288505B2 (en) | 2011-08-11 | 2016-03-15 | Qualcomm Incorporated | Three-dimensional video with asymmetric spatial resolution |
EP2670139A1 (fr) * | 2012-06-01 | 2013-12-04 | Alcatel Lucent | Procédé et appareil permettant de coder un flux vidéo ayant un canal d'informations de transparence |
GB2503875B (en) * | 2012-06-29 | 2015-06-10 | Canon Kk | Method and device for encoding or decoding an image |
WO2014011103A1 (fr) * | 2012-07-10 | 2014-01-16 | Telefonaktiebolaget L M Ericsson (Publ) | Procédés et dispositifs pour permettre la synthèse de vues |
JP6150277B2 (ja) | 2013-01-07 | 2017-06-21 | 国立研究開発法人情報通信研究機構 | 立体映像符号化装置、立体映像復号化装置、立体映像符号化方法、立体映像復号化方法、立体映像符号化プログラム及び立体映像復号化プログラム |
RU2667605C2 (ru) * | 2013-05-10 | 2018-09-21 | Конинклейке Филипс Н.В. | Способ кодирования сигнала видеоданных для использования с многовидовым устройством визуализации |
WO2015015058A1 (fr) * | 2013-07-31 | 2015-02-05 | Nokia Corporation | Procédé et appareil de codage et de décodage vidéo |
US9906804B2 (en) * | 2014-01-16 | 2018-02-27 | Qualcomm Incorporated | Reference layer sample position derivation for scalable video coding |
US9716884B2 (en) * | 2014-03-20 | 2017-07-25 | Hfi Innovation Inc. | Method of signaling for mode selection in 3D and multi-view video coding |
FR3019384B1 (fr) * | 2014-03-25 | 2018-01-12 | Universite Grenoble Alpes | Reacteur implantable biocompatible |
US20170195677A1 (en) * | 2014-05-22 | 2017-07-06 | Mediatek Inc. | Method of Intra Block Copy with Flipping for Image and Video Coding |
US10075640B2 (en) * | 2015-12-31 | 2018-09-11 | Sony Corporation | Motion compensation for image sensor with a block based analog-to-digital converter |
JP6597385B2 (ja) | 2016-02-24 | 2019-10-30 | 沖電気工業株式会社 | 符号化装置及びプログラム、復号装置及びプログラム、並びに、配信システム |
FR3058019A1 (fr) * | 2016-10-21 | 2018-04-27 | Orange | Procede de codage et de decodage de parametres d'image, dispositif de codage et de decodage de parametres d'image et programmes d'ordinateur correspondants |
US11200675B2 (en) | 2017-02-20 | 2021-12-14 | Sony Corporation | Image processing apparatus and image processing method |
CN107371032B (zh) * | 2017-07-10 | 2019-11-22 | 上海国茂数字技术有限公司 | 一种多视点立体视频分层编码方法和解码方法 |
-
2018
- 2018-10-01 FR FR1859067A patent/FR3086831A1/fr not_active Withdrawn
-
2019
- 2019-09-25 KR KR1020217009472A patent/KR20210066825A/ko unknown
- 2019-09-25 CN CN202410246938.5A patent/CN118317113A/zh active Pending
- 2019-09-25 US US17/281,463 patent/US11653025B2/en active Active
- 2019-09-25 EP EP19795265.8A patent/EP3861751A1/fr active Pending
- 2019-09-25 CN CN201980064753.XA patent/CN112806015B/zh active Active
- 2019-09-25 JP JP2021517806A patent/JP7520823B2/ja active Active
- 2019-09-25 BR BR112021005927-0A patent/BR112021005927A2/pt unknown
- 2019-09-25 WO PCT/FR2019/052254 patent/WO2020070409A1/fr active Application Filing
-
2023
- 2023-04-05 US US18/296,073 patent/US11973981B2/en active Active
Patent Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20100259595A1 (en) * | 2009-04-10 | 2010-10-14 | Nokia Corporation | Methods and Apparatuses for Efficient Streaming of Free View Point Video |
Non-Patent Citations (5)
Title |
---|
A VETRO ET AL: "Overview of the Stereo and Multiview Video Coding Extensions of the H.264/MPEG-4 AVC Standard", PROCEEDINGS OF THE IEEE, vol. 99, no. 4, 1 April 2011 (2011-04-01), pages 626 - 642, XP055132298, ISSN: 0018-9219, DOI: 10.1109/JPROC.2010.2098830 * |
CHRISTOPH FEHN ET AL: "Asymmetric Coding of Stereoscopic Video for Transmission Over T-DMB", 3DTV CONFERENCE, 2007, IEEE, PI, 1 May 2007 (2007-05-01), pages 1 - 4, XP031158203, ISBN: 978-1-4244-0721-7 * |
JONGBEOM JEONG ET AL: "3DoF+ 360 Video Location-Based Asymmetric Down-Sampling for View Synthesis to Immersive VR Video Streaming", SENSORS, vol. 18, no. 9, 18 September 2018 (2018-09-18), pages 3148, XP055591982, DOI: 10.3390/s18093148 * |
MAREK DOMANSKI ET AL: "3D video compression by coding of disoccluded regions", IMAGE PROCESSING (ICIP), 2012 19TH IEEE INTERNATIONAL CONFERENCE ON, IEEE, 30 September 2012 (2012-09-30), pages 1317 - 1320, XP032333422, ISBN: 978-1-4673-2534-9, DOI: 10.1109/ICIP.2012.6467110 * |
RAY BAPPADITYA ET AL: "On the possibility to achieve 6-DoF for 360 video using divergent multi- view content", 2018 26TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO), EURASIP, 3 September 2018 (2018-09-03), pages 211 - 215, XP033461782, DOI: 10.23919/EUSIPCO.2018.8553397 * |
Also Published As
Publication number | Publication date |
---|---|
US20210409766A1 (en) | 2021-12-30 |
JP2022503963A (ja) | 2022-01-12 |
CN112806015B (zh) | 2024-03-26 |
US11973981B2 (en) | 2024-04-30 |
CN118317113A (zh) | 2024-07-09 |
KR20210066825A (ko) | 2021-06-07 |
US20230239503A1 (en) | 2023-07-27 |
EP3861751A1 (fr) | 2021-08-11 |
FR3086831A1 (fr) | 2020-04-03 |
US11653025B2 (en) | 2023-05-16 |
BR112021005927A2 (pt) | 2021-06-29 |
JP7520823B2 (ja) | 2024-07-23 |
CN112806015A (zh) | 2021-05-14 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP0547696B1 (fr) | Système de transmission et/ou stockage de signaux correspondant à des images texturées | |
EP3878170B1 (fr) | Synthèse de vues | |
WO2012150407A1 (fr) | Procédé de codage et de décodage d'images intégrales, dispositif de codage et de décodage d'images intégrales et programmes d'ordinateur correspondants | |
EP3788789A2 (fr) | Procede et dispositif de traitement d'images et procede et dispositif de decodage d'une video multi-vue adaptés | |
EP2347590A1 (fr) | Prediction d'une image par compensation en mouvement en avant | |
EP3649778B1 (fr) | Procédé de codage et décodage d'images, dispositif de codage et décodage et programmes d'ordinateur correspondants | |
EP3198876B1 (fr) | Génération et codage d'images intégrales résiduelles | |
WO2020070409A1 (fr) | Codage et décodage d'une vidéo omnidirectionnelle | |
WO2021214395A1 (fr) | Procédés et dispositifs de codage et de décodage d'une séquence vidéo multi-vues | |
EP3725080B1 (fr) | Procédés et dispositifs de codage et de décodage d'une séquence vidéo multi-vues représentative d'une vidéo omnidirectionnelle | |
WO2018073523A1 (fr) | Procédé de codage et de décodage de paramètres d'image, dispositif de codage et de décodage de paramètres d'image et programmes d'ordinateur correspondants | |
WO2020188172A1 (fr) | Procédés et dispositifs de codage et de décodage d'une séquence vidéo multi-vues | |
WO2019008253A1 (fr) | Procédé de codage et décodage d'images, dispositif de codage et décodage et programmes d'ordinateur correspondants | |
EP4085613A1 (fr) | Synthese iterative de vues a partir de donnees d'une video multi-vues | |
EP3991401A1 (fr) | Procédé et dispositif de traitement de données de video multi-vues | |
WO2021160955A1 (fr) | Procédé et dispositif de traitement de données de vidéo multi-vues | |
EP4360319A1 (fr) | Procédé de construction d'une image de profondeur d'une vidéo multi-vues, procédé de décodage d'un flux de données représentatif d'une vidéo multi-vues, procédé de codage, dispositifs, système, équipement terminal, signal et programmes d'ordinateur correspondants | |
WO2022069809A1 (fr) | Codage et decodage d'une video multi-vues | |
WO2013156718A1 (fr) | Procédé de codage et de décodage d'images intégrales, dispositif de codage et de décodage d'images intégrales et programmes d'ordinateur correspondants |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 19795265 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2021517806 Country of ref document: JP Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2101001890 Country of ref document: TH |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
REG | Reference to national code |
Ref country code: BR Ref legal event code: B01A Ref document number: 112021005927 Country of ref document: BR |
|
ENP | Entry into the national phase |
Ref document number: 2019795265 Country of ref document: EP Effective date: 20210503 |
|
ENP | Entry into the national phase |
Ref document number: 112021005927 Country of ref document: BR Kind code of ref document: A2 Effective date: 20210326 |