US11308575B2 - Omnidirectional image processing method and device - Google Patents

Omnidirectional image processing method and device Download PDF

Info

Publication number
US11308575B2
US11308575B2 US17/044,265 US201917044265A US11308575B2 US 11308575 B2 US11308575 B2 US 11308575B2 US 201917044265 A US201917044265 A US 201917044265A US 11308575 B2 US11308575 B2 US 11308575B2
Authority
US
United States
Prior art keywords
omnidirectional image
region
sphere
dimensional coordinate
format information
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US17/044,265
Other versions
US20210073938A1 (en
Inventor
Lu Yu
Xuchang HUANGFU
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang University ZJU
Original Assignee
Zhejiang University ZJU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang University ZJU filed Critical Zhejiang University ZJU
Assigned to ZHEJIANG UNIVERSITY reassignment ZHEJIANG UNIVERSITY ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: HUANGFU, Xuchang, YU, LU
Publication of US20210073938A1 publication Critical patent/US20210073938A1/en
Application granted granted Critical
Publication of US11308575B2 publication Critical patent/US11308575B2/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/12Panospheric to cylindrical image transformations
    • G06T3/0062
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/597Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T9/00Image coding
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T9/00Image coding
    • G06T9/001Model-based coding, e.g. wire frame
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/85Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression

Definitions

  • the present invention relates to the technical field of omnidirectional image processing, and more particularly relates to an omnidirectional image processing method and device.
  • a 360° omnidirectional image refers to all the scenes around an observation point in space which are composed of all the light that can be received by this observation point, and a sphere can describe all the scenes around the observation point. Since spherical images are difficult to store, and an existing image codec is designed for common non-omnidirectional images and is not ideal for the coding effect of the spherical images, it is necessary to convert the spherical images into omnidirectional image formats through projection, and the common projection formats include Equirectangular Projection (ERP), Cubemap Projection (CMP), etc.
  • ERP Equirectangular Projection
  • CMP Cubemap Projection
  • an existing coding mode for common non-omnidirectional images such as HEVC, AVS2, AV1, etc. may be adopted for coding, and a locally adjusted coding method may also be adopted for different formats.
  • ERP sampling is shown in FIG. 1 , and points on a sphere may be expressed with longitude ⁇ and latitude ⁇ .
  • the sphere is mapped to a rectangle with an aspect ratio of 2:1 in a ⁇ - ⁇ coordinate system, and then the rectangle is uniformly sampled.
  • the ERP format has oversampling and image texture overstretching issues in the high-latitude areas.
  • a sphere is projected to six faces of a cube through a perspective projection, and then the cube is uniformly sampled.
  • the CMP samples the sphere more uniformly, and has higher coding efficiency.
  • the corners of each face of the cube have a higher sampling density than the central region.
  • the sampling uniformity of the sphere is insufficient.
  • the omnidirectional image may be sampled, so as to generate a pixel value for each pixel position, and then the omnidirectional image is coded and decoded.
  • the position mapping relationship between the sphere and the omnidirectional image may also be utilized to perform format conversion of the omnidirectional image, and the ERP format may be converted to the CMP format or other formats.
  • the present invention provides a novel omnidirectional image processing method and device.
  • CMP Cubemap Projection
  • a main idea of the present invention is to enable the arc length of the unit interval at each latitude in an omnidirectional image to be as equal as possible on the sphere by adjusting a stretching relationship between the omnidirectional image and the sphere in different positions, thereby reducing information redundancy in the marginal regions in Cubemap projection, and improving the representation efficiency.
  • a first objective of the present invention is to provide an omnidirectional image processing method, including:
  • mapping relationship between a position (x, y) in at least one region in an omnidirectional image obtained by decoding the code stream and a position ( ⁇ , ⁇ ) on a sphere according to the format information, wherein the mapping relationship is:
  • x represents a first dimensional coordinate position of the region in the decoded omnidirectional image
  • y represents a second dimensional coordinate position of the region in the decoded omnidirectional image
  • represents a longitude position of the sphere
  • represents a latitude position of the sphere
  • a line mapped to the sphere by a first dimensional coordinate axis is an equator line
  • a line mapped to the sphere by a second dimensional coordinate axis is a prime meridian.
  • the omnidirectional image format information is one of the following:
  • the format information is a format number, and the format number specifies a direction of a first dimensional coordinate of the region in the decoded omnidirectional image by default.
  • the format information includes a format number and direction information of the first dimensional coordinate of the region in the decoded omnidirectional image.
  • a second objective of the present invention is to provide an omnidirectional image processing method, including the following steps:
  • mapping relationship expressing at least one image region in an omnidirectional image by the following mapping relationship, wherein the mapping relationship between a position (x, y) in the image region and a position ( ⁇ , ⁇ ) on a sphere is:
  • x represents a first dimensional coordinate position of the region in the omnidirectional image
  • y represents a second dimensional coordinate position of the region in the omnidirectional image
  • represents a longitude position of the sphere
  • represents a latitude position of the sphere
  • a line mapped to the sphere by a first dimensional coordinate axis is an equator line
  • a line mapped to the sphere by a second dimensional coordinate axis is a prime meridian
  • the omnidirectional image format information is one of the following:
  • the format information is a format number, and the format number specifies a direction of a first dimensional coordinate of the region in the omnidirectional image by default.
  • the format information includes a format number and direction information of the first dimensional coordinate of the region in the omnidirectional image.
  • a third objective of the present invention is to provide an omnidirectional image processing method, including the following steps:
  • mapping relationships between a position (x1, y1) in a region 1 and a position (x2, y2) in a region 2 and their respective corresponding positions ( ⁇ 1, ⁇ 1) and ( ⁇ 2, ⁇ 2) on a sphere according to the format information since at least one pair of adjacent regions, i.e. the region 1 and the region 2, is present in a decoded omnidirectional image, wherein the mapping relationships are:
  • x1 and ⁇ 1 are in linear relationship, and y1 and
  • x2 and ⁇ 2 are in linear relationship, and y2 and
  • x1 represents a first dimensional coordinate position of the region 1 in the omnidirectional image
  • y1 represents a second dimensional coordinate position of the image region 1
  • ⁇ 1 represents a longitude position of the sphere corresponding to the position (x1, y1) in the region 1 in the omnidirectional image
  • ⁇ 1 represents a latitude position of the sphere corresponding to the position (x1, y1) in the region 1 in the omnidirectional image
  • x2 represents a first dimensional coordinate position of the region 2 in the omnidirectional image
  • y2 represents a second dimensional coordinate position of the region 2 in the omnidirectional image
  • ⁇ 2 represents a longitude position of the sphere corresponding to the position (x2, y2) in the image region 2
  • ⁇ 2 represents a latitude position of the sphere corresponding to the position (x2, y2) in the region 2 in the omnidirectional image
  • a line mapped to the sphere by a first dimensional coordinate axis of the region 1 is the same as
  • a fourth objective of the present invention is to provide an omnidirectional image processing device, including the following modules:
  • a format information extraction module wherein input of the format information extraction module is a coding code stream, output of the format information extraction module is omnidirectional image format information, and the module decodes the coding code stream to obtain the omnidirectional image format information;
  • a position mapping module wherein input of the position mapping module is the omnidirectional image format information, output of the position mapping module is a mapping relationship between a position (x, y) in at least one region in an omnidirectional image obtained by decoding the code stream and a position ( ⁇ , ⁇ ) on a sphere, and the mapping relationship is:
  • x represents a first dimensional coordinate position of the region in the omnidirectional image
  • y represents a second dimensional coordinate position of the region in the omnidirectional image
  • represents a longitude position of the sphere
  • represents a latitude position of the sphere
  • a line mapped to the sphere by a first dimensional coordinate axis is an equator line
  • a line mapped to the sphere by a second dimensional coordinate axis is a prime meridian.
  • the omnidirectional image format information is one of the following:
  • the format information is a format number, and the format number specifies a direction of a first dimensional coordinate of the region in the omnidirectional image by default.
  • the format information includes a format number and direction information of the first dimensional coordinate of the region in the omnidirectional image.
  • a fifth objective of the present invention is to provide an omnidirectional image processing device, including the following modules:
  • a position mapping module wherein output of the position mapping module is omnidirectional image format information, and a mapping relationship between a position (x, y) in at least one region in an omnidirectional image and a position ( ⁇ , ⁇ ) on a sphere is determined in the module as:
  • x represents a first dimensional coordinate position of the region in the omnidirectional image
  • y represents a second dimensional coordinate position of the region in the omnidirectional image
  • represents a longitude position of the sphere
  • represents a latitude position of the sphere
  • a line mapped to the sphere by a first dimensional coordinate axis is an equator line
  • a line mapped to the sphere by a second dimensional coordinate axis is a prime meridian
  • a format information and image coding module wherein input of the format information and image coding module is the omnidirectional image format information and the omnidirectional image expressed by the mapping relationship determined by the format information, output of the format information and image coding module is a code stream including the omnidirectional image and the format information of the omnidirectional image, and the module codes and writes the omnidirectional image and the format information to the code stream.
  • the omnidirectional image format information is one of the following:
  • the format information is a format number, and the format number specifies a direction of a first dimensional coordinate of the region in the omnidirectional image by default.
  • the format information includes a format number and direction information of the first dimensional coordinate of the region in the omnidirectional image.
  • the present invention enables the arc length of the unit interval at each latitude in the omnidirectional image to be as equal as possible on the sphere, and the omnidirectional image to be distributed more uniformly on the sphere, so that the spherical uniformity of expression is improved, the sampling loss is reduced under the same sample number, and the coding efficiency is improved.
  • FIG. 1 is an Equirectangular Projection (ERP) format
  • FIG. 2 is a schematic diagram of a mapping relationship of projecting a position on a sphere to a cube
  • FIG. 3 is position relationships on an unfolded face of a cube in an embodiment of the present invention.
  • FIG. 4 is a coordinate system on a face of a cube in an embodiment of the present invention.
  • FIG. 5 is a coordinate system on a face of a cube in an embodiment of the present invention.
  • FIG. 6 is an effective region corresponding to a face of a cube on a sphere in an embodiment of the present invention.
  • FIG. 7 is a schematic diagram of sampling with equal arc length in a horizontal direction on the equator
  • FIG. 8 is an unfolded image of a cube generated by using a processing method of this patent in an embodiment of the present invention.
  • FIG. 9 is a schematic diagram of two adjacent regions arranged horizontally in an embodiment of the present invention.
  • FIG. 10 is a schematic diagram of arrangement of six faces on an unfolded face of a cube in an embodiment of the present invention.
  • FIG. 11 is a schematic diagram of two adjacent regions arranged vertically in an embodiment of the present invention.
  • FIG. 12 is a source ERP image for generating an unfolded image of a cube in an embodiment of the present invention.
  • FIG. 13 is an unfolded image of a cube generated through format conversion from a source ERP image by using a processing method of this patent in an embodiment of the present invention
  • FIG. 14 is a schematic diagram of a stretching ratio of different mapping relations in an embodiment of the present invention.
  • FIG. 15 is a processing device in an embodiment of the present invention.
  • FIG. 16 is a processing device in an embodiment of the present invention.
  • FIG. 17 is a schematic diagram of arrangement of each face and an effective region of a CMP format in an embodiment of the present invention.
  • FIG. 18 is a schematic diagram of arrangement of each face and an effective region of a CMP format in an embodiment of the present invention.
  • the embodiment of the present invention provides an omnidirectional image processing method.
  • a code stream is decoded to obtain omnidirectional image coding format information.
  • a mapping relationship between a position (x, y) in at least one region in an omnidirectional image obtained by decoding the code stream and a position ( ⁇ , ⁇ ) on a sphere is determined according to the format information.
  • the mapping relationship is: x and ⁇ are in linear relationship, and y and
  • x represents a first dimensional coordinate position of the region in the decoded omnidirectional image
  • y represents a second dimensional coordinate position of the region in the decoded omnidirectional image
  • represents a longitude position of the sphere
  • represents a latitude position of the sphere
  • a line mapped to the sphere by a first dimensional coordinate axis is an equator line
  • a line mapped to the sphere by a second dimensional coordinate axis is a prime meridian.
  • the decoded omnidirectional image is in a Cubemap Projection (CMP) format
  • the region is a front face of the cube in FIG. 3
  • the mapping relationship between (x, y) and ( ⁇ , ⁇ ) is:
  • the format information is a format number, and this format number specifies by default that a direction of a first dimensional coordinate of the region in the omnidirectional image is a horizontal direction, as shown in FIG. 4 .
  • the embodiment of the present invention provides an omnidirectional image processing method.
  • a code stream is decoded to obtain omnidirectional image coding format information.
  • a mapping relationship between a position (x, y) in at least one region in an omnidirectional image obtained by decoding the code stream and a position ( ⁇ , ⁇ ) on a sphere is determined according to the format information.
  • the mapping relationship is:
  • x represents a first dimensional coordinate position of the region in the decoded omnidirectional image
  • y represents a second dimensional coordinate position of the region in the decoded omnidirectional image
  • represents a longitude position of the sphere
  • represents a latitude position of the sphere
  • a line mapped to the sphere by a first dimensional coordinate axis is an equator line
  • a line mapped to the sphere by a second dimensional coordinate axis is a prime meridian.
  • the decoded omnidirectional image is in a CMP format
  • the region is a front face of the cube in FIG. 3
  • the mapping relationship between (x, y) and ( ⁇ , p) is:
  • x 4 ⁇ ⁇ ⁇
  • y ⁇ tan - 1 ⁇ ⁇ cos ⁇ ⁇ ⁇ .
  • the format information includes a format number and direction information of a first dimensional coordinate of the region, and this direction information specifies that a direction of the first dimensional coordinate of the region in the omnidirectional image is a vertical direction, as shown in FIG. 5 .
  • the embodiment of the present invention provides an omnidirectional image processing method.
  • At least one image region in an omnidirectional image is expressed by the following mapping relationship.
  • the mapping relationship between a position (x, y) in the image region and a position ( ⁇ , ⁇ ) on a sphere is:
  • x represents a first dimensional coordinate position of the region in the omnidirectional image
  • y represents a second dimensional coordinate position of the region in the omnidirectional image
  • represents a longitude position of the sphere
  • represents a latitude position of the sphere
  • a line mapped to the sphere by a first dimensional coordinate axis is an equator line
  • a line mapped to the sphere by a second dimensional coordinate axis is a prime meridian.
  • the omnidirectional image is in a CMP format
  • the region is a front face of the cube in FIG. 3
  • the mapping relationship between (x, y) and ( ⁇ , ⁇ ) is:
  • x 4 ⁇ ⁇ ⁇
  • y ⁇ tan - 1 ⁇ ⁇ cos ⁇ ⁇ ⁇ .
  • format information is a format number, and this format number specifies by default that a direction of a first dimensional coordinate of the region in the omnidirectional image is a horizontal direction, as shown in FIG. 4 .
  • the omnidirectional image expressed by the mapping relationship is coded and written to a code stream.
  • the format information of the mapping relationship is also written to the code stream.
  • the sequence of writing the omnidirectional image and the format information to the code stream may be arbitrary. That is, the omnidirectional image may be coded and written to the code stream first, and then the format information of the mapping relationship is also written to the code stream. Or, the format information of the mapping relationship may be written to the code stream first, and then the omnidirectional image expressed by the mapping relationship is coded and written to the code stream.
  • the embodiment of the present invention provides an omnidirectional image processing method.
  • At least one image region in an omnidirectional image is expressed by the following mapping relationship.
  • the mapping relationship between a position (x, y) in the image region and a position ( ⁇ , ⁇ ) on a sphere is:
  • x represents a first dimensional coordinate position of the region in the omnidirectional image
  • y represents a second dimensional coordinate position of the region in the omnidirectional image
  • represents a longitude position of the sphere
  • represents a latitude position of the sphere
  • a line mapped to the sphere by a first dimensional coordinate axis is an equator line
  • a line mapped to the sphere by a second dimensional coordinate axis is a prime meridian.
  • the omnidirectional image is in a CMP format
  • the region in the omnidirectional image is a front face of the cube in FIG. 3
  • the mapping relationship between (x, y) and ( ⁇ , ⁇ ) is:
  • x 4 ⁇ ⁇ ⁇
  • y ⁇ tan - 1 ⁇ ⁇ cos ⁇ ⁇ ⁇ .
  • format information includes a format number and direction information of a first dimensional coordinate of the region, and this direction information specifies that a direction of the first dimensional coordinate of the region in the omnidirectional image is a vertical direction, as shown in FIG. 5 .
  • the omnidirectional image expressed by the mapping relationship is coded and written to a code stream.
  • the format information of the mapping relationship is also written to the code stream.
  • the embodiment of the present invention provides a method of sampling by using a mapping relationship.
  • At least one image region is included in a to-be-generated omnidirectional image.
  • the mapping relationship between a position (x, y) in the image region and a position ( ⁇ , ⁇ ) on a sphere is.
  • x represents a first dimensional coordinate position of the region in the omnidirectional image
  • y represents a second dimensional coordinate position of the region in the omnidirectional image
  • represents a longitude position of the sphere
  • represents a latitude position of the sphere
  • a line mapped to the sphere by a first dimensional coordinate axis is an equator line
  • a line mapped to the sphere by a second dimensional coordinate axis is a prime meridian.
  • the to-be-generated omnidirectional image is in a CMP format
  • the region is a front face of the cube in FIG. 3
  • the mapping relationship between (x, y) and ( ⁇ , ⁇ ) is:
  • x 4 ⁇ ⁇ ⁇
  • y ⁇ tan - 1 ⁇ ⁇ cos ⁇ ⁇ ⁇ .
  • Pixel sampling positions in a vertical direction are shown in FIG. 6 , each meridian is sampled with equal arc length, but the arc lengths between adjacent sampling points on different meridians are different.
  • Pixel sampling positions in a horizontal direction are shown in FIG. 7 , the equator is sampled with equal arc length, and other positions are sampled with approximately equal arc length.
  • the equator is sampled with equal arc length in the horizontal direction, and then meridians corresponding to all the horizontal sampling positions are sampled with equal arc length in the vertical direction.
  • each pixel point in the to-be-generated omnidirectional image is interpolated on the sphere to obtain its corresponding pixel value.
  • FIG. 8 shows an image generated according to the position mapping relationship.
  • the embodiment of the present invention provides an omnidirectional image processing method.
  • a code stream is decoded to obtain omnidirectional image coding format information.
  • mapping relationships between a position (x1, y1) in a region 1 and a position (x2, y2) in a region 2 and their respective corresponding positions ( ⁇ 1, ⁇ 1) and ( ⁇ 2, ⁇ 2) on a sphere are determined according to the format information since at least one pair of adjacent regions, i.e. the region 1 and the region 2, is present in a decoded omnidirectional image.
  • the mapping relationships are:
  • x1 and ⁇ 1 are in linear relationship, and y1 and
  • x2 and ⁇ 2 are in linear relationship, and y2 and
  • x1 represents a first dimensional coordinate position of the region 1 in the omnidirectional image
  • y1 represents a second dimensional coordinate position of the image region 1
  • ⁇ 1 represents a longitude position of the sphere corresponding to the position (x1, y1) in the region 1 in the omnidirectional image
  • ⁇ 1 represents a latitude position of the sphere corresponding to the position (x1, y1) in the region 1 in the omnidirectional image
  • x2 represents a first dimensional coordinate position of the region 2 in the omnidirectional image
  • y2 represents a second dimensional coordinate position of the region 2 in the omnidirectional image
  • ⁇ 2 represents a longitude position of the sphere corresponding to the position (x2, y2) in the image region 2
  • ⁇ 2 represents a latitude position of the sphere corresponding to the position (x2, y2) in the region 2 in the omnidirectional image
  • a line mapped to the sphere by a first dimensional coordinate axis of the region 1 is the same as
  • a front face and a right face in FIG. 3 are taken as the adjacent regions in the omnidirectional image, the front face is the region 1, and the right face is the region 2.
  • the relationship between (x1, y1) and ( ⁇ 1, ⁇ 1) is:
  • x ⁇ 1 4 ⁇ ⁇ ⁇ ⁇ ⁇ 1
  • y ⁇ ⁇ 1 ⁇ ⁇ 1 tan - 1 ⁇ ⁇ cos ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ 1 .
  • x ⁇ ⁇ 2 4 ⁇ ⁇ ⁇ ⁇ ⁇ 2
  • y ⁇ ⁇ 2 ⁇ 2 tan - 1 ⁇ ⁇ cos ⁇ ⁇ ⁇ ⁇ ⁇ 2 .
  • first dimensional directions of the region 1 and the region 2 are both horizontal directions, as shown in FIG. 9 .
  • the embodiment of the present invention provides an omnidirectional image processing method.
  • a code stream is decoded to obtain omnidirectional image coding format information.
  • mapping relationships between a position (x1, y1) in a region 1 and a position (x2, y2) in a region 2 and their respective corresponding positions ( ⁇ 1 , ⁇ 1) and ( ⁇ 2 , ⁇ 2) on a sphere are determined according to the format information since at least one pair of adjacent regions, i.e. the region 1 and the region 2, is present in a decoded omnidirectional image.
  • the mapping relationships are:
  • x1 and ⁇ 1 are in linear relationship, and y1 and
  • x2 and ⁇ 2 are in linear relationship, and y2 and
  • x1 represents a first dimensional coordinate position of the region 1 in the omnidirectional image
  • y1 represents a second dimensional coordinate position of the image region 1
  • ⁇ 1 represents a longitude position of the sphere corresponding to the position (x1, y1) in the region 1 in the omnidirectional image
  • ⁇ 1 represents a latitude position of the sphere corresponding to the position (x1, y1) in the region 1 in the omnidirectional image
  • x2 represents a first dimensional coordinate position of the region 2 in the omnidirectional image
  • y2 represents a second dimensional coordinate position of the region 2 in the omnidirectional image
  • ⁇ 2 represents a longitude position of the sphere corresponding to the position (x2, y2) in the image region 2
  • ⁇ 2 represents a latitude position of the sphere corresponding to the position (x2, y2) in the region 2 in the omnidirectional image
  • a line mapped to the sphere by a first dimensional coordinate axis of the region 1 is the same as
  • a front face and a right face in FIG. 10 are taken as the adjacent regions in the omnidirectional image, the front face is the region 1, and the right face is the region 2.
  • the relationship between (x1, y1) and ( ⁇ 1, ⁇ 1) is:
  • x ⁇ ⁇ 1 4 ⁇ ⁇ ⁇ ⁇ ⁇ 1
  • y ⁇ ⁇ 1 ⁇ ⁇ 1 tan - 1 ⁇ ⁇ cos ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ 1 .
  • x ⁇ ⁇ 2 4 ⁇ ⁇ ⁇ 2
  • y ⁇ ⁇ 2 ⁇ ⁇ 2 tan - 1 ⁇ ⁇ cos ⁇ ⁇ ⁇ ⁇ ⁇ 2 .
  • first dimensional directions of the region 1 and the region 2 are both vertical directions, as shown in FIG. 11 .
  • the embodiment of the present invention provides an omnidirectional image format conversion method.
  • At least one image region is included in a to-be-generated omnidirectional image.
  • a mapping relationship between a position (x, y) in the image region and a position ( ⁇ , ⁇ ) on a sphere is:
  • x represents a first dimensional coordinate position of the region in the omnidirectional image
  • y represents a second dimensional coordinate position of the region in the omnidirectional image
  • represents a longitude position of the sphere
  • represents a latitude position of the sphere
  • a line mapped to the sphere by a first dimensional coordinate axis is an equator line
  • a line mapped to the sphere by a second dimensional coordinate axis is a prime meridian.
  • the omnidirectional image is in a CMP format
  • the region is a front face of the cube in FIG. 3
  • the mapping relationship between (x, y) and ( ⁇ , ⁇ ) is:
  • x 4 ⁇ ⁇ ⁇
  • y ⁇ tan - 1 ⁇ ⁇ cos ⁇ ⁇ ⁇ .
  • a first dimensional direction is a horizontal direction, as shown in FIG. 4 .
  • a position on the sphere is mapped to a source format.
  • Interpolation is performed on the ERP to generate pixel values in the to-be-generated omnidirectional image.
  • FIG. 12 shows a source format image of the ERP
  • FIG. 13 show an omnidirectional image in an unfolded CMP format.
  • the embodiment of the present invention provides an omnidirectional image processing method.
  • a code stream is decoded to obtain omnidirectional image coding format information.
  • a mapping relationship between a position (x, y) in at least one region in an omnidirectional image obtained by decoding the code stream and a position ( ⁇ , ⁇ ) on a sphere is determined according to the format information.
  • the mapping relationship is:
  • x represents a first dimensional coordinate position of the region in the decoded omnidirectional image
  • y represents a second dimensional coordinate position of the region in the decoded omnidirectional image
  • represents a longitude position of the sphere
  • represents a latitude position of the sphere
  • a line mapped to the sphere by a first dimensional coordinate axis is an equator line
  • a line mapped to the sphere by a second dimensional coordinate axis is a prime meridian.
  • mapping relationship between (x, y) and ( ⁇ , ⁇ ) is:
  • mapping relationship between (x′, y′) and ( ⁇ , ⁇ ) is:
  • x ′ 4 ⁇ ⁇ ⁇
  • y ′ ⁇ tan - 1 ⁇ ⁇ cos ⁇ ⁇ ⁇ .
  • mapping relationship between the position (x′, y′) in the omnidirectional image and the position ( ⁇ , ⁇ ) on the sphere can be derived as:
  • x ′ 4 ⁇ ⁇ ⁇
  • y ′ ⁇ tan - 1 ⁇ ⁇ cos ⁇ ⁇ ⁇ ⁇ 2 + 1 2 .
  • the embodiment of the present invention provides an omnidirectional image processing device.
  • a format information extraction module wherein input of the format information extraction module is a coding code stream, output of the format information extraction module is omnidirectional image format information, and the module decodes the coding code stream to obtain the omnidirectional image format information;
  • a position mapping module wherein input of the position mapping module is the omnidirectional image format information, output of the position mapping module is a mapping relationship between a position (x, y) in at least one region in an omnidirectional image obtained by decoding the code stream and a position ( ⁇ , ⁇ ) on a sphere, and the mapping relationship is:
  • x represents a first dimensional coordinate position of the region in the omnidirectional image
  • y represents a second dimensional coordinate position of the region in the omnidirectional image
  • represents a longitude position of the sphere
  • represents a latitude position of the sphere
  • a line mapped to the sphere by a first dimensional coordinate axis is an equator line
  • a line mapped to the sphere by a second dimensional coordinate axis is a prime meridian.
  • the omnidirectional image is in a CMP format
  • the region is a front face of the cube in FIG. 3
  • the mapping relationship between (x, y) and ( ⁇ , ⁇ ) is:
  • x 4 ⁇ ⁇ ⁇
  • y ⁇ tan - 1 ⁇ ⁇ cos ⁇ ⁇ ⁇ .
  • the format information is a format number, and this format number specifies by default that a direction of a first dimensional coordinate of the region in the omnidirectional image is a horizontal direction, as shown in FIG. 4 .
  • FIG. 15 shows a schematic diagram of this processing device.
  • the embodiment of the present invention provides an omnidirectional image processing device.
  • a format information extraction module wherein input of the format information extraction module is a coding code stream, output of the format information extraction module is omnidirectional image format information, and the module decodes the coding code stream to obtain the omnidirectional image format information;
  • a position mapping module wherein input of the position mapping module is the omnidirectional image format information, output of the position mapping module is a mapping relationship between a position (x, y) in at least one region in an omnidirectional image obtained by decoding the code stream and a position ( ⁇ , ⁇ ) on a sphere, and the mapping relationship is:
  • x represents a first dimensional coordinate position of the region in the omnidirectional image
  • y represents a second dimensional coordinate position of the region in the omnidirectional image
  • represents a longitude position of the sphere
  • represents a latitude position of the sphere
  • a line mapped to the sphere by a first dimensional coordinate axis is an equator line
  • a line mapped to the sphere by a second dimensional coordinate axis is a prime meridian.
  • the omnidirectional image is in a CMP format
  • the region is a front face of the cube in FIG. 3
  • the mapping relationship between (x, y) and ( ⁇ , ⁇ ) is:
  • the format information includes a format number and direction information of a first dimensional coordinate of the region in the omnidirectional image, and this direction information specifies that a direction of the first dimensional coordinate of the region in the omnidirectional image is a vertical direction, as shown in FIG. 5 .
  • FIG. 15 shows a schematic diagram of this processing device.
  • the embodiment of the present invention provides an omnidirectional image processing device.
  • a position mapping module wherein output of the position mapping module is omnidirectional image format information, and a mapping relationship between a position (x, y) in at least one region in an omnidirectional image and a position ( ⁇ , ⁇ ) on a sphere is determined in the module as:
  • x represents a first dimensional coordinate position of the region in the omnidirectional image
  • y represents a second dimensional coordinate position of the region in the omnidirectional image
  • represents a longitude position of the sphere
  • represents a latitude position of the sphere
  • a line mapped to the sphere by a first dimensional coordinate axis is an equator line
  • a line mapped to the sphere by a second dimensional coordinate axis is a prime meridian
  • a format information and image coding module wherein input of the format information and image coding module is the omnidirectional image format information and the omnidirectional image expressed by the mapping relationship determined by the format information, output of the format information and image coding module is a code stream including the omnidirectional image and the format information of the omnidirectional image, and the module codes and writes the omnidirectional image and the format information to the code stream.
  • the omnidirectional image is in a CMP format
  • the region is a front face of the cube in FIG. 3
  • the mapping relationship between (x, y) and ( ⁇ , ⁇ ) is:
  • x 4 ⁇ ⁇ ⁇
  • y ⁇ tan - 1 ⁇ ⁇ cos ⁇ ⁇ ⁇ .
  • the format information is a format number, and this format number specifies by default that a direction of a first dimensional coordinate of the region in the omnidirectional image is a horizontal direction, as shown in FIG. 4 .
  • FIG. 16 shows a schematic diagram of this processing device.
  • the embodiment of the present invention provides an omnidirectional image processing device.
  • a position mapping module wherein output of the position mapping module is omnidirectional image format information, and a mapping relationship between a position (x, y) in at least one region in an omnidirectional image and a position ( ⁇ , ⁇ ) on a sphere is determined in the module as:
  • x represents a first dimensional coordinate position of the region in the omnidirectional image
  • y represents a second dimensional coordinate position of the region in the omnidirectional image
  • represents a longitude position of the sphere
  • represents a latitude position of the sphere
  • a line mapped to the sphere by a first dimensional coordinate axis is an equator line
  • a line mapped to the sphere by a second dimensional coordinate axis is a prime meridian
  • a format information and image coding module wherein input of the format information and image coding module is the omnidirectional image format information and the omnidirectional image expressed by the mapping relationship determined by the format information, output of the format information and image coding module is a code stream including the omnidirectional image and the format information of the omnidirectional image, and the module codes and writes the omnidirectional image and the format information to the code stream.
  • the omnidirectional image is in a CMP format
  • the region is a front face of the cube in FIG. 3
  • the mapping relationship between (x, y) and ( ⁇ , ⁇ ) is:
  • x 4 ⁇ ⁇ ⁇
  • y ⁇ tan - 1 ⁇ ⁇ cos ⁇ ⁇ ⁇ .
  • the format information includes a format number and direction information of a first dimensional coordinate of the region, and this direction information specifies that a direction of the first dimensional coordinate of the region in the omnidirectional image is a vertical direction, as shown in FIG. 5 .
  • FIG. 16 shows a schematic diagram of this processing device.
  • the embodiment of the present invention provides an omnidirectional image generation method.
  • At least one image region is included in a to-be-generated omnidirectional image.
  • a mapping relationship between a position (x, y) in the image region and a position ( ⁇ , ⁇ ) on a sphere is specified as:
  • x represents a first dimensional coordinate position of the region in the omnidirectional image
  • y represents a second dimensional coordinate position of the region in the omnidirectional image
  • represents a longitude position of the sphere
  • represents a latitude position of the sphere
  • a line mapped to the sphere by a first dimensional coordinate axis is an equator line
  • a line mapped to the sphere by a second dimensional coordinate axis is a prime meridian.
  • the to-be-generated omnidirectional image is in a CMP format.
  • coordinate systems are respectively established on six faces of the cube.
  • a first dimensional direction is a horizontal direction.
  • mapping relationship between (x, y) and ( ⁇ , ⁇ ) is:
  • the region enclosed by thick lines represents an effective region in the face of the cube, the height of this region is L′, and the side length of the face of the cube is L.
  • a first dimensional direction is the horizontal direction.
  • mapping relationship between (x, y) and ( ⁇ , ⁇ ) is:
  • the embodiment of the present invention provides an omnidirectional image generation method.
  • At least one image region is included in a to-be-generated omnidirectional image.
  • a mapping relationship between a position (x, y) in the image region and a position ( ⁇ , ⁇ ) on a sphere is:
  • x represents a first dimensional coordinate position of the region in the omnidirectional image
  • y represents a second dimensional coordinate position of the region in the omnidirectional image
  • represents a longitude position of the sphere
  • represents a latitude position of the sphere
  • a line mapped to the sphere by a first dimensional coordinate axis is an equator line
  • a line mapped to the sphere by a second dimensional coordinate axis is a prime meridian.
  • the to-be-generated omnidirectional image is in a CMP format.
  • coordinate systems are respectively established on six faces of the cube.
  • a first dimensional direction is a horizontal direction.
  • mapping relationship between (x, y) and ( ⁇ , ⁇ ) is:
  • the region enclosed by thick lines represents an effective region in the face of the cube, the height of this region is L′, and the side length of the face of the cube is L.
  • the region outside the thick lines is a non-effective region, the pixels in this region are generated by interpolation on the sphere according to the same mapping relationship as in the effective region, and at this time, the value of y is less than ⁇ 0.8.
  • the pixels in non-effective regions of the left and right faces are generated by interpolation on the sphere according to the same mapping relationship as in the corresponding effective regions.
  • a first dimensional direction is a vertical direction.
  • mapping relationship between (x, y) and ( ⁇ , ⁇ ) is:
  • x 3.6 ⁇ ⁇ ⁇ + 0.1
  • y ⁇ tan - 1 ⁇ cos ⁇ ⁇ ⁇ .
  • the region outside the thick lines is a non-effective region
  • the pixels in this region are generated by interpolation on the sphere according to the same mapping relationship as in the effective region, and at this time, the value of x is less than ⁇ 0.8.
  • the pixels in non-effective regions of the upper and lower faces are also generated by interpolation on the sphere according to the same mapping relationship as in the corresponding effective regions.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Signal Processing (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)
  • Image Processing (AREA)
  • Studio Devices (AREA)

Abstract

The present invention discloses an omnidirectional image processing method and device. Format information is extracted from a code stream, positions in a decoded omnidirectional image is mapped to positions on a sphere by using a special mapping relationship according to the format information, and the decoded omnidirectional image at least includes one region which meets the special mapping relationship. The present invention can enable the regions in the decoded omnidirectional image to be distributed more uniformly than the sphere, thereby improving representation efficiency, and reducing conversion distortion. While designing the omnidirectional image processing method, the present invention also designs the corresponding device.

Description

TECHNICAL FIELD
The present invention relates to the technical field of omnidirectional image processing, and more particularly relates to an omnidirectional image processing method and device.
BACKGROUND ART
A 360° omnidirectional image refers to all the scenes around an observation point in space which are composed of all the light that can be received by this observation point, and a sphere can describe all the scenes around the observation point. Since spherical images are difficult to store, and an existing image codec is designed for common non-omnidirectional images and is not ideal for the coding effect of the spherical images, it is necessary to convert the spherical images into omnidirectional image formats through projection, and the common projection formats include Equirectangular Projection (ERP), Cubemap Projection (CMP), etc. For omnidirectional images in different formats, an existing coding mode for common non-omnidirectional images such as HEVC, AVS2, AV1, etc. may be adopted for coding, and a locally adjusted coding method may also be adopted for different formats.
ERP sampling is shown in FIG. 1, and points on a sphere may be expressed with longitude φ and latitude θ. The sphere is mapped to a rectangle with an aspect ratio of 2:1 in a φ-θ coordinate system, and then the rectangle is uniformly sampled. The ERP format has oversampling and image texture overstretching issues in the high-latitude areas.
As shown in FIG. 2, a sphere is projected to six faces of a cube through a perspective projection, and then the cube is uniformly sampled. Compared with the ERP, the CMP samples the sphere more uniformly, and has higher coding efficiency. However, due to the characteristics of the perspective projection, the corners of each face of the cube have a higher sampling density than the central region. Correspondingly, for the sphere, the sampling uniformity of the sphere is insufficient.
After determining a position mapping relationship between the sphere and the omnidirectional image, the omnidirectional image may be sampled, so as to generate a pixel value for each pixel position, and then the omnidirectional image is coded and decoded. The position mapping relationship between the sphere and the omnidirectional image may also be utilized to perform format conversion of the omnidirectional image, and the ERP format may be converted to the CMP format or other formats.
SUMMARY OF THE INVENTION
In view of the defects in the prior art, the present invention provides a novel omnidirectional image processing method and device.
For the same area, in a Cubemap Projection (CMP) format, the central region of each face corresponds to a larger region on a sphere than marginal regions, which leads to non-uniform sampling on the sphere and redundant information in the marginal regions of the cube, and reduces the representation efficiency.
A main idea of the present invention is to enable the arc length of the unit interval at each latitude in an omnidirectional image to be as equal as possible on the sphere by adjusting a stretching relationship between the omnidirectional image and the sphere in different positions, thereby reducing information redundancy in the marginal regions in Cubemap projection, and improving the representation efficiency.
For this purpose, the present invention adopts the following technical solution:
A first objective of the present invention is to provide an omnidirectional image processing method, including:
decoding a code stream to obtain omnidirectional image coding format information; and determining a mapping relationship between a position (x, y) in at least one region in an omnidirectional image obtained by decoding the code stream and a position (θ, φ) on a sphere according to the format information, wherein the mapping relationship is:
x and φ are in linear relationship, and y and
θ tan - 1 cos φ
are in linear relationship,
where x represents a first dimensional coordinate position of the region in the decoded omnidirectional image, y represents a second dimensional coordinate position of the region in the decoded omnidirectional image, φ represents a longitude position of the sphere, θ represents a latitude position of the sphere, a line mapped to the sphere by a first dimensional coordinate axis is an equator line, and a line mapped to the sphere by a second dimensional coordinate axis is a prime meridian.
Preferably, the omnidirectional image format information is one of the following:
1) The format information is a format number, and the format number specifies a direction of a first dimensional coordinate of the region in the decoded omnidirectional image by default.
2) The format information includes a format number and direction information of the first dimensional coordinate of the region in the decoded omnidirectional image.
A second objective of the present invention is to provide an omnidirectional image processing method, including the following steps:
expressing at least one image region in an omnidirectional image by the following mapping relationship, wherein the mapping relationship between a position (x, y) in the image region and a position (θ, φ) on a sphere is:
x and φ are in linear relationship, and y and
θ tan - 1 cos φ
are in linear relationship,
where x represents a first dimensional coordinate position of the region in the omnidirectional image, y represents a second dimensional coordinate position of the region in the omnidirectional image, φ represents a longitude position of the sphere, θ represents a latitude position of the sphere, a line mapped to the sphere by a first dimensional coordinate axis is an equator line, and a line mapped to the sphere by a second dimensional coordinate axis is a prime meridian;
coding and writing the omnidirectional image expressed by the mapping relationship to a code stream; and
writing format information of the mapping relationship to the code stream.
Preferably, the omnidirectional image format information is one of the following:
1) The format information is a format number, and the format number specifies a direction of a first dimensional coordinate of the region in the omnidirectional image by default.
2) The format information includes a format number and direction information of the first dimensional coordinate of the region in the omnidirectional image.
A third objective of the present invention is to provide an omnidirectional image processing method, including the following steps:
decoding a code stream to obtain omnidirectional image coding format information; and
determining mapping relationships between a position (x1, y1) in a region 1 and a position (x2, y2) in a region 2 and their respective corresponding positions (θ1, φ1) and (θ2, φ2) on a sphere according to the format information since at least one pair of adjacent regions, i.e. the region 1 and the region 2, is present in a decoded omnidirectional image, wherein the mapping relationships are:
x1 and φ1 are in linear relationship, and y1 and
θ 1 tan - 1 cos φ 1
are in linear relationship;
x2 and φ2 are in linear relationship, and y2 and
θ 2 tan - 1 cos φ2
are in linear relationship,
where x1 represents a first dimensional coordinate position of the region 1 in the omnidirectional image, y1 represents a second dimensional coordinate position of the image region 1, φ1 represents a longitude position of the sphere corresponding to the position (x1, y1) in the region 1 in the omnidirectional image, and θ1 represents a latitude position of the sphere corresponding to the position (x1, y1) in the region 1 in the omnidirectional image; x2 represents a first dimensional coordinate position of the region 2 in the omnidirectional image, y2 represents a second dimensional coordinate position of the region 2 in the omnidirectional image, φ2 represents a longitude position of the sphere corresponding to the position (x2, y2) in the image region 2, and θ2 represents a latitude position of the sphere corresponding to the position (x2, y2) in the region 2 in the omnidirectional image; and a line mapped to the sphere by a first dimensional coordinate axis of the region 1 is the same as a line mapped to the sphere by a first dimensional coordinate axis of the region 2, and is an equator line of the sphere.
A fourth objective of the present invention is to provide an omnidirectional image processing device, including the following modules:
a format information extraction module, wherein input of the format information extraction module is a coding code stream, output of the format information extraction module is omnidirectional image format information, and the module decodes the coding code stream to obtain the omnidirectional image format information; and
a position mapping module, wherein input of the position mapping module is the omnidirectional image format information, output of the position mapping module is a mapping relationship between a position (x, y) in at least one region in an omnidirectional image obtained by decoding the code stream and a position (θ, φ) on a sphere, and the mapping relationship is:
x and φ are in linear relationship, and y and
θ tan - 1 cos φ
are in linear relationship,
where x represents a first dimensional coordinate position of the region in the omnidirectional image, y represents a second dimensional coordinate position of the region in the omnidirectional image, φ represents a longitude position of the sphere, θ represents a latitude position of the sphere, a line mapped to the sphere by a first dimensional coordinate axis is an equator line, and a line mapped to the sphere by a second dimensional coordinate axis is a prime meridian.
Preferably, the omnidirectional image format information is one of the following:
1) The format information is a format number, and the format number specifies a direction of a first dimensional coordinate of the region in the omnidirectional image by default.
2) The format information includes a format number and direction information of the first dimensional coordinate of the region in the omnidirectional image.
A fifth objective of the present invention is to provide an omnidirectional image processing device, including the following modules:
a position mapping module, wherein output of the position mapping module is omnidirectional image format information, and a mapping relationship between a position (x, y) in at least one region in an omnidirectional image and a position (θ, φ) on a sphere is determined in the module as:
x and φ are in linear relationship, and y and
θ tan - 1 cos φ
are in linear relationship,
where x represents a first dimensional coordinate position of the region in the omnidirectional image, y represents a second dimensional coordinate position of the region in the omnidirectional image, φ represents a longitude position of the sphere, θ represents a latitude position of the sphere, a line mapped to the sphere by a first dimensional coordinate axis is an equator line, and a line mapped to the sphere by a second dimensional coordinate axis is a prime meridian; and
a format information and image coding module, wherein input of the format information and image coding module is the omnidirectional image format information and the omnidirectional image expressed by the mapping relationship determined by the format information, output of the format information and image coding module is a code stream including the omnidirectional image and the format information of the omnidirectional image, and the module codes and writes the omnidirectional image and the format information to the code stream.
Preferably, the omnidirectional image format information is one of the following:
1) The format information is a format number, and the format number specifies a direction of a first dimensional coordinate of the region in the omnidirectional image by default.
2) The format information includes a format number and direction information of the first dimensional coordinate of the region in the omnidirectional image.
Compared with the prior art, by designing the special position mapping relationship between the sphere and the omnidirectional image, the present invention enables the arc length of the unit interval at each latitude in the omnidirectional image to be as equal as possible on the sphere, and the omnidirectional image to be distributed more uniformly on the sphere, so that the spherical uniformity of expression is improved, the sampling loss is reduced under the same sample number, and the coding efficiency is improved.
BRIEF DESCRIPTION OF THE DRAWINGS
In combination with the accompanying drawings, the principle of the present invention can be explained from embodiments given below.
The accompanying drawings illustrated here are to provide further understanding of the present invention and constitute a part of this application. Various embodiments of the present invention are merely to explain the present invention and are only some particular cases, and the application range of the present invention is not limited to these embodiments. In the drawings:
FIG. 1 is an Equirectangular Projection (ERP) format;
FIG. 2 is a schematic diagram of a mapping relationship of projecting a position on a sphere to a cube;
FIG. 3 is position relationships on an unfolded face of a cube in an embodiment of the present invention;
FIG. 4 is a coordinate system on a face of a cube in an embodiment of the present invention;
FIG. 5 is a coordinate system on a face of a cube in an embodiment of the present invention;
FIG. 6 is an effective region corresponding to a face of a cube on a sphere in an embodiment of the present invention;
FIG. 7 is a schematic diagram of sampling with equal arc length in a horizontal direction on the equator;
FIG. 8 is an unfolded image of a cube generated by using a processing method of this patent in an embodiment of the present invention;
FIG. 9 is a schematic diagram of two adjacent regions arranged horizontally in an embodiment of the present invention;
FIG. 10 is a schematic diagram of arrangement of six faces on an unfolded face of a cube in an embodiment of the present invention;
FIG. 11 is a schematic diagram of two adjacent regions arranged vertically in an embodiment of the present invention;
FIG. 12 is a source ERP image for generating an unfolded image of a cube in an embodiment of the present invention;
FIG. 13 is an unfolded image of a cube generated through format conversion from a source ERP image by using a processing method of this patent in an embodiment of the present invention;
FIG. 14 is a schematic diagram of a stretching ratio of different mapping relations in an embodiment of the present invention;
FIG. 15 is a processing device in an embodiment of the present invention;
FIG. 16 is a processing device in an embodiment of the present invention;
FIG. 17 is a schematic diagram of arrangement of each face and an effective region of a CMP format in an embodiment of the present invention; and
FIG. 18 is a schematic diagram of arrangement of each face and an effective region of a CMP format in an embodiment of the present invention.
DETAILED DESCRIPTION OF THE INVENTION Embodiment 1
The embodiment of the present invention provides an omnidirectional image processing method.
The omnidirectional image processing method provided by the present embodiment includes the following steps:
A code stream is decoded to obtain omnidirectional image coding format information. A mapping relationship between a position (x, y) in at least one region in an omnidirectional image obtained by decoding the code stream and a position (θ, φ) on a sphere is determined according to the format information. The mapping relationship is: x and φ are in linear relationship, and y and
θ tan - 1 cos φ
are in linear relationship.
x represents a first dimensional coordinate position of the region in the decoded omnidirectional image, y represents a second dimensional coordinate position of the region in the decoded omnidirectional image, φ represents a longitude position of the sphere, θ represents a latitude position of the sphere, a line mapped to the sphere by a first dimensional coordinate axis is an equator line, and a line mapped to the sphere by a second dimensional coordinate axis is a prime meridian.
In the present embodiment, the decoded omnidirectional image is in a Cubemap Projection (CMP) format, the region is a front face of the cube in FIG. 3, and the mapping relationship between (x, y) and (θ, φ) is:
x = 4 π × φ , and y = θ tan - 1 cos φ .
In this embodiment, the format information is a format number, and this format number specifies by default that a direction of a first dimensional coordinate of the region in the omnidirectional image is a horizontal direction, as shown in FIG. 4.
Embodiment 2
The embodiment of the present invention provides an omnidirectional image processing method.
The omnidirectional image processing method provided by the present embodiment includes the following steps:
A code stream is decoded to obtain omnidirectional image coding format information. A mapping relationship between a position (x, y) in at least one region in an omnidirectional image obtained by decoding the code stream and a position (θ, φ) on a sphere is determined according to the format information. The mapping relationship is:
x and φ are in linear relationship, and y and
θ tan - 1 cos φ
are in linear relationship.
x represents a first dimensional coordinate position of the region in the decoded omnidirectional image, y represents a second dimensional coordinate position of the region in the decoded omnidirectional image, φ represents a longitude position of the sphere, θ represents a latitude position of the sphere, a line mapped to the sphere by a first dimensional coordinate axis is an equator line, and a line mapped to the sphere by a second dimensional coordinate axis is a prime meridian.
In the present embodiment, the decoded omnidirectional image is in a CMP format, the region is a front face of the cube in FIG. 3, and the mapping relationship between (x, y) and (θ, p) is:
x = 4 π × φ , and y = θ tan - 1 cos φ .
In this embodiment, the format information includes a format number and direction information of a first dimensional coordinate of the region, and this direction information specifies that a direction of the first dimensional coordinate of the region in the omnidirectional image is a vertical direction, as shown in FIG. 5.
Embodiment 3
The embodiment of the present invention provides an omnidirectional image processing method.
The omnidirectional image processing method provided by the present embodiment includes the following steps:
At least one image region in an omnidirectional image is expressed by the following mapping relationship. The mapping relationship between a position (x, y) in the image region and a position (θ, φ) on a sphere is:
x and φ are in linear relationship, and y and
θ tan - 1 cos φ
are in linear relationship.
x represents a first dimensional coordinate position of the region in the omnidirectional image, y represents a second dimensional coordinate position of the region in the omnidirectional image, φ represents a longitude position of the sphere, θ represents a latitude position of the sphere, a line mapped to the sphere by a first dimensional coordinate axis is an equator line, and a line mapped to the sphere by a second dimensional coordinate axis is a prime meridian.
In the present embodiment, the omnidirectional image is in a CMP format, the region is a front face of the cube in FIG. 3, and the mapping relationship between (x, y) and (θ, φ) is:
x = 4 π × φ , and y = θ tan - 1 cos φ .
In this embodiment, format information is a format number, and this format number specifies by default that a direction of a first dimensional coordinate of the region in the omnidirectional image is a horizontal direction, as shown in FIG. 4.
The omnidirectional image expressed by the mapping relationship is coded and written to a code stream.
The format information of the mapping relationship is also written to the code stream.
The sequence of writing the omnidirectional image and the format information to the code stream may be arbitrary. That is, the omnidirectional image may be coded and written to the code stream first, and then the format information of the mapping relationship is also written to the code stream. Or, the format information of the mapping relationship may be written to the code stream first, and then the omnidirectional image expressed by the mapping relationship is coded and written to the code stream.
Embodiment 4
The embodiment of the present invention provides an omnidirectional image processing method.
The omnidirectional image processing method provided by the present embodiment includes the following steps:
At least one image region in an omnidirectional image is expressed by the following mapping relationship. The mapping relationship between a position (x, y) in the image region and a position (θ, φ) on a sphere is:
x and φ are in linear relationship, and y and
θ tan - 1 cos φ
are in linear relationship.
x represents a first dimensional coordinate position of the region in the omnidirectional image, y represents a second dimensional coordinate position of the region in the omnidirectional image, φ represents a longitude position of the sphere, θ represents a latitude position of the sphere, a line mapped to the sphere by a first dimensional coordinate axis is an equator line, and a line mapped to the sphere by a second dimensional coordinate axis is a prime meridian.
In the present embodiment, the omnidirectional image is in a CMP format, the region in the omnidirectional image is a front face of the cube in FIG. 3, and the mapping relationship between (x, y) and (θ, φ) is:
x = 4 π × φ , and y = θ tan - 1 cos φ .
In this embodiment, format information includes a format number and direction information of a first dimensional coordinate of the region, and this direction information specifies that a direction of the first dimensional coordinate of the region in the omnidirectional image is a vertical direction, as shown in FIG. 5.
The omnidirectional image expressed by the mapping relationship is coded and written to a code stream.
The format information of the mapping relationship is also written to the code stream.
Embodiment 5
The embodiment of the present invention provides a method of sampling by using a mapping relationship.
The sampling method provided by the present embodiment includes the following steps:
At least one image region is included in a to-be-generated omnidirectional image. The mapping relationship between a position (x, y) in the image region and a position (θ, φ) on a sphere is.
x and φ are in linear relationship, and y and
θ tan - 1 cos φ
are in linear relationship.
x represents a first dimensional coordinate position of the region in the omnidirectional image, y represents a second dimensional coordinate position of the region in the omnidirectional image, φ represents a longitude position of the sphere, θ represents a latitude position of the sphere, a line mapped to the sphere by a first dimensional coordinate axis is an equator line, and a line mapped to the sphere by a second dimensional coordinate axis is a prime meridian.
In the present embodiment, the to-be-generated omnidirectional image is in a CMP format, the region is a front face of the cube in FIG. 3, and the mapping relationship between (x, y) and (θ, φ) is:
x = 4 π × φ , and y = θ tan - 1 cos φ .
Pixel sampling positions in a vertical direction are shown in FIG. 6, each meridian is sampled with equal arc length, but the arc lengths between adjacent sampling points on different meridians are different. Pixel sampling positions in a horizontal direction are shown in FIG. 7, the equator is sampled with equal arc length, and other positions are sampled with approximately equal arc length.
During sampling, first, the equator is sampled with equal arc length in the horizontal direction, and then meridians corresponding to all the horizontal sampling positions are sampled with equal arc length in the vertical direction.
According to the pixel sampling positions in the to-be-generated omnidirectional image, each pixel point in the to-be-generated omnidirectional image is interpolated on the sphere to obtain its corresponding pixel value.
FIG. 8 shows an image generated according to the position mapping relationship.
Embodiment 6
The embodiment of the present invention provides an omnidirectional image processing method.
The omnidirectional image processing method provided by the present embodiment includes the following steps:
A code stream is decoded to obtain omnidirectional image coding format information.
Mapping relationships between a position (x1, y1) in a region 1 and a position (x2, y2) in a region 2 and their respective corresponding positions (θ1, φ1) and (θ2, φ2) on a sphere are determined according to the format information since at least one pair of adjacent regions, i.e. the region 1 and the region 2, is present in a decoded omnidirectional image. The mapping relationships are:
x1 and φ1 are in linear relationship, and y1 and
θ 1 tan - 1 cos φ 1
are in linear relationship;
x2 and φ2 are in linear relationship, and y2 and
θ2 tan - 1 cos φ 2
are in linear relationship.
x1 represents a first dimensional coordinate position of the region 1 in the omnidirectional image, y1 represents a second dimensional coordinate position of the image region 1, φ1 represents a longitude position of the sphere corresponding to the position (x1, y1) in the region 1 in the omnidirectional image, and θ1 represents a latitude position of the sphere corresponding to the position (x1, y1) in the region 1 in the omnidirectional image; x2 represents a first dimensional coordinate position of the region 2 in the omnidirectional image, y2 represents a second dimensional coordinate position of the region 2 in the omnidirectional image, φ2 represents a longitude position of the sphere corresponding to the position (x2, y2) in the image region 2, and θ2 represents a latitude position of the sphere corresponding to the position (x2, y2) in the region 2 in the omnidirectional image; and a line mapped to the sphere by a first dimensional coordinate axis of the region 1 is the same as a line mapped to the sphere by a first dimensional coordinate axis of the region 2, and is an equator line of the sphere.
In this embodiment, a front face and a right face in FIG. 3 are taken as the adjacent regions in the omnidirectional image, the front face is the region 1, and the right face is the region 2. The relationship between (x1, y1) and (φ1, θ1) is:
x 1 = 4 π × φ 1 , and y 1 = θ 1 tan - 1 cos φ 1 .
The relationship between (x2, y2) and (φ2, θ2) is:
x 2 = 4 π × φ 2 , and y 2 = θ2 tan - 1 cos φ 2 .
In this embodiment, first dimensional directions of the region 1 and the region 2 are both horizontal directions, as shown in FIG. 9.
Embodiment 7
The embodiment of the present invention provides an omnidirectional image processing method.
The omnidirectional image processing method provided by the present embodiment includes the following steps:
A code stream is decoded to obtain omnidirectional image coding format information.
Mapping relationships between a position (x1, y1) in a region 1 and a position (x2, y2) in a region 2 and their respective corresponding positions (θ1, φ1) and (θ2, φ2) on a sphere are determined according to the format information since at least one pair of adjacent regions, i.e. the region 1 and the region 2, is present in a decoded omnidirectional image. The mapping relationships are:
x1 and φ1 are in linear relationship, and y1 and
θ 1 tan - 1 cos φ 1
are in linear relationship;
x2 and φ2 are in linear relationship, and y2 and
θ2 tan - 1 cos φ 2
are in linear relationship.
x1 represents a first dimensional coordinate position of the region 1 in the omnidirectional image, y1 represents a second dimensional coordinate position of the image region 1, φ1 represents a longitude position of the sphere corresponding to the position (x1, y1) in the region 1 in the omnidirectional image, and θ1 represents a latitude position of the sphere corresponding to the position (x1, y1) in the region 1 in the omnidirectional image; x2 represents a first dimensional coordinate position of the region 2 in the omnidirectional image, y2 represents a second dimensional coordinate position of the region 2 in the omnidirectional image, φ2 represents a longitude position of the sphere corresponding to the position (x2, y2) in the image region 2, and θ2 represents a latitude position of the sphere corresponding to the position (x2, y2) in the region 2 in the omnidirectional image; and a line mapped to the sphere by a first dimensional coordinate axis of the region 1 is the same as a line mapped to the sphere by a first dimensional coordinate axis of the region 2, and is an equator line of the sphere.
In this embodiment, a front face and a right face in FIG. 10 are taken as the adjacent regions in the omnidirectional image, the front face is the region 1, and the right face is the region 2. The relationship between (x1, y1) and (φ1, θ1) is:
x 1 = 4 π × φ 1 , and y 1 = θ 1 tan - 1 cos φ 1 .
The relationship between (x2, y2) and (φ2, θ2) is:
x 2 = 4 π × φ2 , and y 2 = θ 2 tan - 1 cos φ 2 .
In this embodiment, first dimensional directions of the region 1 and the region 2 are both vertical directions, as shown in FIG. 11.
Embodiment 8
The embodiment of the present invention provides an omnidirectional image format conversion method.
The omnidirectional image format conversion method provided by the present embodiment includes the following steps:
At least one image region is included in a to-be-generated omnidirectional image. A mapping relationship between a position (x, y) in the image region and a position (θ, φ) on a sphere is:
x and φ are in linear relationship, and y and
θ tan - 1 cos φ
are in linear relationship.
x represents a first dimensional coordinate position of the region in the omnidirectional image, y represents a second dimensional coordinate position of the region in the omnidirectional image, φ represents a longitude position of the sphere, θ represents a latitude position of the sphere, a line mapped to the sphere by a first dimensional coordinate axis is an equator line, and a line mapped to the sphere by a second dimensional coordinate axis is a prime meridian.
In the present embodiment, the omnidirectional image is in a CMP format, the region is a front face of the cube in FIG. 3, and the mapping relationship between (x, y) and (θ, φ) is:
x = 4 π × φ , and y = θ tan - 1 cos φ .
In this embodiment, a first dimensional direction is a horizontal direction, as shown in FIG. 4.
A position on the sphere is mapped to a source format. An Equirectangular Projection (ERP) is taken as the source format, and a mapping relationship between the position (u, v) in the ERP and the position (θ, φ) on the sphere is:
ϕ=(u−0.5)*(2*π),and
θ=(0.5−v)*π.
The values of u and v are both in [0,1].
Interpolation is performed on the ERP to generate pixel values in the to-be-generated omnidirectional image.
FIG. 12 shows a source format image of the ERP, and FIG. 13 show an omnidirectional image in an unfolded CMP format.
Embodiment 9
The embodiment of the present invention provides an omnidirectional image processing method.
The omnidirectional image processing method provided by the present embodiment includes the following steps:
A code stream is decoded to obtain omnidirectional image coding format information. A mapping relationship between a position (x, y) in at least one region in an omnidirectional image obtained by decoding the code stream and a position (θ, φ) on a sphere is determined according to the format information. The mapping relationship is:
x and φ are in linear relationship, and y and
θ tan - 1 cos φ
are in linear relationship.
x represents a first dimensional coordinate position of the region in the decoded omnidirectional image, y represents a second dimensional coordinate position of the region in the decoded omnidirectional image, φ represents a longitude position of the sphere, θ represents a latitude position of the sphere, a line mapped to the sphere by a first dimensional coordinate axis is an equator line, and a line mapped to the sphere by a second dimensional coordinate axis is a prime meridian.
In this embodiment, the region in the omnidirectional image is a rear face of the cube in FIG. 3. Coordinate systems shown in FIG. 14 are established on this face. In the X-Y coordinate system, when y=−1,
θ = - π 4 ,
and when y=1,
θ = π 2 ,
then
the mapping relationship between (x, y) and (θ, φ) is:
x = 4 π × φ , and y = 2 θ tan - 1 cos φ - 3 .
In the X′-Y′ coordinate system, when y′=−1,
θ = - π 4 ,
and when y′=1,
θ = π 4 ,
then the mapping relationship between (x′, y′) and (θ, φ) is:
x = 4 π × φ , and y = θ tan - 1 cos φ .
The region where y=±1 is the effective content corresponding to the sphere in the omnidirectional image. Conversion between (x′, y′) and (x, y) is as follows:
x=x′, and
y=(y′−½)×2.
Through the above conversion, the mapping relationship between the position (x′, y′) in the omnidirectional image and the position (θ, φ) on the sphere can be derived as:
x = 4 π × φ , and y = θ tan - 1 cos φ × 2 + 1 2 .
Embodiment 10
The embodiment of the present invention provides an omnidirectional image processing device.
The omnidirectional image processing device provided by the present embodiment includes the following modules:
a format information extraction module, wherein input of the format information extraction module is a coding code stream, output of the format information extraction module is omnidirectional image format information, and the module decodes the coding code stream to obtain the omnidirectional image format information; and
a position mapping module, wherein input of the position mapping module is the omnidirectional image format information, output of the position mapping module is a mapping relationship between a position (x, y) in at least one region in an omnidirectional image obtained by decoding the code stream and a position (θ, φ) on a sphere, and the mapping relationship is:
x and φ are in linear relationship, and y and
θ tan - 1 cos φ
are in linear relationship,
where x represents a first dimensional coordinate position of the region in the omnidirectional image, y represents a second dimensional coordinate position of the region in the omnidirectional image, φ represents a longitude position of the sphere, θ represents a latitude position of the sphere, a line mapped to the sphere by a first dimensional coordinate axis is an equator line, and a line mapped to the sphere by a second dimensional coordinate axis is a prime meridian.
In the present embodiment, the omnidirectional image is in a CMP format, the region is a front face of the cube in FIG. 3, and the mapping relationship between (x, y) and (θ, φ) is:
x = 4 π × φ , and y = θ tan - 1 cos φ .
In this embodiment, the format information is a format number, and this format number specifies by default that a direction of a first dimensional coordinate of the region in the omnidirectional image is a horizontal direction, as shown in FIG. 4.
FIG. 15 shows a schematic diagram of this processing device.
Embodiment 11
The embodiment of the present invention provides an omnidirectional image processing device.
The omnidirectional image processing device provided by the present embodiment includes the following modules:
a format information extraction module, wherein input of the format information extraction module is a coding code stream, output of the format information extraction module is omnidirectional image format information, and the module decodes the coding code stream to obtain the omnidirectional image format information; and
a position mapping module, wherein input of the position mapping module is the omnidirectional image format information, output of the position mapping module is a mapping relationship between a position (x, y) in at least one region in an omnidirectional image obtained by decoding the code stream and a position (θ, φ) on a sphere, and the mapping relationship is:
x and φ are in linear relationship, and y and
θ tan - 1 cos φ
are in linear relationship,
where x represents a first dimensional coordinate position of the region in the omnidirectional image, y represents a second dimensional coordinate position of the region in the omnidirectional image, φ represents a longitude position of the sphere, θ represents a latitude position of the sphere, a line mapped to the sphere by a first dimensional coordinate axis is an equator line, and a line mapped to the sphere by a second dimensional coordinate axis is a prime meridian.
In the present embodiment, the omnidirectional image is in a CMP format, the region is a front face of the cube in FIG. 3, and the mapping relationship between (x, y) and (θ, φ) is:
x = 4 π × φ , and y = θ tan - 1 cos φ .
In this embodiment, the format information includes a format number and direction information of a first dimensional coordinate of the region in the omnidirectional image, and this direction information specifies that a direction of the first dimensional coordinate of the region in the omnidirectional image is a vertical direction, as shown in FIG. 5.
FIG. 15 shows a schematic diagram of this processing device.
Embodiment 12
The embodiment of the present invention provides an omnidirectional image processing device.
The omnidirectional image processing device provided by the present embodiment includes the following modules:
a position mapping module, wherein output of the position mapping module is omnidirectional image format information, and a mapping relationship between a position (x, y) in at least one region in an omnidirectional image and a position (θ, φ) on a sphere is determined in the module as:
x and φ are in linear relationship, and y and
θ tan - 1 cos φ
are in linear relationship,
where x represents a first dimensional coordinate position of the region in the omnidirectional image, y represents a second dimensional coordinate position of the region in the omnidirectional image, φ represents a longitude position of the sphere, θ represents a latitude position of the sphere, a line mapped to the sphere by a first dimensional coordinate axis is an equator line, and a line mapped to the sphere by a second dimensional coordinate axis is a prime meridian; and
a format information and image coding module, wherein input of the format information and image coding module is the omnidirectional image format information and the omnidirectional image expressed by the mapping relationship determined by the format information, output of the format information and image coding module is a code stream including the omnidirectional image and the format information of the omnidirectional image, and the module codes and writes the omnidirectional image and the format information to the code stream.
In the present embodiment, the omnidirectional image is in a CMP format, the region is a front face of the cube in FIG. 3, and the mapping relationship between (x, y) and (θ, φ) is:
x = 4 π × φ , and y = θ tan - 1 cos φ .
In this embodiment, the format information is a format number, and this format number specifies by default that a direction of a first dimensional coordinate of the region in the omnidirectional image is a horizontal direction, as shown in FIG. 4.
FIG. 16 shows a schematic diagram of this processing device.
Embodiment 13
The embodiment of the present invention provides an omnidirectional image processing device.
The omnidirectional image processing device provided by the present embodiment includes the following modules:
a position mapping module, wherein output of the position mapping module is omnidirectional image format information, and a mapping relationship between a position (x, y) in at least one region in an omnidirectional image and a position (θ, φ) on a sphere is determined in the module as:
x and φ are in linear relationship, and y and
θ tan - 1 cos φ
are in linear relationship,
where x represents a first dimensional coordinate position of the region in the omnidirectional image, y represents a second dimensional coordinate position of the region in the omnidirectional image, φ represents a longitude position of the sphere, θ represents a latitude position of the sphere, a line mapped to the sphere by a first dimensional coordinate axis is an equator line, and a line mapped to the sphere by a second dimensional coordinate axis is a prime meridian; and
a format information and image coding module, wherein input of the format information and image coding module is the omnidirectional image format information and the omnidirectional image expressed by the mapping relationship determined by the format information, output of the format information and image coding module is a code stream including the omnidirectional image and the format information of the omnidirectional image, and the module codes and writes the omnidirectional image and the format information to the code stream.
In the present embodiment, the omnidirectional image is in a CMP format, the region is a front face of the cube in FIG. 3, and the mapping relationship between (x, y) and (θ, φ) is:
x = 4 π × φ , and y = θ tan - 1 cos φ .
In this embodiment, the format information includes a format number and direction information of a first dimensional coordinate of the region, and this direction information specifies that a direction of the first dimensional coordinate of the region in the omnidirectional image is a vertical direction, as shown in FIG. 5.
FIG. 16 shows a schematic diagram of this processing device.
Embodiment 14
The embodiment of the present invention provides an omnidirectional image generation method.
The omnidirectional image generation method provided by the present embodiment includes the following steps:
At least one image region is included in a to-be-generated omnidirectional image. A mapping relationship between a position (x, y) in the image region and a position (θ, φ) on a sphere is specified as:
x and φ are in linear relationship, and y and
θ tan - 1 cos φ
are in linear relationship.
x represents a first dimensional coordinate position of the region in the omnidirectional image, y represents a second dimensional coordinate position of the region in the omnidirectional image, φ represents a longitude position of the sphere, θ represents a latitude position of the sphere, a line mapped to the sphere by a first dimensional coordinate axis is an equator line, and a line mapped to the sphere by a second dimensional coordinate axis is a prime meridian.
In the present embodiment, the to-be-generated omnidirectional image is in a CMP format. As shown in FIG. 17, coordinate systems are respectively established on six faces of the cube. For the faces on the first line of the CMP format, namely, the left, front and right faces, a first dimensional direction is a horizontal direction. When y=−0.8 and x=0
θ = - π 4 ,
and when y=1 and x=0,
θ = π 4 ,
then the mapping relationship between (x, y) and (θ, φ) is:
x = 4 π × φ , and y = 0 . 9 θ tan - 1 cos φ + 0 . 1 .
In FIG. 17, the region enclosed by thick lines represents an effective region in the face of the cube, the height of this region is L′, and the side length of the face of the cube is L. In the front face, the region outside the thick lines is a non-effective region, the pixels in this region are filled with the pixels in the line where y=−0.8, and the pixels in non-effective regions of the left and right faces are also filled with the pixels in the lines where y=−0.8 in the corresponding regions.
For the faces on the second line of the CMP format, namely, the upper, lower and rear faces, a first dimensional direction is the horizontal direction. When y=0.8 and x=0,
θ = π 4 ,
and when y=−1 and x=0,
θ = - π 4 ,
then the mapping relationship between (x, y) and (θ, φ) is:
x = 4 π × φ , and y = 0 . 9 θ tan - 1 cos φ - 0 . 1 .
In the rear face, the region outside the thick lines is a non-effective region, the pixels in this region are filled with the pixels in the line where y=0.8, and the pixels in non-effective regions of the upper and lower faces are also filled with the pixels in the lines where y=0.8 in the corresponding regions.
Embodiment 15
The embodiment of the present invention provides an omnidirectional image generation method.
The omnidirectional image generation method provided by the present embodiment includes the following steps:
At least one image region is included in a to-be-generated omnidirectional image. A mapping relationship between a position (x, y) in the image region and a position (θ, φ) on a sphere is:
x and φ are in linear relationship, and y and
θ tan - 1 cos φ
are in linear relationship.
x represents a first dimensional coordinate position of the region in the omnidirectional image, y represents a second dimensional coordinate position of the region in the omnidirectional image, φ represents a longitude position of the sphere, θ represents a latitude position of the sphere, a line mapped to the sphere by a first dimensional coordinate axis is an equator line, and a line mapped to the sphere by a second dimensional coordinate axis is a prime meridian.
In the present embodiment, the to-be-generated omnidirectional image is in a CMP format. As shown in FIG. 18, coordinate systems are respectively established on six faces of the cube. For the faces on the first line of the CMP format, namely, the left, front and right faces, a first dimensional direction is a horizontal direction. When y=−0.8 and x=0,
θ = - π 4 ,
and when y=1 and x=0,
θ = π 4 ,
then the mapping relationship between (x, y) and (θ, φ) is:
x = 4 π × φ , and y = 0 . 9 θ tan - 1 cos φ + 0 . 1 .
In FIG. 18, the region enclosed by thick lines represents an effective region in the face of the cube, the height of this region is L′, and the side length of the face of the cube is L. In the front face, the region outside the thick lines is a non-effective region, the pixels in this region are generated by interpolation on the sphere according to the same mapping relationship as in the effective region, and at this time, the value of y is less than −0.8. The pixels in non-effective regions of the left and right faces are generated by interpolation on the sphere according to the same mapping relationship as in the corresponding effective regions.
For the faces on the second line of the CMP format, namely, the upper, lower and rear faces, a first dimensional direction is a vertical direction. When x=−0.8,
φ = - π 4 ,
and when x=1,
φ = π 4 ,
then the mapping relationship between (x, y) and (θ, φ) is:
x = 3.6 π × φ + 0.1 , and y = θ tan - 1 cos φ .
In the rear face, the region outside the thick lines is a non-effective region, the pixels in this region are generated by interpolation on the sphere according to the same mapping relationship as in the effective region, and at this time, the value of x is less than −0.8. The pixels in non-effective regions of the upper and lower faces are also generated by interpolation on the sphere according to the same mapping relationship as in the corresponding effective regions.
Finally, it should be noted that the above embodiments are merely to illustrate the technical solutions of the present invention, and not to limit them. Although the present invention has been illustrated in detail with reference to the foregoing embodiments, those of ordinary skill in the art should understand that they can still make modifications on the technical solutions recorded in various embodiments, or equivalent replacements on some of the technical features. However, these modifications or replacements do not make the nature of the corresponding technical solutions depart from the scope of the technical solutions of various embodiments in the present invention.

Claims (9)

The invention claimed is:
1. An omnidirectional image processing method, comprising:
decoding a code stream to obtain omnidirectional image coding format information; and
determining a mapping relationship between a position (x, y) in at least one region in an omnidirectional image obtained by decoding the code stream and a position (θ, φ) on a sphere according to the format information, wherein the mapping relationship is:
x and φ are in linear relationship, and y and
θ tan - 1 cos φ
are in linear relationship,
where x represents a first dimensional coordinate position of the region in the decoded omnidirectional image, y represents a second dimensional coordinate position of the region in the decoded omnidirectional image, φ represents a longitude position of the sphere, θ represents a latitude position of the sphere, a line mapped to the sphere by a first dimensional coordinate axis is an equator line, and a line mapped to the sphere by a second dimensional coordinate axis is a prime meridian.
2. The omnidirectional image processing method according to claim 1, wherein the format information is one of the following:
1) the format information is a format number, and the format number specifies a direction of a first dimensional coordinate of the region in the decoded omnidirectional image by default; and
2) the format information comprises a format number and direction information of the first dimensional coordinate of the region in the decoded omnidirectional image.
3. An omnidirectional image processing method, comprising the following steps:
expressing at least one image region in an omnidirectional image by the following mapping relationship, wherein the mapping relationship between a position (x, y) in the image region and a position (θ, φ) on a sphere is:
x and φ are in linear relationship, and y and
θ tan - 1 cos φ
are in linear relationship,
where x represents a first dimensional coordinate position of the region in the omnidirectional image, y represents a second dimensional coordinate position of the region in the omnidirectional image, φ represents a longitude position of the sphere, θ represents a latitude position of the sphere, a line mapped to the sphere by a first dimensional coordinate axis is an equator line, and a line mapped to the sphere by a second dimensional coordinate axis is a prime meridian;
coding and writing the omnidirectional image expressed by the mapping relationship to a code stream; and
writing format information of the mapping relationship to the code stream.
4. The omnidirectional image processing method according to claim 3, wherein the format information is one of the following:
1) the format information is a format number, and the format number specifies a direction of a first dimensional coordinate of the region in the omnidirectional image by default; and
2) the format information comprises a format number and direction information of the first dimensional coordinate of the region in the omnidirectional image.
5. An omnidirectional image processing method, comprising:
decoding a code stream to obtain omnidirectional image coding format information; and
determining mapping relationships between a position (x1, y1) in a region 1 and a position (x2, y2) in a region 2 and their respective corresponding positions (θ1, φ1) and (θ2, φ2) on a sphere according to the format information since at least one pair of adjacent regions, i.e. the region 1 and the region 2, is present in a decoded omnidirectional image, wherein the mapping relationships are:
x1 and φ1 are in linear relationship, and y1 and
θ 1 tan - 1 cos φ 1
are in linear relationship;
x2 and φ2 are in linear relationship, and y2 and
θ 2 tan - 1 cos φ2
are in linear relationship,
where x1 represents a first dimensional coordinate position of the region 1 in the omnidirectional image, y1 represents a second dimensional coordinate position of the image region 1, φ1 represents a longitude position of the sphere corresponding to the position (x1, y1) in the region 1 in the omnidirectional image, and θ1 represents a latitude position of the sphere corresponding to the position (x1, y1) in the region 1 in the omnidirectional image; x2 represents a first dimensional coordinate position of the region 2 in the omnidirectional image, y2 represents a second dimensional coordinate position of the region 2 in the omnidirectional image, φ2 represents a longitude position of the sphere corresponding to the position (x2, y2) in the image region 2, and θ2 represents a latitude position of the sphere corresponding to the position (x2, y2) in the region 2 in the omnidirectional image; and a line mapped to the sphere by a first dimensional coordinate axis of the region 1 is the same as a line mapped to the sphere by a first dimensional coordinate axis of the region 2, and is an equator line of the sphere.
6. An omnidirectional image processing device, comprising the following modules:
a format information extraction module, wherein input of the format information extraction module is a coding code stream, output of the format information extraction module is omnidirectional image format information, and the module decodes the coding code stream to obtain the omnidirectional image format information; and
a position mapping module, wherein input of the position mapping module is the omnidirectional image format information, output of the position mapping module is a mapping relationship between a position (x, y) in at least one region in an omnidirectional image obtained by decoding the code stream and a position (θ, φ) on a sphere, and the mapping relationship is:
x and φ are in linear relationship, and y and
θ tan - 1 cos φ
are in linear relationship,
where x represents a first dimensional coordinate position of the region in the omnidirectional image, y represents a second dimensional coordinate position of the region in the omnidirectional image, φ represents a longitude position of the sphere, θ represents a latitude position of the sphere, a line mapped to the sphere by a first dimensional coordinate axis is an equator line, and a line mapped to the sphere by a second dimensional coordinate axis is a prime meridian.
7. The omnidirectional image processing device according to claim 6, wherein the format information is one of the following:
1) the format information is a format number, and the format number specifies a direction of a first dimensional coordinate of the region in the omnidirectional image by default; and
2) the format information comprises a format number and direction information of the first dimensional coordinate of the region in the omnidirectional image.
8. An omnidirectional image processing device, comprising the following modules:
a position mapping module, wherein output of the position mapping module is omnidirectional image format information, and a mapping relationship between a position (x, y) in at least one region in an omnidirectional image and a position (θ, φ) on a sphere is determined in the module as:
x and φ are in linear relationship, and y and
θ tan - 1 cos φ
are in linear relationship,
where x represents a first dimensional coordinate position of the region in the omnidirectional image, y represents a second dimensional coordinate position of the region in the omnidirectional image, φ represents a longitude position of the sphere, θ represents a latitude position of the sphere, a line mapped to the sphere by a first dimensional coordinate axis is an equator line, and a line mapped to the sphere by a second dimensional coordinate axis is a prime meridian; and
a format information and image coding module, wherein input of the format information and image coding module is the omnidirectional image format information and the omnidirectional image expressed by the mapping relationship determined by the format information, output of the format information and image coding module is a code stream comprising the omnidirectional image and the format information of the omnidirectional image, and the module codes and writes the omnidirectional image and the format information to the code stream.
9. The omnidirectional image processing device according to claim 8, wherein the format information is one of the following:
1) the format information is a format number, and the format number specifies a direction of a first dimensional coordinate of the region in the omnidirectional image by default; and
2) the format information comprises a format number and direction information of the first dimensional coordinate of the region in the omnidirectional image.
US17/044,265 2018-04-01 2019-03-28 Omnidirectional image processing method and device Active 2039-04-12 US11308575B2 (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
CN201810280029.8A CN110349226B (en) 2018-04-01 2018-04-01 Panoramic image processing method and device
CN201810280029.8 2018-04-01
PCT/CN2019/080042 WO2019192377A1 (en) 2018-04-01 2019-03-28 Method and device for processing panoramic image

Publications (2)

Publication Number Publication Date
US20210073938A1 US20210073938A1 (en) 2021-03-11
US11308575B2 true US11308575B2 (en) 2022-04-19

Family

ID=68100035

Family Applications (1)

Application Number Title Priority Date Filing Date
US17/044,265 Active 2039-04-12 US11308575B2 (en) 2018-04-01 2019-03-28 Omnidirectional image processing method and device

Country Status (4)

Country Link
US (1) US11308575B2 (en)
EP (1) EP3779885A4 (en)
CN (1) CN110349226B (en)
WO (1) WO2019192377A1 (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3881225A1 (en) * 2018-11-15 2021-09-22 Qualcomm Incorporated Coordinate estimation on n-spheres with spherical regression
EP3895425A4 (en) * 2018-12-14 2021-12-15 ZTE Corporation Immersive video bitstream processing

Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102298771A (en) 2011-08-16 2011-12-28 淮安盈科伟力科技有限公司 Fish-eye image rapid correction method of panoramic parking auxiliary system
CN106384367A (en) 2016-08-26 2017-02-08 深圳拍乐科技有限公司 Method for automatically stabilizing view angle of panoramic camera
CN107018336A (en) 2017-04-11 2017-08-04 腾讯科技(深圳)有限公司 The method and apparatus of image procossing and the method and apparatus of Video processing
US20170339391A1 (en) * 2016-05-19 2017-11-23 Avago Technologies General Ip (Singapore) Pte. Ltd. 360 degree video system with coordinate compression
US20170336705A1 (en) * 2016-05-19 2017-11-23 Avago Technologies General Ip (Singapore) Pte. Ltd. 360 degree video capture and playback
US20180199065A1 (en) * 2014-10-20 2018-07-12 Google Llc Streaming the visible parts of a spherical video
US20190200023A1 (en) * 2016-09-02 2019-06-27 Vid Scale, Inc. Method and system for signaling of 360-degree video information
US20190238861A1 (en) * 2016-10-12 2019-08-01 Koninklijke Kpn N.V. Processing Spherical Video Data on the Basis of a Region of Interest
US10620441B2 (en) * 2016-12-14 2020-04-14 Qualcomm Incorporated Viewport-aware quality metric for 360-degree video
US10643301B2 (en) * 2017-03-20 2020-05-05 Qualcomm Incorporated Adaptive perturbed cube map projection
US10839480B2 (en) * 2017-03-22 2020-11-17 Qualcomm Incorporated Sphere equator projection for efficient compression of 360-degree video
US10887621B2 (en) * 2016-07-08 2021-01-05 Vid Scale, Inc. 360-degree video coding using geometry projection
US10957044B2 (en) * 2017-03-22 2021-03-23 Qualcomm Incorporated Sphere pole projections for efficient compression of 360-degree video

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102404571A (en) * 2011-11-22 2012-04-04 浙江大学 Binary method and device for video image coding and decoding
KR20140100656A (en) * 2013-02-06 2014-08-18 한국전자통신연구원 Point video offer device using omnidirectional imaging and 3-dimensional data and method
CN104568983B (en) * 2015-01-06 2017-03-15 浙江工业大学 Pipeline Inner Defect Testing device and method based on active panoramic vision
CN106899840B (en) * 2017-03-01 2018-06-05 北京大学深圳研究生院 Panoramic picture mapping method

Patent Citations (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102298771A (en) 2011-08-16 2011-12-28 淮安盈科伟力科技有限公司 Fish-eye image rapid correction method of panoramic parking auxiliary system
US20180199065A1 (en) * 2014-10-20 2018-07-12 Google Llc Streaming the visible parts of a spherical video
US20170339391A1 (en) * 2016-05-19 2017-11-23 Avago Technologies General Ip (Singapore) Pte. Ltd. 360 degree video system with coordinate compression
US20170336705A1 (en) * 2016-05-19 2017-11-23 Avago Technologies General Ip (Singapore) Pte. Ltd. 360 degree video capture and playback
US10887621B2 (en) * 2016-07-08 2021-01-05 Vid Scale, Inc. 360-degree video coding using geometry projection
CN106384367A (en) 2016-08-26 2017-02-08 深圳拍乐科技有限公司 Method for automatically stabilizing view angle of panoramic camera
US20190200023A1 (en) * 2016-09-02 2019-06-27 Vid Scale, Inc. Method and system for signaling of 360-degree video information
US20190238861A1 (en) * 2016-10-12 2019-08-01 Koninklijke Kpn N.V. Processing Spherical Video Data on the Basis of a Region of Interest
US10805614B2 (en) * 2016-10-12 2020-10-13 Koninklijke Kpn N.V. Processing spherical video data on the basis of a region of interest
US10620441B2 (en) * 2016-12-14 2020-04-14 Qualcomm Incorporated Viewport-aware quality metric for 360-degree video
US10643301B2 (en) * 2017-03-20 2020-05-05 Qualcomm Incorporated Adaptive perturbed cube map projection
US10839480B2 (en) * 2017-03-22 2020-11-17 Qualcomm Incorporated Sphere equator projection for efficient compression of 360-degree video
US10957044B2 (en) * 2017-03-22 2021-03-23 Qualcomm Incorporated Sphere pole projections for efficient compression of 360-degree video
CN107018336A (en) 2017-04-11 2017-08-04 腾讯科技(深圳)有限公司 The method and apparatus of image procossing and the method and apparatus of Video processing

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
Search Report dated Jun. 28, 2019 issued in PCT Application No. PCT/CN2019/080042.

Also Published As

Publication number Publication date
WO2019192377A1 (en) 2019-10-10
US20210073938A1 (en) 2021-03-11
CN110349226A (en) 2019-10-18
EP3779885A1 (en) 2021-02-17
CN110349226B (en) 2021-06-01
EP3779885A4 (en) 2021-05-19

Similar Documents

Publication Publication Date Title
US10819971B2 (en) Mapping of spherical image data into rectangular faces for transport and decoding across networks
CN107454468B (en) Method, apparatus and stream for formatting immersive video
WO2018095087A1 (en) Deblocking filter method and terminal
US20180152663A1 (en) View-dependent operations during playback of panoramic video
KR101007679B1 (en) Apparatus of warping image generation for curved display and method thereof
US6567086B1 (en) Immersive video system using multiple video streams
US20180152682A1 (en) Re-projecting flat projections of pictures of panoramic video for rendering by application
US11308575B2 (en) Omnidirectional image processing method and device
KR20220047284A (en) Systems and methods for foveated rendering
US20190230368A1 (en) Encoding and decoding a video signal
US20200068205A1 (en) Geodesic intra-prediction for panoramic video coding
US20110194024A1 (en) Content adaptive and art directable scalable video coding
US11270502B2 (en) Method and apparatus for image processing of three-dimensional model maps
CN111246208B (en) Video processing method and device and electronic equipment
WO2019037656A1 (en) Method and apparatus of signalling syntax for immersive video coding
US20220108420A1 (en) Method and system of efficient image rendering for near-eye light field displays
WO2021169817A1 (en) Video processing method and electronic device
CN114257733A (en) Method and system for image processing of omni-directional image with viewpoint offset
EP2839437B1 (en) View synthesis using low resolution depth maps
US20190272617A1 (en) Cube-based projection method that applies different mapping functions to different square projection faces, different axes, and/or different locations of axis
WO2024140684A1 (en) End-cloud collaboration system, encoding method, decoding method, and electronic device
US12126840B2 (en) Video processing method and apparatus, and device, decoder, system and storage medium
US11190801B2 (en) Video encoding method with syntax element signaling of mapping function employed by cube-based projection and associated video decoding method
US20230100615A1 (en) Video processing method and apparatus, and device, decoder, system and storage medium
CN118175354A (en) Live broadcast method and device, electronic equipment and storage medium

Legal Events

Date Code Title Description
FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY

AS Assignment

Owner name: ZHEJIANG UNIVERSITY, CHINA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:YU, LU;HUANGFU, XUCHANG;REEL/FRAME:053955/0127

Effective date: 20200929

FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO SMALL (ORIGINAL EVENT CODE: SMAL); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY

STPP Information on status: patent application and granting procedure in general

Free format text: APPLICATION DISPATCHED FROM PREEXAM, NOT YET DOCKETED

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

STPP Information on status: patent application and granting procedure in general

Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED

STCF Information on status: patent grant

Free format text: PATENTED CASE