CN111340695A - Super-resolution reconstruction method of dome screen video - Google Patents

Super-resolution reconstruction method of dome screen video Download PDF

Info

Publication number
CN111340695A
CN111340695A CN202010083678.6A CN202010083678A CN111340695A CN 111340695 A CN111340695 A CN 111340695A CN 202010083678 A CN202010083678 A CN 202010083678A CN 111340695 A CN111340695 A CN 111340695A
Authority
CN
China
Prior art keywords
spherical screen
resolution reconstruction
screen video
image
neural network
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202010083678.6A
Other languages
Chinese (zh)
Inventor
秦永进
李艳玲
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shanghai Zhihuan Software Technology Co ltd
Original Assignee
Shanghai Zhihuan Software Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shanghai Zhihuan Software Technology Co ltd filed Critical Shanghai Zhihuan Software Technology Co ltd
Priority to CN202010083678.6A priority Critical patent/CN111340695A/en
Publication of CN111340695A publication Critical patent/CN111340695A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformation in the plane of the image
    • G06T3/40Scaling the whole image or part thereof
    • G06T3/4053Super resolution, i.e. output image resolution higher than sensor resolution
    • G06T5/73
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]

Abstract

The invention relates to a super-resolution reconstruction method of a spherical screen video, which comprises the following steps: introducing distortion correction parameters into the spherical screen video image to perform 3D correction processing; and (4) extracting the features of the spherical screen video image, and performing super-resolution reconstruction on the extracted features through a wide activation WDSR-B neural network. The jump connection and splicing operation between the shallow layer and the deep layer is introduced, based on a residual SISR network, the simple feature expansion before Relu can be obviously improved on the premise of no additional parameters and computation amount, and the feature expansion before Relu allows more information to pass through, and meanwhile, the high nonlinearity of the deep neural network is still maintained. The low-level SISR characteristics from the shallow layer can be more easily transmitted to the last layer to obtain better dense pixel value prediction, and the technology is applied to the dome cinema, so that the definition of images is improved, and the viewing experience of audiences is better.

Description

Super-resolution reconstruction method of dome screen video
Technical Field
The invention relates to a super-resolution reconstruction technology of a dome screen video image, in particular to a super-resolution reconstruction method of a dome screen video.
Background
With the continuous development of pattern recognition and artificial intelligence, the super-resolution reconstruction technology has high application value in many fields such as video monitoring, medical images, satellite remote sensing and the like. The super-resolution reconstruction technique of images refers to restoring a given low-resolution image into a corresponding high-resolution image through a specific algorithm. The spherical screen video image applied to the spherical screen cinema keeps the scenes except the center of the picture unchanged, and other scenes which are supposed to be horizontal or vertical are correspondingly changed, so that the resolutions of different areas of the imaging surface of the spherical screen lens are different, the closer to the center of the picture, the higher the resolution, the more detailed information, the more deviated from the center of the picture, the lower the resolution, the less detailed information and the more serious deformation are caused.
The former image super-resolution network uses relatively shallow convolutional neural networks, which have poor accuracy, and the increase of depth can more effectively improve the resolution, but does not effectively utilize the characteristic information from the shallow layer, thereby increasing the time complexity.
At present, image processing is directly carried out on a distorted spherical screen video image, but because the spherical screen video image data are stored in a nonlinear mode and cannot be directly processed, the processing mode cannot obtain a good image deblurring effect.
In the hyper-resolution reconstruction algorithm, shallow information is not effectively utilized in the deep layer at present, the training network is very complex, and even a plurality of redundant convolutional layers exist, so that the influence on the hyper-resolution reconstruction effect of only considering a single image is not large, but when the hyper-resolution reconstruction algorithm is applied to a video, the speed is very low, the consumed time is long, and the real-time performance is poor under the condition that the network is complex.
Disclosure of Invention
In order to solve the problem of the definition of a spherical screen video image, the invention provides a hyper-resolution reconstruction method of a spherical screen video.
In order to solve the problems, the invention adopts the following technical scheme:
a method for hyper-resolution reconstruction of a dome screen video comprises the following steps:
s1, extracting an image from a spherical screen video by frames, and introducing distortion correction parameters into the image to perform 3D correction processing;
s2, extracting the characteristics of the corrected image;
and S3, carrying out hyper-resolution reconstruction on the extracted features through a wide-activation WDSR-B neural network.
Preferably, in the 3D correction in S1, a projection plane is set, and a mapping relationship is established between a distance r from a spherical screen pixel point to an optical axis center and a projection plane light incident angle θ'; and calculating a projection plane light ray incidence angle according to the distance meter from the spherical screen pixel point to the center of the optical axis through the mapping relation, and performing back projection on the projection plane light ray incidence angle to the plane to obtain a corrected point.
Preferably, the 3D correction is calculated according to different equations for different angles of incidence:
when the light incidence angle is less than or equal to a first threshold value, an orthogonal model calculation mode is adopted:
r=fθ′;
when the first threshold value is larger than the light ray incidence angle and smaller than or equal to the second threshold value, an equiangular model calculation mode is adopted:
r=2·f sinθ′/2;
when the light incidence angle is larger than a second threshold value, adopting an equidistant model calculation mode:
r=f sinθ′;
f is the distance from the center of the optical axis to the plane of the spherical screen.
Preferably, the S2 performing feature extraction on the image after the rectification processing includes performing feature extraction on brightness, edges, textures and/or colors.
Preferably, the wide-activation WDSR-B neural network comprises a feature extraction module, a residual module, and an output module; the characteristic extraction module adopts the image characteristics extracted by the convolution layer; extracting the characteristics of the potential region by a residual error module; and the output module performs convolution on the extracted features and the residual error output by the residual error module through the convolution layer and then outputs the convolution result.
Preferably, the residual module includes sequentially arranged 1 × 1 convolutional layers for expanding the number of the same tracks, convolutional layers for feature extraction, a Relu activation link for adding a nonlinear feature, 1 × 1 convolutional layer for shallow feature extraction, and two 3 × 3 convolutional layers.
Preferably, the S3 further includes weight normalizing the WDSR-B neural network, the weight normalized parameterized weight vector being:
Figure BDA0002381330180000031
v is a k-dimensional vector, k being the dimension of the extracted feature; g is a scalar, w | | | is g; i v i is the euclidean norm of v, which will result in g, independent of the parameter v.
The technical scheme of the invention has the following beneficial technical effects:
(1) the jump connection and splicing operation between the shallow layer and the deep layer is introduced, based on a residual SISR network, the simple feature expansion before Relu can be obviously improved on the premise of no additional parameters and computation amount, and the feature expansion before Relu allows more information to pass through, and meanwhile, the high nonlinearity of the deep neural network is still maintained. The low-level SISR characteristics from the shallow layer can be more easily transmitted to the last layer to obtain better dense pixel value prediction, and the technology is applied to the dome cinema, so that the definition of images is improved, and the viewing experience of audiences is better.
(2) And 3D correction processing is carried out on the fisheye image of the spherical screen video, distortion correction parameters are introduced, and then characteristic extraction is carried out to obtain a better image deblurring effect. Reflecting each point of the spherical curtain to a plane, so that pixel holes or discontinuity cannot be formed; the correction method is introduced into the spherical screen video, and the problem that the spherical screen video and the common video are different in spatial special distribution is solved.
Drawings
FIG. 1 is a schematic diagram of a hyper-resolution reconstruction method for a dome screen video;
FIG. 2 is a schematic diagram of a spherical screen video image 3D correction model of a spherical screen video hyper-resolution reconstruction method;
fig. 3 is a flowchart of the hyper-resolution reconstruction of the dome video.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, the present invention will be described in further detail with reference to the accompanying drawings in conjunction with the following detailed description. It should be understood that the description is intended to be exemplary only, and is not intended to limit the scope of the present invention. Moreover, in the following description, descriptions of well-known structures and techniques are omitted so as to not unnecessarily obscure the concepts of the present invention.
Referring to fig. 1, a method for super-resolution reconstruction of a dome screen video includes: 3D correction processing of a dome video image and wide activation WDSR-B processing.
Referring to fig. 2, the dome video image 3D correction process:
the spherical screen video image is from a plane video image, and each frame of video image in the video is obtained. Images of certain frames may also be selected. And 3D correction processing is carried out on the image to obtain a corresponding projection mode from the plane to the spherical screen.
Introducing distortion correction parameters into the spherical screen video image to perform 3D correction processing, establishing a 3D correction model, establishing a mapping relation between the distance from a spherical screen pixel point P to an optical axis center 0 and a space object light incident angle, then giving a projection plane 0', and reversely projecting the projection plane along the solved incident angle to obtain a corrected point. And further establishing a mapping relation between each point on the spherical screen and each point on the plane, and calculating projection points on the spherical screen corresponding to each point of the image according to the mapping relation to finish 3D correction. And performing feature extraction on brightness, edges, textures, colors and other data with features of the corrected image.
The 3D correction model is used for establishing a mapping relation between the distance from the spherical screen video image point to the center of the optical axis and the incident angle of the light rays of the space object.
D is a correction model;
r is the distance from the spherical screen pixel point to the center of the optical axis;
theta is an included angle between the incident ray and the plumb line; theta' is an included angle between the incident ray of the projection plane and the plumb line;
psi, psi' are the azimuth angles (radians) before and after correction of the incident ray;
p is a pixel point of the space object light ray on the spherical screen;
p' is the point where the incident ray is back projected onto the correction plane.
In one embodiment, in conjunction with FIG. 2, the 3D correction is calculated using different equations based on different angles of incidence.
When the light incidence angle is less than or equal to a first threshold value, an orthogonal model calculation mode is adopted:
r=fθ′;
when the first threshold value is larger than the light ray incidence angle and smaller than or equal to the second threshold value, an equiangular model calculation mode is adopted:
r=2·f sinθ′/2;
when the light incidence angle is larger than a second threshold value, adopting an equidistant model calculation mode:
r=f sinθ′;
r is the distance from the spherical screen pixel point to the center of the optical axis; theta' is an included angle between the incident ray and the main shaft; f is the focal length of the spherical screen; the 3D correction is performed using different calculation methods by different incident angles. The first threshold value is, for example, 60 °, and the second threshold value is, for example, 65 °. And adjusting according to the angle of the actual plane video shooting lens.
The azimuth angle ψ' of the incident ray is substantially not distorted in the actual screen video image or is distorted to a negligible extent, i.e., the equation: since ψ 'is always true, the model d can be expressed as θ'd (r). The formula of theta ═ D (r) only contains one parameter of the distance r between the point on the spherical screen video image and the optical axis, and the point of the mapping relation between r and theta' after 3D correction through the formula is located on the projection plane to form a plane image.
For the wide-activation WDSR-B neural network, features are expanded before a ReLU activation layer, and the wide activation effect in a residual block is considered, the most basic mode is to directly increase the number of channels in all the features, namely, the number of channels is increased by using convolution of 3X3, then an activation function is used, and then the number of channels is reduced by using convolution of 3X3, wherein the structure mainly aims at the condition that the expansion multiple is small (the value of the structure is between 2 and 4). When the value is large, the effect is rapidly reduced, and the method is not applicable.
As shown in fig. 1, in order to solve the above limitation, in A WDSR-B neural network, A large convolution kernel in an original WDSR- A neural network is split into two low-rank convolution kernels while keeping the number of equivalent mapping path channels unchanged, so as to save parameters. The WDSR-B neural network comprises a feature extraction module, a residual error module and an output module; the characteristic extraction module adopts the image characteristics extracted by the convolution layer; extracting the characteristics of the potential region by a residual error module; and the output module performs convolution on the extracted features and the residual error output by the residual error module through the convolution layer and then outputs the convolution result. The residual module includes 1 × 1 convolutional layer, nonlinear (Relu) activation, 1 × 1 convolutional layer, two 3 × 3 convolutional layers.
First the number of channels is enlarged with 1 x 1 convolutional layers (Conv), then a non-linear function (Relu) is applied after the convolutional layers, further proposing an efficient linear low-rank convolution, which decomposes a large convolution kernel into two low-rank convolution kernels. The method is a combination of 1 × 1Conv for reducing the number of channels and 3 × 3Conv for extracting spatial features, further increases the number of feature map channels before Relu under the condition of the same parameters, and can improve the resolution ratio without increasing the time complexity so as to make the dome screen video image clearer.
On the premise of not increasing the calculation overhead, the number of filters of the convolution kernel before the Relu activation layer is increased so as to increase the width of the feature mapping.
The residual network of the wide-active WDSR-B effectively reduces the time complexity. And a plurality of redundant convolution layers are removed, the time complexity is reduced, and the real-time performance of the dome screen video is met.
Weight normalization is the reparameterization of weight vectors in neural networks, and the length of those weight vectors and their direction can be decoupled. It has no dependency on the samples in the mini-batch, and has the same formula when training and testing.
When the model reaches a certain depth (180 layers), the model is difficult to train, and therefore weight normalization operation is provided. The method aims to recalculate the distribution of the network in the network, introduce weight normalization to allow training with higher learning rate, and improve the accuracy of training and testing.
The weight normalization algorithm is as follows:
y=w.x+b (1)
where w is a k-dimensional vector matrix; x is a vector of input features in k dimensions; y is a vector of output features; b is a bias, a parameter for adjusting balance; the k is a dimension;
weight normalization parameterized weight vector algorithm:
Figure BDA0002381330180000061
where v is a k-dimensional vector, k being the dimension of the extracted feature; g is a scalar quantity, has no direction and only represents the size; if v is the euclidean norm of v, then we will get w, which is independent of the parameter v; w ranges between 0 and 1.
W in the formula (2) affects y in the formula (1), and y represents the output feature vector, so that weight normalization w affects the y output feature vector, and the accuracy of the output feature vector is improved by adding a weight normalization method.
And training the neural network model by using a sample, packaging and packaging after meeting the precision requirement, and performing over-resolution reconstruction after correcting the image.
Referring to fig. 3, the invention provides a method for hyper-resolution reconstruction of a dome video, which is mainly implemented by the following steps:
s1, extracting an image from a spherical screen video by frames, and introducing distortion correction parameters into the image to perform 3D correction processing;
s2, extracting the characteristics of the corrected image;
and S3, carrying out hyper-resolution reconstruction on the extracted features through a wide-activation WDSR-B neural network.
And in the process of wide activation of the WDSR-B neural network, weight normalization is adopted to improve the accuracy of the spherical screen video image.
Through the steps, various jump connection and splicing operations between the shallow layer and the deep layer are realized, based on the residual SISR network, the simple feature expansion before Relu can be obviously improved on the premise of no additional parameters and calculation amount, and the feature expansion before Relu allows more information to pass through, and meanwhile, the high nonlinearity of the deep neural network is still maintained. Low-level SISR features from shallow layers can be more easily passed to the last layer for better dense pixel value prediction.
In summary, the present invention relates to a method for super-resolution reconstruction of a dome screen video, including: introducing distortion correction parameters into the spherical screen video image to perform 3D correction processing; and (4) extracting the features of the spherical screen video image, and performing super-resolution reconstruction on the extracted features through a wide activation WDSR-B neural network. The jump connection and splicing operation between the shallow layer and the deep layer is introduced, based on a residual SISR network, the simple feature expansion before Relu can be obviously improved on the premise of no additional parameters and computation amount, and the feature expansion before Relu allows more information to pass through, and meanwhile, the high nonlinearity of the deep neural network is still maintained. The low-level SISR characteristics from the shallow layer can be more easily transmitted to the last layer to obtain better dense pixel value prediction, and the technology is applied to the dome cinema, so that the definition of images is improved, and the viewing experience of audiences is better.
It is to be understood that the above-described embodiments of the present invention are merely illustrative of or explaining the principles of the invention and are not to be construed as limiting the invention. Therefore, any modification, equivalent replacement, improvement and the like made without departing from the spirit and scope of the present invention should be included in the protection scope of the present invention. Further, it is intended that the appended claims cover all such variations and modifications as fall within the scope and boundaries of the appended claims or the equivalents of such scope and boundaries.

Claims (7)

1. A method for hyper-resolution reconstruction of a dome screen video is characterized by comprising the following steps:
s1, extracting an image from a spherical screen video by frames, and introducing distortion correction parameters into the image to perform 3D correction processing;
s2, extracting the characteristics of the corrected image;
and S3, carrying out hyper-resolution reconstruction on the extracted features through a wide-activation WDSR-B neural network.
2. The method for super-resolution reconstruction of dome video according to claim 1, wherein the 3D correction in S1 includes: setting a projection plane, and calculating the distance r from a projection point on the spherical screen to the center of an optical axis through a light incident angle theta' passing through the point on the projection screen; the azimuth angles of the projection points on the spherical screen and the projection points on the projection screen are the same, and then the corresponding projection points on the spherical screen are obtained.
3. The method of claim 2, wherein the 3D correction is calculated according to different formula based on different light incidence angles θ':
when the light incidence angle is less than or equal to a first threshold value, an orthogonal model calculation mode is adopted:
r=fθ′;
when the first threshold value is larger than the light ray incidence angle and smaller than or equal to the second threshold value, an equiangular model calculation mode is adopted:
r=2·f sinθ′/2;
when the light incidence angle is larger than a second threshold value, adopting an equidistant model calculation mode:
r=f sinθ′;
f is the distance from the center of the optical axis to the plane of the spherical screen.
4. The method for reconstructing the dome screen video according to claim 1 or 2, wherein the S2 performs feature extraction on the rectified image, including performing feature extraction on brightness, edge, texture and/or color.
5. The method of claim 1 or 2, wherein the wide-activation WDSR-B neural network comprises a feature extraction module, a residual error module and an output module; the characteristic extraction module adopts the image characteristics extracted by the convolution layer; extracting the characteristics of the potential region by a residual error module; and the output module performs convolution on the extracted features and the residual error output by the residual error module through the convolution layer and then outputs the convolution result.
6. The method according to claim 5, wherein the residual module comprises sequentially arranged 1 x 1 convolutional layers for increasing the number of the same channel, convolutional layers for feature extraction, Relu activation links for adding nonlinear features, 1 x 1 convolutional layers for shallow feature extraction, and two 3x3 convolutional layers.
7. The method of super-resolution reconstruction of dome video according to claim 1 or 2, wherein said S3 further comprises performing weight normalization on WDSR-B neural network, the weight normalization parameterized weight vector is:
Figure FDA0002381330170000021
v is a k-dimensional vector, k being the dimension of the extracted feature; g is a scalar, w | | | is g; i v i is the euclidean norm of v, which will result in g, independent of the parameter v.
CN202010083678.6A 2020-02-10 2020-02-10 Super-resolution reconstruction method of dome screen video Pending CN111340695A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010083678.6A CN111340695A (en) 2020-02-10 2020-02-10 Super-resolution reconstruction method of dome screen video

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010083678.6A CN111340695A (en) 2020-02-10 2020-02-10 Super-resolution reconstruction method of dome screen video

Publications (1)

Publication Number Publication Date
CN111340695A true CN111340695A (en) 2020-06-26

Family

ID=71183852

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010083678.6A Pending CN111340695A (en) 2020-02-10 2020-02-10 Super-resolution reconstruction method of dome screen video

Country Status (1)

Country Link
CN (1) CN111340695A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111832463A (en) * 2020-07-07 2020-10-27 哈尔滨理工大学 Deep learning-based traffic sign detection method

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102663734A (en) * 2012-03-15 2012-09-12 天津理工大学 Fish eye lens calibration and fish eye image distortion correction method
CN103533235A (en) * 2013-09-17 2014-01-22 北京航空航天大学 Quick digital panoramic device based on linear array charge coupled device (CCD) for great case/event scene
CN104067372A (en) * 2012-01-27 2014-09-24 塞莫费雪科学(不来梅)有限公司 Multi-reflection mass spectrometer
CN106469448A (en) * 2015-06-26 2017-03-01 康耐视公司 Carry out automatic industrial inspection using 3D vision
CN106842520A (en) * 2017-03-30 2017-06-13 中山联合光电科技股份有限公司 A kind of high definition panorama looks around optical imaging system
CN107240066A (en) * 2017-04-28 2017-10-10 天津大学 Image super-resolution rebuilding algorithm based on shallow-layer and deep layer convolutional neural networks
CN108961151A (en) * 2018-05-08 2018-12-07 中德(珠海)人工智能研究院有限公司 A method of the three-dimensional large scene that ball curtain camera obtains is changed into sectional view
CN108983394A (en) * 2018-08-23 2018-12-11 上海帛视光电科技有限公司 A kind of fish eye lens system, image acquisition device

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104067372A (en) * 2012-01-27 2014-09-24 塞莫费雪科学(不来梅)有限公司 Multi-reflection mass spectrometer
CN102663734A (en) * 2012-03-15 2012-09-12 天津理工大学 Fish eye lens calibration and fish eye image distortion correction method
CN103533235A (en) * 2013-09-17 2014-01-22 北京航空航天大学 Quick digital panoramic device based on linear array charge coupled device (CCD) for great case/event scene
CN106469448A (en) * 2015-06-26 2017-03-01 康耐视公司 Carry out automatic industrial inspection using 3D vision
CN106842520A (en) * 2017-03-30 2017-06-13 中山联合光电科技股份有限公司 A kind of high definition panorama looks around optical imaging system
CN107240066A (en) * 2017-04-28 2017-10-10 天津大学 Image super-resolution rebuilding algorithm based on shallow-layer and deep layer convolutional neural networks
CN108961151A (en) * 2018-05-08 2018-12-07 中德(珠海)人工智能研究院有限公司 A method of the three-dimensional large scene that ball curtain camera obtains is changed into sectional view
CN108983394A (en) * 2018-08-23 2018-12-11 上海帛视光电科技有限公司 A kind of fish eye lens system, image acquisition device

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
JIAHUI YU.ETC: ""Wide Activation for Efficient and Accurate Image Super-Resolution"", <ARXIV>, 21 December 2018 (2018-12-21), pages 3 - 4 *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111832463A (en) * 2020-07-07 2020-10-27 哈尔滨理工大学 Deep learning-based traffic sign detection method

Similar Documents

Publication Publication Date Title
US11037278B2 (en) Systems and methods for transforming raw sensor data captured in low-light conditions to well-exposed images using neural network architectures
US10257501B2 (en) Efficient canvas view generation from intermediate views
US10708525B2 (en) Systems and methods for processing low light images
US11880939B2 (en) Embedding complex 3D objects into an augmented reality scene using image segmentation
US20210124985A1 (en) System and method for deep machine learning for computer vision applications
Chen et al. Deep exposure fusion with deghosting via homography estimation and attention learning
US11871110B2 (en) Single image ultra-wide fisheye camera calibration via deep learning
CN111553841B (en) Real-time video splicing method based on optimal suture line updating
KR20230146649A (en) Color and infrared 3D reconstruction using implicit radiance functions.
EP3886044A1 (en) Robust surface registration based on parameterized perspective of image templates
CN115375536A (en) Image processing method and apparatus
Chang et al. Deep learning based image Super-resolution for nonlinear lens distortions
CN115115516A (en) Real-world video super-resolution algorithm based on Raw domain
CN111340695A (en) Super-resolution reconstruction method of dome screen video
JP2006221599A (en) Method and apparatus for generating mapping function, and compound picture develop method, and its device
Bergmann et al. Gravity alignment for single panorama depth inference
Hong et al. PAR $^{2} $ Net: End-to-end Panoramic Image Reflection Removal
Liang et al. Multi-scale and multi-patch transformer for sandstorm image enhancement
Yue et al. Rvideformer: Efficient raw video denoising transformer with a larger benchmark dataset
Okura et al. Aerial full spherical HDR imaging and display
CN114257733A (en) Method and system for image processing of omni-directional image with viewpoint offset
CN112991174A (en) Method and system for improving resolution of single-frame infrared image
CN111709880B (en) Multi-path picture splicing method based on end-to-end neural network
US11651475B2 (en) Image restoration method and device
Jung et al. Deep low-contrast image enhancement using structure tensor representation

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination