CN114387395A - A fast generation method of hologram based on phase-dual resolution network - Google Patents

A fast generation method of hologram based on phase-dual resolution network Download PDF

Info

Publication number
CN114387395A
CN114387395A CN202210029238.1A CN202210029238A CN114387395A CN 114387395 A CN114387395 A CN 114387395A CN 202210029238 A CN202210029238 A CN 202210029238A CN 114387395 A CN114387395 A CN 114387395A
Authority
CN
China
Prior art keywords
phase
convolution
network
hologram
dual
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202210029238.1A
Other languages
Chinese (zh)
Inventor
王帅
于挺
田子建
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
China University of Mining and Technology Beijing CUMTB
Original Assignee
China University of Mining and Technology Beijing CUMTB
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by China University of Mining and Technology Beijing CUMTB filed Critical China University of Mining and Technology Beijing CUMTB
Priority to CN202210029238.1A priority Critical patent/CN114387395A/en
Publication of CN114387395A publication Critical patent/CN114387395A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Software Systems (AREA)
  • General Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Computing Systems (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Biomedical Technology (AREA)
  • Molecular Biology (AREA)
  • Biophysics (AREA)
  • General Engineering & Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • Mathematical Physics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Computer Graphics (AREA)
  • Geometry (AREA)
  • Holo Graphy (AREA)

Abstract

本发明公开了一种基于相位‑双分辨率网络的全息图快速生成方法,该方法包括:无需制作真实的纯相位全息图掩模,利用角谱法的可微性,使用自然图像实现卷积神经网络的无监督训练;卷积层在同一空间学习特征映射,而不是跨距离学习特征映射;使用空洞卷积和组卷积降低网络模型计算量和GPU的内存占用率;使用MS‑SSIM损失和MSE损失的组合作为一致性损失函数,以生成更加符合人类视觉系统的重建图。本申请计算生成一幅1080P分辨率的全息图仅需57毫秒,且最优的数值重建图与目标强度图的峰值信噪比达到了31.17dB。

Figure 202210029238

The invention discloses a method for rapidly generating a hologram based on a phase-double-resolution network. The method comprises the following steps: without making a real pure-phase hologram mask, utilizing the differentiability of an angular spectrum method, and using a natural image to realize convolution Unsupervised training of neural networks; convolutional layers learn feature maps in the same space instead of across distances; use atrous convolution and group convolution to reduce network model computation and GPU memory usage; use MS‑SSIM loss and the MSE loss are combined as a consistency loss function to generate reconstruction maps that are more in line with the human visual system. This application calculates and generates a hologram with a resolution of 1080P in only 57 milliseconds, and the peak signal-to-noise ratio between the optimal numerical reconstruction image and the target intensity image reaches 31.17dB.

Figure 202210029238

Description

一种基于相位-双分辨率网络的全息图快速生成方法A fast generation method of hologram based on phase-dual resolution network

技术领域technical field

本发明涉及计算生成全息领域和计算机视觉领域,特别是涉及一种基于相位-双分辨率网络的全息图快速生成方法。The invention relates to the field of computationally generated holography and computer vision, in particular to a method for rapidly generating a hologram based on a phase-dual resolution network.

背景技术Background technique

全息显示被认为是一种先进的显示技术,逐渐应用于三维场景重建、虚拟现实和增强显示系统中,使无需眼镜的三维显示模式成为可能。在实现实时全息显示的同时,如何保证全息显示影像的高保真度仍然是一个难题。相位型的空间光调制器光学效率较高,重构过程中无共轭图像的干扰,纯相位全息图已成为计算生成全息图的主要编码方法。因此,开发一种能够实时计算生成高质量相位型全息图的算法,对于全息显示技术的发展具有重要意义。Holographic display is considered to be an advanced display technology, and is gradually applied in 3D scene reconstruction, virtual reality and augmented display systems, enabling a 3D display mode without glasses. While realizing the real-time holographic display, how to ensure the high fidelity of the holographic display image is still a difficult problem. The phase-type spatial light modulator has high optical efficiency, and there is no interference of conjugate images in the reconstruction process. Pure-phase holograms have become the main coding method for computationally generated holograms. Therefore, it is of great significance for the development of holographic display technology to develop an algorithm that can calculate and generate high-quality phase holograms in real time.

半个多世纪以来,人们推出了多种计算生成全息图的算法,主要包括迭代优化算法和非迭代优化算法。1972年提出的Gercheberg-Saxton算法是迭代优化计算生成全息图的典型代表。随后也有一些基于Gercheberg-Saxton迭代优化算法的改进算法,如,在Gercheberg-Saxton 迭代优化算法中添加随机噪声优化生成全息图;双向误差扩散算法生成全息图;通过梯度下降或Wirtinger导数求解非凸优化问题间接计算生成全息图等。由于迭代优化算法需要花费较长计算时间生成全息图,不适用于大规模、高分辨率全息图的实时计算,因此,有研究学者提出非迭代优化算法,如,双相位幅值编码和一步相位提取算法等。虽然非迭代优化算法计算速度得到大幅度提升,能够满足实时计算,但全息图的重建图像包含较多的散斑噪声。因此,非迭代优化算法无法保证计算生成全息图的质量。For more than half a century, a variety of algorithms for computationally generating holograms have been introduced, mainly including iterative optimization algorithms and non-iterative optimization algorithms. The Gercheberg-Saxton algorithm proposed in 1972 is a typical representative of iterative optimization to generate holograms. Then there are some improved algorithms based on the Gercheberg-Saxton iterative optimization algorithm, such as adding random noise to the Gercheberg-Saxton iterative optimization algorithm to generate holograms; bidirectional error diffusion algorithm to generate holograms; gradient descent or Wirtinger derivatives to solve non-convex optimization The problem is indirect computationally generating holograms, etc. Since the iterative optimization algorithm takes a long time to generate holograms, it is not suitable for the real-time calculation of large-scale and high-resolution holograms. Therefore, some researchers have proposed non-iterative optimization algorithms, such as two-phase amplitude encoding and one-step phase encoding. Extraction algorithms, etc. Although the calculation speed of the non-iterative optimization algorithm has been greatly improved and can meet the real-time calculation, the reconstructed image of the hologram contains more speckle noise. Therefore, non-iterative optimization algorithms cannot guarantee the quality of computationally generated holograms.

近年来,随着算法研究的不断深入,深度学习技术逐渐被引入光学领域,深度学习相关算法的出现也为计算生成全息图提供了新方法。卷积神经网络作为广义近似函数,可以学习输入和输出之间的映射。2021年,Shi等人引入了大规模菲涅耳全息数据集,训练卷积神经网络生成逼真的三维全息图,并且能够变换焦距。也有研究人员利用菲涅尔方法生成大量的全息图作为数据集,训练对抗生成网络。这类方法所使用的训练集均是通过传统的迭代优化算法计算生成的全息图。因此,网络模型的性能直接受限于训练集的质量,间接受限于传统迭代优化算法。如何打破这些限制成为深度学习应用于计算生成全息图的一个关键挑战。更重要的是,卷积层作用于输入的空间维度,最适合于建模和计算输入与输出之间的空间关系映射。卷积神经网络将在物平面中定义的图像跨域映射到在全息图平面中定义的纯相位全息图掩模,存在着空间对应关系不能得到保留,无法发挥卷积神经网络强大的特征映射能力的问题。In recent years, with the deepening of algorithm research, deep learning technology has been gradually introduced into the field of optics, and the emergence of deep learning-related algorithms has also provided a new method for computationally generating holograms. Convolutional neural networks act as generalized approximation functions that learn the mapping between input and output. In 2021, Shi et al. introduced a large-scale Fresnel holography dataset to train a convolutional neural network to generate realistic 3D holograms with the ability to change the focal length. Some researchers also use the Fresnel method to generate a large number of holograms as a dataset to train a confrontational generative network. The training sets used in these methods are all holograms generated by traditional iterative optimization algorithms. Therefore, the performance of the network model is directly limited by the quality of the training set, and indirectly limited by the traditional iterative optimization algorithm. How to break these limitations becomes a key challenge for the application of deep learning to computationally generated holograms. More importantly, convolutional layers act on the spatial dimension of the input and are best suited for modeling and computing the spatial relationship mapping between input and output. The convolutional neural network maps the image defined in the object plane to the pure phase hologram mask defined in the hologram plane. There is a spatial correspondence that cannot be preserved, and the powerful feature mapping capability of the convolutional neural network cannot be exerted. The problem.

发明内容SUMMARY OF THE INVENTION

为了解决上述背景技术中的问题,本申请提供一种基于相位-双分辨率网络的全息图快速生成方法,主要技术方案如下:In order to solve the above-mentioned problems in the background technology, the present application provides a method for rapidly generating a hologram based on a phase-dual resolution network. The main technical solutions are as follows:

构建相位-双分辨率网络模型fnet1和fnet2Build phase-dual resolution network models f net1 and f net2 ;

经过训练的相位-双分辨率网络fnet1,根据输入的目标强度图I计算出初始相位φ0The trained phase-dual-resolution network f net1 calculates the initial phase φ 0 according to the input target intensity map I;

根据初始相位φ0与目标强度图I,计算出复值波场Uz;根据角谱法,计算复值波场Uz在自由空间传播-z后得到的复值波场U0Calculate the complex-valued wave field U z according to the initial phase φ 0 and the target intensity map I; calculate the complex-valued wave field U 0 obtained after the complex-valued wave field U z propagates in free space -z according to the angular spectrum method;

经过训练的相位-双分辨率网络fnet2,根据输入的复值波场U0计算出纯相位的全息图。The trained phase-dual-resolution network fnet2 computes a phase-only hologram from the input complex-valued wavefield U 0 .

优选地,所述构建相位-双分辨率网络模型包括:Preferably, the building a phase-dual resolution network model includes:

相位-双分辨率网络主要分为编码器和解码器两部分;The phase-dual resolution network is mainly divided into two parts: encoder and decoder;

在相位-双分辨率网络的编码器中,前两层网络使用组卷积进行特征提取;编码器中剩余的网络层均使用空洞卷积进行特征提取,其中空洞卷积中的空洞因子,遵循卷积中两个非零值之间的最大距离计算公式:In the encoder of the phase-dual resolution network, the first two layers of the network use group convolution for feature extraction; the remaining network layers in the encoder use atrous convolution for feature extraction, where the atrous factor in the atrous convolution follows the The formula for calculating the maximum distance between two non-zero values in convolution:

Mi=max[Mi+1-2ri,Mi+1-2(Mi+1-ri),ri],Mn=rnM i =max[M i+ 1-2r i ,M i+ 1-2(M i+1 -r i ),r i ],M n =r n ;

在解码器中,通过sub-pixel卷积对编码器提取的特征进行上采样,直至与输入图像分辨率相同;编码器的第三层网络输出使用跳跃连接至解码器的第一层进行特征拼接;将1×1 卷积作为解码器的最后一层,使得输出为单通道图像;In the decoder, the features extracted by the encoder are upsampled by sub-pixel convolution until it is the same resolution as the input image; the third layer network output of the encoder is connected to the first layer of the decoder for feature stitching using skip connections ;Use 1×1 convolution as the last layer of the decoder, so that the output is a single-channel image;

将单通道图像的灰度值归一化至[-π,π]。Normalize the grayscale values of a single-channel image to [-π,π].

优选地,所述经过训练的相位-双分辨率网络的训练过程包括:Preferably, the training process of the trained phase-dual resolution network includes:

已构建的相位-双分辨率网络fnet2计算生成纯相位的全息图,根据角谱法计算出全息图的重建图

Figure BDA0003464893540000028
其中重建距离为z;The constructed phase-dual-resolution network f net2 is calculated to generate a pure phase hologram, and the reconstructed image of the hologram is calculated according to the angular spectrum method
Figure BDA0003464893540000028
where the reconstruction distance is z;

计算输入的目标强度图I和角谱法计算的重建图

Figure BDA0003464893540000029
之间的MS-SSIM损失和MSE损失;Calculate the input target intensity map I and the reconstructed map calculated by the angular spectrum method
Figure BDA0003464893540000029
between MS-SSIM loss and MSE loss;

将MS-SSIM损失和MSE损失按照一定的权重比例进行求和作为整体的一致性损失函数,实现相位-双分辨率网络模型fnet1和fnet2的同时训练。The MS-SSIM loss and the MSE loss are summed according to a certain weight ratio as the overall consistency loss function to realize the simultaneous training of the phase-dual-resolution network models f net1 and f net2 .

优选地,所述复值波场Uz的计算公式包括:Preferably, the calculation formula of the complex-valued wave field U z includes:

Figure BDA0003464893540000021
Figure BDA0003464893540000021

优选地,所述MS-SSIM损失和MSE损失按照一定的权重比例进行求和作为整体损失函数包括:Preferably, the MS-SSIM loss and the MSE loss are summed according to a certain weight ratio as an overall loss function including:

Figure BDA0003464893540000022
Figure BDA0003464893540000022

Figure BDA0003464893540000023
Figure BDA0003464893540000023

Figure BDA0003464893540000024
Figure BDA0003464893540000024

其中,I表示目标强度图,

Figure BDA0003464893540000025
表示重建图,
Figure BDA0003464893540000026
用于计算I和
Figure BDA0003464893540000027
之间的多尺度结构相似性,p表示像素,用于逐像素计算误差,依据经验设置α=0.84。where I represents the target intensity map,
Figure BDA0003464893540000025
represents the reconstructed graph,
Figure BDA0003464893540000026
used to calculate I and
Figure BDA0003464893540000027
The multi-scale structural similarity between , p represents the pixel, and is used to calculate the error pixel by pixel, and α = 0.84 is set empirically.

附图说明Description of drawings

图1示出了本申请的基于相位-双分辨率网络生成纯相位全息图的整体结构图;Fig. 1 shows the overall structure diagram of the pure phase hologram generated based on the phase-dual resolution network of the present application;

图2示出了本申请的相位-双分辨率网络结构图;Fig. 2 shows the phase-dual resolution network structure diagram of the present application;

图3示出了本申请搭建的全息显示结构示意图和仿真实物图FIG. 3 shows a schematic diagram of the holographic display structure and a simulated physical diagram of the holographic display constructed in the present application.

具体实施方式Detailed ways

下面对照附图并结合优选的实施方式对本发明进行清楚、完整地描述。The present invention will be clearly and completely described below with reference to the accompanying drawings and in conjunction with the preferred embodiments.

全息显示技术逐渐应用于三维场景重建、虚拟现实和增强显示系统中,使无需眼镜的三维显示模式成为可能。在全息显示技术中,全息图作为信息的载体,包含了物体的深度和强度信息,因此计算生成全息图是全息显示中关键一步。目前计算生成全息图的方法主要包括迭代优化算法和非迭代优化算法。然而两种算法多年来一直在全息图的优化质量和计算时间之间进行权衡。随着深度学习技术逐渐引入光学研究领域,并取得了诸多令人满意的成果。发明一种能够满足实时地计算生成高质量全息图的神经网络算法,对于实时全息显示技术的发展具有重要意义。Holographic display technology is gradually applied in 3D scene reconstruction, virtual reality and augmented display systems, making 3D display mode without glasses possible. In the holographic display technology, the hologram, as the carrier of information, contains the depth and intensity information of the object, so the computationally generated hologram is a key step in the holographic display. The current methods for generating holograms by computation mainly include iterative optimization algorithms and non-iterative optimization algorithms. However, both algorithms have traded off optimal quality of holograms and computation time for many years. With the gradual introduction of deep learning technology into the field of optics research, many satisfactory results have been achieved. The invention of a neural network algorithm capable of generating high-quality holograms by real-time calculation is of great significance to the development of real-time holographic display technology.

为此,本申请提供了一种基于相位-双分辨率网络的全息图快速生成方法。首先引入光场在自由空间中传播的角谱法,其计算公式如下:To this end, the present application provides a method for fast generation of holograms based on a phase-dual resolution network. First, the angular spectrum method of light field propagation in free space is introduced, and its calculation formula is as follows:

uz(x1,y1)=fAS{φ(x,y)}=IFFT{FFT{usrc(x,y)·eiφ(x,y)}·H(fx,fy)} (1)u z (x 1 , y 1 )=f AS {φ(x,y)}=IFFT{FFT{u src (x,y)·e iφ(x,y) }·H(f x ,f y ) } (1)

其中usrc表示相干光源产生的复数波场,uz表示衍射距离z处的复值波场,f(x,y)表示加载到空间光调制器中的纯相位全息图,FFT和IFFT分别是傅里叶变换和逆傅里叶变换算子,H(fx, fy)是角谱变换函数。H(fx,fy)的表达式表示为:where u src represents the complex wave field generated by the coherent light source, u z represents the complex valued wave field at the diffraction distance z, f(x, y) represents the pure phase hologram loaded into the spatial light modulator, FFT and IFFT are respectively Fourier transform and inverse Fourier transform operators, H(f x , f y ) is the angular spectral transform function. The expression of H(f x , f y ) is expressed as:

Figure BDA0003464893540000031
Figure BDA0003464893540000031

其中fx和fy是空间频率;λ是光的波长,z是衍射距离,H(fx,fy)是角光谱法的传递函数。由式(1)可知,usrc(x,y)是一个与内容无关的相干光源,因此角光谱法的可导性仅与H(fx,fy) 有关。根据式(2),角光谱法的传递函数与波长、传播距离、像素间距和空间分辨率有关。当光学系统固定时,四个参数是固定的,传递函数在卷积神经网络的反向传播中可以看作是一个常数。因此,角谱分析法的计算公式满足梯度传播。根据神经网络的通用近似定理,利用卷积神经网络正向传播产生纯相位全息图,并用角谱法计算全息图的数值重建。这个过程可以用以下形式表示:where f x and f y are the spatial frequencies; λ is the wavelength of light, z is the diffraction distance, and H(f x , f y ) is the transfer function of angular spectroscopy. It can be known from equation (1) that u src (x, y) is a content-independent coherent light source, so the derivability of angular spectroscopy is only related to H(f x , f y ). According to equation (2), the transfer function of angular spectroscopy is related to wavelength, propagation distance, pixel pitch and spatial resolution. When the optical system is fixed, the four parameters are fixed, and the transfer function can be regarded as a constant in the backpropagation of the convolutional neural network. Therefore, the calculation formula of the angular spectrum analysis method satisfies the gradient propagation. According to the general approximation theorem of neural network, the pure phase hologram is generated by forward propagation of convolutional neural network, and the numerical reconstruction of the hologram is calculated by the angular spectrum method. This process can be expressed in the following form:

Figure BDA0003464893540000032
Figure BDA0003464893540000032

其中I表示目标振幅,fnet表示卷积神经网络的近似函数,φholo表示纯相位全息图,

Figure BDA00034648935400000310
表示再现图像。通过损失函数和角谱法计算梯度的过程可以表示为:where I represents the target amplitude, fnet represents the approximate function of the convolutional neural network, φ holo represents the pure phase hologram,
Figure BDA00034648935400000310
Indicates the reproduced image. The process of calculating the gradient through the loss function and the angle spectrum method can be expressed as:

Figure BDA0003464893540000033
Figure BDA0003464893540000033

其中wk和bk表示卷积神经网络的第k层可训练参数,φholo表示全息图。where w k and b k represent the kth layer trainable parameters of the convolutional neural network, and φ holo represents the hologram.

图1示出了基于相位-双分辨率网络的全息图快速生成方法的整体结构。输入为任意一张强度图,经过训练的相位-双分辨率网络fnet1将其映射为一张相位图;将该相位图和输入的目标强度图通过式(5)计算为复值波场Uz;根据角谱法使复值波场在自由空间传播-z距离后得到复值波场U0;由于相位-双分辨率网络fnet2不能够进行复数运算,因此根据欧拉公式将复值波场U0分解为实部和虚部,并进行矩阵拼接操作;经过矩阵拼接的特征矩阵输入相位-双分辨率网络fnet2计算得到全息图。Figure 1 shows the overall structure of a fast generation method for holograms based on a phase-dual resolution network. The input is any intensity map, and the trained phase-dual-resolution network f net1 maps it to a phase map; the phase map and the input target intensity map are calculated as a complex-valued wave field U by formula (5). z ; According to the angular spectrum method, the complex-valued wavefield U 0 is obtained after the complex-valued wavefield propagates in the free space -z distance; since the phase-dual resolution network f net2 cannot perform complex number operations, the complex-valued wavefield is calculated according to Euler's formula. The wave field U 0 is decomposed into real and imaginary parts, and the matrix splicing operation is performed; the eigenmatrix after matrix splicing is input into the phase-dual resolution network f net2 to calculate the hologram.

Figure BDA0003464893540000034
Figure BDA0003464893540000034

上述过程为使用经过训练的相位-双分辨率网络fnet1和fnet2的生成全息图的过程。由于高质量的全息掩模较难获取,训练过程是通过无全息图掩模的卷积神经网络训练来完成。在训练阶段将相位-双分辨率网络fnet2计算得到的全息图,通过角谱法计算其重建图

Figure BDA0003464893540000035
此时式 (3)可进一步表示为:The above process is the process of generating holograms using the trained phase-dual resolution networks f net1 and f net2 . Since high-quality holographic masks are difficult to obtain, the training process is done by convolutional neural network training without holographic masks. In the training phase, the hologram calculated by the phase-double-resolution network f net2 is calculated by the angular spectrum method.
Figure BDA0003464893540000035
At this time, formula (3) can be further expressed as:

Figure BDA0003464893540000036
Figure BDA0003464893540000036

式中,

Figure BDA0003464893540000037
表示再现图像,φholo表示纯相位全息图,fAS表示角谱法,I表示目标振幅。分别计算输入的目标强度图I和角谱法计算的重建图
Figure BDA0003464893540000038
之间的MS-SSIM损失和MSE损失,并将其按照一定的比例求和作为整体的一致性损失函数,实现相位-双分辨率网络模型fnet1和fnet2的同时训练。一致性损失函数可表示为:In the formula,
Figure BDA0003464893540000037
represents the reproduced image, φ holo represents the pure phase hologram, f AS represents the angular spectrum method, and I represents the target amplitude. Calculate the input target intensity map I and the reconstructed map calculated by the angular spectrum method respectively
Figure BDA0003464893540000038
The MS-SSIM loss and the MSE loss between them are summed according to a certain ratio as the overall consistency loss function to realize the simultaneous training of the phase-dual-resolution network models f net1 and f net2 . The consistency loss function can be expressed as:

Figure BDA0003464893540000039
Figure BDA0003464893540000039

其中I表示目标振幅,

Figure BDA0003464893540000041
表示重建图,
Figure BDA0003464893540000042
用于计算I和
Figure BDA0003464893540000043
之间的多尺度结构相似性,p表示像素,用于逐像素计算误差,依据经验设置α=0.84。where I is the target amplitude,
Figure BDA0003464893540000041
represents the reconstructed graph,
Figure BDA0003464893540000042
used to calculate I and
Figure BDA0003464893540000043
The multi-scale structural similarity between , p represents the pixel, and is used to calculate the error pixel by pixel, and α = 0.84 is set empirically.

图2示出了相位-双分辨率网络结构图。其中CBP表示卷积模块,包含卷积层、批归一化层和PReLU激活函数。RB表示残差模块,RBB表示残差瓶颈模块。DAPPM代表深度聚合金字塔池化模块。相位-双分辨率网络主要由编码器和解码器组成,其中编码器是由高分辨率分支和低分辨率分支组成。Figure 2 shows a phase-dual resolution network structure diagram. where CBP represents the convolution module, which includes a convolution layer, a batch normalization layer, and a PReLU activation function. RB stands for residual block, and RBB stands for residual bottleneck block. DAPPM stands for Deep Aggregation Pyramid Pooling Module. The phase-dual resolution network is mainly composed of an encoder and a decoder, where the encoder is composed of a high-resolution branch and a low-resolution branch.

编码器的高分辨率分支可以保留较多的高频特征,而低分辨率分支可以提取目标振幅的高级抽象特征。编码器的前两层网络使用组卷积进行特征提取。编码器中剩余的网络层均使用空洞卷积进行特征提取,其中空洞卷积中的空洞因子,遵循卷积中两个非零值之间的最大距离计算公式:The high-resolution branch of the encoder can retain more high-frequency features, while the low-resolution branch can extract high-level abstract features of the target amplitude. The first two layers of the encoder network use group convolutions for feature extraction. The remaining network layers in the encoder use atrous convolution for feature extraction, where the atrous factor in atrous convolution follows the formula for calculating the maximum distance between two non-zero values in the convolution:

Mi=max[Mi+1-2ri,Mi+1-2(Mi+1-ri),ri],Mn=rn (8)M i =max[M i+1 -2r i ,M i+1 -2(M i+1 -r i ), r i ],M n =rn (8)

其中i表示第i层卷积,Mi表示第i层卷积输出中非零值之间的最大距离。ri表示卷积第i层的空洞因子。通过使用组卷积可以减少相位-双分辨率网络生成高分辨率全息图时的计算量和GPU内存占用,使用空洞卷积可以扩大卷积操作的感受野增强特征提取能力。where i represents the i-th layer convolution, and M i represents the maximum distance between non-zero values in the i-th layer convolution output. r i represents the dilation factor of the i-th layer of the convolution. The use of group convolution can reduce the amount of computation and GPU memory usage when generating high-resolution holograms by the phase-dual resolution network, and the use of atrous convolution can expand the receptive field of convolution operations to enhance feature extraction capabilities.

在解码器中,通过sub-pixel卷积对编码器提取的特征进行上采样,直至与输入图像分辨率相同。编码器的第三层网络输出使用跳跃连接至解码器的第一层进行特征拼接,增强特征解码能力。将1×1卷积作为解码器的最后一层,使得输出为单通道图像。将单通道图像的灰度值归一化至[-π,π],使得网络的最终输出为相位图。In the decoder, the features extracted by the encoder are upsampled by sub-pixel convolutions until the same resolution as the input image. The encoder's third-layer network output uses skip connections to the decoder's first layer for feature concatenation, enhancing feature decoding capabilities. A 1×1 convolution is used as the last layer of the decoder so that the output is a single-channel image. The grayscale values of the single-channel images are normalized to [-π,π] so that the final output of the network is a phase map.

图3示出了本申请搭建的全息显示结构示意图和仿真实物图。空间光调制器的型号为 Holoeye Photonics AG,像素间距为8μm,激光波长为532nm,传播距离为40cm。FIG. 3 shows a schematic diagram of a holographic display structure and a simulated physical diagram of the holographic display constructed in the present application. The model of the spatial light modulator is Holoeye Photonics AG, the pixel pitch is 8 μm, the laser wavelength is 532 nm, and the propagation distance is 40 cm.

Claims (5)

1.一种基于相位-双分辨率网络的全息图快速生成方法,其特征在于:1. a kind of hologram fast generation method based on phase-double resolution network, it is characterized in that: (1)构建相位-双分辨率网络模型fnet1和fnet2(1) Build phase-dual resolution network models f net1 and f net2 ; (2)经过训练的相位-双分辨率网络fnet1,根据输入的目标强度图I计算出初始相位φ0(2) The trained phase-dual resolution network f net1 calculates the initial phase φ 0 according to the input target intensity map I; (3)根据初始相位φ0与目标强度图I,计算出复值波场Uz;根据角谱法,计算复值波场Uz在自由空间传播-z后得到的复值波场U0(3) Calculate the complex-valued wave field U z according to the initial phase φ 0 and the target intensity map I; according to the angular spectrum method, calculate the complex-valued wave field U 0 obtained after the complex-valued wave field U z propagates in free space -z ; (4)经过训练的相位-双分辨率网络fnet2,根据输入的复值波场U0计算出纯相位的全息图。(4) The trained phase-dual-resolution network f net2 calculates a phase-only hologram according to the input complex-valued wave field U 0 . 2.根据权利要求1所述方法,其特征在于,构建相位-双分辨率网络模型包括:2. The method according to claim 1, wherein building a phase-dual resolution network model comprises: (1)在相位-双分辨率网络的编码器中,前两层网络使用组卷积进行特征提取;编码器中剩余的网络层均使用空洞卷积进行特征提取,其中空洞卷积中的空洞因子,遵循卷积中两个非零值之间的最大距离计算公式:(1) In the encoder of the phase-dual resolution network, the first two layers of the network use group convolution for feature extraction; the remaining network layers in the encoder use atrous convolution for feature extraction, and the holes in atrous convolution are used for feature extraction. factor, following the formula for calculating the maximum distance between two non-zero values in the convolution: Mi=max[Mi+1-2ri,Mi+1-2(Mi+1-ri),ri],Mn=rnM i =max[M i+ 1-2r i ,M i+ 1-2(M i+1 -r i ),r i ],M n =r n ; (2)在解码器中,通过sub-pixel卷积对编码器提取的特征进行上采样,直至与输入图像分辨率相同;编码器的第三层网络输出使用跳跃连接至解码器的第一层进行特征拼接;将1×1卷积作为解码器的最后一层,使得输出为单通道图像;(2) In the decoder, the features extracted by the encoder are up-sampled by sub-pixel convolution until the resolution is the same as the input image; the third layer network output of the encoder is connected to the first layer of the decoder using skips Perform feature stitching; use 1×1 convolution as the last layer of the decoder, so that the output is a single-channel image; (3)将单通道图像的灰度值归一化至[-π,π]。(3) Normalize the gray value of the single-channel image to [-π,π]. 3.根据权利要求1所述的方法,其特征在于,所述经过训练的相位-双分辨率网络的训练,包括:3. The method according to claim 1, wherein the training of the trained phase-dual resolution network comprises: (1)已构建的相位-双分辨率网络fnet2计算生成纯相位的全息图,根据角谱法计算出全息图的重建图
Figure FDA0003464893530000011
其中重建距离为z;
(1) The constructed phase-dual resolution network f net2 calculates and generates a pure phase hologram, and calculates the reconstructed image of the hologram according to the angular spectrum method
Figure FDA0003464893530000011
where the reconstruction distance is z;
(2)分别计算输入的目标强度图I和角谱法计算的重建图
Figure FDA0003464893530000012
之间的MS-SSIM损失和MSE损失;
(2) Calculate the input target intensity map I and the reconstruction map calculated by the angular spectrum method respectively
Figure FDA0003464893530000012
between MS-SSIM loss and MSE loss;
(3)将MS-SSIM损失和MSE损失按照一定的权重比例进行求和作为整体的一致性损失函数,实现相位-双分辨率网络模型fnet1和fnet2的同时训练。(3) The MS-SSIM loss and the MSE loss are summed according to a certain weight ratio as the overall consistency loss function to realize the simultaneous training of the phase-dual-resolution network models f net1 and f net2 .
4.根据权利要求1所述的方法,其特征在于,所述复值波场Uz,计算公式包括:4. The method according to claim 1, wherein, for the complex-valued wave field U z , the calculation formula comprises:
Figure FDA0003464893530000013
Figure FDA0003464893530000013
5.根据权利要求3所述方法,其特征在于,所述整体的一致性损失函数,计算公式包括:5. The method according to claim 3, wherein the overall consistency loss function, the calculation formula comprises:
Figure FDA0003464893530000014
Figure FDA0003464893530000014
Figure FDA0003464893530000015
Figure FDA0003464893530000015
Figure FDA0003464893530000016
Figure FDA0003464893530000016
其中,I表示目标振幅,
Figure FDA0003464893530000017
表示重构振幅,
Figure FDA0003464893530000018
用于计算I和
Figure FDA0003464893530000019
之间的多尺度结构相似性,p表示像素,用于逐像素计算误差,依据经验设置α=0.84。
where I is the target amplitude,
Figure FDA0003464893530000017
represents the reconstructed amplitude,
Figure FDA0003464893530000018
used to calculate I and
Figure FDA0003464893530000019
The multi-scale structural similarity between , p represents the pixel, and is used to calculate the error pixel by pixel, and α = 0.84 is set empirically.
CN202210029238.1A 2022-01-11 2022-01-11 A fast generation method of hologram based on phase-dual resolution network Pending CN114387395A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210029238.1A CN114387395A (en) 2022-01-11 2022-01-11 A fast generation method of hologram based on phase-dual resolution network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210029238.1A CN114387395A (en) 2022-01-11 2022-01-11 A fast generation method of hologram based on phase-dual resolution network

Publications (1)

Publication Number Publication Date
CN114387395A true CN114387395A (en) 2022-04-22

Family

ID=81201867

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210029238.1A Pending CN114387395A (en) 2022-01-11 2022-01-11 A fast generation method of hologram based on phase-dual resolution network

Country Status (1)

Country Link
CN (1) CN114387395A (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114967398A (en) * 2022-05-13 2022-08-30 安徽大学 Large-size two-dimensional calculation hologram real-time generation method based on deep learning
CN115097709A (en) * 2022-07-05 2022-09-23 东南大学 A holographic coding method based on complex number optimizer or complex number solver
CN115173227A (en) * 2022-07-28 2022-10-11 深圳博升光电科技有限公司 Super-surface structure, laser array, optical imaging chip and preparation method
CN115294412A (en) * 2022-10-10 2022-11-04 临沂大学 Real-time coal and rock segmentation network generation method based on deep learning

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106054570A (en) * 2016-04-26 2016-10-26 上海大学 Method for realizing large-phase reconstruction of single digital hologram by adopting intensity transmission equation
US20190294108A1 (en) * 2018-03-21 2019-09-26 The Regents Of The University Of California Method and system for phase recovery and holographic image reconstruction using a neural network
WO2021216747A1 (en) * 2020-04-21 2021-10-28 Massachusetts Institute Of Technology Real-Time Photorealistic 3D Holography with Deep Neural Networks
CN113658330A (en) * 2021-08-17 2021-11-16 东南大学 A holographic coding method based on neural network

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106054570A (en) * 2016-04-26 2016-10-26 上海大学 Method for realizing large-phase reconstruction of single digital hologram by adopting intensity transmission equation
US20190294108A1 (en) * 2018-03-21 2019-09-26 The Regents Of The University Of California Method and system for phase recovery and holographic image reconstruction using a neural network
WO2021216747A1 (en) * 2020-04-21 2021-10-28 Massachusetts Institute Of Technology Real-Time Photorealistic 3D Holography with Deep Neural Networks
CN113658330A (en) * 2021-08-17 2021-11-16 东南大学 A holographic coding method based on neural network

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
TING YU, ET AL: "Phase dual-resolution networks for a computer-generated hologram", 《OPTICS EXPRESS》, 10 January 2022 (2022-01-10), pages 2378 - 2389 *
王帅;陈波: "数字全息深度学习重建方法", 现代计算机, no. 034, 31 December 2020 (2020-12-31), pages 68 - 72 *

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114967398A (en) * 2022-05-13 2022-08-30 安徽大学 Large-size two-dimensional calculation hologram real-time generation method based on deep learning
CN114967398B (en) * 2022-05-13 2024-05-31 安徽大学 Large-size two-dimensional calculation hologram real-time generation method based on deep learning
CN115097709A (en) * 2022-07-05 2022-09-23 东南大学 A holographic coding method based on complex number optimizer or complex number solver
CN115097709B (en) * 2022-07-05 2023-11-17 东南大学 Holographic coding method based on complex optimizer or complex solver
CN115173227A (en) * 2022-07-28 2022-10-11 深圳博升光电科技有限公司 Super-surface structure, laser array, optical imaging chip and preparation method
CN115294412A (en) * 2022-10-10 2022-11-04 临沂大学 Real-time coal and rock segmentation network generation method based on deep learning

Similar Documents

Publication Publication Date Title
CN114387395A (en) A fast generation method of hologram based on phase-dual resolution network
CN105976408A (en) Digital holographic compression transmission method of quantum backward propagation nerve network
CN112598602B (en) A mask-based deep learning video moiré removal method
CN113467211B (en) Holographic encoding method based on gradient reduction of spectral loss function
CN114067015B (en) A method and system for generating pure phase holograms combined with DNN
CN115100300B (en) Computer hologram compression method based on attention-based heuristic neural network
CN115797231A (en) Real-time hologram generation method based on Fourier-inspired neural network
CN115690252A (en) Hologram reconstruction method and system based on convolutional neural network
CN117876591B (en) Real fuzzy three-dimensional hologram reconstruction method for combined training of multiple neural networks
CN113658330B (en) Holographic encoding method based on neural network
Rong et al. Reconstruction efficiency enhancement of amplitude-type holograms by using Single-Scale Retinex algorithm
Chen et al. Convolutional neural network for phase-only hologram optimization based on the point source method with the holographic viewing-window
CN114255308B (en) A method for generating blendshape of human face based on single-view 3D reconstruction
Lee et al. Complex hologram generation of multi-depth images using deep neural network
KR20210095433A (en) Deep learning model for generating phase hologram at high speed and learning method thereof
CN117474819A (en) Low-light image enhancement method based on semantic information guidance and light curve estimation
Imtiaz et al. Advanced deep learning model for direct phase-only hologram generation using complex-valued neural networks
Liu et al. Rapid computer-generated hologram with lightweight local and global self-attention network
Yan et al. Generation of Multiple‐Depth 3D Computer‐Generated Holograms from 2D‐Image‐Datasets Trained CNN
Jia et al. Deep learning-based approach for efficient generation and transmission of high-definition computer-generated holography
Dong et al. Divide-conquer-and-merge: memory-and time-efficient holographic displays
Fang et al. Generating high-quality phase-only holograms of binary images using global loss and stochastic homogenization training strategy
Shevkunov et al. Deep convolutional neural network-based lensless quantitative phase retrieval
CN111176091A (en) 3D holographic animal model projection method
Liang et al. A time-multiplexed neural network framework for enhancing focal cues in multi-depth 3D holographic displays from 2D image input

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination