CN112132959B - Digital rock core image processing method and device, computer equipment and storage medium - Google Patents

Digital rock core image processing method and device, computer equipment and storage medium Download PDF

Info

Publication number
CN112132959B
CN112132959B CN202011027637.1A CN202011027637A CN112132959B CN 112132959 B CN112132959 B CN 112132959B CN 202011027637 A CN202011027637 A CN 202011027637A CN 112132959 B CN112132959 B CN 112132959B
Authority
CN
China
Prior art keywords
image
dimensional
loss
digital core
model
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202011027637.1A
Other languages
Chinese (zh)
Other versions
CN112132959A (en
Inventor
刘彦昌
单立群
孙明港
戚先锋
祁妍嫣
汤敏
孔维芳
白雪原
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Northeast Petroleum University
Original Assignee
Northeast Petroleum University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Northeast Petroleum University filed Critical Northeast Petroleum University
Priority to CN202011027637.1A priority Critical patent/CN112132959B/en
Publication of CN112132959A publication Critical patent/CN112132959A/en
Application granted granted Critical
Publication of CN112132959B publication Critical patent/CN112132959B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformation in the plane of the image
    • G06T3/40Scaling the whole image or part thereof
    • G06T3/4007Interpolation-based scaling, e.g. bilinear interpolation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformation in the plane of the image
    • G06T3/40Scaling the whole image or part thereof
    • G06T3/4053Super resolution, i.e. output image resolution higher than sensor resolution
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformation in the plane of the image
    • G06T3/60Rotation of a whole image or part thereof
    • G06T5/73
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10072Tomographic images
    • G06T2207/10081Computed x-ray tomography [CT]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Abstract

The embodiment of the invention discloses a digital rock core image processing method, a digital rock core image processing device, computer equipment and a storage medium, wherein the digital rock core image processing method comprises the following steps: acquiring a digital core training image; the digital core training images comprise digital core three-dimensional CT images with different resolutions; respectively training a generator model and a discriminator model for generating a confrontation network according to the digital core training images to obtain a three-dimensional digital core image generation confrontation network model; and generating a generator model of a confrontation network model according to the three-dimensional digital core image, and reconstructing the input digital core image to be optimized to obtain a target three-dimensional digital core image. The embodiment of the invention can improve the image quality of the three-dimensional digital core reconstructed image.

Description

Digital rock core image processing method and device, computer equipment and storage medium
Technical Field
The embodiment of the invention relates to the technical field of image processing, in particular to a digital core image processing method and device, computer equipment and a storage medium.
Background
Computed Tomography (CT) techniques can generate 3D images detailing the internal microstructure of porous rock samples, which can help determine the petrophysical properties and flow characteristics of the rock. CT can resolve features as low as a few microns across a Field of view (FOV) of a few millimeters, is sufficient to characterize the microstructure of conventional rock, can assist geological researchers in analyzing the physical properties of rock, and plays an important role in the fields of geological and oil exploration. In fact, a complete 3D digital core CT image is composed of two-dimensional (2D) slice images, and due to the inherent limitations of CT equipment, setting high resolution not only requires high cost, but also results in a reduction in FOV, thereby resulting in a loss of the long-range properties of reservoir rock. In many cases, only low resolution CT images can be obtained. Therefore, the super-resolution reconstruction algorithm is an effective method for improving the resolution of the digital core CT image, and can provide clearer sample data for subsequent geological research.
Image super-resolution reconstruction is one of the classical tasks in computer vision, with the goal of reconstructing a high resolution image from a low resolution input image. The method mainly comprises three types of implementation methods: interpolation-based algorithms, reconstruction-based algorithms, and learning-based algorithms. The interpolation algorithm for obtaining the target pixel value by using the neighborhood pixels to perform weight combination is widely applied due to the characteristics of simple form and high processing speed. However, such algorithms lose high frequency information of the image, lose image details, and cause image blurring. On the basis, an edge-based interpolation algorithm appears, which retains high-frequency information of the image to a certain extent, however, the edge-based interpolation algorithm is difficult to process texture regions of the image, and the application range is greatly limited. The reconstruction-based method is to constrain the consistency between high and low resolution image variables by modeling a degradation model of the image, and then estimate the high resolution image. This kind of method obtains stable solution through regular term constraint, but adds mandatory prior information and destroys the original structural feature of the image, leads to the distortion of the reconstructed image. And the calculation complexity is high, and the requirement of real-time processing in reconstruction cannot be met.
The learning-based method utilizes supervised machine learning to establish a nonlinear mapping relation between a low-resolution image and a high-resolution image for image reconstruction. The method can extract the prior information of the image from the sample set, thereby obtaining higher reconstruction precision. However, the conventional reconstruction method based on learning can only extract simpler features of the image, which is not enough to fully represent the image information, so that the reconstruction quality of the image is limited. In recent years, a Deep Learning method is applied to Image Super-Resolution reconstruction, and a Super-Resolution reconstruction method (SRCNN) based on a Convolutional neural Network is proposed, in which an end-to-end mapping function from a low-Resolution Image to a high-Resolution Image is directly learned in an end-to-end method, and a very clear generated Image can be generated. However, due to the influence of the gradient diffusion phenomenon in the convolutional neural network, the network degradation phenomenon occurs in the SRCNN under the condition of a large convolution depth, that is, the reconstruction quality of the image is reduced, and the reconstruction performance of the algorithm is limited. Subsequently, a Generative countermeasure network is used for image Super-Resolution reconstruction, and a Super-Resolution Generative adaptive network (SRGAN) is proposed, which can recover more high-frequency details. At present, on the basis of SRGAN, an ESRGAN network is proposed to further improve the confrontation loss and the perception loss, and a residual error dense block is introduced to construct the network, so that the reconstruction precision of an image is improved.
Although the super-resolution method based on the deep neural network has great breakthrough in image reconstruction quality, the method has many defects. The lack of a deep neural network (e.g., SRCNN) trained against this training can make the reconstructed image too smooth to fit human perception of natural pictures. By adopting a method of resisting a neural Network, such as an SRGAN (simplified Super-Resolution adaptive Network) and an ESRGAN (Enhanced Super-Resolution generation resisting Network), the problem of too smooth reconstructed images can be solved, and the problem is more suitable for human perception of natural pictures. In addition, the algorithm is mainly applied to two-dimensional image reconstruction at present and is not applied to three-dimensional core CT image reconstruction.
Disclosure of Invention
The embodiment of the invention provides a digital core image processing method, a digital core image processing device, computer equipment and a storage medium, which are used for improving the image quality of a three-dimensional digital core reconstruction image.
In a first aspect, an embodiment of the present invention provides a digital core image processing method, including:
acquiring a digital core training image; the digital core training images comprise digital core three-dimensional CT images with different resolutions;
respectively training a generator model and a discriminator model for generating a confrontation network according to the digital core training images to obtain a three-dimensional digital core image generation confrontation network model;
and generating a generator model of a confrontation network model according to the three-dimensional digital core image, and reconstructing the input digital core image to be optimized to obtain a target three-dimensional digital core image.
In a second aspect, an embodiment of the present invention further provides a digital core image processing apparatus, including:
the training image acquisition module is used for acquiring a digital core training image; the digital core training images comprise digital core three-dimensional CT images with different resolutions;
the network model acquisition module is used for respectively training a generator model and a discriminator model for generating a confrontation network according to the digital core training image to obtain a three-dimensional digital core image generation confrontation network model;
and the image reconstruction module is used for reconstructing the input digital core image to be optimized according to a generator model of the three-dimensional digital core image generation countermeasure network model to obtain a target three-dimensional digital core image.
In a third aspect, an embodiment of the present invention further provides a computer device, where the computer device includes:
one or more processors;
storage means for storing one or more programs;
when executed by the one or more processors, cause the one or more processors to implement a method for digital core image processing as provided by any of the embodiments of the present invention.
In a fourth aspect, an embodiment of the present invention further provides a computer storage medium, on which a computer program is stored, where the computer program, when executed by a processor, implements the digital core image processing method provided in any embodiment of the present invention.
According to the embodiment of the invention, the digital core three-dimensional CT images with different resolutions are used as the digital core training images, the generator model and the discriminator model of the countermeasure network are respectively generated by training according to the digital core training images, the three-dimensional digital core image generation countermeasure network model is obtained, the generator model of the countermeasure network model is generated according to the three-dimensional digital core image, the input digital core image to be optimized is reconstructed, the target three-dimensional digital core image is obtained, the problem of low image quality of the existing three-dimensional digital core image reconstruction image is solved, and the image quality of the three-dimensional digital core reconstruction image is improved.
Drawings
Fig. 1 is a flowchart of a digital core image processing method according to an embodiment of the present invention;
fig. 2 is a flowchart of a digital core image processing method according to a second embodiment of the present disclosure;
fig. 3 is a schematic flowchart of model training and digital core image processing according to a second embodiment of the present disclosure;
FIG. 4 is a schematic structural diagram of a generator model according to a second embodiment of the present invention;
fig. 5 is a schematic structural diagram of a discriminator model according to a second embodiment of the present invention;
fig. 6 is a schematic diagram illustrating an effect of generating a target three-dimensional digital core image by using a generator model for generating a confrontation network model by using a three-dimensional digital core image according to a second embodiment of the present invention;
fig. 7 is a schematic diagram of a comparative effect of an evaluation index provided in the second embodiment of the present invention;
fig. 8 is a schematic diagram of a digital core image processing apparatus according to a third embodiment of the present disclosure;
fig. 9 is a schematic structural diagram of a computer device according to a fourth embodiment of the present invention.
Detailed Description
The present invention will be described in further detail with reference to the accompanying drawings and examples. It is to be understood that the specific embodiments described herein are merely illustrative of the invention and are not limiting of the invention.
It should be further noted that, for the convenience of description, only some but not all of the relevant aspects of the present invention are shown in the drawings. Before discussing exemplary embodiments in more detail, it should be noted that some exemplary embodiments are described as processes or methods depicted as flowcharts. Although a flowchart may describe the operations (or steps) as a sequential process, many of the operations can be performed in parallel, concurrently or simultaneously. In addition, the order of the operations may be re-arranged. The process may be terminated when its operations are completed, but may have additional steps not included in the figure. The processes may correspond to methods, functions, procedures, subroutines, and the like.
Example one
Fig. 1 is a flowchart of a digital core image processing method according to an embodiment of the present invention, where the present embodiment is applicable to a case where a three-dimensional digital core image to be optimized is reconstructed to improve image quality, and the method may be executed by a digital core image processing apparatus, and the apparatus may be implemented by software and/or hardware, and may be generally integrated in a computer device. Accordingly, as shown in fig. 1, the method comprises the following operations:
s110, acquiring a digital core training image; the digital core training image comprises digital core three-dimensional CT images with different resolutions.
The digital core training image can be a training image constructed according to an existing digital core image and is used for training a three-dimensional digital core image generation countermeasure network model. And the digital core three-dimensional CT image is also the three-dimensional CT image of the core.
In an embodiment of the present disclosure, the digital core training images may include digital core three-dimensional CT images of different resolutions. Optionally, the digital core training image may include two digital core three-dimensional CT images with different resolutions, such as a low resolution and a high resolution. The resolution corresponding to the digital core three-dimensional CT image with the low resolution is lower than a first set resolution threshold, and the resolution corresponding to the digital core three-dimensional CT image with the high resolution is higher than a second set resolution threshold. The first set resolution threshold and the second set resolution threshold may be set according to actual requirements, and the embodiment of the present invention does not limit specific values of the first set resolution threshold and the second set resolution threshold.
And S120, respectively training a generator model and a discriminator model for generating the confrontation network according to the digital core training image to obtain a three-dimensional digital core image generation confrontation network model.
The generation of the countermeasure network consists of two neural networks, namely a generator model and a discriminator model. The generator model is mainly used to generate new data from acquired content (such as images) and the discriminator model is mainly used to discriminate whether the data generated by the generator model is legitimate. The three-dimensional digital core image generation confrontation network model is a generated confrontation network model obtained by successfully training the generator model and the discriminator model by utilizing the digital core training image.
In the embodiment of the invention, the countermeasure network is trained by using the digital core three-dimensional CT images with different resolutions as the digital core training images, so that the three-dimensional digital core images obtained by training can be effectively reconstructed by generating the countermeasure network model.
S130, a generator model of a confrontation network model is generated according to the three-dimensional digital core image, and the input digital core image to be optimized is reconstructed to obtain a target three-dimensional digital core image.
The digital core image to be optimized may be a low-resolution three-dimensional digital core image. The target three-dimensional digital core image may be a three-dimensional digital core image having a particular resolution.
Correspondingly, when the training of the three-dimensional digital core image generation countermeasure network model is successful, the digital core image to be optimized is input into the generator model of the three-dimensional digital core image generation countermeasure network model, and the generator model of the three-dimensional digital core image generation countermeasure network model is used for reconstructing the input digital core image to be optimized to obtain the target three-dimensional digital core image. The resolution ratio of the target three-dimensional digital core image is higher than that of the digital core image to be optimized, so that the reconstructed target three-dimensional digital core image can recover detail information in the digital core image to be optimized, the definition of the image is higher, and the image quality of the reconstructed three-dimensional digital core image is improved.
According to the embodiment of the invention, the digital core three-dimensional CT images with different resolutions are used as the digital core training images, the generator model and the discriminator model of the countermeasure network are respectively generated by training according to the digital core training images, the three-dimensional digital core image generation countermeasure network model is obtained, the generator model of the countermeasure network model is generated according to the three-dimensional digital core image, the input digital core image to be optimized is reconstructed, the target three-dimensional digital core image is obtained, the problem of low image quality of the existing three-dimensional digital core image reconstruction image is solved, and the image quality of the three-dimensional digital core reconstruction image is improved.
Example two
Fig. 2 is a flowchart of a digital core image processing method according to a second embodiment of the present invention, and fig. 3 is a flowchart of a model training and digital core image processing according to a second embodiment of the present invention. Accordingly, as shown in fig. 2 and 3, the method of the present embodiment may include:
s210, acquiring a two-dimensional CT image of at least one known rock.
The known rock may be a rock type such as sandstone, carbonate, or coal geologic rock (i.e., coal rock), and the embodiment of the present invention is not limited to the specific rock type of the known rock.
In an embodiment of the present invention, when acquiring the digital core training image, a two-dimensional CT image of at least one known rock may be acquired first. For example, two-dimensional CT images from multiple sandstone, carbonate, and coal geological cores may be gathered.
S220, performing image preprocessing and image enhancement processing on the two-dimensional CT image to obtain a two-dimensional CT processed image.
The image preprocessing is to perform preprocessing operation on the image, and the image enhancement processing is to perform data enhancement on the image. The two-dimensional CT processed image can be a two-dimensional CT image obtained by performing image preprocessing and image enhancement on the two-dimensional CT image.
Correspondingly, after the two-dimensional CT image of at least one known rock is obtained, image preprocessing operations such as drying removal and the like can be carried out on the two-dimensional CT image, and then image enhancement processing such as rotation, turning and blurring operations and the like can be carried out on the two-dimensional CT image obtained through image preprocessing, so that the number of samples in the digital core training image is expanded.
And S230, constructing a three-dimensional CT image according to the two-dimensional CT processing image.
Furthermore, after the two-dimensional CT processing image is obtained, a three-dimensional CT image can be constructed according to the two-dimensional CT processing image.
Illustratively, the collected two-dimensional CT images may be processed and the resulting images stacked to form a three-dimensional CT image.
S240, constructing the digital core training image according to the three-dimensional CT image.
Correspondingly, after the three-dimensional CT image is constructed according to the two-dimensional CT processing image, the digital core training image can be constructed by utilizing the three-dimensional CT image.
In an optional embodiment of the present invention, the constructing the digital core training image according to the three-dimensional CT image may include: segmenting the three-dimensional CT image to obtain a three-dimensional CT segmented image with a first resolution; carrying out scaling processing on the three-dimensional CT segmentation image according to a set scaling value to obtain a three-dimensional CT scaling image with a second resolution ratio; and constructing the digital core training image according to the three-dimensional CT segmentation image and the three-dimensional CT scaling image.
The three-dimensional CT segmented image may be a three-dimensional CT image of a small module obtained by segmenting the three-dimensional CT image, and the first resolution is an original image resolution of the three-dimensional CT image constructed according to the two-dimensional CT processed image. The setting of the scaling value may be set according to actual requirements, such as 1/2, 1/4, or 1/8, and the embodiment of the present invention does not limit the specific value of the setting of the scaling value. The second resolution is also the resolution corresponding to the zoomed image obtained by zooming the three-dimensional CT segmented image. It will be appreciated that the second resolution is a lower numerical value than the first resolution.
Specifically, when the digital core training image is constructed according to the three-dimensional CT image, the acquired three-dimensional CT image may be segmented to obtain a three-dimensional CT segmented image with a first resolution, for example, the acquired three-dimensional CT image is segmented into a cube with a fixed size, so as to save training time and calculation overhead. Then, the three-dimensional CT segmented image may be scaled according to a set scaling value, for example, by a fixed factor using bicubic interpolation. Thus, the generated three-dimensional CT zoom image is a low-resolution image, and the three-dimensional CT segmentation image is a high-resolution image of the original scale. Correspondingly, the three-dimensional CT segmentation image and the three-dimensional CT scaling image construct a digital core training image.
In one specific example, assume that a digital core training image is constructed using a disclosed digital core data set. The data set contains ten thousand original images of various digital rocks (sandstone, carbonate, coal, etc.) at 500 x 500 high resolution, with image resolutions ranging from 2.7 to 25 μm. After a two-dimensional CT image of a known rock is obtained, image denoising pretreatment is firstly carried out, and then data enhancement is carried out by adopting rotation, turning and fuzzy operations, so that the number of samples of a training image data set is expanded. The 500 x 500 two-dimensional CT images in the data set may then be two-dimensional slice stacked to form a 500 x 500 three-dimensional CT image. Further, the three-dimensional CT image is divided into 80 × 80 × 80 cubes to obtain a high-resolution three-dimensional CT divided image, and the three-dimensional CT divided image is reduced to 1/2 (40 × 40 × 40 pixel size) and 1/4 (20 × 20 × 20 pixel size) of the original size using bicubic interpolation to generate a low-resolution three-dimensional CT scaled image. Correspondingly, the three-dimensional CT segmentation image of the original scale obtained by segmentation is used as a high resolution image, the three-dimensional CT scaling image is used as a low resolution image, and the two images form a digital core training image.
And S250, inputting the three-dimensional CT scaling image with the second resolution in the digital core training image into the generator model, and taking the image output by the generator model as a third resolution image.
Wherein the third resolution image is an image output by the generator model according to the low-resolution three-dimensional CT scaling image, and the resolution of the third resolution image is higher than the second resolution of the three-dimensional CT scaling image.
In the embodiment of the invention, when the generator model and the discriminator model of the countermeasure network are generated by training with the digital core training image, the three-dimensional CT scaling image of the second resolution can be input into the generator model, and the image output by the generator model is taken as the third resolution image. The third resolution image may also be referred to as a super-resolution image.
S260, simultaneously inputting the third resolution image and the three-dimensional CT segmentation image with the first resolution in the digital core training image into the discriminator model, and determining whether the third resolution image is matched with the three-dimensional CT segmentation image with the first resolution according to an output value of the discriminator model so as to train the discriminator model.
Correspondingly, after the third resolution image is obtained, the third resolution image and the three-dimensional CT segmentation image can be simultaneously input into the discriminator model, and whether the third resolution image is matched with the three-dimensional CT segmentation image or not is determined according to an output value of the discriminator model, so that the discriminator model is trained. The output value of the discriminator model may be the probability of matching the third resolution image with the three-dimensional CT segmented image, or the same value or different value of the third resolution image and the three-dimensional CT segmented image, which is not limited in the embodiment of the present invention.
When the generation countermeasure network is trained, the discriminator model can be trained firstly, and after the discriminator model is trained successfully, the generator model is trained by using the successfully trained discriminator model. The specific process of training the discriminator model may be: if the third resolution image is determined to be not matched with the three-dimensional CT segmented image according to the output value of the discriminator model, if the probability of matching the third resolution image with the three-dimensional CT segmented image is lower than a set probability threshold (for example, 90% and the like can be set according to actual requirements), or the third resolution image is directly output to be different from the three-dimensional CT segmented image, the generator model is continuously optimized until the third resolution image generated by the generator model can "cheat" the discriminator, namely when the third resolution image is determined to be matched with the three-dimensional CT segmented image according to the output value of the discriminator model, the success of the training of the discriminator model can be determined.
S270, after the discriminant model is successfully trained, fixing the discriminant model, and returning to execute the operation of inputting the three-dimensional CT scaling image with the second resolution in the character core training image into the generator model to continue training the generator model until the generator model is successfully trained.
Accordingly, after the training of the discriminator model is successful, the discriminator model may be fixed, and the operation of inputting the three-dimensional CT scaled image into the generator model is returned to continue training the generator model until it is determined that the training of the generator model is successful. That is, the training process of the generator model and the discriminator model belongs to a process in which the two optimize training alternately.
In an optional embodiment of the present invention, the generator model is a fusion attention mechanism residual U-net network, the fusion attention mechanism residual U-net network includes a set number of network structures, and an encoder and a decoder of the fusion attention mechanism residual U-net network perform multi-scale image fusion through an attention gate structure; wherein the attention gate structure may be implemented based on the following formula:
Figure BDA0002702578530000121
Figure BDA0002702578530000122
Figure BDA0002702578530000123
wherein the content of the first and second substances,
Figure BDA0002702578530000124
representing a transfer function, # T 、/>
Figure BDA0002702578530000125
And &>
Figure BDA0002702578530000126
Represents a 1 × 1 × 1 convolution operation, b g And b ψ Representing the offset coefficient, σ 1 Represents the ReLU function->
Figure BDA0002702578530000127
Input representing the attention gate structure, g i A gate signal, representing the provision of high level context information>
Figure BDA0002702578530000128
Denotes the attention coefficient, σ 2 Representing sigmoid activation function, Θ att Denotes a parameter set, which may be psi T
Figure BDA0002702578530000129
b g And b ψ ,/>
Figure BDA00027025785300001210
Representing the output of the attention gate structure.
The set number may be, for example, 5, and the specific numerical value of the set number is not limited in the embodiment of the present invention.
Optionally, in the embodiment of the present invention, a fusion attention mechanism residual U-net network may be constructed as a generator model, a residual depth structure is adopted to ensure data fitting capability of the network, the network is easier to train to alleviate the problem of gradient disappearance, an attention mechanism is introduced to inhibit feature response in an irrelevant background region in an image, and a U-net network structure is adopted to ensure transmission of image features of different levels by adopting skip connection, so as to realize fusion of image features of different scales.
Fig. 4 is a schematic structural diagram of a generator model according to a second embodiment of the present invention, and in a specific example, as shown in fig. 4, the entire generator model may be a U-net structure with a five-layer network depth. At the encoder side: the first layer (E1) contains 64 1 × 1 × 1 convolutions, batch normalization, prilu activation and max pooling; the second layer (E2) comprises 128 1 × 1 × 1 convolutions, PReLU activation, a residual module and maximum pooling, wherein two 128 3 × 3 × 3 convolution layers, batch normalization and PReLU activation structures are connected in series in the residual module, the bottom layer input and the top layer output of the whole feature learning layer are connected by global jump connection, and global residual learning is carried out; the third layer (E3), the fourth layer (E4) and the fifth layer (E5) have the same structure as the second layer, but the number of convolution kernels is 256, 512 and 1024, respectively. E5 is followed by an upsampling layer to become the fifth layer of the decoder (D5). The decoder adopts a structure similar to that of the encoder, removes the maximum pooling in the encoder, amplifies the scale of the restored image from bottom to top (D4, D3, D2, D1) layer by layer through upsampling, and simultaneously performs multi-scale image fusion through the attention gate structure.
The discriminator model contains two inputs, one being the original high resolution image (HR, i.e. the three-dimensional CT segmented image) and the other being the super resolution image (SR) generated by the generator. Fig. 5 is a schematic structural diagram of a discriminator model according to a second embodiment of the present invention, and in a specific example, as shown in fig. 5, an SR image to be discriminated is input to the discriminator model, features are extracted by 5 convolutional layers, in order to increase a local receptive field, a convolutional kernel with a size of 4 × 4 × 4 is adopted, the number of convolutional kernels starts from 64 in the first layer, and the number of each layer increases by 2 times until 1024 in the fifth layer. Then, the sixth layer performs dimensionality reduction using the convolution layer of the 1 × 1 × 1 convolution kernel, followed by a convolution operation using the 3 × 3 × 3 convolution kernels of the two layers, and feeds the output of the fifth layer and the output of the eighth layer to the next layer. And at the end of the discriminator model, smoothing the feature mapping, flattening the dimensionality of the image data, and then sequentially sending the image data into the full connection layer, the LReLU activation layer and the full connection layer and obtaining a discrimination result Logits through Sigmoid. As shown in fig. 5, the discriminator model may also include a residual module.
In an optional embodiment of the present invention, the determining that the discriminant model is successfully trained may include: calculating a discriminator loss of the third resolution image on the discriminator model; determining that the discriminant model is successfully trained under the condition that the discriminant loss is determined to meet the discriminant stability condition; the determining that training of the generator model is successful may include: calculating pixel loss and feature loss between the three-dimensional CT segmentation image with the third resolution and the three-dimensional CT segmentation image with the first resolution, and the contrast loss of the third resolution image on the discriminator model; determining that the generator model training was successful if it is determined that the pixel loss, the feature loss, and the antagonistic loss satisfy a generator stability condition.
The discriminator loss is the loss generated when the discriminator discriminates the third resolution image. The discriminator-stable condition may be a condition for judging that the loss function of the discriminator model tends to be stable, and the generator-stable condition may be a condition for judging that the loss function of the generator model tends to be stable. The trend toward stability may be that the output value of the loss function tends toward a stable value, or the output value of the loss function is smaller than a set threshold, and the like, which is not limited by the embodiment of the present invention. The pixel loss is the pixel difference between the third resolution image and the three-dimensional CT segmented image, and the feature loss is the feature difference between the third resolution image and the three-dimensional CT segmented image. The countermeasure loss is the loss generated after the third resolution image is discriminated by the discriminator.
In an optional embodiment of the present invention, the calculating a discriminator loss of the third resolution image on the discriminator model may include: calculating the discriminator loss based on a network loss function:
Figure BDA0002702578530000141
wherein D is loss Representing the discriminator loss, N representing a natural number, and p representing a discriminator function; i denotes the number of training times, SR i Representing the third resolution image, HR i A three-dimensional CT segmentation image representing the first resolution.
The determining that the discriminator loss satisfies a discriminator stabilizing condition may include: and when the loss of the discriminator reaches a first set target value, determining that the loss of the discriminator meets a stable condition of the discriminator.
The calculating of pixel loss between the image of the third resolution and the three-dimensional CT segmented image of the first resolution may include: calculating the pixel loss based on a pixel loss function:
Figure BDA0002702578530000151
wherein L is 1 Indicating pixel loss.
The calculating of the feature loss between the third resolution image and the three-dimensional CT segmentation image of the first resolution may include: calculating the characteristic loss based on a characteristic loss function:
Figure BDA0002702578530000152
wherein, VGG loss Is indicative of the loss of the characteristic,
Figure BDA0002702578530000153
representing a deep learning network function, such as the VGG-19 network function, the VGG-19 penalty corresponds to the fourth convolution output before the fifth largest pooling layer in the VGG-19 network. That is, after VGG (Visual Geometry Group) feature extraction is performed on the third-resolution image and the three-dimensional CT segmented image of the first resolution, the feature loss between the third-resolution image and the three-dimensional CT segmented image of the first resolution may be calculated by using a feature loss function.
Alternatively, a 19-layer VGG network can be used to compensate the image pixel level L 1 And (4) loss perception information, calculating content loss of the third resolution image and the three-dimensional CT segmentation image on the VGG-19, and comparing semantic information of the super-resolution image and the high-resolution image.
The calculating the countermeasure loss of the third resolution image on the discriminator model may include: calculating the challenge loss based on a challenge loss function as follows:
Figure BDA0002702578530000161
wherein Adv loss Representing the challenge loss.
The determining that the pixel loss, the characteristic loss, and the antagonistic loss satisfy a generator stabilization condition may include: constructing a generator total loss from the pixel loss, the feature loss, and the antagonistic loss based on the following formula:
G loss =L 1 +α·VGG loss +β·Adv loss
wherein G is loss Representing the total loss of the generator, and alpha and beta represent weighting coefficients.
And when the total loss of the generator reaches a second set target value, determining that the pixel loss, the characteristic loss and the antagonistic loss meet a generator stability condition.
The first set target value and the second set target value may be set according to an actual requirement, such as a minimum value, which is not limited in the embodiment of the present invention.
In particular, the discriminant model and the generator model may be trained in conjunction with a variety of loss types each time D loss And G loss When the minimum value of (c) is reached, it can be determined that the discriminant model and the generator model are successfully trained. During the training, if D loss And G loss If the minimum is not reached, a back-propagation algorithm may be employed to update the network parameters that adjust the generation of the counterpoise network. Optionally, an Adam optimizer may be employed to generate an antagonistic network for antagonistic training until it converges.
S280, a generator model of a confrontation network model is generated according to the three-dimensional digital core image, and the input digital core image to be optimized is reconstructed to obtain a target three-dimensional digital core image.
Wherein the image resolution of the target three-dimensional digital core image is related to the resolution of the digital core training image. For example, when the three-dimensional CT zoom image is 40 × 40 × 40 pixels in size and 20 × 20 × 20 pixels in size, the corresponding target three-dimensional digital core image may be 80 × 80 × 80 pixels in size. That is, the image resolution of the target three-dimensional digital core image can be 2 times or 4 times of the image reconstruction.
Fig. 6 is a schematic diagram illustrating an effect of generating a target three-dimensional digital core image by using a generator model for generating a countering network model by using a three-dimensional digital core image according to a second embodiment of the present invention, and as shown in fig. 6, a low-resolution image (i.e., a digital core image to be optimized) is input to the generator model for generating the countering network model by using a three-dimensional digital core image, and a super-resolution image (i.e., a target three-dimensional digital core image) generated by reconstructing the low-resolution image has a higher definition, so that a disadvantage of an existing image reconstruction method that is too smooth is eliminated, and the super-resolution image has a clear texture structure.
According to the embodiment of the invention, the network depth can be deepened by using the residual U-net network, so that the image reconstruction effect of the generated countermeasure network is improved. Meanwhile, redundant information is suppressed by adopting an attention mechanism, and the visual perception effect of the reconstructed image is enhanced by adopting a mixed loss function, so that the target three-dimensional digital core image can be blurred and is more real.
Optionally, the Peak Signal-to-Noise Ratio (PSNR) may be used to evaluate the effect of the target three-dimensional digital core image obtained by the digital core image processing method provided by the embodiment of the present invention. Fig. 7 is a schematic diagram of a comparison effect of the evaluation index provided in the second embodiment of the present invention, as shown in fig. 7, whether the scale of the digital core image to be optimized is increased by 2 times, that is, the resolution is increased by 2 times, or the scale of the digital core image to be optimized is increased by 4 times, that is, the resolution is increased by 4 times. Compared with Bicubic (Bicubic), A + reconstruction (3D-A +), super-resolution reconstruction (3D-SRCNN) based on a convolutional neural network and a super-resolution generation countermeasure network (3D-SRGAN), the image quality of the target three-dimensional digital core image obtained by the digital core image processing method provided by the invention is optimal.
According to the technical scheme, the network structure of the generator model is designed into the residual U-net structure, so that the depth of the neural network is greatly increased, multi-scale image information can be fused, and the plurality of local jump connections can help important feature information to cross different modules and layers to be transmitted. Global jump connection is introduced into the whole feature learning layer, residual errors between high-resolution images and low-resolution images are learned, and the problems of gradient loss and network degradation are effectively solved. An attention gate structure is introduced between the encoder and the decoder, so that the generated network pays more attention to the region with more high-frequency information, the characteristic weight containing rich high-frequency information is amplified, the weight containing redundant low-frequency information is reduced, the network convergence is accelerated, and the network performance is improved. Meanwhile, measurement between the distribution of the reconstructed image and the original high-resolution image is carried out by utilizing the mixed loss, the visual perception effect of the reconstructed image is enhanced by constructing the characteristic matching loss, and the details and the definition of the reconstructed image of the three-dimensional digital core are greatly improved, so that the image quality is improved.
It should be noted that any permutation and combination between the technical features in the above embodiments also belong to the scope of the present invention.
EXAMPLE III
Fig. 8 is a schematic diagram of a digital core image processing apparatus according to a third embodiment of the present invention, and as shown in fig. 8, the apparatus includes: a training image acquisition module 310, a network model acquisition module 320, and an image reconstruction module 330, wherein:
a training image obtaining module 310, configured to obtain a digital core training image; the digital core training images comprise digital core three-dimensional CT images with different resolutions;
a network model obtaining module 320, configured to respectively train a generator model and a discriminator model for generating a confrontation network according to the digital core training image, so as to obtain a three-dimensional digital core image generation confrontation network model;
and the image reconstruction module 330 is configured to reconstruct the input digital core image to be optimized according to a generator model of the three-dimensional digital core image generation countermeasure network model, so as to obtain a target three-dimensional digital core image.
According to the embodiment of the invention, the digital core three-dimensional CT images with different resolutions are used as the digital core training images, the generator model and the discriminator model of the countermeasure network are respectively generated by training according to the digital core training images, the three-dimensional digital core image generation countermeasure network model is obtained, the generator model of the countermeasure network model is generated according to the three-dimensional digital core image, the input digital core image to be optimized is reconstructed, the target three-dimensional digital core image is obtained, the problem of low image quality of the existing three-dimensional digital core image reconstruction image is solved, and the image quality of the three-dimensional digital core reconstruction image is improved.
Optionally, the training image obtaining module 310 is configured to: acquiring a two-dimensional CT image of at least one known rock; performing image preprocessing and image enhancement processing on the two-dimensional CT image to obtain a two-dimensional CT processed image; constructing a three-dimensional CT image according to the two-dimensional CT processing image; and constructing the digital core training image according to the three-dimensional CT image.
Optionally, the training image obtaining module 310 is configured to: segmenting the three-dimensional CT image to obtain a three-dimensional CT segmented image with a first resolution; carrying out zooming processing on the three-dimensional CT segmentation image according to a set zooming value to obtain a three-dimensional CT zoomed image with a second resolution; and constructing the digital core training image according to the three-dimensional CT segmentation image and the three-dimensional CT scaling image.
Optionally, the network model obtaining module 320 is configured to input a three-dimensional CT scaled image with a second resolution in the digital core training image into the generator model, and use an image output by the generator model as a third resolution image; inputting the three-dimensional CT segmentation image of the first resolution in the third resolution image and the digital core training image into the discriminator model at the same time, and determining whether the three-dimensional CT segmentation image of the third resolution is matched with the three-dimensional CT segmentation image of the first resolution according to the output value of the discriminator model so as to train the discriminator model; and after the discriminant model is successfully trained, fixing the discriminant model, and returning to the generator model to perform the operation of inputting the three-dimensional CT scaling image with the second resolution in the digital core training image so as to continue training the generator model until the generator model is successfully trained.
Optionally, the generator model is a fusion attention mechanism residual U-net network, the fusion attention mechanism residual U-net network includes a set number of network structures, and an encoder and a decoder of the fusion attention mechanism residual U-net network perform multi-scale image fusion through an attention gate structure; wherein the attention gate structure is implemented based on the following formula:
Figure BDA0002702578530000191
Figure BDA0002702578530000192
Figure BDA0002702578530000193
wherein the content of the first and second substances,
Figure BDA0002702578530000194
representing a transfer function, # T 、/>
Figure BDA0002702578530000195
And &>
Figure BDA0002702578530000196
Represents a 1 × 1 × 1 convolution operation, b g And b ψ Representing the offset coefficient, σ 1 Represents the ReLU function->
Figure BDA0002702578530000197
Input representing the attention gate structure, g i A gate signal, asserted on a high level of context information>
Figure BDA0002702578530000198
Denotes the attention coefficient, σ 2 Representing sigmoid activation function, Θ att Represents a parameter set, is selected>
Figure BDA0002702578530000199
Representing the output of the attention gate structure.
Optionally, the network model obtaining module 320 is configured to calculate a discriminator loss of the third resolution image on the discriminator model; determining that the discriminant model is successfully trained under the condition that the discriminant loss is determined to meet the discriminant stability condition; calculating pixel loss and feature loss between the third-resolution image and the three-dimensional CT segmentation image of the first resolution, and the confrontation loss of the third-resolution image on the discriminator model; determining that the generator model training was successful if it is determined that the pixel loss, the feature loss, and the antagonistic loss satisfy a generator stability condition.
Optionally, the network model obtaining module 320 is configured to calculate the arbiter loss based on the following network loss function:
Figure BDA00027025785300001910
wherein D is loss Representing the discriminator loss, N representing a natural number, and p representing a discriminator function; i denotes the number of training times, SR i Representing the third resolution image, HR i A three-dimensional CT segmented image representing the first resolution;
when the loss of the discriminator reaches a first set target value, determining that the loss of the discriminator meets a stable condition of the discriminator;
calculating the pixel loss based on a pixel loss function:
Figure BDA0002702578530000201
wherein L is 1 Represents pixel loss;
calculating the characteristic loss based on a characteristic loss function:
Figure BDA0002702578530000202
wherein, VGG loss Is indicative of a loss of said characteristic,
Figure BDA0002702578530000203
representing deep learning netA complex function;
calculating the challenge loss based on a challenge loss function as follows:
Figure BDA0002702578530000204
wherein Adv loss Representing the challenge loss;
constructing a generator total loss from the pixel loss, the feature loss, and the counter loss based on the following formula:
G loss =L 1 +α·VGG loss +β·Adv loss
wherein G is loss Representing the total loss of the generator, alpha and beta representing weight coefficients;
and when the total loss of the generator reaches a second set target value, determining that the pixel loss, the characteristic loss and the antagonistic loss meet a generator stability condition.
The digital core image processing device can execute the digital core image processing method provided by any embodiment of the invention, and has corresponding functional modules and beneficial effects of the execution method. For technical details that are not described in detail in this embodiment, reference may be made to the digital core image processing method provided in any embodiment of the present invention.
Since the digital core image processing apparatus described above is an apparatus capable of executing the digital core image processing method in the embodiment of the present invention, based on the digital core image processing method described in the embodiment of the present invention, a person skilled in the art can understand a specific implementation manner of the digital core image processing apparatus of the embodiment and various variations thereof, and therefore, a detailed description of how the digital core image processing method in the embodiment of the present invention is implemented by the digital core image processing apparatus is not described here. The apparatus used by those skilled in the art to implement the digital core image processing method in the embodiments of the present invention is within the scope of the present application.
Example four
Fig. 9 is a schematic structural diagram of a computer device according to a fourth embodiment of the present invention. FIG. 9 illustrates a block diagram of a computer device 412 suitable for use in implementing embodiments of the present invention. The computer device 412 shown in FIG. 9 is only one example and should not impose any limitations on the functionality or scope of use of embodiments of the present invention.
As shown in fig. 9, the computer device 412 is in the form of a general purpose computing device. Components of computer device 412 may include, but are not limited to: one or more processors 416, a storage device 428, and a bus 418 that couples the various system components including the storage device 428 and the processors 416.
Bus 418 represents one or more of any of several types of bus structures, including a memory bus or memory controller, a peripheral bus, an accelerated graphics port, and a processor or local bus using any of a variety of bus architectures. By way of example, such architectures include, but are not limited to, an Industry Standard Architecture (ISA) bus, a Micro Channel Architecture (MCA) bus, an enhanced ISA bus, a Video Electronics Standards Association (VESA) local bus, and a Peripheral Component Interconnect (PCI) bus.
Computer device 412 typically includes a variety of computer system readable media. Such media can be any available media that is accessible by computer device 412 and includes both volatile and nonvolatile media, removable and non-removable media.
Storage 428 may include computer system readable media in the form of volatile Memory, such as Random Access Memory (RAM) 430 and/or cache Memory 432. The computer device 412 may further include other removable/non-removable, volatile/nonvolatile computer system storage media. By way of example only, storage system 434 may be used to read from and write to non-removable, nonvolatile magnetic media (not shown in FIG. 9, commonly referred to as a "hard drive"). Although not shown in FIG. 9, a magnetic disk drive for reading from and writing to a removable, nonvolatile magnetic disk (e.g., a "floppy disk") and an optical disk drive for reading from or writing to a removable, nonvolatile optical disk (e.g., a Compact disk-Read Only Memory (CD-ROM), digital Video disk (DVD-ROM), or other optical media) may be provided. In these cases, each drive may be connected to bus 418 by one or more data media interfaces. Storage 428 may include at least one program product having a set (e.g., at least one) of program modules that are configured to carry out the functions of embodiments of the invention.
Program 436 having a set (at least one) of program modules 426 may be stored, for example, in storage 428, such program modules 426 including, but not limited to, an operating system, one or more application programs, other program modules, and program data, each of which examples or some combination may comprise an implementation of a network environment. Program modules 426 generally perform the functions and/or methodologies of embodiments of the invention as described herein.
The computer device 412 may also communicate with one or more external devices 414 (e.g., keyboard, pointing device, camera, display 424, etc.), with one or more devices that enable a user to interact with the computer device 412, and/or with any devices (e.g., network card, modem, etc.) that enable the computer device 412 to communicate with one or more other computing devices. Such communication may be through an Input/Output (I/O) interface 422. Also, computer device 412 may communicate with one or more networks (e.g., a Local Area Network (LAN), wide Area Network (WAN)) and/or a public Network, such as the Internet, through Network adapter 420. As shown, network adapter 420 communicates with the other modules of computer device 412 over bus 418. It should be appreciated that although not shown in the figures, other hardware and/or software modules may be used in conjunction with the computer device 412, including but not limited to: microcode, device drivers, redundant processing units, external disk drive Arrays, disk array (RAID) systems, tape drives, and data backup storage systems, to name a few.
The processor 416 executes programs stored in the storage 428 to perform various functional applications and data processing, such as implementing the digital core image processing methods provided by the above-described embodiments of the present invention.
That is, the processing unit implements, when executing the program: acquiring a digital core training image; the digital core training images comprise digital core three-dimensional CT images with different resolutions; respectively training a generator model and a discriminator model for generating a confrontation network according to the digital core training images to obtain a three-dimensional digital core image generation confrontation network model; and generating a generator model of a confrontation network model according to the three-dimensional digital core image, and reconstructing the input digital core image to be optimized to obtain a target three-dimensional digital core image.
EXAMPLE five
An embodiment of the fifth invention further provides a computer storage medium storing a computer program, where the computer program is used to execute the digital core image processing method according to any one of the above embodiments of the present invention when executed by a computer processor: acquiring a digital core training image; the digital core training images comprise digital core three-dimensional CT images with different resolutions; respectively training a generator model and a discriminator model for generating a confrontation network according to the digital core training images to obtain a three-dimensional digital core image generation confrontation network model; and generating a generator model of a confrontation network model according to the three-dimensional digital core image, and reconstructing the input digital core image to be optimized to obtain a target three-dimensional digital core image.
Computer storage media for embodiments of the invention may employ any combination of one or more computer-readable media. The computer readable medium may be a computer readable signal medium or a computer readable storage medium. A computer readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any combination of the foregoing. More specific examples (a non-exhaustive list) of the computer readable storage medium would include the following: an electrical connection having one or more wires, a portable computer diskette, a hard disk, a Random Access Memory (RAM), a Read-Only Memory (ROM), an Erasable Programmable Read-Only Memory (EPROM) or flash Memory), an optical fiber, a portable compact disc Read-Only Memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. In the context of this document, a computer readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device.
A computer readable signal medium may include a propagated data signal with computer readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated data signal may take many forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof. A computer readable signal medium may also be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device.
Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to wireless, wireline, optical fiber cable, radio Frequency (RF), etc., or any suitable combination of the foregoing.
Computer program code for carrying out operations for aspects of the present invention may be written in any combination of one or more programming languages, including an object oriented programming language such as Java, smalltalk, C + + or the like and conventional procedural programming languages, such as the "C" programming language or similar programming languages. The program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In the case of a remote computer, the remote computer may be connected to the user's computer through any type of network, including a Local Area Network (LAN) or a Wide Area Network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet service provider).
It is to be noted that the foregoing is only illustrative of the preferred embodiments of the present invention and the technical principles employed. It will be understood by those skilled in the art that the present invention is not limited to the particular embodiments described herein, but is capable of various obvious changes, rearrangements and substitutions as will now become apparent to those skilled in the art without departing from the scope of the invention. Therefore, although the present invention has been described in greater detail by the above embodiments, the present invention is not limited to the above embodiments, and may include other equivalent embodiments without departing from the spirit of the present invention, and the scope of the present invention is determined by the scope of the appended claims.

Claims (8)

1. A digital core image processing method is characterized by comprising the following steps:
acquiring a digital core training image; the digital core training images comprise digital core three-dimensional CT images with different resolutions;
respectively training a generator model and a discriminator model for generating a confrontation network according to the digital core training images to obtain a three-dimensional digital core image generation confrontation network model;
a generator model for generating a confrontation network model according to the three-dimensional digital core image carries out reconstruction processing on the input digital core image to be optimized to obtain a target three-dimensional digital core image;
wherein, the acquiring of the digital core training image comprises:
acquiring a two-dimensional CT image of at least one known rock;
performing image preprocessing and image enhancement processing on the two-dimensional CT image to obtain a two-dimensional CT processed image;
constructing a three-dimensional CT image according to the two-dimensional CT processing image;
constructing the digital core training image according to the three-dimensional CT image;
wherein, the training respectively according to the digital rock core training image to generate a generator model and a discriminator model of the confrontation network comprises:
inputting a three-dimensional CT (computed tomography) zooming image with a second resolution in the digital core training image into the generator model, and taking an image output by the generator model as a third resolution image;
inputting the three-dimensional CT segmentation image of the first resolution in the third resolution image and the digital core training image into the discriminator model at the same time, and determining whether the three-dimensional CT segmentation image of the third resolution is matched with the three-dimensional CT segmentation image of the first resolution according to the output value of the discriminator model so as to train the discriminator model;
and after the discriminant model is successfully trained, fixing the discriminant model, and returning to perform the operation of inputting the three-dimensional CT scaling image with the second resolution in the digital core training image into the generator model to continue training the generator model until the generator model is successfully trained.
2. The method of claim 1, wherein constructing the digital core training image from the three-dimensional CT image comprises:
segmenting the three-dimensional CT image to obtain a three-dimensional CT segmented image with a first resolution;
carrying out zooming processing on the three-dimensional CT segmentation image according to a set zooming value to obtain a three-dimensional CT zoomed image with a second resolution;
and constructing the digital core training image according to the three-dimensional CT segmentation image and the three-dimensional CT scaling image.
3. The method of claim 1, wherein the generator model is a converged attention mechanism residual U-net network comprising a set number of network structures, and wherein an encoder and decoder of the converged attention mechanism residual U-net network performs multi-scale image fusion through an attention gate structure;
wherein the attention gate structure is implemented based on the following formula:
Figure FDA0003898298310000021
Figure FDA0003898298310000022
Figure FDA0003898298310000023
wherein the content of the first and second substances,
Figure FDA0003898298310000024
representing a transfer function, # T
Figure FDA0003898298310000025
And
Figure FDA0003898298310000026
represents a 1 × 1 × 1 convolution operation, b g And b ψ Representing the offset coefficient, σ 1 Denotes the ReLU function, g i A gate signal indicating the provision of high-level context information,
Figure FDA0003898298310000027
denotes the attention coefficient, σ 2 Denotes the sigmoid activation function, Θ att Representing a parameter set.
4. The method of claim 1, wherein the determining that the discriminant model training was successful comprises:
calculating a discriminator loss of the third resolution image on the discriminator model;
determining that the discriminant model is successfully trained under the condition that the discriminant loss is determined to meet the discriminant stability condition;
the determining that the training of the generator model is successful comprises:
calculating pixel loss and feature loss between the third-resolution image and the three-dimensional CT segmentation image of the first resolution, and the confrontation loss of the third-resolution image on the discriminator model;
determining that the generator model training was successful if it is determined that the pixel loss, the feature loss, and the antagonistic loss satisfy a generator stability condition.
5. The method of claim 4, wherein the calculating a discriminant loss for the third resolution image on the discriminant model comprises:
calculating the discriminator loss based on a network loss function:
Figure FDA0003898298310000031
wherein D is loss Representing the discriminator loss, N representing a natural number, and p representing a discriminator function; i denotes the number of training times, SR i Representing the third resolution image, HR i A three-dimensional CT segmented image representing the first resolution;
the determining that the arbiter loss satisfies an arbiter stabilizing condition includes:
when the loss of the discriminator reaches a first set target value and is stable, determining that the loss of the discriminator meets the stability condition of the discriminator;
the calculating pixel loss between the third resolution image and the three-dimensional CT segmented image of the first resolution comprises:
calculating the pixel loss based on a pixel loss function:
Figure FDA0003898298310000041
wherein L is 1 Represents pixel loss;
the calculating of the feature loss between the third resolution image and the three-dimensional CT segmented image of the first resolution comprises:
calculating the characteristic loss based on a characteristic loss function:
Figure FDA0003898298310000042
wherein, VGG loss Is indicative of the loss of the characteristic,
Figure FDA0003898298310000043
representing a deep learning network function;
the calculating the confrontation loss of the third resolution image on the discriminator model includes:
calculating the challenge loss based on a challenge loss function as follows:
Figure FDA0003898298310000044
wherein Adv loss Representing the challenge loss;
the determining that the pixel loss, the characteristic loss, and the antagonistic loss satisfy a generator stabilization condition includes:
constructing a generator total loss from the pixel loss, the feature loss, and the antagonistic loss based on the following formula:
G loss =L 1 +α·VGG loss +β·Adv loss
wherein G is loss Representing the total loss of the generator, alpha and beta representing weight coefficients;
determining that the pixel loss, the characteristic loss, and the antagonistic loss satisfy a generator stabilization condition when the total generator loss reaches a second set target value and stabilizes.
6. A digital core image processing device is characterized by comprising:
the training image acquisition module is used for acquiring a digital core training image; the digital core training images comprise digital core three-dimensional CT images with different resolutions;
the network model acquisition module is used for respectively training a generator model and a discriminator model for generating a confrontation network according to the digital core training image to obtain a three-dimensional digital core image generation confrontation network model;
the image reconstruction module is used for reconstructing the input digital core image to be optimized according to a generator model of the three-dimensional digital core image generation countermeasure network model to obtain a target three-dimensional digital core image;
the training image acquisition module is used for acquiring a two-dimensional CT image of at least one known rock;
performing image preprocessing and image enhancement processing on the two-dimensional CT image to obtain a two-dimensional CT processed image;
constructing a three-dimensional CT image according to the two-dimensional CT processing image; constructing the digital core training image according to the three-dimensional CT image;
the network model acquisition module is used for inputting a three-dimensional CT (computed tomography) zoom image of a second resolution in the digital core training image into the generator model and taking the image output by the generator model as a third resolution image;
inputting the three-dimensional CT segmentation image of the first resolution in the third resolution image and the digital core training image into the discriminator model at the same time, and determining whether the three-dimensional CT segmentation image of the third resolution is matched with the three-dimensional CT segmentation image of the first resolution according to the output value of the discriminator model so as to train the discriminator model;
and after the discriminant model is successfully trained, fixing the discriminant model, and returning to perform the operation of inputting the three-dimensional CT scaling image with the second resolution in the digital core training image into the generator model to continue training the generator model until the generator model is successfully trained.
7. A computer device, characterized in that the computer device comprises:
one or more processors;
storage means for storing one or more programs;
when executed by the one or more processors, cause the one or more processors to implement the digital core image processing method of any of claims 1-5.
8. A computer storage medium having a computer program stored thereon, wherein the program, when executed by a processor, implements a digital core image processing method as recited in any of claims 1-5.
CN202011027637.1A 2020-09-25 2020-09-25 Digital rock core image processing method and device, computer equipment and storage medium Active CN112132959B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011027637.1A CN112132959B (en) 2020-09-25 2020-09-25 Digital rock core image processing method and device, computer equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011027637.1A CN112132959B (en) 2020-09-25 2020-09-25 Digital rock core image processing method and device, computer equipment and storage medium

Publications (2)

Publication Number Publication Date
CN112132959A CN112132959A (en) 2020-12-25
CN112132959B true CN112132959B (en) 2023-03-24

Family

ID=73840593

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011027637.1A Active CN112132959B (en) 2020-09-25 2020-09-25 Digital rock core image processing method and device, computer equipment and storage medium

Country Status (1)

Country Link
CN (1) CN112132959B (en)

Families Citing this family (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112801912B (en) * 2021-02-09 2023-10-31 华南理工大学 Face image restoration method, system, device and storage medium
CN113570505B (en) * 2021-09-24 2022-01-04 中国石油大学(华东) Shale three-dimensional super-resolution digital core grading reconstruction method and system
CN114169457B (en) * 2021-12-13 2023-05-23 成都理工大学 Storm rock logging identification method based on core reconstruction
US20230184087A1 (en) * 2021-12-13 2023-06-15 Saudi Arabian Oil Company Multi-modal and Multi-dimensional Geological Core Property Prediction using Unified Machine Learning Modeling
CN115049781B (en) * 2022-05-11 2023-05-23 西南石油大学 Shale digital rock core three-dimensional reconstruction method based on deep learning
CN115115783B (en) * 2022-07-08 2023-08-15 西南石油大学 Digital rock core construction method and system for simulating shale matrix nano-micro pores
CN115272156B (en) * 2022-09-01 2023-06-30 中国海洋大学 High-resolution wellbore imaging characterization method for oil and gas reservoirs based on cyclic generation countermeasure network
CN117152373B (en) * 2023-11-01 2024-02-02 中国石油大学(华东) Core-level pore network model construction method considering cracks

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109978762A (en) * 2019-02-27 2019-07-05 南京信息工程大学 A kind of super resolution ratio reconstruction method generating confrontation network based on condition
CN111402266A (en) * 2020-03-13 2020-07-10 中国石油大学(华东) Method and system for constructing digital core
CN111461303A (en) * 2020-03-31 2020-07-28 中国石油大学(北京) Digital core reconstruction method and system based on generation of antagonistic neural network
CN111583148A (en) * 2020-05-07 2020-08-25 苏州闪掣智能科技有限公司 Rock core image reconstruction method based on generation countermeasure network

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11105942B2 (en) * 2018-03-27 2021-08-31 Schlumberger Technology Corporation Generative adversarial network seismic data processor

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109978762A (en) * 2019-02-27 2019-07-05 南京信息工程大学 A kind of super resolution ratio reconstruction method generating confrontation network based on condition
CN111402266A (en) * 2020-03-13 2020-07-10 中国石油大学(华东) Method and system for constructing digital core
CN111461303A (en) * 2020-03-31 2020-07-28 中国石油大学(北京) Digital core reconstruction method and system based on generation of antagonistic neural network
CN111583148A (en) * 2020-05-07 2020-08-25 苏州闪掣智能科技有限公司 Rock core image reconstruction method based on generation countermeasure network

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
Super-resolution of real-world rock microcomputed tomography images using cycle-consistent generative adversarial networks;Honggang Chen 等;《PHYSICAL REVIEW E》;20200211;第101卷(第2期);第023305-1-023305-15页 *
双判别器生成对抗网络图像的超分辨率重建方法;袁飘逸 等;《激光与光电子学进展》;20191231;第56卷(第23期);第231010-1-231010-10页 *

Also Published As

Publication number Publication date
CN112132959A (en) 2020-12-25

Similar Documents

Publication Publication Date Title
CN112132959B (en) Digital rock core image processing method and device, computer equipment and storage medium
CN110189255B (en) Face detection method based on two-stage detection
Li et al. Survey of single image super‐resolution reconstruction
CN115661144B (en) Adaptive medical image segmentation method based on deformable U-Net
Li et al. Multifocus image fusion using wavelet-domain-based deep CNN
CN111275686B (en) Method and device for generating medical image data for artificial neural network training
CN112541864A (en) Image restoration method based on multi-scale generation type confrontation network model
Sun et al. Multiscale generative adversarial network for real‐world super‐resolution
CN115375711A (en) Image segmentation method of global context attention network based on multi-scale fusion
CN116739899A (en) Image super-resolution reconstruction method based on SAUGAN network
CN115601299A (en) Intelligent liver cirrhosis state evaluation system and method based on images
Zhang et al. Dense haze removal based on dynamic collaborative inference learning for remote sensing images
CN111507950B (en) Image segmentation method and device, electronic equipment and computer-readable storage medium
CN112396605B (en) Network training method and device, image recognition method and electronic equipment
CN116416221A (en) Ultrasonic image analysis method
Chen et al. Image denoising via generative adversarial networks with detail loss
CN116543246A (en) Training method of image denoising model, image denoising method, device and equipment
CN109242797B (en) Image denoising method, system and medium based on homogeneous and heterogeneous region fusion
CN114862739B (en) Intelligent medical image enhancement method and system
CN110570417A (en) Pulmonary nodule classification method and device and image processing equipment
CN116912345B (en) Portrait cartoon processing method, device, equipment and storage medium
Patel et al. Deep Learning in Medical Image Super-Resolution: A Survey
Zin Improvement of Image Denoising with Interpolation and RAISR
Chao et al. Instance-aware image dehazing
CN117710295A (en) Image processing method, device, apparatus, medium, and program product

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant