CN115424119B - Image generation training method and device capable of explaining GAN based on semantic fractal - Google Patents

Image generation training method and device capable of explaining GAN based on semantic fractal Download PDF

Info

Publication number
CN115424119B
CN115424119B CN202211373030.8A CN202211373030A CN115424119B CN 115424119 B CN115424119 B CN 115424119B CN 202211373030 A CN202211373030 A CN 202211373030A CN 115424119 B CN115424119 B CN 115424119B
Authority
CN
China
Prior art keywords
fractal
gan
semantic
interpretable
feature map
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202211373030.8A
Other languages
Chinese (zh)
Other versions
CN115424119A (en
Inventor
李超
王劲
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang Lab
Original Assignee
Zhejiang Lab
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang Lab filed Critical Zhejiang Lab
Priority to CN202211373030.8A priority Critical patent/CN115424119B/en
Publication of CN115424119A publication Critical patent/CN115424119A/en
Application granted granted Critical
Publication of CN115424119B publication Critical patent/CN115424119B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/74Image or video pattern matching; Proximity measures in feature spaces
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Evolutionary Computation (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Computing Systems (AREA)
  • General Health & Medical Sciences (AREA)
  • General Physics & Mathematics (AREA)
  • Software Systems (AREA)
  • Databases & Information Systems (AREA)
  • Medical Informatics (AREA)
  • Multimedia (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Molecular Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The invention discloses a semantic fractal-based interpretable GAN image generation training method and a device, wherein a traditional GAN model is modified into an interpretable GAN, so that the high-level feature representation of the model is clear and consistent; in interpretable GAN, all filters at the same layer "learn" to be activated by the same part, so that features inside the GAN represent clear and consistent semantic information; according to the method, a fractal loss function is designed under the condition that no additional marking is needed and only a normal training sample is used, and the characteristic expression of the GAN is restrained, so that the GAN can automatically learn important parts of an attention object; by optimizing the original loss function and the fractal loss function of the GAN at the same time, the interpretability of internal features of the GAN is improved, and meanwhile, the quality of the generated picture of the GAN is guaranteed.

Description

Image generation training method and device capable of explaining GAN based on semantic fractal
Technical Field
The invention relates to the technical field of artificial intelligence, in particular to an image generation training method and device capable of explaining GAN based on semantic fractal.
Background
In recent years, with the rise and development of generation of countermeasure Networks (GANs), it has become possible to directly generate high-fidelity and diversified pictures from noise data. The generation countermeasure network consists of generators (G) whose goal is to map random noise to samples, and discriminators (D) whose purpose is to discriminate between real samples and generate samples. The training goal of GAN is to find nash equilibrium for the D and G training processes. Traditional GAN research has focused on solving the problem of poor picture generation effect of original GAN. For this reason, various improvements have been proposed by the predecessors, which can be divided into two main categories: firstly, modifications to the GAN model structure, such as SAGAN introducing attention mechanism, proGAN progressively generating pictures, etc., these structural improvements greatly improve the resolution and fidelity of GAN generated pictures; secondly, the improvement of the model training process, which is sensitive to the original GAN training process without the training aid skill, requires fine tuning and super-parameter adjustment to stabilize the training process, and thus, it is an important research direction to improve the training process of the original GAN to enhance its stability, which includes both theoretical exploration and empirical attempts, such as modification of GAN loss function, adding a gradient penalty term, or normalization technique to limit updating of the D parameter of the discriminator.
However, although GAN generation picture effect is enough to compare with real picture, the mechanism of GAN model internal generation picture is still a "black box", and research on GAN interpretability is also drawing attention of researchers.
On the one hand, the content of the picture generated by GAN is uncertain, and people cannot correlate the input random noise with the content of the generated picture. For this purpose, interfaceGAN proposes a method of performing a linear transformation on input random noise to perform a specific semantic editing on a generated picture. The InfoGAN provides a mode of maximizing mutual information between random noise and generated picture attributes, so that a certain dimension of the noise can correspondingly control a certain attribute of a picture. The GANspace adopts a principal component analysis mode to find a change direction having semantic information for GAN in a noise space or a feature space. On the other hand, the mechanism of generating pictures inside GAN is unknown, and the features of the middle layer inside GAN are often complex and disordered. In particular, for a certain high-level convolutional layer, its filters may be activated by different sites. For example, when generating a picture of a human face, filter a is activated by the eyes and nose, filter B is activated by both ears, and filter C is activated by the mouth. The high-level feature representation becomes complex, fuzzy, and inconsistent. However, there is relatively little interpretative research on the internal characteristics of GAN, which also presents significant challenges for its further development.
Disclosure of Invention
In order to solve the defects of the prior art, realize the automatic attention to the important part of an object in an image, improve the interpretability of the internal image characteristics and ensure the quality of the generated image, the invention adopts the following technical scheme:
an image generation training method capable of explaining GAN based on semantic fractal comprises the following steps:
s1, generating a countermeasure network based on loss function training through an image training set to obtain a generator of the countermeasure network;
s2, acquiring a first feature map output by the image generator in the middle layer, and performing semantic fractal on the first feature map;
s3, constructing fractal loss for the middle-layer characteristics of the generation countermeasure network according to the semantic fractal result, and comprising the following steps:
s3.1, obtaining a fractal result of the first feature graph divided based on a semantic fractal rule;
s3.2, acquiring a second feature map output by the image generator in the middle layer, and taking the matching degree of the second feature map and the fractal result as fractal loss so as to enable the fractal of the second feature map to be close to the first feature map;
and S4, generating the countermeasure network through joint optimization based on the fractal loss and the original loss function of the generated countermeasure network, and training image generation.
Further, the step S2 includes the steps of:
s2.1, designing a semantic fractal rule of the feature map; dividing the feature graph into a plurality of fractal by taking each fractal as a class based on the principles of minimizing intra-class variance and maximizing inter-class distance, wherein the fractal is not overlapped;
and S2.2, carrying out iterative division on the first characteristic diagram based on a fractal rule.
Further, in step S2, the first feature map is an activated feature map, and the fractal rule is to divide the feature map into a plurality of fractal shapes according to distribution of activation values of the feature map.
Further, in the step S2.2, in each iteration, a fractal with a large area and a large variance of an activation value is found in the current fractal, and the fractal is divided into two parts according to the principle of minimizing the intra-class variance.
Further, the principle of minimizing the intra-class variance is divided into two, and the formula is as follows:
Figure 100002_DEST_PATH_IMAGE002
wherein, the first and the second end of the pipe are connected with each other,
Figure 100002_DEST_PATH_IMAGE004
and &>
Figure 100002_DEST_PATH_IMAGE006
Respectively representing the variance of the activation values of the two divided sub-fractal parts; />
Figure 100002_DEST_PATH_IMAGE008
And &>
Figure 100002_DEST_PATH_IMAGE010
Respectively showing the area ratio of the two sub-fractal parts after division.
Further, the fractal result in step S3.1 includes a fractal position of the first feature map based on a fractal rule; in step S3.2, the matching degree of the fractal positions of the second feature map and the first feature map is calculated as a fractal loss function, so that all filters in the same middle layer of the generator are activated by the same part.
Further, in step S4, the generator is reinitialized as the parameters saved in step S1, a generator image is generated according to the training set image, the discriminator is initialized randomly, the generator image and the training set image are discriminated, the countermeasure network is generated by using the loss function fine tuning of step S4, in the training process, the fractal loss and the original loss function of the countermeasure network are used for optimizing the network parameters alternately, so that the convolution kernels in the same layer can express consistent features, the generation logic of the generator is further explained, and the construction of the interpretable GAN is realized.
In order to train the GAN end-to-end so that the GAN can automatically learn feature maps with consistent fractal, based on the original GAN loss function, the fractal loss function (frac loss) is added, and network parameters are alternately optimized between the two loss functions, so as to encourage all feature maps to have consistent fractal while ensuring the picture generation quality.
The device comprises a memory and one or more processors, wherein the memory stores executable codes, and the one or more processors are used for realizing the semantic fractal-based interpretable GAN image generation training method when executing the executable codes.
An image generation method capable of explaining GAN based on semantic fractal is characterized in that an image to be generated is generated through a trained discriminator.
An image generation device based on semantic fractal interpretable GAN comprises a memory and one or more processors, wherein the memory is stored with executable codes, and the one or more processors are used for realizing the image generation method based on semantic fractal interpretable GAN when executing the executable codes.
The invention has the advantages and beneficial effects that:
in the invention, the problem of complex and chaotic characteristics of the layers in the GAN is considered, so that the filters in the same layer represent clear and consistent characteristics; by designing a fractal loss function, the filters of the same layer can be activated by the same part without adding extra artificial labels; by optimizing the original loss function and the fractal loss function of the GAN at the same time, the interpretability of internal features of the GAN is improved, and meanwhile, the quality of the generated picture of the GAN is guaranteed. The method provided by the invention can be widely applied to various GAN models and has a certain application prospect.
Drawings
Fig. 1 is a flowchart of an image generation training method for interpretable GAN based on semantic fractal in an embodiment of the present invention.
FIG. 2 is a flowchart of a method for calculating and generating layer feature semantic fractal in a countermeasure network according to an embodiment of the present invention.
FIG. 3 is a flowchart of a method for constructing and generating a fractal constraint for layer feature semantics in an anti-adversarial network in an embodiment of the present invention.
Fig. 4 is an overview of the structure of the generated countermeasure network in the embodiment of the present invention.
Fig. 5 is a schematic structural diagram of an image generation apparatus capable of interpreting GAN based on semantic fractal according to an embodiment of the present invention.
Detailed Description
The following detailed description of embodiments of the invention refers to the accompanying drawings. It should be understood that the detailed description and specific examples, while indicating the present invention, are given by way of illustration and explanation only, not limitation.
The present invention aims to modify the conventional GAN model into an interpretable GAN such that its high-level representation of features is unambiguous and consistent. In interpretable GAN, all filters at the same layer "learn" to be activated by the same part, so that features inside the GAN represent clear and consistent semantic information.
The key technology of the invention is that the GAN is automatically made to learn the important parts of the object of interest without additional input/labeling, and only using normal training samples, and all the interest is consistent among different filters of the same high level, i.e. the filters of the same level are activated by the same parts.
The hardware used in the embodiment is a workstation capable of deep learning, and the auxiliary software used in the embodiment is a deep learning training frame pytorech.
Taking DCGAN as an example, the image generation training method for interpretable GAN based on semantic fractal provided by the invention, as shown in fig. 1 and fig. 2, comprises the following steps:
s1, generating a countermeasure network based on loss function training through an image training set to obtain a generator of the countermeasure network;
in the embodiment of the invention, an FFHQ data set is selected, only the original loss function of the DCGAN is used for training one DCGAN, and the model parameters of a generator (G) are stored;
s2, acquiring a first feature map output by the middle layer of the image generator, and performing semantic fractal on the first feature map;
in the embodiment of the invention, the input graph is divided in an iterative way based on a fractal rule; the selected generator outputs the feature map with the size of 32 × 32 and activated by the RELU function, and performs semantic shape processing on the feature map. Specifically, as shown in fig. 3, the generator has n layers in total, where the mth layer is a semantic layering layer, and the semantic layering includes the following steps:
s2.1, designing a semantic fractal rule of the feature map; dividing the feature graph into a plurality of fractal by taking each fractal as a class based on the principles of minimizing intra-class variance and maximizing inter-class distance, wherein the fractal is not overlapped;
the first characteristic diagram is an activated characteristic diagram, and the fractal rule is to divide the characteristic diagram into a plurality of fractal according to the distribution of the activation values of the characteristic diagram.
In the embodiment of the invention, a feature diagram A is randomly selected from output feature diagrams of a target convolutional layer, and a fractal rule is designed;
the semantic fractal of the feature graph means that the feature graph is divided according to a certain rule, and a complete feature graph is divided into a plurality of parts, namely a fractal, wherein the fractal does not overlap with each other;
in order to divide a feature graph into a plurality of fractal according to the distribution of the activation value of the feature graph, each fractal is regarded as a class, and then the feature graph is divided into a plurality of fractal on the basis of the principle of minimizing intra-class variance and maximizing inter-class distance.
S2.2, performing iterative division on the first characteristic diagram based on a fractal rule;
finding a fractal with a large area and a large variance of an activation value in the current fractal every iteration, and dividing the fractal into two parts according to the principle of minimizing the variance in the class; the formula is as follows:
Figure DEST_PATH_IMAGE002A
finding the position that minimizes the result of the above formula, divides the current fractal into two, wherein,
Figure 154354DEST_PATH_IMAGE004
and &>
Figure 868232DEST_PATH_IMAGE006
Respectively representing the variance of the activation values of the two divided sub-fractal parts; />
Figure 461019DEST_PATH_IMAGE008
And &>
Figure 856228DEST_PATH_IMAGE010
Respectively showing the area ratio of the two sub-fractal parts after division.
Step S3, constructing a fractal loss for the middle-layer feature of the generation countermeasure network according to the semantic fractal result, specifically, as shown in fig. 4, including the following steps:
s3.1, obtaining a fractal result of the first feature graph divided based on a semantic fractal rule; the fractal result comprises a fractal position of the first characteristic diagram based on a fractal rule;
in the embodiment of the invention, a fractal position P of a feature diagram A based on a fractal rule is stored A
S3.2, acquiring a second feature map output by the image generator through a middle layer, and taking the matching degree of the second feature map and a fractal result as fractal loss so as to enable the fractal of the second feature map to be close to the first feature map;
and calculating the matching degree of the second characteristic diagram and the fractal position of the first characteristic diagram as a fractal loss function so that all filters in the same layer of the generator are activated by the same part.
In the embodiment of the invention, a characteristic diagram B is selected from the output characteristic diagram of the target convolutional layer, and the characteristic diagram B and the fractal position P are calculated A As fractal loss function (frac loss):
Figure DEST_PATH_IMAGE012
wherein
Figure DEST_PATH_IMAGE014
Representing a fractal position P A The fractal index of (2). By the fractal loss function (frac loss), the fractal of the feature diagram B can be encouraged to approach the feature diagram A, so that the feature diagram fractal constraint is obtained, and the aim of activating all filters on the same layer by the same part is fulfilled.
S4, based on the fractal loss and the original loss function of the generated countermeasure network, jointly optimizing the generated countermeasure network, and training the image generation;
in order to train the GAN end-to-end so that the GAN can automatically learn feature maps with consistent fractal, based on the original GAN loss function, the fractal loss function (frac loss) is added, and network parameters are alternately optimized between the two loss functions, so as to encourage all feature maps to have consistent fractal while ensuring the picture generation quality.
In the embodiment of the invention, a fractal loss function (frac loss) is added on the basis of the loss function of the DCGAN. Specifically, a generator (G) is reinitialized as the parameters stored in the step S1, a generator image is generated according to the training set image, a discriminator (D) is initialized randomly to discriminate the generator image and the training set image, the countermeasure network (DCGAN) is generated by using the loss function fine tuning of the step S4, and in the training process, the fractal loss and the original loss function of the countermeasure network are used for optimizing the network parameters alternately, so that convolution kernels in the same layer can express consistent characteristics, the generation logic of the generator (G) is further explained, and the construction of the DCGAN can be explained.
An image generation training device based on semantic fractal interpretable GAN comprises a memory and one or more processors, wherein executable codes are stored in the memory, and the one or more processors are used for realizing the image generation training method based on semantic fractal interpretable GAN when executing the executable codes.
An image generation method capable of explaining GAN based on semantic fractal is used for generating an image to be generated through a trained discriminator.
Referring to fig. 5, an image generation apparatus based on semantic fractal interpretable GAN according to an embodiment of the present invention includes a memory and one or more processors, where the memory stores executable code, and the one or more processors execute the executable code to implement the image generation method based on semantic fractal interpretable GAN according to the above embodiment.
The embodiment of the image generation device based on semantic fractal interpretable GAN of the invention can be applied to any equipment with data processing capability, such as computers and other equipment or devices. The device embodiments may be implemented by software, or by hardware, or by a combination of hardware and software. The software implementation is taken as an example, and as a logical device, the device is formed by reading corresponding computer program instructions in the nonvolatile memory into the memory for running through the processor of any device with data processing capability. In terms of hardware, as shown in fig. 5, a hardware structure diagram of any device with data processing capability where the image generating apparatus based on semantic fractal and interpretable GAN of the present invention is located is shown, except for the processor, the memory, the network interface, and the nonvolatile memory shown in fig. 5, in the embodiment, any device with data processing capability where the apparatus is located may also include other hardware according to the actual function of the any device with data processing capability, which is not described again.
The specific details of the implementation process of the functions and actions of each unit in the above device are the implementation processes of the corresponding steps in the above method, and are not described herein again.
For the device embodiment, since it basically corresponds to the method embodiment, reference may be made to the partial description of the method embodiment for relevant points. The above-described embodiments of the apparatus are merely illustrative, and the units described as separate parts may or may not be physically separate, and parts displayed as units may or may not be physical units, may be located in one place, or may be distributed on a plurality of network units. Some or all of the modules can be selected according to actual needs to achieve the purpose of the scheme of the invention. One of ordinary skill in the art can understand and implement it without inventive effort.
An embodiment of the present invention further provides a computer-readable storage medium, on which a program is stored, and when the program is executed by a processor, the method for generating an image based on semantic fractal interpretable GAN in the above embodiments is implemented.
The computer readable storage medium may be an internal storage unit, such as a hard disk or a memory, of any data processing capability device described in any of the foregoing embodiments. The computer readable storage medium may also be any external storage device of a device with data processing capabilities, such as a plug-in hard disk, a Smart Media Card (SMC), an SD Card, a Flash memory Card (Flash Card), etc. provided on the device. Further, the computer readable storage medium may include both an internal storage unit and an external storage device of any data processing capable device. The computer-readable storage medium is used for storing the computer program and other programs and data required by the arbitrary data processing-capable device, and may also be used for temporarily storing data that has been output or is to be output.
The above examples are only intended to illustrate the technical solution of the present invention, but not to limit it; although the present invention has been described in detail with reference to the foregoing embodiments, it will be understood by those of ordinary skill in the art that: the technical solutions described in the foregoing embodiments may still be modified, or some or all of the technical features may be equivalently replaced; and the modifications or the substitutions do not make the essence of the corresponding technical solutions depart from the scope of the technical solutions of the embodiments of the present invention.

Claims (9)

1. An image generation training method capable of explaining GAN based on semantic fractal is characterized by comprising the following steps:
s1, generating a countermeasure network based on loss function training through an image training set to obtain a generator of the countermeasure network;
s2, acquiring a first feature map output by the image generator in the middle layer, and performing semantic fractal on the first feature map, wherein the method comprises the following steps:
s2.1, designing a semantic fractal rule of the feature map; dividing the feature graph into a plurality of fractal by taking each fractal as a class based on the principles of minimizing intra-class variance and maximizing inter-class distance, wherein the fractal is not overlapped;
s2.2, performing iterative division on the first characteristic diagram based on a fractal rule;
s3, constructing fractal loss for the middle-layer characteristics of the generation countermeasure network according to the semantic fractal result, and comprising the following steps of:
s3.1, obtaining a fractal result of the first feature graph divided based on a semantic fractal rule;
s3.2, acquiring a second feature map output by the image generator in the middle layer, and taking the matching degree of the second feature map and the fractal result as fractal loss so as to enable the fractal of the second feature map to be close to the first feature map;
and S4, generating the countermeasure network through joint optimization based on the fractal loss and the original loss function of the generated countermeasure network, and training image generation.
2. The image generation training method based on semantic fractal interpretable GAN of claim 1, wherein: in step S2, the first feature map is an activated feature map, and the fractal rule is to divide the feature map into a plurality of fractal according to the distribution of the activation values of the feature map.
3. The image generation training method based on semantic fractal interpretable GAN of claim 2, wherein: in the step S2.2, a fractal with a large area and a large variance of an activation value is found in the current fractal in each iteration, and the fractal is divided into two parts according to the principle of minimizing the intra-class variance.
4. The image generation training method based on semantic fractal interpretable GAN of claim 3, wherein: the principle of minimizing the intra-class variance is divided into two, and the formula is as follows:
Figure DEST_PATH_IMAGE002
wherein the content of the first and second substances,
Figure DEST_PATH_IMAGE004
and
Figure DEST_PATH_IMAGE006
respectively representing the variance of the activation values of the two divided sub-fractal parts;
Figure DEST_PATH_IMAGE008
and
Figure DEST_PATH_IMAGE010
respectively showing the area ratio of the two sub-fractal parts after division.
5. The image generation training method based on semantic fractal interpretable GAN of claim 1, wherein: the fractal result in the step S3.1 comprises a fractal position of the first characteristic diagram based on a fractal rule; in step S3.2, the matching degree between the second feature map and the fractal position of the first feature map is calculated as a fractal loss, so that all filters in the same middle layer of the generator are activated by the same part.
6. The image generation training method based on semantic fractal interpretable GAN of claim 1, wherein: in the step S4, the generator is reinitialized as the parameters stored in the step S1, a generator image is generated according to the training set image, the discriminator is initialized randomly to discriminate the generator image and the training set image, the countermeasure network is generated by fine tuning of the loss function in the step S4, and in the training process, the network parameters are optimized alternately by the fractal loss and the original loss function of the generation countermeasure network, so that the convolution kernels in the same layer can express consistent features, and the generation logic of the generator is further explained.
7. An image generation training device based on semantic fractal interpretable GAN, comprising a memory and one or more processors, wherein the memory stores executable code, and the one or more processors execute the executable code to implement the image generation training method based on semantic fractal interpretable GAN as claimed in any one of claims 1 to 6.
8. The semantic fractal-based interpretable GAN image generating method according to claim 1, wherein: and generating an image to be generated through the trained discriminator.
9. An image generation device based on semantic fractal interpretable GAN, comprising a memory and one or more processors, wherein the memory stores executable code, and the one or more processors execute the executable code to implement the image generation method based on semantic fractal interpretable GAN as claimed in claim 8.
CN202211373030.8A 2022-11-04 2022-11-04 Image generation training method and device capable of explaining GAN based on semantic fractal Active CN115424119B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202211373030.8A CN115424119B (en) 2022-11-04 2022-11-04 Image generation training method and device capable of explaining GAN based on semantic fractal

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202211373030.8A CN115424119B (en) 2022-11-04 2022-11-04 Image generation training method and device capable of explaining GAN based on semantic fractal

Publications (2)

Publication Number Publication Date
CN115424119A CN115424119A (en) 2022-12-02
CN115424119B true CN115424119B (en) 2023-03-24

Family

ID=84207824

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202211373030.8A Active CN115424119B (en) 2022-11-04 2022-11-04 Image generation training method and device capable of explaining GAN based on semantic fractal

Country Status (1)

Country Link
CN (1) CN115424119B (en)

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111476294A (en) * 2020-04-07 2020-07-31 南昌航空大学 Zero sample image identification method and system based on generation countermeasure network
EP3742346A2 (en) * 2019-05-23 2020-11-25 HTC Corporation Method for training generative adversarial network (gan), method for generating images by using gan, and computer readable storage medium
CN115017337A (en) * 2022-08-03 2022-09-06 中国电子科技集团公司第五十四研究所 Latent semantic interpretation method for deep learning model

Family Cites Families (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB201809604D0 (en) * 2018-06-12 2018-07-25 Tom Tom Global Content B V Generative adversarial networks for image segmentation
CN110689086B (en) * 2019-10-08 2020-09-25 郑州轻工业学院 Semi-supervised high-resolution remote sensing image scene classification method based on generating countermeasure network
CN111489287B (en) * 2020-04-10 2024-02-09 腾讯科技(深圳)有限公司 Image conversion method, device, computer equipment and storage medium
CN112308862A (en) * 2020-06-04 2021-02-02 北京京东尚科信息技术有限公司 Image semantic segmentation model training method, image semantic segmentation model training device, image semantic segmentation model segmentation method, image semantic segmentation model segmentation device and storage medium
CN111832570A (en) * 2020-07-02 2020-10-27 北京工业大学 Image semantic segmentation model training method and system
CN112132149B (en) * 2020-09-10 2023-09-05 武汉汉达瑞科技有限公司 Semantic segmentation method and device for remote sensing image
CN112101463A (en) * 2020-09-17 2020-12-18 成都数之联科技有限公司 Image semantic segmentation network training method, segmentation device and medium
CN113450313B (en) * 2021-06-04 2022-03-15 电子科技大学 Image significance visualization method based on regional contrast learning
CN114359526B (en) * 2021-12-29 2024-05-28 中山大学 Cross-domain image style migration method based on semantic GAN
CN114757864A (en) * 2022-04-21 2022-07-15 西安交通大学 Multi-level fine-grained image generation method based on multi-scale feature decoupling
CN114998124A (en) * 2022-05-23 2022-09-02 北京航空航天大学 Image sharpening processing method for target detection

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3742346A2 (en) * 2019-05-23 2020-11-25 HTC Corporation Method for training generative adversarial network (gan), method for generating images by using gan, and computer readable storage medium
CN111476294A (en) * 2020-04-07 2020-07-31 南昌航空大学 Zero sample image identification method and system based on generation countermeasure network
CN115017337A (en) * 2022-08-03 2022-09-06 中国电子科技集团公司第五十四研究所 Latent semantic interpretation method for deep learning model

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
基于图像云模型语义标注的条件生成对抗网络;杜秋平等;《模式识别与人工智能》;20180415(第04期);全文 *

Also Published As

Publication number Publication date
CN115424119A (en) 2022-12-02

Similar Documents

Publication Publication Date Title
Hui et al. Linguistic structure guided context modeling for referring image segmentation
Gwak et al. Generative sparse detection networks for 3d single-shot object detection
CN111754596B (en) Editing model generation method, device, equipment and medium for editing face image
US20220172518A1 (en) Image recognition method and apparatus, computer-readable storage medium, and electronic device
WO2022105125A1 (en) Image segmentation method and apparatus, computer device, and storage medium
CN111275107A (en) Multi-label scene image classification method and device based on transfer learning
US9031894B2 (en) Parsing and rendering structured images
DE102022107186A1 (en) GENERATOR UTILIZATION FOR DEEPFAKE DETECTION
DE102022106057A1 (en) AUTHENTICATOR-INTEGRATED GENERATIVE ADVERSARIAL NETWORK (GAN) FOR SECURE DEEPFAKE GENERATION
Douillard et al. Tackling catastrophic forgetting and background shift in continual semantic segmentation
Rombach et al. Making sense of cnns: Interpreting deep representations and their invariances with inns
CN116958323A (en) Image generation method, device, electronic equipment, storage medium and program product
CN115131803A (en) Document word size identification method and device, computer equipment and storage medium
Xu et al. A novel image feature extraction algorithm based on the fusion AutoEncoder and CNN
Asperti et al. Image embedding for denoising generative models
CN117197292A (en) Method, apparatus, device and storage medium for generating image
CN117392293A (en) Image processing method, device, electronic equipment and storage medium
CN115424119B (en) Image generation training method and device capable of explaining GAN based on semantic fractal
CN112069412A (en) Information recommendation method and device, computer equipment and storage medium
Yu et al. Filling gaps of cartographic polylines by using an encoder–decoder model
CN112395834B (en) Brain graph generation method, device and equipment based on picture input and storage medium
CN114863000A (en) Method, device, medium and equipment for generating hairstyle
CN114580510A (en) Bone marrow cell fine-grained classification method, system, computer device and storage medium
Jiang et al. Image inpainting based on cross-hierarchy global and local aware network
CN116740540B (en) Data processing method, device, equipment and computer readable storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant