CN111126599A - Neural network weight initialization method based on transfer learning - Google Patents

Neural network weight initialization method based on transfer learning Download PDF

Info

Publication number
CN111126599A
CN111126599A CN201911321102.2A CN201911321102A CN111126599A CN 111126599 A CN111126599 A CN 111126599A CN 201911321102 A CN201911321102 A CN 201911321102A CN 111126599 A CN111126599 A CN 111126599A
Authority
CN
China
Prior art keywords
model
student
student model
teacher
neural network
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201911321102.2A
Other languages
Chinese (zh)
Other versions
CN111126599B (en
Inventor
范益波
刘超
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Fudan University
Original Assignee
Fudan University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Fudan University filed Critical Fudan University
Priority to CN201911321102.2A priority Critical patent/CN111126599B/en
Publication of CN111126599A publication Critical patent/CN111126599A/en
Application granted granted Critical
Publication of CN111126599B publication Critical patent/CN111126599B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/042Knowledge-based neural networks; Logical representations of neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T9/00Image coding
    • G06T9/002Image coding using neural networks
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/80Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation
    • H04N19/82Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation involving filtering within a prediction loop
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Evolutionary Computation (AREA)
  • General Physics & Mathematics (AREA)
  • Artificial Intelligence (AREA)
  • General Health & Medical Sciences (AREA)
  • General Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Computational Linguistics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Image Analysis (AREA)

Abstract

The invention belongs to the technical field of neural network models, and particularly relates to a neural network weight initialization method based on transfer learning. In the method, for a specified target task, a neural network model with higher complexity, namely a teacher model, is designed, the teacher model is trained, and after the training is finished, the generated characteristic diagram is used for guiding the weight initialization of the student model; calculating the difference of the characteristic maps in a regeneration core Hilbert space by calculating the difference between the characteristic maps or mapping the characteristic maps to the regeneration core Hilbert space, and simplifying the calculation by adopting a kernel function method; the simple student model achieves a better weight initialization effect, and after the weight initialization is completed, the student model is generally trained, so that the student model achieves a better global convergence point, and the performance of the student model is more excellent. The invention can effectively improve the performance of the student model on the premise of not increasing the complexity of the student model.

Description

Neural network weight initialization method based on transfer learning
Technical Field
The invention belongs to the technical field of neural network models, and particularly relates to a neural network weight initialization method based on knowledge transfer learning.
Background
The neural network has been developed in recent years, especially in the computer vision field and the natural language processing field, many performances of the neural network exceed those of human beings, however, the high calculation amount and the large training requirement of the neural network cause great obstacles for the practical application of the neural network. Therefore, how to make a lightweight model perform better becomes a hot spot problem to be solved.
Over the past few years, many researchers have proposed various schemes to help neural networks achieve a better convergence. The method mainly comprises the following classes, one class is based on knowledge distillation and knowledge migration, and the trained teacher model is used for helping the student model to express better by adding some additional loss functions in the training process of the student model, so that the performance of the student model is improved on the basis of not increasing the complexity of the student model. The second type is quantization pruning based on a model, and the original 32-bit addition and subtraction is changed into 8-bit or even 1-bit addition and subtraction by quantizing the weight of the neural network, so that the weight complexity of the neural network is greatly reduced. Thereby reducing the amount of computation. Pruning is to directly delete some connecting edges of the neural network connection, and then to evaluate whether the loss of the model caused by pruning is negligible, thereby achieving the effect of effectively reducing the complexity of the model.
Disclosure of Invention
The invention aims to provide a neural network weight initialization method for effectively improving the performance of a model on the basis of not increasing the complexity of the model.
The neural network weight initialization method provided by the invention is based on the knowledge transfer learning technology, namely a neural network model with higher complexity is called a teacher model (the teacher model with higher complexity is difficult to be applied in practical engineering); the prior knowledge is learned from a relatively complex teacher model, so that a neural network model with low complexity (called a student model which has good balance between complexity and performance in practical application) is helped to have a good initialization state, and a local optimal point in the training process is removed, so that a better training effect is achieved.
Firstly, designing a neural network model with higher complexity, namely a teacher model, for a specified task, training the teacher model, and guiding the weight initialization of the student model by utilizing a generated characteristic diagram after the training is finished; calculating the difference of the characteristic maps in a regeneration core Hilbert space by calculating the difference between the characteristic maps or mapping the characteristic maps into the regeneration core Hilbert space, and simplifying the calculation by adopting a kernel function method; the simple student model achieves a better weight initialization effect, and after the weight initialization is completed, the student model is generally trained, so that the student model achieves a better global convergence point and is more excellent in performance.
The invention can effectively lead the student model to avoid the problem of converging to the local optimal solution caused by parameter dependence in the initial training.
The invention provides a neural network weight initialization method, which comprises the following specific steps:
(1) for a specific learning task, a conventional loss function and a model structure are often possessed, firstly, a teacher model is designed aiming at a target task, and the teacher model is trained by using the conventional loss function;
(2) then exporting the middle layer output of the trained teacher model, and obtaining a characteristic diagram in a mapping mode; wherein the mapping method may be attention migration [ Sergey Zagoruyko and Nikos Komodakis, "riding movement to engagement of the performance of the connected neural networks vision engagement transfer," arXiv prediction arXiv:1612.03928,2016 ], or mapping into the regeneration nuclear Hilbert space using a kernel function [ Zehao Huang and Naiyan Wang, "Like world shift of vision selection," arXiv prediction arXiv:1707.01219,2017 ]; specifically, the formula is shown as (2) and (3);
(3) designing a student model with a simpler structure, wherein the student model and the teacher model are required to have the same network structure; that is, the basic network layers constituting the network should be consistent, for example, the network structure adopts a serial connection network constituted based on convolutional layers, and the teacher model has more convolutional layers, more feature maps, fewer convolutional layers of the student model, and fewer feature maps;
(4) training the student model by taking the mean square error between the feature map calculated in the step (2) and the feature map obtained by mapping in the same way as the student model as a loss function; after training is finished, the obtained weights of the student models are not initialization of normal distribution or uniform distribution in the traditional sense, but the weights of the student models are adjusted by learning knowledge from the teacher model, so that the weights of the student models are specifically initialized, and the student models have the capability of approaching the performance of the teacher model;
(5) after the initialization is completed, the student model is finally trained by using a conventional loss function, and a usable student model is obtained.
The conventional loss function refers to a mean square error that is often used in the current task.
In the invention, the complexity of the teacher model is higher than that of the student model, so that the student model can well learn the characteristics of the teacher model.
Drawings
FIG. 1 is a schematic view of the process of the present invention.
Training loss and test loss (QP 22) for the experiment of fig. 2.
Training loss and testing loss (QP 37) for the experiment of fig. 3.
Detailed Description
The present invention is further described below by taking the neural network-based loop filtering task in video coding as an example.
For the target task, a neural network module is required to be added in the traditional video encoder, such as HEVC, and the function of the neural network module is loop filtering. The performance of the video encoder is improved by a loop filtering method based on a neural network. It can be understood as a noise reduction filtering problem to remove the artificial imprint and noise brought by the conventional video encoder. Firstly, designing a teacher model with higher complexity, wherein the complexity of the teacher model is significantly higher than that of the teacher model in the actual target application of the final target, for example, the computational complexity and the consumed computational resources of the teacher model are more than 2 times of those of the expected design model; and training the teacher model by using a conventional loss function to obtain a trained teacher model. Aiming at a loop filtering task based on a neural network, a convolutional neural network is designed, the structure of the convolutional neural network is shown in figure 1, the upper half part is a designed teacher model structure, and the lower half part is a designed student model structure. The conventional loss function is referred to herein as the mean square error, which is often used in current tasks. This loss function is used to train teacher and student models.
In the aspect of model structure, the depth separable convolution and batch normalization are adopted as main layers of the teacher model, the characteristic layer number is 64, and the convolution kernel size is 3x 3. Wherein 24 depth-separable convolutional layers are used as the backbone of the teacher's model, and are divided into three parts, the first part being composed of 10 depth-separable convolutional layers, the second part being composed of 8 depth-separable convolutional layers, and the third part being composed of 6 depth-separable convolutional layers. The last layer of the model is a common convolution layer, the number of the characteristic layers is 1, and the size of a convolution kernel is 1x 1. All depth separable convolutions have a ReLU activation function. The input of the model is connected to the final output through a direct edge, so that the neural network is in a residual learning state, and the model converges more quickly.
The input and output of the model are respectively a reconstructed pixel map and a filtered pixel map of a video encoder, and a loss function LTOptionally the output of the neural network
Figure BDA0002327174760000031
And original pixel
Figure BDA0002327174760000032
Mean square error between:
Figure BDA0002327174760000033
after the teacher model training is finished, the middle layer outputs of three sub-parts of the teacher model are obtained from the trained data set, and the output results F of the teacher model at the positions are calculated from the neural network attention map or the mapping result of the regenerative nuclear Hilbert spaceT. Intermediate layer calculation result F using teacher modelTAnd the data set input by the teacher model form a new data set to train the student model.
The construction of the student model needs a similar teacher model so as to ensure the success of knowledge migration. We have also adopted a similar network architecture using 9 depth separable convolutional layers as the backbone of the student model, dividing it into three parts, each of which is composed of 3 depth separable convolutional layers. The number of characteristic layers is 32, and the convolution kernel size is 3x 3. Meanwhile, because the designed targets of the student model and the teacher model are consistent, the input and the output of the student model are also consistent, and a structure of a direct connection edge is adopted for well learning residual errors of the student model.
The middle layer output of the teacher model and the middle layer output of the student model are subjected to the same mapping, and the mean square error between the middle layer output of the teacher model and the middle layer output of the student model is calculated. The loss function is expressed as equations (2) and (3). Calculating an attention diagram of the system by using a linear kernel function k (x, y) ═ xTy to approximate the result of the hubert space mapping at the regenerating kernel. The calculation formula is as follows:
Figure BDA0002327174760000041
Figure BDA0002327174760000042
here, FT,FSRepresenting the attention diagrams of the teacher and student models respectively,
Figure BDA0002327174760000043
and
Figure BDA0002327174760000044
the ith feature map, CTAnd CSThe number of feature layers of the teacher model and the student model are respectively represented. For practical applications, p is a positive integer. Thus, an initialized student model is obtained.
Then, the standard mean square error L shown in the formula (4) is usedSAnd finally training the student model to obtain the trained student model. Corresponding to
Figure BDA0002327174760000045
Representing the output of the student model.
Figure BDA0002327174760000046
After such initialization, lightweight models may tend to perform better than without the initialization method. After the student model is obtained, the student model is led back to the video encoder, and therefore loop filtering based on the neural network in video encoding can be achieved. The training and testing penalty in the experiment is shown in fig. 3, and it is evident that the penalty function decreases faster after initialization is used.

Claims (4)

1. A neural network weight initialization algorithm based on transfer learning is characterized in that for a specified target task, a neural network model with high complexity, namely a teacher model, is designed, the teacher model is trained, and after the training is finished, a generated characteristic diagram is used for guiding the weight initialization of a student model; calculating the difference of the characteristic maps in a regeneration core Hilbert space by calculating the difference between the characteristic maps or mapping the characteristic maps to the regeneration core Hilbert space, and simplifying the calculation by adopting a kernel function method; the simple student model achieves a better weight initialization effect, and after the weight initialization is completed, the student model is generally trained, so that the student model achieves a better global convergence point and has more excellent performance; the method comprises the following specific steps:
(1) for a specific learning task, a conventional loss function and a model structure are possessed, firstly, a teacher model is designed aiming at a target task, and the teacher model is trained by using the conventional loss function;
(2) then exporting the middle layer output of the trained teacher model, and obtaining a characteristic diagram in a mapping mode; wherein, the mapping mode has attention migration, or the mapping mode uses kernel function to map to the regeneration kernel Hilbert space;
(3) designing a student model with a simpler structure, wherein the student model and the teacher model are required to have the same network structure, namely the basic network layer forming the network is consistent; when the network structures all adopt the series connection network formed by the convolution layers, the teacher model has more convolution layers and more feature graphs, and the student model has fewer convolution layers and fewer feature graphs;
(4) training the student model by taking the mean square error between the feature map calculated in the step (2) and the feature map obtained by mapping in the same way as the student model as a loss function; after training is finished, the obtained weight of the student model is adjusted by learning knowledge from the teacher model, so that the weight of the student model is specifically initialized, and the student model has the capability of approaching the performance of the teacher model;
(5) after initialization is completed, the student model is trained by using a conventional loss function, and a usable student model is obtained.
2. The neural network weight initialization algorithm based on transfer learning of claim 1, characterized in that a deep separable convolution and batch normalization is adopted as a main layer of a teacher model, the number of characteristic layers is 64, and the convolution kernel size is 3x 3; wherein 24 depth-separable convolutional layers are used as the backbone of the teacher model, and are divided into three parts, the first part is composed of 10 depth-separable convolutional layers, the second part is composed of 8 depth-separable convolutional layers, and the third part is composed of 6 depth-separable convolutional layers; the last layer of the model is a common convolution layer, the number of the characteristic layers is 1, and the size of a convolution kernel is 1x 1; all depth separable convolutions have a ReLU activation function; the input of the model is connected to the final output through a straight edge, so that the neural network is in a residual learning state, and the convergence is faster.
3. The neural network weight initialization algorithm based on transfer learning of claim 2, wherein the input and output of the model are respectively a reconstructed pixel map and a filtered pixel map of a video encoder, and the loss function L isTSelecting the output of a neural network
Figure FDA0002327174750000011
And original pixel
Figure FDA0002327174750000012
Mean square error between:
Figure FDA0002327174750000021
after the teacher model training is finished, the middle layer outputs of three sub-parts of the teacher model are obtained from the trained data set, and the output results F of the teacher model at the positions are calculated from the neural network attention map or the mapping result of the regenerative nuclear Hilbert spaceT(ii) a Intermediate layer calculation result F using teacher modelTForming a new data set with the data set input by the teacher model to train the student model;
the student model adopts a network structure similar to a teacher model, uses 9 depth separable convolutional layers as a backbone of the student model, and divides the student model into three parts, wherein each part is composed of 3 depth separable convolutional layers, the number of the characteristic layers is 32, and the size of a convolutional kernel is 3x 3; the designed targets of the student model and the teacher model are consistent, the input and the output of the student model are also consistent, and a structure of a direct connection edge is also adopted for better learning residual errors of the student model;
the middle layer output of the teacher model and the middle layer output of the student model are subjected to the same mapping, and the mean square error between the middle layer output of the teacher model and the middle layer output of the student model is calculated; the loss function is expressed as formulas (2) and (3); calculating an attention diagram of the system by using a linear kernel function k (x, y) ═ xTy to approximate the result of the hilbert space mapping at the regenerating kernel; the calculation formula is as follows:
Figure FDA0002327174750000022
Figure FDA0002327174750000023
here, FT,FSRepresenting the attention diagrams of the teacher and student models respectively,
Figure FDA0002327174750000024
and
Figure FDA0002327174750000025
the ith feature map, CTAnd CSRespectively representing the number of characteristic layers of a teacher model and a student model, and taking a positive integer as p; thus, an initialized student model is obtained.
4. The neural network weight initialization algorithm based on transfer learning of claim 3, wherein a standard mean square error L as shown in formula (4) is usedSCarrying out final training on the initialized student model to obtain a trained student model; corresponding to
Figure FDA0002327174750000026
Representing the output of the student model;
Figure FDA0002327174750000027
and leading the student model back to a video encoder, namely realizing the loop filtering based on the neural network in the video encoding.
CN201911321102.2A 2019-12-20 2019-12-20 Neural network weight initialization method based on transfer learning Active CN111126599B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911321102.2A CN111126599B (en) 2019-12-20 2019-12-20 Neural network weight initialization method based on transfer learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911321102.2A CN111126599B (en) 2019-12-20 2019-12-20 Neural network weight initialization method based on transfer learning

Publications (2)

Publication Number Publication Date
CN111126599A true CN111126599A (en) 2020-05-08
CN111126599B CN111126599B (en) 2023-09-05

Family

ID=70500352

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911321102.2A Active CN111126599B (en) 2019-12-20 2019-12-20 Neural network weight initialization method based on transfer learning

Country Status (1)

Country Link
CN (1) CN111126599B (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111554268A (en) * 2020-07-13 2020-08-18 腾讯科技(深圳)有限公司 Language identification method based on language model, text classification method and device
CN112464959A (en) * 2020-12-12 2021-03-09 中南民族大学 Plant phenotype detection system and method based on attention and multiple knowledge migration
CN112929663A (en) * 2021-04-08 2021-06-08 中国科学技术大学 Knowledge distillation-based image compression quality enhancement method
CN113469977A (en) * 2021-07-06 2021-10-01 浙江霖研精密科技有限公司 Flaw detection device and method based on distillation learning mechanism and storage medium
WO2022001805A1 (en) * 2020-06-30 2022-01-06 华为技术有限公司 Neural network distillation method and device

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2017167347A (en) * 2016-03-16 2017-09-21 日本電信電話株式会社 Acoustic signal analysis device, method, and program
CN109087303A (en) * 2018-08-15 2018-12-25 中山大学 The frame of semantic segmentation modelling effect is promoted based on transfer learning
CN110163110A (en) * 2019-04-23 2019-08-23 中电科大数据研究院有限公司 A kind of pedestrian's recognition methods again merged based on transfer learning and depth characteristic
US20190287515A1 (en) * 2018-03-16 2019-09-19 Microsoft Technology Licensing, Llc Adversarial Teacher-Student Learning for Unsupervised Domain Adaptation

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2017167347A (en) * 2016-03-16 2017-09-21 日本電信電話株式会社 Acoustic signal analysis device, method, and program
US20190287515A1 (en) * 2018-03-16 2019-09-19 Microsoft Technology Licensing, Llc Adversarial Teacher-Student Learning for Unsupervised Domain Adaptation
CN109087303A (en) * 2018-08-15 2018-12-25 中山大学 The frame of semantic segmentation modelling effect is promoted based on transfer learning
CN110163110A (en) * 2019-04-23 2019-08-23 中电科大数据研究院有限公司 A kind of pedestrian's recognition methods again merged based on transfer learning and depth characteristic

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
张振宇;: "稳健的多支持向量机自适应提升算法", 大连交通大学学报 *

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2022001805A1 (en) * 2020-06-30 2022-01-06 华为技术有限公司 Neural network distillation method and device
EP4163831A4 (en) * 2020-06-30 2023-12-06 Huawei Technologies Co., Ltd. Neural network distillation method and device
CN111554268A (en) * 2020-07-13 2020-08-18 腾讯科技(深圳)有限公司 Language identification method based on language model, text classification method and device
CN111554268B (en) * 2020-07-13 2020-11-03 腾讯科技(深圳)有限公司 Language identification method based on language model, text classification method and device
CN112464959A (en) * 2020-12-12 2021-03-09 中南民族大学 Plant phenotype detection system and method based on attention and multiple knowledge migration
CN112464959B (en) * 2020-12-12 2023-12-19 中南民族大学 Plant phenotype detection system and method based on attention and multiple knowledge migration
CN112929663A (en) * 2021-04-08 2021-06-08 中国科学技术大学 Knowledge distillation-based image compression quality enhancement method
CN113469977A (en) * 2021-07-06 2021-10-01 浙江霖研精密科技有限公司 Flaw detection device and method based on distillation learning mechanism and storage medium
CN113469977B (en) * 2021-07-06 2024-01-12 浙江霖研精密科技有限公司 Flaw detection device, method and storage medium based on distillation learning mechanism

Also Published As

Publication number Publication date
CN111126599B (en) 2023-09-05

Similar Documents

Publication Publication Date Title
CN111126599A (en) Neural network weight initialization method based on transfer learning
CN113159173B (en) Convolutional neural network model compression method combining pruning and knowledge distillation
CN111784602A (en) Method for generating countermeasure network for image restoration
CN113313644B (en) Underwater image enhancement method based on residual double-attention network
CN112489164B (en) Image coloring method based on improved depth separable convolutional neural network
CN109005398B (en) Stereo image parallax matching method based on convolutional neural network
CN113595993B (en) Vehicle-mounted sensing equipment joint learning method for model structure optimization under edge calculation
CN113947680A (en) Image semantic segmentation method based on cascade multi-scale vision Transformer
CN112017116B (en) Image super-resolution reconstruction network based on asymmetric convolution and construction method thereof
CN113163203A (en) Deep learning feature compression and decompression method, system and terminal
CN115330620A (en) Image defogging method based on cyclic generation countermeasure network
CN111861886A (en) Image super-resolution reconstruction method based on multi-scale feedback network
CN112651360A (en) Skeleton action recognition method under small sample
CN116958534A (en) Image processing method, training method of image processing model and related device
CN113989283B (en) 3D human body posture estimation method and device, electronic equipment and storage medium
CN108629374A (en) A kind of unsupervised multi-modal Subspace clustering method based on convolutional neural networks
CN109448039B (en) Monocular vision depth estimation method based on deep convolutional neural network
CN109063834B (en) Neural network pruning method based on convolution characteristic response graph
CN110752894A (en) CNN-based LDPC code blind channel decoding method and decoder
CN113938254A (en) Attention mechanism-based layered source-channel joint coding transmission system and transmission method thereof
CN110223224A (en) A kind of Image Super-resolution realization algorithm based on information filtering network
CN113807497A (en) Non-paired image translation method for enhancing texture details
CN116030537B (en) Three-dimensional human body posture estimation method based on multi-branch attention-seeking convolution
CN117078539A (en) CNN-transducer-based local global interactive image restoration method
CN116600119A (en) Video encoding method, video decoding method, video encoding device, video decoding device, computer equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant