CN111126599A - Neural network weight initialization method based on transfer learning - Google Patents
Neural network weight initialization method based on transfer learning Download PDFInfo
- Publication number
- CN111126599A CN111126599A CN201911321102.2A CN201911321102A CN111126599A CN 111126599 A CN111126599 A CN 111126599A CN 201911321102 A CN201911321102 A CN 201911321102A CN 111126599 A CN111126599 A CN 111126599A
- Authority
- CN
- China
- Prior art keywords
- model
- student
- student model
- teacher
- neural network
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/042—Knowledge-based neural networks; Logical representations of neural networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T9/00—Image coding
- G06T9/002—Image coding using neural networks
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/80—Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation
- H04N19/82—Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation involving filtering within a prediction loop
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02T—CLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
- Y02T10/00—Road transport of goods or passengers
- Y02T10/10—Internal combustion engine [ICE] based vehicles
- Y02T10/40—Engine management systems
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Evolutionary Computation (AREA)
- General Physics & Mathematics (AREA)
- Artificial Intelligence (AREA)
- General Health & Medical Sciences (AREA)
- General Engineering & Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Biophysics (AREA)
- Biomedical Technology (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Computational Linguistics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Image Analysis (AREA)
Abstract
The invention belongs to the technical field of neural network models, and particularly relates to a neural network weight initialization method based on transfer learning. In the method, for a specified target task, a neural network model with higher complexity, namely a teacher model, is designed, the teacher model is trained, and after the training is finished, the generated characteristic diagram is used for guiding the weight initialization of the student model; calculating the difference of the characteristic maps in a regeneration core Hilbert space by calculating the difference between the characteristic maps or mapping the characteristic maps to the regeneration core Hilbert space, and simplifying the calculation by adopting a kernel function method; the simple student model achieves a better weight initialization effect, and after the weight initialization is completed, the student model is generally trained, so that the student model achieves a better global convergence point, and the performance of the student model is more excellent. The invention can effectively improve the performance of the student model on the premise of not increasing the complexity of the student model.
Description
Technical Field
The invention belongs to the technical field of neural network models, and particularly relates to a neural network weight initialization method based on knowledge transfer learning.
Background
The neural network has been developed in recent years, especially in the computer vision field and the natural language processing field, many performances of the neural network exceed those of human beings, however, the high calculation amount and the large training requirement of the neural network cause great obstacles for the practical application of the neural network. Therefore, how to make a lightweight model perform better becomes a hot spot problem to be solved.
Over the past few years, many researchers have proposed various schemes to help neural networks achieve a better convergence. The method mainly comprises the following classes, one class is based on knowledge distillation and knowledge migration, and the trained teacher model is used for helping the student model to express better by adding some additional loss functions in the training process of the student model, so that the performance of the student model is improved on the basis of not increasing the complexity of the student model. The second type is quantization pruning based on a model, and the original 32-bit addition and subtraction is changed into 8-bit or even 1-bit addition and subtraction by quantizing the weight of the neural network, so that the weight complexity of the neural network is greatly reduced. Thereby reducing the amount of computation. Pruning is to directly delete some connecting edges of the neural network connection, and then to evaluate whether the loss of the model caused by pruning is negligible, thereby achieving the effect of effectively reducing the complexity of the model.
Disclosure of Invention
The invention aims to provide a neural network weight initialization method for effectively improving the performance of a model on the basis of not increasing the complexity of the model.
The neural network weight initialization method provided by the invention is based on the knowledge transfer learning technology, namely a neural network model with higher complexity is called a teacher model (the teacher model with higher complexity is difficult to be applied in practical engineering); the prior knowledge is learned from a relatively complex teacher model, so that a neural network model with low complexity (called a student model which has good balance between complexity and performance in practical application) is helped to have a good initialization state, and a local optimal point in the training process is removed, so that a better training effect is achieved.
Firstly, designing a neural network model with higher complexity, namely a teacher model, for a specified task, training the teacher model, and guiding the weight initialization of the student model by utilizing a generated characteristic diagram after the training is finished; calculating the difference of the characteristic maps in a regeneration core Hilbert space by calculating the difference between the characteristic maps or mapping the characteristic maps into the regeneration core Hilbert space, and simplifying the calculation by adopting a kernel function method; the simple student model achieves a better weight initialization effect, and after the weight initialization is completed, the student model is generally trained, so that the student model achieves a better global convergence point and is more excellent in performance.
The invention can effectively lead the student model to avoid the problem of converging to the local optimal solution caused by parameter dependence in the initial training.
The invention provides a neural network weight initialization method, which comprises the following specific steps:
(1) for a specific learning task, a conventional loss function and a model structure are often possessed, firstly, a teacher model is designed aiming at a target task, and the teacher model is trained by using the conventional loss function;
(2) then exporting the middle layer output of the trained teacher model, and obtaining a characteristic diagram in a mapping mode; wherein the mapping method may be attention migration [ Sergey Zagoruyko and Nikos Komodakis, "riding movement to engagement of the performance of the connected neural networks vision engagement transfer," arXiv prediction arXiv:1612.03928,2016 ], or mapping into the regeneration nuclear Hilbert space using a kernel function [ Zehao Huang and Naiyan Wang, "Like world shift of vision selection," arXiv prediction arXiv:1707.01219,2017 ]; specifically, the formula is shown as (2) and (3);
(3) designing a student model with a simpler structure, wherein the student model and the teacher model are required to have the same network structure; that is, the basic network layers constituting the network should be consistent, for example, the network structure adopts a serial connection network constituted based on convolutional layers, and the teacher model has more convolutional layers, more feature maps, fewer convolutional layers of the student model, and fewer feature maps;
(4) training the student model by taking the mean square error between the feature map calculated in the step (2) and the feature map obtained by mapping in the same way as the student model as a loss function; after training is finished, the obtained weights of the student models are not initialization of normal distribution or uniform distribution in the traditional sense, but the weights of the student models are adjusted by learning knowledge from the teacher model, so that the weights of the student models are specifically initialized, and the student models have the capability of approaching the performance of the teacher model;
(5) after the initialization is completed, the student model is finally trained by using a conventional loss function, and a usable student model is obtained.
The conventional loss function refers to a mean square error that is often used in the current task.
In the invention, the complexity of the teacher model is higher than that of the student model, so that the student model can well learn the characteristics of the teacher model.
Drawings
FIG. 1 is a schematic view of the process of the present invention.
Training loss and test loss (QP 22) for the experiment of fig. 2.
Training loss and testing loss (QP 37) for the experiment of fig. 3.
Detailed Description
The present invention is further described below by taking the neural network-based loop filtering task in video coding as an example.
For the target task, a neural network module is required to be added in the traditional video encoder, such as HEVC, and the function of the neural network module is loop filtering. The performance of the video encoder is improved by a loop filtering method based on a neural network. It can be understood as a noise reduction filtering problem to remove the artificial imprint and noise brought by the conventional video encoder. Firstly, designing a teacher model with higher complexity, wherein the complexity of the teacher model is significantly higher than that of the teacher model in the actual target application of the final target, for example, the computational complexity and the consumed computational resources of the teacher model are more than 2 times of those of the expected design model; and training the teacher model by using a conventional loss function to obtain a trained teacher model. Aiming at a loop filtering task based on a neural network, a convolutional neural network is designed, the structure of the convolutional neural network is shown in figure 1, the upper half part is a designed teacher model structure, and the lower half part is a designed student model structure. The conventional loss function is referred to herein as the mean square error, which is often used in current tasks. This loss function is used to train teacher and student models.
In the aspect of model structure, the depth separable convolution and batch normalization are adopted as main layers of the teacher model, the characteristic layer number is 64, and the convolution kernel size is 3x 3. Wherein 24 depth-separable convolutional layers are used as the backbone of the teacher's model, and are divided into three parts, the first part being composed of 10 depth-separable convolutional layers, the second part being composed of 8 depth-separable convolutional layers, and the third part being composed of 6 depth-separable convolutional layers. The last layer of the model is a common convolution layer, the number of the characteristic layers is 1, and the size of a convolution kernel is 1x 1. All depth separable convolutions have a ReLU activation function. The input of the model is connected to the final output through a direct edge, so that the neural network is in a residual learning state, and the model converges more quickly.
The input and output of the model are respectively a reconstructed pixel map and a filtered pixel map of a video encoder, and a loss function LTOptionally the output of the neural networkAnd original pixelMean square error between:
after the teacher model training is finished, the middle layer outputs of three sub-parts of the teacher model are obtained from the trained data set, and the output results F of the teacher model at the positions are calculated from the neural network attention map or the mapping result of the regenerative nuclear Hilbert spaceT. Intermediate layer calculation result F using teacher modelTAnd the data set input by the teacher model form a new data set to train the student model.
The construction of the student model needs a similar teacher model so as to ensure the success of knowledge migration. We have also adopted a similar network architecture using 9 depth separable convolutional layers as the backbone of the student model, dividing it into three parts, each of which is composed of 3 depth separable convolutional layers. The number of characteristic layers is 32, and the convolution kernel size is 3x 3. Meanwhile, because the designed targets of the student model and the teacher model are consistent, the input and the output of the student model are also consistent, and a structure of a direct connection edge is adopted for well learning residual errors of the student model.
The middle layer output of the teacher model and the middle layer output of the student model are subjected to the same mapping, and the mean square error between the middle layer output of the teacher model and the middle layer output of the student model is calculated. The loss function is expressed as equations (2) and (3). Calculating an attention diagram of the system by using a linear kernel function k (x, y) ═ xTy to approximate the result of the hubert space mapping at the regenerating kernel. The calculation formula is as follows:
here, FT,FSRepresenting the attention diagrams of the teacher and student models respectively,andthe ith feature map, CTAnd CSThe number of feature layers of the teacher model and the student model are respectively represented. For practical applications, p is a positive integer. Thus, an initialized student model is obtained.
Then, the standard mean square error L shown in the formula (4) is usedSAnd finally training the student model to obtain the trained student model. Corresponding toRepresenting the output of the student model.
After such initialization, lightweight models may tend to perform better than without the initialization method. After the student model is obtained, the student model is led back to the video encoder, and therefore loop filtering based on the neural network in video encoding can be achieved. The training and testing penalty in the experiment is shown in fig. 3, and it is evident that the penalty function decreases faster after initialization is used.
Claims (4)
1. A neural network weight initialization algorithm based on transfer learning is characterized in that for a specified target task, a neural network model with high complexity, namely a teacher model, is designed, the teacher model is trained, and after the training is finished, a generated characteristic diagram is used for guiding the weight initialization of a student model; calculating the difference of the characteristic maps in a regeneration core Hilbert space by calculating the difference between the characteristic maps or mapping the characteristic maps to the regeneration core Hilbert space, and simplifying the calculation by adopting a kernel function method; the simple student model achieves a better weight initialization effect, and after the weight initialization is completed, the student model is generally trained, so that the student model achieves a better global convergence point and has more excellent performance; the method comprises the following specific steps:
(1) for a specific learning task, a conventional loss function and a model structure are possessed, firstly, a teacher model is designed aiming at a target task, and the teacher model is trained by using the conventional loss function;
(2) then exporting the middle layer output of the trained teacher model, and obtaining a characteristic diagram in a mapping mode; wherein, the mapping mode has attention migration, or the mapping mode uses kernel function to map to the regeneration kernel Hilbert space;
(3) designing a student model with a simpler structure, wherein the student model and the teacher model are required to have the same network structure, namely the basic network layer forming the network is consistent; when the network structures all adopt the series connection network formed by the convolution layers, the teacher model has more convolution layers and more feature graphs, and the student model has fewer convolution layers and fewer feature graphs;
(4) training the student model by taking the mean square error between the feature map calculated in the step (2) and the feature map obtained by mapping in the same way as the student model as a loss function; after training is finished, the obtained weight of the student model is adjusted by learning knowledge from the teacher model, so that the weight of the student model is specifically initialized, and the student model has the capability of approaching the performance of the teacher model;
(5) after initialization is completed, the student model is trained by using a conventional loss function, and a usable student model is obtained.
2. The neural network weight initialization algorithm based on transfer learning of claim 1, characterized in that a deep separable convolution and batch normalization is adopted as a main layer of a teacher model, the number of characteristic layers is 64, and the convolution kernel size is 3x 3; wherein 24 depth-separable convolutional layers are used as the backbone of the teacher model, and are divided into three parts, the first part is composed of 10 depth-separable convolutional layers, the second part is composed of 8 depth-separable convolutional layers, and the third part is composed of 6 depth-separable convolutional layers; the last layer of the model is a common convolution layer, the number of the characteristic layers is 1, and the size of a convolution kernel is 1x 1; all depth separable convolutions have a ReLU activation function; the input of the model is connected to the final output through a straight edge, so that the neural network is in a residual learning state, and the convergence is faster.
3. The neural network weight initialization algorithm based on transfer learning of claim 2, wherein the input and output of the model are respectively a reconstructed pixel map and a filtered pixel map of a video encoder, and the loss function L isTSelecting the output of a neural networkAnd original pixelMean square error between:
after the teacher model training is finished, the middle layer outputs of three sub-parts of the teacher model are obtained from the trained data set, and the output results F of the teacher model at the positions are calculated from the neural network attention map or the mapping result of the regenerative nuclear Hilbert spaceT(ii) a Intermediate layer calculation result F using teacher modelTForming a new data set with the data set input by the teacher model to train the student model;
the student model adopts a network structure similar to a teacher model, uses 9 depth separable convolutional layers as a backbone of the student model, and divides the student model into three parts, wherein each part is composed of 3 depth separable convolutional layers, the number of the characteristic layers is 32, and the size of a convolutional kernel is 3x 3; the designed targets of the student model and the teacher model are consistent, the input and the output of the student model are also consistent, and a structure of a direct connection edge is also adopted for better learning residual errors of the student model;
the middle layer output of the teacher model and the middle layer output of the student model are subjected to the same mapping, and the mean square error between the middle layer output of the teacher model and the middle layer output of the student model is calculated; the loss function is expressed as formulas (2) and (3); calculating an attention diagram of the system by using a linear kernel function k (x, y) ═ xTy to approximate the result of the hilbert space mapping at the regenerating kernel; the calculation formula is as follows:
here, FT,FSRepresenting the attention diagrams of the teacher and student models respectively,andthe ith feature map, CTAnd CSRespectively representing the number of characteristic layers of a teacher model and a student model, and taking a positive integer as p; thus, an initialized student model is obtained.
4. The neural network weight initialization algorithm based on transfer learning of claim 3, wherein a standard mean square error L as shown in formula (4) is usedSCarrying out final training on the initialized student model to obtain a trained student model; corresponding toRepresenting the output of the student model;
and leading the student model back to a video encoder, namely realizing the loop filtering based on the neural network in the video encoding.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201911321102.2A CN111126599B (en) | 2019-12-20 | 2019-12-20 | Neural network weight initialization method based on transfer learning |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201911321102.2A CN111126599B (en) | 2019-12-20 | 2019-12-20 | Neural network weight initialization method based on transfer learning |
Publications (2)
Publication Number | Publication Date |
---|---|
CN111126599A true CN111126599A (en) | 2020-05-08 |
CN111126599B CN111126599B (en) | 2023-09-05 |
Family
ID=70500352
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201911321102.2A Active CN111126599B (en) | 2019-12-20 | 2019-12-20 | Neural network weight initialization method based on transfer learning |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111126599B (en) |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111554268A (en) * | 2020-07-13 | 2020-08-18 | 腾讯科技(深圳)有限公司 | Language identification method based on language model, text classification method and device |
CN112464959A (en) * | 2020-12-12 | 2021-03-09 | 中南民族大学 | Plant phenotype detection system and method based on attention and multiple knowledge migration |
CN112929663A (en) * | 2021-04-08 | 2021-06-08 | 中国科学技术大学 | Knowledge distillation-based image compression quality enhancement method |
CN113469977A (en) * | 2021-07-06 | 2021-10-01 | 浙江霖研精密科技有限公司 | Flaw detection device and method based on distillation learning mechanism and storage medium |
WO2022001805A1 (en) * | 2020-06-30 | 2022-01-06 | 华为技术有限公司 | Neural network distillation method and device |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2017167347A (en) * | 2016-03-16 | 2017-09-21 | 日本電信電話株式会社 | Acoustic signal analysis device, method, and program |
CN109087303A (en) * | 2018-08-15 | 2018-12-25 | 中山大学 | The frame of semantic segmentation modelling effect is promoted based on transfer learning |
CN110163110A (en) * | 2019-04-23 | 2019-08-23 | 中电科大数据研究院有限公司 | A kind of pedestrian's recognition methods again merged based on transfer learning and depth characteristic |
US20190287515A1 (en) * | 2018-03-16 | 2019-09-19 | Microsoft Technology Licensing, Llc | Adversarial Teacher-Student Learning for Unsupervised Domain Adaptation |
-
2019
- 2019-12-20 CN CN201911321102.2A patent/CN111126599B/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2017167347A (en) * | 2016-03-16 | 2017-09-21 | 日本電信電話株式会社 | Acoustic signal analysis device, method, and program |
US20190287515A1 (en) * | 2018-03-16 | 2019-09-19 | Microsoft Technology Licensing, Llc | Adversarial Teacher-Student Learning for Unsupervised Domain Adaptation |
CN109087303A (en) * | 2018-08-15 | 2018-12-25 | 中山大学 | The frame of semantic segmentation modelling effect is promoted based on transfer learning |
CN110163110A (en) * | 2019-04-23 | 2019-08-23 | 中电科大数据研究院有限公司 | A kind of pedestrian's recognition methods again merged based on transfer learning and depth characteristic |
Non-Patent Citations (1)
Title |
---|
张振宇;: "稳健的多支持向量机自适应提升算法", 大连交通大学学报 * |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2022001805A1 (en) * | 2020-06-30 | 2022-01-06 | 华为技术有限公司 | Neural network distillation method and device |
EP4163831A4 (en) * | 2020-06-30 | 2023-12-06 | Huawei Technologies Co., Ltd. | Neural network distillation method and device |
CN111554268A (en) * | 2020-07-13 | 2020-08-18 | 腾讯科技(深圳)有限公司 | Language identification method based on language model, text classification method and device |
CN111554268B (en) * | 2020-07-13 | 2020-11-03 | 腾讯科技(深圳)有限公司 | Language identification method based on language model, text classification method and device |
CN112464959A (en) * | 2020-12-12 | 2021-03-09 | 中南民族大学 | Plant phenotype detection system and method based on attention and multiple knowledge migration |
CN112464959B (en) * | 2020-12-12 | 2023-12-19 | 中南民族大学 | Plant phenotype detection system and method based on attention and multiple knowledge migration |
CN112929663A (en) * | 2021-04-08 | 2021-06-08 | 中国科学技术大学 | Knowledge distillation-based image compression quality enhancement method |
CN113469977A (en) * | 2021-07-06 | 2021-10-01 | 浙江霖研精密科技有限公司 | Flaw detection device and method based on distillation learning mechanism and storage medium |
CN113469977B (en) * | 2021-07-06 | 2024-01-12 | 浙江霖研精密科技有限公司 | Flaw detection device, method and storage medium based on distillation learning mechanism |
Also Published As
Publication number | Publication date |
---|---|
CN111126599B (en) | 2023-09-05 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN111126599A (en) | Neural network weight initialization method based on transfer learning | |
CN113159173B (en) | Convolutional neural network model compression method combining pruning and knowledge distillation | |
CN111784602A (en) | Method for generating countermeasure network for image restoration | |
CN113313644B (en) | Underwater image enhancement method based on residual double-attention network | |
CN112489164B (en) | Image coloring method based on improved depth separable convolutional neural network | |
CN109005398B (en) | Stereo image parallax matching method based on convolutional neural network | |
CN113595993B (en) | Vehicle-mounted sensing equipment joint learning method for model structure optimization under edge calculation | |
CN113947680A (en) | Image semantic segmentation method based on cascade multi-scale vision Transformer | |
CN112017116B (en) | Image super-resolution reconstruction network based on asymmetric convolution and construction method thereof | |
CN113163203A (en) | Deep learning feature compression and decompression method, system and terminal | |
CN115330620A (en) | Image defogging method based on cyclic generation countermeasure network | |
CN111861886A (en) | Image super-resolution reconstruction method based on multi-scale feedback network | |
CN112651360A (en) | Skeleton action recognition method under small sample | |
CN116958534A (en) | Image processing method, training method of image processing model and related device | |
CN113989283B (en) | 3D human body posture estimation method and device, electronic equipment and storage medium | |
CN108629374A (en) | A kind of unsupervised multi-modal Subspace clustering method based on convolutional neural networks | |
CN109448039B (en) | Monocular vision depth estimation method based on deep convolutional neural network | |
CN109063834B (en) | Neural network pruning method based on convolution characteristic response graph | |
CN110752894A (en) | CNN-based LDPC code blind channel decoding method and decoder | |
CN113938254A (en) | Attention mechanism-based layered source-channel joint coding transmission system and transmission method thereof | |
CN110223224A (en) | A kind of Image Super-resolution realization algorithm based on information filtering network | |
CN113807497A (en) | Non-paired image translation method for enhancing texture details | |
CN116030537B (en) | Three-dimensional human body posture estimation method based on multi-branch attention-seeking convolution | |
CN117078539A (en) | CNN-transducer-based local global interactive image restoration method | |
CN116600119A (en) | Video encoding method, video decoding method, video encoding device, video decoding device, computer equipment and storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |