WO2022160980A1 - 一种超分辨率方法、装置、终端设备及存储介质 - Google Patents
一种超分辨率方法、装置、终端设备及存储介质 Download PDFInfo
- Publication number
- WO2022160980A1 WO2022160980A1 PCT/CN2021/137582 CN2021137582W WO2022160980A1 WO 2022160980 A1 WO2022160980 A1 WO 2022160980A1 CN 2021137582 W CN2021137582 W CN 2021137582W WO 2022160980 A1 WO2022160980 A1 WO 2022160980A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- resolution
- image
- super
- sub
- low
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 81
- 238000003860 storage Methods 0.000 title claims abstract description 11
- 238000012545 processing Methods 0.000 claims abstract description 75
- 238000013145 classification model Methods 0.000 claims abstract description 71
- 230000008569 process Effects 0.000 claims abstract description 28
- 238000012549 training Methods 0.000 claims description 68
- 230000006870 function Effects 0.000 claims description 67
- 238000004590 computer program Methods 0.000 claims description 22
- 230000009467 reduction Effects 0.000 claims description 6
- 238000005520 cutting process Methods 0.000 claims description 4
- 238000010923 batch production Methods 0.000 claims description 2
- 238000013135 deep learning Methods 0.000 abstract description 5
- 230000000694 effects Effects 0.000 description 26
- 238000004364 calculation method Methods 0.000 description 21
- 238000010586 diagram Methods 0.000 description 18
- 238000012360 testing method Methods 0.000 description 13
- 238000011084 recovery Methods 0.000 description 5
- 230000004044 response Effects 0.000 description 5
- 238000004422 calculation algorithm Methods 0.000 description 4
- 238000013507 mapping Methods 0.000 description 4
- 238000013528 artificial neural network Methods 0.000 description 3
- 230000008878 coupling Effects 0.000 description 3
- 238000010168 coupling process Methods 0.000 description 3
- 238000005859 coupling reaction Methods 0.000 description 3
- 238000004891 communication Methods 0.000 description 2
- 238000013527 convolutional neural network Methods 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 208000037170 Delayed Emergence from Anesthesia Diseases 0.000 description 1
- 210000004556 brain Anatomy 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 230000007423 decrease Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 238000011156 evaluation Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000003062 neural network model Methods 0.000 description 1
- 238000010606 normalization Methods 0.000 description 1
- 238000011176 pooling Methods 0.000 description 1
- 230000008521 reorganization Effects 0.000 description 1
- 238000005070 sampling Methods 0.000 description 1
- 238000012795 verification Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/40—Scaling of whole images or parts thereof, e.g. expanding or contracting
- G06T3/4053—Scaling of whole images or parts thereof, e.g. expanding or contracting based on super-resolution, i.e. the output image resolution being higher than the sensor resolution
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
- G06F18/2415—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on parametric or probabilistic models, e.g. based on likelihood ratio or false acceptance rate versus a false rejection rate
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/084—Backpropagation, e.g. using gradient descent
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/40—Scaling of whole images or parts thereof, e.g. expanding or contracting
- G06T3/4038—Image mosaicing, e.g. composing plane images from plane sub-images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
Definitions
- the present application relates to the technical field of deep learning, and in particular, to a super-resolution method, apparatus, terminal device and storage medium.
- Super-resolution technology refers to the technology of reconstructing low-resolution images into high-resolution images.
- the super-resolution algorithm based on deep learning is the most commonly used super-resolution method at present.
- the super-resolution algorithm based on deep learning is to cut the low-resolution image into sub-images, and then input each sub-image into the super-resolution network model for processing to obtain a reconstructed image, and then stitch the reconstructed images of each sub-image. Get high-resolution images.
- the more commonly used super-resolution network models include Accelerating the Super-Resolution Convolutional Neural Network (FSRCNN), fast, accurate, lightweight super-resolution and cascade residual network (Fast, Accurate , and Lightweight Super-Resolution with Cascading Residual Network, CARN), Photo-Realistic Single Image Super-Resolution Using a Generative Adversarial Network, SRResNet, Image Super-Resolution Use very deep residual channel attention network (Image Super-Resolution Using Very Deep Residual Channel Attention Networks, RCAN) and so on.
- FSRCNN Super-Resolution Convolutional Neural Network
- Fast Accurate
- CARN Lightweight Super-Resolution with Cascading Residual Network
- CARN Lightweight Super-Resolution with Cascading Residual Network
- SRResNet Photo-Realistic Single Image Super-Resolution Using a Generative Adversarial Network
- SRResNet Image Super-Resolution Use very deep
- the present application provides a super-resolution method, apparatus, terminal device, and storage medium, which can reduce the amount of computation for super-resolution processing.
- the present application provides a super-resolution method, comprising: inputting a low-resolution image to be processed into a trained classification super-resolution network model for processing, and outputting a high-resolution image corresponding to the low-resolution image;
- the classification super-resolution network model includes a classification model and multiple super-resolution network models with different complexities.
- the processing process of the classification super-resolution network model for low-resolution images includes:
- the method further includes: using a preset first loss function, a second loss function, a third loss function and a training set to train a preset initial network model to obtain a classification super-score network model.
- the initial classification model includes an initial classification model and multiple initial super-resolution network models with different complexities
- the training set includes multiple low-resolution image samples and high-resolution image samples corresponding to each low-resolution image sample
- the first loss function is used to reduce the error between the high-resolution image corresponding to the low-resolution image sample output by the initial classification model and the high-resolution image sample corresponding to the low-resolution image sample in the training set
- the second loss The function is used to increase the difference between the maximum probability value and other probability values among the multiple probability values output by the initial classification model
- the gap in the number of sub-image samples.
- the initial network model processing the low-resolution image samples in the training set includes:
- the sub-image samples are respectively input into multiple initial super-resolution network models for processing, and the first reconstructed image samples respectively output by the multiple initial super-resolution network models are obtained; the classification results are used to classify the multiple first reconstructed image samples
- a weighted summation is performed to obtain a second reconstructed image sample; the second reconstructed image samples of the plurality of sub-image samples are spliced to obtain a high-resolution image corresponding to the low-resolution image sample.
- the second loss function is:
- L c is the negative of the sum of the distances between the probability values belonging to each complexity category output by the sub-image sample x after being processed by the initial classification model
- M is the number of complexity categories
- P i (x) is the The probability value of the image sample x being assigned to the ith complexity class.
- the third loss function is:
- L a is the number of sub-image samples assigned to each complexity category by the initial classification model in the batch process and The sum of the distances between.
- B is the batch size
- P i (x j ) represents the probability that the j-th sub-image sample is assigned to the i-th complexity class in a batch
- the multiple super-resolution network models include a preset first super-resolution network model and at least one first super-resolution network model that has undergone network parameter reduction processing.
- the present application provides a super-resolution device, comprising:
- An acquisition unit for acquiring processed low-resolution images An acquisition unit for acquiring processed low-resolution images.
- the processing unit is used for inputting the low-resolution image into the trained classification super-segmentation network model for processing, and outputting the high-resolution image corresponding to the low-resolution image.
- the classification super-resolution network model includes a classification model and multiple super-resolution network models with different complexities.
- the processing process of the classification super-resolution network model for low-resolution images includes:
- the super-resolution device further includes a training unit:
- the training unit is used for training the preset initial network model by using the preset first loss function, the second loss function, the third loss function and the training set to obtain a classification super-score network model.
- the initial classification model includes an initial classification model and multiple initial super-resolution network models with different complexities
- the training set includes multiple low-resolution image samples and high-resolution image samples corresponding to each low-resolution image sample
- the first loss function is used to reduce the error between the high-resolution image corresponding to the low-resolution image sample output by the initial classification model and the high-resolution image sample corresponding to the low-resolution image sample in the training set
- the second loss The function is used to increase the difference between the maximum probability value and other probability values among the multiple probability values output by the initial classification model
- the gap in the number of sub-image samples.
- the present application provides a terminal device, including: a memory and a processor, where the memory is used for storing a computer program; the processor is used for executing the method described in any one of the above-mentioned first aspect when the computer program is invoked.
- the present application provides a computer-readable storage medium on which a computer program is stored, and when the computer program is executed by a processor, implements the method described in any one of the foregoing first aspects.
- an embodiment of the present application provides a computer program product that, when the computer program product runs on a processor, causes the processor to execute the method described in any one of the foregoing first aspects.
- the complexity of each sub-image of a low-resolution image is identified by using a classification model, and then super-resolution network models of different complexity are used to process different Subimage of complexity.
- the sub-image with relatively small complexity is processed by the super-resolution network model with relatively small complexity, so as to reduce the calculation amount of the sub-image with relatively small complexity under the condition of ensuring the restoration effect, Speed up processing.
- the sub-image with relatively large complexity is processed by the super-resolution network model with relatively large complexity, so as to ensure the restoration effect of the sub-image with relatively large complexity. Therefore, for a complete low-resolution image, the super-resolution method provided in this application can reduce the amount of calculation in the super-resolution processing and speed up the processing while ensuring the restoration effect of the high-resolution image. .
- FIG. 1 is a schematic flowchart of an embodiment of a super-resolution method provided by an embodiment of the present application
- FIG. 2 is a schematic flowchart of processing a low-resolution image by a classification super-resolution network model according to an embodiment of the present application
- FIG. 3 is a schematic diagram of a network structure of a classification model provided by an embodiment of the present application.
- FIG. 4 is a schematic diagram of the network structure of a plurality of FSRCNNs with different complexities provided by an embodiment of the present application;
- FIG. 5 is a schematic diagram of the network structure of multiple SRResNets of different complexity provided by an embodiment of the present application
- FIG. 6 is a schematic diagram 1 of experimental data comparison provided by an embodiment of the present application.
- FIG. 7 is a schematic diagram 2 of experimental data comparison provided by an embodiment of the present application.
- FIG. 8 is a schematic diagram of a training process of an initial network model provided by an embodiment of the present application.
- FIG. 9 is a schematic diagram three of experimental data comparison provided by an embodiment of the present application.
- FIG. 10 is a schematic diagram four of experimental data comparison provided by an embodiment of the application.
- FIG. 11 is a schematic diagram five of experimental data comparison provided by an embodiment of the application.
- FIG. 12 is a schematic structural diagram of a super-resolution device according to an embodiment of the application.
- FIG. 13 is a schematic structural diagram of a terminal device according to an embodiment of the application.
- the method of designing a lightweight network model or setting up efficient plug-in modules is usually adopted to reduce the amount of calculation.
- the calculation amount of the entire network model is reduced, and for a sub-image with greater complexity, it will inevitably lead to a poor recovery effect.
- the present application provides a super-resolution method, by designing a classification-super-resolution (Class Super-Resolution, Class SR) network model pair comprising a classification model and a plurality of super-resolution network models with different complexities
- Low-resolution images are subjected to super-resolution processing.
- the processing principle is to identify the complexity of each sub-image of a low-resolution image through a classification model, and then use a super-resolution network model of different complexity to process the sub-images of different complexity.
- the sub-image with relatively small complexity is processed by the super-resolution network model with relatively small complexity, so as to reduce the calculation amount of the sub-image with relatively small complexity under the condition of ensuring the restoration effect, Speed up processing.
- the sub-image with relatively large complexity is processed by the super-resolution network model with relatively large complexity, so as to ensure the restoration effect of the sub-image with relatively large complexity.
- the accelerated processing of super-resolution processing of low-resolution images is realized.
- the execution subject of the method may be an image processing device, such as a mobile terminal such as a smart phone, a smart phone, a tablet computer, a camera, etc., It can also be terminal devices such as desktop computers, robots, and servers.
- the trained classification super-score network model provided by this application is deployed in the image processing device.
- the low-resolution image can be input into the classification super-resolution network model for processing, and a high-resolution image corresponding to the low-resolution image can be output.
- the classification super-resolution network model provided in this application includes a classification model and a plurality of super-resolution network models with different complexities (in FIG. 1, three different complexities of small, medium and large are taken as examples) super-resolution network models.
- the processing process of the classification super-resolution network model for low-resolution images includes:
- the image processing device may cut the low-resolution image according to the preset size of the sub-image.
- the size of the sub-image can be set based on the input requirements of the classification model and the super-resolution network model used in the classification super-resolution network model.
- the classification model may be any neural network model with classification function.
- the classification model can be a convolutional neural network composed of several convolutional layers, pooling layers, and fully connected layers.
- the classification model is used to identify the complexity of the sub-image, which can classify the input sub-image and output the probability value of the sub-image being classified into each complexity category.
- the complexity class with the largest probability value is the complexity class of the sub-image.
- the recognition difficulty of different sub-images is different, and thus the difficulty of restoring to a high-resolution image is also different. Therefore, in this application, the so-called complexity of an image refers to the difficulty of reconstruction to high resolution.
- the output of the classification model is a vector of length M (M ⁇ 2, M is an integer), where M also represents the number of super-resolution network models in the classification super-resolution network model.
- M M is an integer
- M also represents the number of super-resolution network models in the classification super-resolution network model.
- the sub-image After the complexity category of the sub-image is determined according to the classification model, the sub-image can be input into the super-resolution network model corresponding to the complexity category of the sub-image for processing, and the reconstructed image of the sub-image (that is, the sub-image is outputted) high-resolution images).
- the sub-image is input into the super-resolution network model of "small complexity” for high-resolution restoration processing.
- the multiple super-resolution network models of different complexity may include different network models. For example, assuming that three super-resolution network models with different complexities need to be set in the classification super-resolution network model, three super-resolution network models can be selected from the existing and/or reconstructed super-resolution network models to build the classification super-resolution network model.
- FSRCNN in order of complexity of the network model from small to large, currently existing super-resolution network models include FSRCNN, CARN, SRResNet, RCAN and so on. If FSRCNN, CARN, and SRResNet are selected to build a classification super-resolution network model, FSRCNN is used as a super-resolution network model of "small” complexity, corresponding to the "small” complexity category; CARN is used as a super-resolution network model of "medium” complexity, Corresponds to the "medium” complexity category; SRResNet, as a super-resolution network model of "large” complexity, corresponds to the "large” complexity category.
- the multiple super-resolution network models with different complexities may also include a preset first super-resolution network model and at least one first super-resolution network model that has undergone network parameter reduction processing. .
- the first super-resolution network model may be any existing super-resolution network model or a reconstructed super-resolution network model. That is, in this embodiment of the present application, the original version and at least one simplified version of any super-resolution network model can be used to build a classification super-resolution network model.
- the first super-resolution network model is FSRCNN.
- the original version of the FSRCNN used is shown in (a) of Figure 4, and the original version includes convolutional layers a1, convolutional layers a2, 4-layer convolutional layers a3, convolutional layers a4 and 4 layer deconvolution layer.
- the convolutional layer a1 is used to extract the features of sub-images.
- the input channel (input channel) of the convolutional layer a1 is 3, the output channel (output channel) is 56, and the convolution kernel size (kernelsize) is 5.
- the convolutional layer a2 is used to perform dimension reduction processing on the feature map output by the convolutional layer a1, so as to reduce the calculation amount of the subsequent feature mapping process.
- the 4-layer continuous convolutional layer a3 is used for feature mapping, which maps low-resolution features to high-resolution features.
- the convolutional layer a4 is used to increase the dimension of the feature map output by the convolutional layer a3 to restore the dimension of the feature map.
- the 4-layer continuous deconvolution layer is used to perform an upsampling operation to obtain the reconstructed image of the sub-image.
- the original version can be simplified to different degrees according to the number of required simplified versions, that is, the network parameters of the FSRCNN can be deleted in different degrees, so as to obtain the obtained results. required simplified version.
- the complexity of the original version of FSRCNN is "large” by default, and two versions need to be simplified to obtain FSRCNN with complexity "small” and “medium”.
- the network structure of the “medium” complexity FSRCNN can be shown in (b) of FIG. 4 .
- the output channel of the convolutional layer a1 the input channel of the convolutional layer a2, the output channel of the convolutional layer a4, and the input channel of the deconvolutional layer are all reduced. Small is 36.
- the network structure of the "small" complexity FSRCNN can be shown in (c) of Fig. 3.
- the output channel of the convolutional layer a1 the input channel of the convolutional layer a2, the output channel of the convolutional layer a4, and the input channel of the deconvolutional layer are all reduced. Small is 16.
- the first super-resolution network model is SRResNet.
- the original version of FSRCNN is shown in (a) of Figure 5.
- the original version includes convolutional layer a1, convolutional layer a2, 4-layer convolutional layer a3, convolutional layer a4 and 4-layer deconvolution Laminate.
- the convolutional layer a1 is used to extract the features of sub-images.
- the input channel (input channel) of the convolutional layer a1 is 3, the output channel (output channel) is 56, and the convolution kernel size (kernelsize) is 5.
- the convolutional layer a2 is used to perform dimension reduction processing on the feature map output by the convolutional layer a1, so as to reduce the calculation amount of the subsequent feature mapping process.
- the 4-layer continuous convolutional layer a3 is used for feature mapping, which maps low-resolution features to high-resolution features.
- the convolutional layer a4 is used to increase the dimension of the feature map output by the convolutional layer a3 to restore the dimension of the feature map.
- the 4-layer continuous deconvolution layer is used to perform an upsampling operation to obtain the reconstructed image of the sub-image.
- the first super-resolution network model is SRResNet.
- the original version of the obtained SRResNet is shown in (a) in Figure 5.
- the original version includes convolutional layers b1, 16 residual layers, 2 convolutional layers b2, 2 pixel reorganization layers (pixel_shuffle), convolutional layers layer b3 and convolutional layer b4.
- the convolutional layer b1 and the residual layer are used to extract the features of the sub-image.
- the 2-layer convolutional layer b2 and the 2-layer pixel_shuffle are alternately arranged to map low-resolution features to high-resolution features.
- the convolutional layer b3 and the convolutional layer b4 are used to perform an up-sampling operation to obtain a reconstructed image of the sub-image.
- the complexity of the original version of SRResNet is "large” by default, and two versions need to be simplified to obtain SRResNet with "small” and “medium” complexity.
- the network structure of the "medium” complexity SRResNet can be shown in (b) of FIG. 5 .
- the output channel of the convolutional layer b1 the input channel and output channel of the residual layer, the input channel of the convolutional layer b2, the input channel of the convolutional layer b3 And the output channel, the input channel of the convolutional layer b4 are reduced to 48, and the output channel of the convolutional layer b2 is reduced to 48*4.
- the network structure of the "small" complexity SRResNet can be shown in (c) of Figure 5.
- the output channel of the convolutional layer b1 the input channel and output channel of the residual layer, the input channel of the convolutional layer b2, and the input channel of the convolutional layer b3
- the input channel of the output channel and the convolutional layer b4 are reduced to 32, and the output channel of the convolutional layer b2 is reduced to 32*4.
- the network parameters required to be calculated are reduced due to the reduction of the channel of the feature map in the network layer. Therefore, the amount of calculation in the process of processing the feature map is reduced, the processing speed is accelerated, and the corresponding complexity can be guaranteed.
- the restoration effect of sub-images of degrees that is to say, compared with using the original version of the single first super-resolution network model, using the original version of the first super-resolution network model and the simplified version of the original version to build the classification super-resolution network model can be used to a certain extent. Reduce the amount of calculation and speed up the processing speed. That is, the classification super-resolution network model provided in this application can be regarded as an accelerated version of the first super-resolution network model.
- step S203 After the reconstructed image of each sub-image is obtained, step S203 can be executed.
- a classification model is used to identify the complexity of each sub-image of a low-resolution image, and then a super-resolution network model of different complexity is used to process the sub-images of different complexity.
- the sub-image with relatively small complexity is processed by the super-resolution network model with relatively small complexity, so as to reduce the calculation amount of the sub-image with relatively small complexity under the condition of ensuring the restoration effect, Speed up processing.
- the sub-image with relatively large complexity is processed by the super-resolution network model with relatively large complexity, so as to ensure the restoration effect of the sub-image with relatively large complexity. Therefore, for a complete low-resolution image, using the classification super-resolution network model provided by the present application to perform super-resolution processing can ensure the restoration effect of the high-resolution image while reducing the amount of computation.
- the selected comparison group includes the original version of FSRCNN-O and the accelerated version of ClassSR-FSRCNN built with the network framework provided by this application, the original version of CARN-O and the accelerated version of ClassSR-CARN, and the original version of SRResNet-O and the accelerated version of ClassSR-SRResNet, the original version of RCAN-O and the accelerated version of ClassSR-RCAN.
- FIG. 6 is a statistical diagram of the obtained experimental data after the original version of each super-resolution network model and the accelerated version built by using the network framework provided by this application are tested on the 8K image test set.
- the ordinate is the peak signal to noise ratio (Peak Signal to Noise Ratio, PSNR) of the high-resolution image
- the unit is dB
- the abscissa is the calculation amount (FLOPs)
- the unit is M.
- the peak signal-to-noise ratio (PSNR) of the obtained high-resolution image can be guaranteed by using the accelerated version for super-resolution processing.
- PSNR peak signal-to-noise ratio
- Even on lightweight super-resolution network models e.g., FSRCNN-O and CARN-O
- the PSNR of high-resolution images obtained by super-resolution processing with the accelerated version is improved compared to the original version.
- the higher the PSNR the better the restoration effect of the network model on low-resolution images.
- the computational load of the accelerated versions of each super-resolution network model is reduced by nearly 50% (-50%, -47%, -48%, -50%, respectively). That is to say, the processing speed of the accelerated version is nearly doubled compared to the original version.
- Test/FLOPs represents the average PSNR (unit is dB) of the reconstructed high-resolution images after the corresponding network model performs super-resolution processing on 100 low-resolution images in the test set, and the average calculation amount (unit is M or G) ). It can be seen that after using the original version and the accelerated version to test on the same test set under different test conditions, the average PSNR of the high-resolution images output by the original version and the accelerated version are basically the same. That is to say, in the accelerated version, although some sub-images are processed by the simplified super-resolution network model, the restoration effect of the final restored high-resolution image is not significantly reduced.
- the calculation amount of the accelerated version for processing low-resolution images is significantly reduced, from 100% to 50% to 71%. It can be seen that under the condition that the restoration effect of high-resolution images is guaranteed, the processing speed of the accelerated version is greatly improved compared with the original version.
- Figure 7 is a schematic diagram of experimental data comparison of any two low-resolution image samples from the 2K image test set, the 4K image test set, and the 8K image test set. Among them, it includes the original version of each super-resolution network and the accelerated version of the reconstructed image sample obtained after super-resolution processing a sub-image sample, and also includes the reconstructed image sample (GT) corresponding to the sub-image sample in the test set. and high-resolution reconstructed image samples recovered using traditional bicubic interpolation.
- GT reconstructed image sample
- the classification super-resolution network model provided by the present application can speed up the processing speed while ensuring the restoration effect of the high-resolution image.
- the preset initial network model can be trained by using the preset first loss function, the second loss function, the third loss function and the training set to obtain a classification super-score network model .
- the initial network model refers to the classification super-score network model whose network parameters have not been optimized. It can be understood that the initial classification model includes an initial classification model and a plurality of initial super-resolution network models with different complexities.
- the training set includes multiple low-resolution image samples and a high-resolution image sample corresponding to each low-resolution image sample.
- the training set may include a 2K image training set, a 4K image training set, and/or an 8K image training set.
- the present application provides a training method.
- the network parameters of the initial classification model are optimized according to the recovery effect of the initial super-resolution network model on the sub-image samples, so that the trained classification model can accurately
- the images are assigned to the appropriate super-resolution network model.
- the initial network model's processing of the low-resolution image samples in the training set includes:
- each sub-image sample input the sub-image sample into the initial classification model for processing to obtain a classification result, where the classification result includes the probability value of the sub-image sample being classified into each complexity category; input the sub-image sample into multiple Perform processing in the initial super-resolution network model to obtain first reconstructed image samples respectively output by multiple initial super-resolution network models; use the classification result to perform weighted summation on multiple first reconstructed image samples to obtain second reconstructed image samples .
- the first loss function is used to calculate the high-resolution image corresponding to the low-resolution image sample output by the initial neural network and the low-resolution image in the training set.
- the error between the high-resolution image samples corresponding to the samples and then adjust the network parameters of multiple initial super-resolution network models and initial classification models according to the error values. Understandably, the smaller the error, the better the recovery effect. In this way, the recovery effect can be back-propagated to the initial classification module to adjust the network parameters.
- the first loss function is used to reduce the error between the high-resolution image corresponding to the low-resolution image sample output by the initial neural network and the high-resolution image sample corresponding to the low-resolution image sample in the training set.
- the first loss function may be a conventional L1 loss function.
- each probability value in the classification result output by the classification module is close in size, resulting in the classification being close to random classification.
- the present application also provides a second loss function for increasing the difference between the largest probability value and other probability values among the plurality of probability values output by the initial classification model during the training process. That is to say, when classifying a sub-image sample, the initial classification model is constrained by the second loss function to ensure that the probability of the sub-image sample being classified into the corresponding complexity category is as large as possible, and tends to be as close to 1 as possible.
- the second loss function may also be referred to as a classification-loss.
- the second loss function can be expressed by the following formula:
- L c is the negative number of the distance sum between the probability values belonging to each complexity category output by the same sub-image sample x after being processed by the initial classification model.
- M is the number of complexity classes
- P i (x) is the probability value of the sub-image sample x being assigned to the ith complexity class. This loss can widen the probability gap between different classification results, making the maximum probability value close to 1.
- the present application in order to ensure that each initial super-resolution network model can be fully trained, so as to ensure the training effect of each initial super-resolution network model, the present application also provides a third loss function, the third The loss function is used to reduce the gap in the number of sub-image samples belonging to multiple complexity classes determined by the initial classification model. That is, the initial classification model is constrained by the third loss function to assign roughly the same number of sub-image samples to each complexity class during training. This ensures that each initial super-resolution network model can be fully trained.
- the third loss function can be expressed by the following formula:
- L a is the number and average number of sub-image samples assigned to each complexity category by the initial classification model in batch processing distance between and.
- B is the batch size, which is the number of sub-image samples processed in a batch.
- P i (x j ) is the probability value of the jth sub-image sample being assigned to the ith complexity class in a batch.
- the third loss function may also be referred to as the average loss (Average-loss).
- FIG. 9 is a schematic diagram of a training curve for training a classification model using the first loss function, the second loss function and the third loss function at the same time.
- (a) in FIG. 9 shows the PSNR of the output high-resolution image samples of the initial classification super-score network model as a function of training time.
- (b) in Fig. 9 shows the variation curve of the calculation amount of the initial classification super-score network model with the training time.
- the PSNR of the initial classification super-score network model increases while the amount of computation decreases. It shows that each sub-image sample of each low-resolution image sample is gradually being assigned to a suitable super-resolution network model.
- Figure 10 is the training curve (the first PSNR curve and the first FLOPs curve) for training the classification model using the first loss function and the second loss function but not using the third loss function, and the classification model using the three loss functions simultaneously Schematic diagram of the comparison between the training curves (the second PSNR curve and the second FLOPs curve) for training.
- FIG. 10 shows the PSNR of the output high-resolution image samples of the initial classification super-score network model as a function of training time.
- (b) in Fig. 10 shows the change curve of the calculation amount of the initial classification super-score network model with the training time.
- Figure 11 is a training curve (the third PSNR curve and the third FLOPs curve) for training the classification model using the first loss function and the third loss function but not using the second loss function, and the classification model using the three loss functions at the same time.
- (a) in Figure 10 represents the PSNR variation curve of the output high-resolution image samples of the initial classification super-score network model with the training time.
- (b) in Fig. 11 shows the variation curve of the calculation amount of the initial classification super-score network model with the training time.
- the joint training method provided by this application in combination with the first loss function, the second loss function, and the third loss function can ensure that each super-resolution network model can be fully trained, and make the classification model based on The restoration effect is effectively optimized, and effective classification results are output. It is ensured that the classification super-score network model obtained by training can greatly improve the processing speed while ensuring the recovery effect.
- the network framework and training method provided in this application are universal. It can be applied to any image restoration task or tasks with image restoration effect as the evaluation index. For example, in addition to super-resolution tasks, it can also be applied to image denoising tasks. It can also greatly reduce the amount of calculation while ensuring PSNR.
- the embodiment of the present application provides an image-driven brain atlas construction device, and the device embodiment corresponds to the foregoing method embodiment.
- the details in the foregoing method embodiments are described one by one, but it should be clear that the apparatus in this embodiment can correspondingly implement all the content in the foregoing method embodiments.
- FIG. 12 is a schematic structural diagram of a super-resolution apparatus provided by an embodiment of the present application. As shown in FIG. 12 , the super-resolution apparatus provided by this embodiment includes an acquisition unit 1201 and a processing unit 1202 .
- the acquiring unit 1201 is used for acquiring the processed low-resolution image.
- the processing unit 1202 is configured to input the low-resolution image into the trained classification super-resolution network model for processing, and output the high-resolution image corresponding to the low-resolution image.
- the super-resolution apparatus further includes a training unit 1203 for training the preset initial network model by using the preset first loss function, second loss function, third loss function and training set to obtain a classification super-resolution method.
- a training unit 1203 for training the preset initial network model by using the preset first loss function, second loss function, third loss function and training set to obtain a classification super-resolution method.
- Sub-network model for training the preset initial network model by using the preset first loss function, second loss function, third loss function and training set to obtain a classification super-resolution method.
- the super-resolution apparatus provided in this embodiment can execute the above-mentioned method embodiments, and the implementation principle and technical effect thereof are similar, and details are not described herein again.
- FIG. 13 is a schematic structural diagram of a terminal device provided by an embodiment of the application.
- the terminal device provided by this embodiment includes: a memory 1301 and a processor 1302.
- the memory 1301 is used for storing computer programs; the processor 1302 is used for The methods described in the above method embodiments are executed when the computer program is invoked.
- the terminal device provided in this embodiment may execute the foregoing method embodiments, and the implementation principle and technical effect thereof are similar, and details are not described herein again.
- Embodiments of the present application further provide a computer-readable storage medium, on which a computer program is stored, and when the computer program is executed by a processor, the method described in the foregoing method embodiment is implemented.
- the embodiments of the present application further provide a computer program product, when the computer program product runs on a terminal device, the terminal device executes the method described in the above method embodiments.
- the above-mentioned integrated units are implemented in the form of software functional units and sold or used as independent products, they may be stored in a computer-readable storage medium.
- the present application realizes all or part of the processes in the methods of the above embodiments, which can be completed by instructing the relevant hardware through a computer program, and the computer program can be stored in a computer-readable storage medium.
- the steps of each of the above method embodiments can be implemented.
- the computer program includes computer program code
- the computer program code may be in the form of source code, object code, executable file or some intermediate form, and the like.
- the computer-readable storage medium may include at least: any entity or device capable of carrying the computer program code to the photographing device/terminal device, recording medium, computer memory, read-only memory (Read-Only Memory, ROM), random access Memory (Random Access Memory, RAM), electrical carrier signal, telecommunication signal and software distribution medium.
- computer readable media may not be electrical carrier signals and telecommunications signals.
- the disclosed apparatus/device and method may be implemented in other manners.
- the apparatus/equipment embodiments described above are only illustrative.
- the division of the modules or units is only a logical function division.
- the shown or discussed mutual coupling or direct coupling or communication connection may be through some interfaces, indirect coupling or communication connection of devices or units, and may be in electrical, mechanical or other forms.
- the term “if” may be contextually interpreted as “when” or “once” or “in response to determining” or “in response to detecting “.
- the phrases “if it is determined” or “if the [described condition or event] is detected” may be interpreted, depending on the context, to mean “once it is determined” or “in response to the determination” or “once the [described condition or event] is detected. ]” or “in response to detection of the [described condition or event]”.
- references in this specification to "one embodiment” or “some embodiments” and the like mean that a particular feature, structure or characteristic described in connection with the embodiment is included in one or more embodiments of the present application.
- appearances of the phrases “in one embodiment,” “in some embodiments,” “in other embodiments,” “in other embodiments,” etc. in various places in this specification are not necessarily All refer to the same embodiment, but mean “one or more but not all embodiments” unless specifically emphasized otherwise.
- the terms “including”, “including”, “having” and their variants mean “including but not limited to” unless specifically emphasized otherwise.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- General Physics & Mathematics (AREA)
- Evolutionary Computation (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- General Engineering & Computer Science (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Molecular Biology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Bioinformatics & Computational Biology (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- General Health & Medical Sciences (AREA)
- Evolutionary Biology (AREA)
- Computing Systems (AREA)
- Mathematical Physics (AREA)
- Probability & Statistics with Applications (AREA)
- Image Analysis (AREA)
Abstract
Description
Claims (10)
- 一种超分辨率方法,其特征在于,所述方法包括:将待处理的低分辨率图像输入已训练的分类超分网络模型中处理,输出得到与所述低分辨率图像对应的高分辨率图像;其中,所述分类超分网络模型包括分类模型和复杂度不同的多个超分辨网络模型,所述分类超分网络模型对所述低分辨率图像的处理过程包括:将所述低分辨率图像切割为多个子图像;针对每个子图像,根据所述分类模型确定所述子图像的复杂度类别,并将所述子图像输入到所述多个超分辨网络模型中与所述复杂度类别对应的超分辨网络模型中处理,输出得到所述子图像的重建图像;将所述多个子图像的重建图像进行拼接,得到所述与所述低分辨率图像对应的高分辨率图像。
- 如权利要求1所述的方法,其特征在于,所述方法还包括:利用预设的第一损失函数、第二损失函数、第三损失函数和训练集对预设的初始网络模型进行训练,得到所述分类超分网络模型;其中,所述初始分类模型包括初始分类模型和复杂度不同的多个初始超分辨网络模型,所述训练集包括多个低分辨率图像样本和分别于每个低分辨率图像样本对应的高分辨率图像样本;所述第一损失函数用于减小所述初始分类模型输出的与低分辨率图像样本对应的高分辨率图像,和所述训练集中与所述低分辨率图像样本对应的高分辨率图像样本之间的误差;所述第二损失函数用于增大所述初始分类模型输出的多个概率值中的最大概率值与其他概率值之间的差值;所述第三损失函数用于减小所述初始分类模型确定的分别属于多个复杂度类别的子图像样本的数量差距。
- 如权利要求2所述的方法,其特征在于,在训练过程中,所述初始网络模型对所述训练集中的低分辨率图像样本的处理过程包括:将所述低分辨率图像样本切割为多个子图像样本;针对每个子图像样本,将所述子图像样本输入所述初始分类模型中处理得到分类结果,所述分类结果包括所述子图像样本被归类到每个复杂度类别的概率值;将所述子图像样本分别输入到所述多个初始超分辨率网络模型中进行处理,得到所述多个初始超分辨率网络模型分别输出的第一重建图像样本;利用所述分类结果对多个所述第一重建图像样本进行加权求和,得到第二重建图像样本;将所述多个子图像样本的第二重建图像样本进行拼接,得到与所述低分辨率图像样本对应的高分辨率图像。
- 根据权利要求1-4任一项所述的方法,其特征在于,所述多个超分辨率网络模型包括预设的第一超分辨网络模型和至少一个经过网络参数删减处理的所述第一超分辨率网络模型。
- 一种超分辨率装置,其特征在于,包括:获取单元,用于获取处理的低分辨率图像;处理单元,用于将所述低分辨率图像输入已训练的分类超分网络模型中处理,输出得到与所述低分辨率图像对应的高分辨率图像;其中,所述分类超分网络模型包括分类模型和复杂度不同的多个超分辨网络模型,所述分类超分网络模型对所述低分辨率图像的处理过程包括:将所述低分辨率图像切割为多个子图像;针对每个子图像,根据所述分类模型确定所述子图像的复杂度类别,并将所述子图像输入到所述多个超分辨网络模型中与所述复杂度类别对应的超分辨网络模型中处理,输出得到所述子图像的重建图像;将所述多个子图像的重建图像进行拼接,得到所述与所述低分辨率图像对应的高分辨率图像。
- 如权利要求7所述的装置,其特征在于,所述装置还包括训练单元:所述训练单元,用于利用预设的第一损失函数、第二损失函数、第三损失函数和训练集对预设的初始网络模型进行训练,得到所述分类超分网络模型;其中,所述初始分类模型包括初始分类模型和复杂度不同的多个初始超分辨网络模型,所述训练集包括多个低分辨率图像样本和分别于每个低分辨率图像样本对应的高分辨率图像样本;所述第一损失函数用于减小所述初始分类模型输出的与低分辨率图像 样本对应的高分辨率图像,和所述训练集中与所述低分辨率图像样本对应的高分辨率图像样本之间的误差;所述第二损失函数用于增大所述初始分类模型输出的多个概率值中的最大概率值与其他概率值之间的差值;所述第三损失函数用于减小所述初始分类模型确定的分别属于多个复杂度类别的子图像样本的数量差距。
- 一种终端设备,其特征在于,包括:存储器和处理器,所述存储器用于存储计算机程序;所述处理器用于在调用所述计算机程序时执行如权利要求1-6任一项所述的方法。
- 一种计算机可读存储介质,其上存储有计算机程序,其特征在于,所述计算机程序被处理器执行时实现如权利要求1-6任一项所述的方法。
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110130561.3 | 2021-01-29 | ||
CN202110130561.3A CN112862681B (zh) | 2021-01-29 | 2021-01-29 | 一种超分辨率方法、装置、终端设备及存储介质 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2022160980A1 true WO2022160980A1 (zh) | 2022-08-04 |
Family
ID=75987330
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2021/137582 WO2022160980A1 (zh) | 2021-01-29 | 2021-12-13 | 一种超分辨率方法、装置、终端设备及存储介质 |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN112862681B (zh) |
WO (1) | WO2022160980A1 (zh) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115392216A (zh) * | 2022-10-27 | 2022-11-25 | 科大讯飞股份有限公司 | 一种虚拟形象生成方法、装置、电子设备及存储介质 |
CN116071238A (zh) * | 2023-03-06 | 2023-05-05 | 武汉人工智能研究院 | 图像超分处理方法、装置、电子设备及存储介质 |
Families Citing this family (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112862681B (zh) * | 2021-01-29 | 2023-04-14 | 中国科学院深圳先进技术研究院 | 一种超分辨率方法、装置、终端设备及存储介质 |
CN113421189A (zh) * | 2021-06-21 | 2021-09-21 | Oppo广东移动通信有限公司 | 图像超分辨率处理方法和装置、电子设备 |
CN113411521B (zh) * | 2021-06-23 | 2022-09-09 | 北京达佳互联信息技术有限公司 | 视频处理方法、装置、电子设备及存储介质 |
CN113313633A (zh) * | 2021-06-25 | 2021-08-27 | 西安紫光展锐科技有限公司 | 超分网络模型的训练方法、装置和电子设备 |
CN113596576A (zh) * | 2021-07-21 | 2021-11-02 | 杭州网易智企科技有限公司 | 一种视频超分辨率的方法及装置 |
CN113674152A (zh) * | 2021-08-03 | 2021-11-19 | Oppo广东移动通信有限公司 | 图像处理方法、装置、电子设备和计算机可读存储介质 |
CN113888411A (zh) * | 2021-09-29 | 2022-01-04 | 豪威科技(武汉)有限公司 | 分辨率提升方法及可读存储介质 |
CN113706390A (zh) * | 2021-10-29 | 2021-11-26 | 苏州浪潮智能科技有限公司 | 图像转换模型训练方法和图像转换方法、设备及介质 |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110136056A (zh) * | 2018-02-08 | 2019-08-16 | 华为技术有限公司 | 图像超分辨率重建的方法和装置 |
CN111598796A (zh) * | 2020-04-27 | 2020-08-28 | Oppo广东移动通信有限公司 | 图像处理方法及装置、电子设备、存储介质 |
CN111598779A (zh) * | 2020-05-14 | 2020-08-28 | Oppo广东移动通信有限公司 | 图像超分辨率处理方法和装置、电子设备及存储介质 |
US20210004935A1 (en) * | 2018-04-04 | 2021-01-07 | Huawei Technologies Co., Ltd. | Image Super-Resolution Method and Apparatus |
CN112862681A (zh) * | 2021-01-29 | 2021-05-28 | 中国科学院深圳先进技术研究院 | 一种超分辨率方法、装置、终端设备及存储介质 |
Family Cites Families (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130177242A1 (en) * | 2012-01-10 | 2013-07-11 | James E. Adams, Jr. | Super-resolution image using selected edge pixels |
US10009587B1 (en) * | 2017-08-14 | 2018-06-26 | Christie Digital Systems Usa, Inc. | Real-time spatial-based resolution enhancement using shifted superposition |
CN109886891B (zh) * | 2019-02-15 | 2022-01-11 | 北京市商汤科技开发有限公司 | 一种图像复原方法及装置、电子设备、存储介质 |
CN111062872B (zh) * | 2019-12-17 | 2021-02-05 | 暨南大学 | 一种基于边缘检测的图像超分辨率重建方法及系统 |
CN111080527B (zh) * | 2019-12-20 | 2023-12-05 | 北京金山云网络技术有限公司 | 一种图像超分辨率的方法、装置、电子设备及存储介质 |
CN111369440B (zh) * | 2020-03-03 | 2024-01-30 | 网易(杭州)网络有限公司 | 模型训练、图像超分辨处理方法、装置、终端及存储介质 |
-
2021
- 2021-01-29 CN CN202110130561.3A patent/CN112862681B/zh active Active
- 2021-12-13 WO PCT/CN2021/137582 patent/WO2022160980A1/zh active Application Filing
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110136056A (zh) * | 2018-02-08 | 2019-08-16 | 华为技术有限公司 | 图像超分辨率重建的方法和装置 |
US20210004935A1 (en) * | 2018-04-04 | 2021-01-07 | Huawei Technologies Co., Ltd. | Image Super-Resolution Method and Apparatus |
CN111598796A (zh) * | 2020-04-27 | 2020-08-28 | Oppo广东移动通信有限公司 | 图像处理方法及装置、电子设备、存储介质 |
CN111598779A (zh) * | 2020-05-14 | 2020-08-28 | Oppo广东移动通信有限公司 | 图像超分辨率处理方法和装置、电子设备及存储介质 |
CN112862681A (zh) * | 2021-01-29 | 2021-05-28 | 中国科学院深圳先进技术研究院 | 一种超分辨率方法、装置、终端设备及存储介质 |
Non-Patent Citations (1)
Title |
---|
KONG XIANGTAO; ZHAO HENGYUAN; QIAO YU; DONG CHAO: "ClassSR: A General Framework to Accelerate Super-Resolution Networks by Data Characteristic", 2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), IEEE, 20 June 2021 (2021-06-20), pages 12011 - 12020, XP034008596, DOI: 10.1109/CVPR46437.2021.01184 * |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115392216A (zh) * | 2022-10-27 | 2022-11-25 | 科大讯飞股份有限公司 | 一种虚拟形象生成方法、装置、电子设备及存储介质 |
CN115392216B (zh) * | 2022-10-27 | 2023-03-14 | 科大讯飞股份有限公司 | 一种虚拟形象生成方法、装置、电子设备及存储介质 |
CN116071238A (zh) * | 2023-03-06 | 2023-05-05 | 武汉人工智能研究院 | 图像超分处理方法、装置、电子设备及存储介质 |
Also Published As
Publication number | Publication date |
---|---|
CN112862681B (zh) | 2023-04-14 |
CN112862681A (zh) | 2021-05-28 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2022160980A1 (zh) | 一种超分辨率方法、装置、终端设备及存储介质 | |
CN109493350B (zh) | 人像分割方法及装置 | |
CN108921806B (zh) | 一种图像处理方法、图像处理装置及终端设备 | |
WO2021164234A1 (zh) | 图像处理方法以及图像处理装置 | |
WO2021104058A1 (zh) | 图像分割方法、装置及终端设备 | |
WO2022105638A1 (zh) | 图像退化处理方法、装置、存储介质及电子设备 | |
CN112602088B (zh) | 提高弱光图像的质量的方法、系统和计算机可读介质 | |
WO2021082819A1 (zh) | 一种图像生成方法、装置及电子设备 | |
WO2022082999A1 (zh) | 一种物体识别方法、装置、终端设备及存储介质 | |
CN112614110B (zh) | 评估图像质量的方法、装置及终端设备 | |
CN113066017A (zh) | 一种图像增强方法、模型训练方法及设备 | |
WO2023082453A1 (zh) | 一种图像处理方法及装置 | |
WO2023179095A1 (zh) | 一种图像分割方法、装置、终端设备及存储介质 | |
CN112308866A (zh) | 图像处理方法、装置、电子设备及存储介质 | |
CN113222855A (zh) | 一种图像恢复方法、装置和设备 | |
WO2021139351A1 (zh) | 图像分割方法、装置、介质及电子设备 | |
JP2024508867A (ja) | 画像クラスタリング方法、装置、コンピュータ機器及びコンピュータプログラム | |
CN110717864A (zh) | 一种图像增强方法、装置、终端设备及计算机可读介质 | |
CN113158773A (zh) | 一种活体检测模型的训练方法及训练装置 | |
CN110738625B (zh) | 图像重采样方法、装置、终端及计算机可读存储介质 | |
GB2587833A (en) | Image modification styles learned from a limited set of modified images | |
WO2020056688A1 (zh) | 提取图像关键点的方法及装置 | |
CN116823869A (zh) | 背景替换的方法和电子设备 | |
CN110619668B (zh) | 一种图像抽象方法、装置及终端设备 | |
CN114764839A (zh) | 动态视频生成方法、装置、可读存储介质及终端设备 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 21922570 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 21922570 Country of ref document: EP Kind code of ref document: A1 |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 21922570 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 16/01/2024) |