CN113822856A - End-to-end no-reference video quality evaluation method based on layered time-space domain feature representation - Google Patents

End-to-end no-reference video quality evaluation method based on layered time-space domain feature representation Download PDF

Info

Publication number
CN113822856A
CN113822856A CN202110945647.1A CN202110945647A CN113822856A CN 113822856 A CN113822856 A CN 113822856A CN 202110945647 A CN202110945647 A CN 202110945647A CN 113822856 A CN113822856 A CN 113822856A
Authority
CN
China
Prior art keywords
video
stage
quality
time
feature
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202110945647.1A
Other languages
Chinese (zh)
Inventor
杨峰
周明亮
沈文昊
咸伟志
江蔚
纪程
隋修宝
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nanjing Zhongke Inverse Entropy Technology Co ltd
Original Assignee
Nanjing Zhongke Inverse Entropy Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanjing Zhongke Inverse Entropy Technology Co ltd filed Critical Nanjing Zhongke Inverse Entropy Technology Co ltd
Priority to CN202110945647.1A priority Critical patent/CN113822856A/en
Publication of CN113822856A publication Critical patent/CN113822856A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30168Image quality inspection

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • General Engineering & Computer Science (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Biomedical Technology (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Quality & Reliability (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)

Abstract

The invention discloses an end-to-end non-reference video quality evaluation method based on layered time-space domain feature representation. Firstly, preprocessing a video: dividing an original video into non-overlapping time segments, and carrying out block cutting on the time segments, wherein the regions at the same position in each segment form a video block which is used as the input of a neural network; secondly, training the neural network: carrying out feature extraction on an input video clip, and outputting a series of space-time feature maps extracted in stages; then, inputting the characteristic diagrams of each stage to a convolutional neural network and a cyclic neural network to obtain stage quality characteristic vectors with the same dimensionality; and finally, respectively calculating the quality scores of all stages, and calculating the global quality score of the video sequence by combining the attention model. The invention utilizes the three-dimensional convolution layer to form a feature extractor, and the network can effectively extract space-time features so as to detect the degradation mode of the video.

Description

End-to-end no-reference video quality evaluation method based on layered time-space domain feature representation
Technical Field
The invention relates to the technical field of video quality evaluation in video coding, in particular to an end-to-end no-reference video quality evaluation method based on layered time-space domain feature representation.
Background
Over the past few years, the demand for video services has increased exponentially. Cisco predicts that in the next few years, video-generated data traffic will be 80-90% of the total network data traffic. With the development of communication technology, 2/3 mobile data is transmitted to various multimedia mobile devices to meet the needs of consumers. Such a flexible digital lifestyle requires consumers to enjoy high-quality multimedia contents at any time regardless of their location.
Digital Video is subject to various distortions during acquisition, processing, compression, storage and transmission, causing a visual quality degradation, and the purpose of Video Quality Assessment (VQA) is to predict the perceived quality of Video. A good quality evaluation method can not only automatically and accurately evaluate the quality of the video, but also monitor and guide parameter updating and optimization algorithms in real time, thereby better serving video transmission.
In a macroscopic view, the video quality evaluation methods are divided into three types: full Reference (FR), half Reference (RR), No Reference (No Reference, NR). The no-reference quality evaluation method of the video is a method for evaluating the objective quality of a distorted video when an original lossless video cannot be used, so that the research difficulty of the method is large: 1) the distortion of the single frame image is poor in average precision; 2) lack of perception of motion-induced spatial distortion; 3) the interaction between spatio-temporal artifacts is difficult to estimate.
Disclosure of Invention
In order to overcome the defects of the prior art, the invention aims to effectively fuse video time-space domain characteristics and establish a non-reference video quality evaluation method with excellent performance based on deep learning and hierarchical time-space domain characteristic representation. The evaluation method is more accurate and efficient, reasonably utilizes semantic information of the neural network middle layer, and can be jointly optimized as an end-to-end whole framework.
In order to achieve the purpose, the technical scheme of the invention is as follows: an end-to-end no-reference video quality evaluation method based on layered time-space domain feature representation comprises the following steps:
preprocessing a video: preprocessing a video: dividing an original video into non-overlapping time segments, and cutting the time segments into blocks, wherein the areas at the same position in each segment form a video block;
training a neural network to obtain a first network model with a video space-time feature extraction function, and performing staged feature extraction on an input video time segment by using the first network model;
training the convolutional neural network and the cyclic neural network to obtain a second network model with a video space-time feature fusion function, and inputting feature graphs of all stages to the second network model to obtain feature vectors of the same dimensionality of all stages;
and respectively calculating the quality scores of all stages by using a third network model which is obtained by training and has the function of calculating the video quality scores, and calculating the global quality scores of the video sequence by combining an attention model in the third network model.
Preferably, the first network model includes J three-dimensional convolutional layers, J generalized divisor normalization layers, and J maximum pooling layers, and each stage includes a three-dimensional convolutional layer, a generalized divisor normalization layer, and a maximum pooling layer connected in sequence; the first network model employs a linear rectification unit as an activation function.
Preferably, the second network model includes J branches, each branch includes a plurality of airspace convolution layers, a plurality of gated circulation units and a uniform pooling layer, the activation function adopts a linear rectification unit, and the features of each stage are respectively input into the branches to obtain a mass feature vector of each branch, where J is the total number of stages.
Preferably, the third network model comprises J fully-connected layers for stage mass regression, an attention model consisting of two fully-connected layers, a linear rectification unit and an S-shaped growth curve unit, and a fully-connected layer for global mass regression.
Compared with the prior art, the invention has the following remarkable advantages:
the method can effectively fuse the video time-space domain characteristics, reasonably utilize the semantic information of the neural network middle layer, and can be used as an end-to-end overall framework for common optimization, and the accuracy and reliability of the method are superior to those of other current video objective quality evaluation methods; the invention utilizes the three-dimensional convolution layer to form a feature extractor, and the network can effectively extract space-time features so as to detect the degradation mode of the video.
Drawings
FIG. 1 is a basic flow diagram of the present invention;
Detailed Description
The following describes the detailed implementation of the present invention with reference to the accompanying drawings.
The invention relates to an end-to-end no-reference video quality evaluation method represented by layered space-time characteristics, as shown in figure 1, comprising the following steps:
step 1: preprocessing a video: dividing an original video into non-overlapping time segments, and cutting the time segments into blocks, wherein the areas at the same position in each segment form a video block which is used as the input of a neural network;
specifically, in order to sufficiently extract temporal and spatial information of videos, each video is segmented into video segments with a time length of 8; meanwhile, each frame of the video is also uniformly and non-overlapping cut into image blocks. According to the difference of the resolution, each video can extract a plurality of video blocks with the time length of 8 and the space size of 256 multiplied by 3, which respectively represent the height, the width and the channel number of the video frame.
Step 2: training a neural network to obtain a first network model with a video space-time feature extraction function, and performing staged feature extraction on an input video time segment by using the first network model;
specifically, the first network model includes J three-dimensional convolution layers, four generalized divisor normalization layers, and four maximum pooling layers, and each stage includes a three-dimensional convolution layer, a generalized divisor normalization layer, and a maximum pooling layer connected in sequence; the first network model employs a linear rectification unit (ReLU) as an activation function. The resulting outputs of each stage are as follows:
Xj=CNNj(Xj-1),j∈[1,2,…,J],
wherein CNNjThree-dimensional convolutional layers representing the j stage, XjRepresents the output of the J stage, J being the total number of stages.
And step 3: training the convolutional neural network and the cyclic neural network to obtain a second network model with a video space-time feature fusion function, and inputting feature graphs of all stages to the second network model to obtain feature vectors of the same dimensionality of all stages;
specifically, each branch of the second network model includes a plurality of airspace convolution layers, a plurality of gating circulation units and a uniform pooling layer, and the activation function adopts a linear rectification unit to finally obtain a quality feature vector of each branch.
Specifically, the specific process of inputting the feature map of each stage into the second network model to obtain the feature vector of the same dimension of each stage is as follows:
step 31: performing spatial feature fusion by using a plurality of sequentially connected airspace convolution layers to obtain features with consistent dimensions:
Figure BDA0003214117020000031
wherein phij(. to) represents a series of spatial convolution layers with a kernel size of 3 x 3, zero padding and a step size of 2 x 2,
Figure BDA0003214117020000032
a characteristic diagram with the time slice serial number k in the j stage is shown,
Figure BDA0003214117020000033
and representing the characteristic vector with the time slice serial number k after the spatial domain information is fused in the j stage.
Step 32: given frame level features
Figure BDA0003214117020000034
Using a Global maximum pooling layer (denoted GP)max) To obtain efficient features and reduce spatial redundancy. Meanwhile, a Gated Round Unit (GRU) is used to refine the frame-level features by integrating the time information:
Figure BDA0003214117020000035
wherein
Figure BDA0003214117020000036
And representing the characteristic vector with the sequence number k of the time slice after the time domain information is fused in the jth stage.
Step 33: obtaining the feature vector of the stage after the spatio-temporal information fusion by using uniform pooling
Figure BDA0003214117020000037
Figure BDA0003214117020000041
Step 34: and repeating the steps 31-33 for each stage to obtain the quality feature vector with the same dimension for each stage.
And 4, step 4: and respectively calculating the quality scores of all stages by using a third network model which is obtained by training and has the function of calculating the video quality scores, and calculating the global quality scores of the video sequence by combining an attention model in the third network model.
Specifically, the third network model comprises four full-connection layers for stage mass regression, an attention model consisting of two full-connection layers, a linear rectification unit and an S-shaped growth curve unit, and a full-connection layer for global mass regression.
Further, the specific process of calculating the global quality score of the video sequence is as follows:
step 41: respectively inputting the quality characteristic vectors of all stages into a full-connection layer to obtain the quality scores of all stages of the video:
Figure BDA0003214117020000042
wherein FCj(. h) a fully connected layer representing the quality feature vector input for stage j, qjIs the mass fraction at this stage.
Step 42: since the learned model tends to overfit a specific scene in the training set, the attention model is used to obtain a corresponding weight vector, and a feature having a greater influence on the perceptual quality is obtained. The attention model consists of two full-connection layers, a linear rectifying unit and an S-shaped growth curve unit, and the calculation mode is as follows:
Figure BDA0003214117020000043
HW=Sigmoid(FCw2(ReLu(FCw1(H)))),
wherein the content of the first and second substances,
Figure BDA0003214117020000044
representing a connection operation, FCw1(. o) and FCw2Both represent fully connected layers, Sigmoid (. cndot.) and ReLu (. cndot.) represent linear rectification functions and S-shaped growth curve functions, respectively, H, HWRespectively representing the feature vector and the weight vector.
Step 43: inputting the global quality feature vector into a full-connection layer to obtain a global quality score:
Q=FC(H⊙HW).
wherein, the lines indicate that the parity elements are correspondingly multiplied.
The above-mentioned embodiment only represents one embodiment of the present invention, and the description is more specific and detailed, but the embodiment does not represent the limitation of the invention patent scope. It should be noted that, for the person skilled in the art, several variations and modifications are possible without departing from the inventive concept, and that such presently unforeseen alternatives or modifications to the present disclosure are intended to be within the scope of the present disclosure.

Claims (7)

1. An end-to-end no-reference video quality evaluation method based on layered time-space domain feature representation is characterized by comprising the following steps:
preprocessing a video: preprocessing a video: dividing an original video into non-overlapping time segments, and cutting the time segments into blocks, wherein the areas at the same position in each segment form a video block;
training a neural network to obtain a first network model with a video space-time feature extraction function, and performing staged feature extraction on an input video time segment by using the first network model;
training the convolutional neural network and the cyclic neural network to obtain a second network model with a video space-time feature fusion function, and inputting feature graphs of all stages to the second network model to obtain feature vectors of the same dimensionality of all stages;
and respectively calculating the quality scores of all stages by using a third network model which is obtained by training and has the function of calculating the video quality scores, and calculating the global quality scores of the video sequence by combining an attention model in the third network model.
2. The end-to-end no-reference video quality evaluation method based on the hierarchical time-space domain feature representation according to claim 1, wherein the first network model comprises J three-dimensional convolutional layers, J generalized divisor normalization layers and J maximum pooling layers, and each stage comprises the three-dimensional convolutional layers, the generalized divisor normalization layers and the maximum pooling layer which are connected in sequence; the first network model adopts a linear rectifying unit as an activation function, and J is the total number of stages.
3. The end-to-end no-reference video quality evaluation method based on the hierarchical time-space domain feature representation according to claim 1 or 2, wherein the output features of the first network model at each stage are as follows:
Xj=CNNj(Xj-1),j∈[1,2,…,J],
wherein CNNjThree-dimensional convolutional layers representing the j stage, XjRepresents the output of the J stage, J being the total number of stages.
4. The method according to claim 1, wherein the second network model comprises J branches, each branch comprises a plurality of space-domain convolutional layers, a plurality of gated cyclic units and a uniform pooling layer, the activation function adopts a linear rectification unit, and the features of each stage are respectively input into the branches to obtain the quality feature vector of each branch, wherein J is the total number of stages.
5. The end-to-end no-reference video quality evaluation method based on the hierarchical time-space domain feature representation according to claim 1 or 4, characterized in that the specific process of inputting the feature map of each stage to the second network model to obtain the feature vector of the same dimension of each stage is as follows:
step 31: performing spatial feature fusion by using a plurality of airspace convolution layers which are connected in sequence to obtain features with consistent dimensions:
Figure FDA0003214117010000011
wherein phij(·) Representing a plurality of spatial convolution layers,
Figure FDA0003214117010000012
a characteristic diagram with the time slice serial number k in the j stage is shown,
Figure FDA0003214117010000013
representing a characteristic vector with the time slice serial number k after the spatial domain information is fused in the jth stage;
step 32: given frame level features
Figure FDA0003214117010000014
Using global max pooling layer GPmaxTo obtain efficient features and reduce spatial redundancy while using gated cyclic units GRU to refine the frame-level features by integrating temporal information:
Figure FDA0003214117010000021
wherein
Figure FDA0003214117010000022
Representing a characteristic vector with the sequence number k of the time slice after time domain information fusion in the jth stage;
step 33: obtaining the feature vector of the stage after the spatio-temporal information fusion by using uniform pooling
Figure FDA0003214117010000023
Figure FDA0003214117010000024
Wherein K is the total number of the time segments;
step 34: and repeating the steps 31-33 for each stage feature to obtain the quality feature vector with the same dimension of each stage.
6. The method according to claim 1, wherein the third network model comprises J fully-connected layers for stage quality regression, an attention model consisting of two fully-connected layers, a linear rectification unit and an S-shaped growth curve unit, and a fully-connected layer for global quality regression.
7. The end-to-end reference-free video quality evaluation method based on the hierarchical time-space domain feature representation according to claim 1 or 6, characterized in that the specific process of calculating the global quality score of the video sequence is as follows:
step 41: respectively inputting the quality characteristic vectors of all stages into a full-connection layer to obtain the quality scores of all stages of the video:
Figure FDA0003214117010000025
wherein FCj(. h) a fully connected layer representing the quality feature vector input for stage j, qjIs the mass fraction of the stage;
step 42: inputting the mass fractions of all stages into an attention model to obtain corresponding weight vectors and obtain characteristics which have larger influence on the perception quality, wherein the calculation mode is as follows:
H=h1⊕h2⊕…⊕hj…⊕hJ,
HW=Sigmoid(FCw2(ReLu(FCw1(H)))),
wherein ≧ represents a connection operation, FCw1(. o) and FCw2(. cndot.) represents the fully connected layer, Sigmoid (. cndot.) and ReLu (. cndot.) represent the linear rectification function and the S-shaped growth curve function, respectively, H, HWRespectively representing the feature vector and the weight vector;
step 43: inputting the global quality feature vector into a full-connection layer to obtain a global quality score:
Q=FC(H⊙HW).
wherein, the lines indicate that the parity elements are correspondingly multiplied.
CN202110945647.1A 2021-08-16 2021-08-16 End-to-end no-reference video quality evaluation method based on layered time-space domain feature representation Pending CN113822856A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110945647.1A CN113822856A (en) 2021-08-16 2021-08-16 End-to-end no-reference video quality evaluation method based on layered time-space domain feature representation

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110945647.1A CN113822856A (en) 2021-08-16 2021-08-16 End-to-end no-reference video quality evaluation method based on layered time-space domain feature representation

Publications (1)

Publication Number Publication Date
CN113822856A true CN113822856A (en) 2021-12-21

Family

ID=78922891

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110945647.1A Pending CN113822856A (en) 2021-08-16 2021-08-16 End-to-end no-reference video quality evaluation method based on layered time-space domain feature representation

Country Status (1)

Country Link
CN (1) CN113822856A (en)

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100316131A1 (en) * 2009-06-12 2010-12-16 Motorola, Inc. Macroblock level no-reference objective quality estimation of video
CN106303507A (en) * 2015-06-05 2017-01-04 江苏惠纬讯信息科技有限公司 Video quality evaluation without reference method based on space-time united information
CN107959848A (en) * 2017-12-08 2018-04-24 天津大学 Universal no-reference video quality evaluation algorithms based on Three dimensional convolution neutral net
CN108235003A (en) * 2018-03-19 2018-06-29 天津大学 Three-dimensional video quality evaluation method based on 3D convolutional neural networks
CN110517237A (en) * 2019-08-20 2019-11-29 西安电子科技大学 No-reference video quality evaluating method based on expansion Three dimensional convolution neural network
CN110677639A (en) * 2019-09-30 2020-01-10 中国传媒大学 Non-reference video quality evaluation method based on feature fusion and recurrent neural network
CN112085102A (en) * 2020-09-10 2020-12-15 西安电子科技大学 No-reference video quality evaluation method based on three-dimensional space-time characteristic decomposition
CN112784698A (en) * 2020-12-31 2021-05-11 杭州电子科技大学 No-reference video quality evaluation method based on deep spatiotemporal information
CN112954312A (en) * 2021-02-07 2021-06-11 福州大学 No-reference video quality evaluation method fusing spatio-temporal characteristics
CN113255786A (en) * 2021-05-31 2021-08-13 西安电子科技大学 Video quality evaluation method based on electroencephalogram signals and target significant characteristics

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100316131A1 (en) * 2009-06-12 2010-12-16 Motorola, Inc. Macroblock level no-reference objective quality estimation of video
CN106303507A (en) * 2015-06-05 2017-01-04 江苏惠纬讯信息科技有限公司 Video quality evaluation without reference method based on space-time united information
CN107959848A (en) * 2017-12-08 2018-04-24 天津大学 Universal no-reference video quality evaluation algorithms based on Three dimensional convolution neutral net
CN108235003A (en) * 2018-03-19 2018-06-29 天津大学 Three-dimensional video quality evaluation method based on 3D convolutional neural networks
CN110517237A (en) * 2019-08-20 2019-11-29 西安电子科技大学 No-reference video quality evaluating method based on expansion Three dimensional convolution neural network
CN110677639A (en) * 2019-09-30 2020-01-10 中国传媒大学 Non-reference video quality evaluation method based on feature fusion and recurrent neural network
CN112085102A (en) * 2020-09-10 2020-12-15 西安电子科技大学 No-reference video quality evaluation method based on three-dimensional space-time characteristic decomposition
CN112784698A (en) * 2020-12-31 2021-05-11 杭州电子科技大学 No-reference video quality evaluation method based on deep spatiotemporal information
CN112954312A (en) * 2021-02-07 2021-06-11 福州大学 No-reference video quality evaluation method fusing spatio-temporal characteristics
CN113255786A (en) * 2021-05-31 2021-08-13 西安电子科技大学 Video quality evaluation method based on electroencephalogram signals and target significant characteristics

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
吴泽民;彭韬频;田畅;胡磊;王露萌;: "融合空时感知特性的无参考视频质量评估算法", 电子学报, no. 03 *
王春峰;苏荔;黄庆明;: "基于卷积神经网络的时空融合的无参考视频质量评价方法", 中国科学院大学学报, no. 04 *

Similar Documents

Publication Publication Date Title
CN107483920B (en) A kind of panoramic video appraisal procedure and system based on multi-layer quality factor
CN113240580A (en) Lightweight image super-resolution reconstruction method based on multi-dimensional knowledge distillation
CN108391121B (en) No-reference stereo image quality evaluation method based on deep neural network
CN108235003B (en) Three-dimensional video quality evaluation method based on 3D convolutional neural network
CN109831664B (en) Rapid compressed stereo video quality evaluation method based on deep learning
CN112954312A (en) No-reference video quality evaluation method fusing spatio-temporal characteristics
CN110674925B (en) No-reference VR video quality evaluation method based on 3D convolutional neural network
CN112291570B (en) Real-time video enhancement method based on lightweight deformable convolutional neural network
CN112507920B (en) Examination abnormal behavior identification method based on time displacement and attention mechanism
Bosse et al. Neural network-based full-reference image quality assessment
CN105046725B (en) Head shoulder images method for reconstructing in low-bit rate video call based on model and object
CN109859166A (en) It is a kind of based on multiple row convolutional neural networks without ginseng 3D rendering method for evaluating quality
CN109685772B (en) No-reference stereo image quality evaluation method based on registration distortion representation
CN116485741A (en) No-reference image quality evaluation method, system, electronic equipment and storage medium
CN105376563A (en) No-reference three-dimensional image quality evaluation method based on binocular fusion feature similarity
CN111160356A (en) Image segmentation and classification method and device
CN109523558A (en) A kind of portrait dividing method and system
CN110782458A (en) Object image 3D semantic prediction segmentation method of asymmetric coding network
CN114598864A (en) Full-reference ultrahigh-definition video quality objective evaluation method based on deep learning
CN115546589A (en) Image generation method based on graph neural network
CN113947538A (en) Multi-scale efficient convolution self-attention single image rain removing method
CN113822954A (en) Deep learning image coding method for man-machine cooperation scene under resource constraint
CN113362239A (en) Deep learning image restoration method based on feature interaction
CN112862675A (en) Video enhancement method and system for space-time super-resolution
CN113822856A (en) End-to-end no-reference video quality evaluation method based on layered time-space domain feature representation

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination