CN113077525A - Image classification method based on frequency domain contrast learning - Google Patents
Image classification method based on frequency domain contrast learning Download PDFInfo
- Publication number
- CN113077525A CN113077525A CN202110164693.8A CN202110164693A CN113077525A CN 113077525 A CN113077525 A CN 113077525A CN 202110164693 A CN202110164693 A CN 202110164693A CN 113077525 A CN113077525 A CN 113077525A
- Authority
- CN
- China
- Prior art keywords
- image
- frequency domain
- learning
- training
- network
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/90—Determination of colour characteristics
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/10—Complex mathematical operations
- G06F17/14—Fourier, Walsh or analogous domain transformations, e.g. Laplace, Hilbert, Karhunen-Loeve, transforms
- G06F17/141—Discrete Fourier transforms
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/10—Complex mathematical operations
- G06F17/16—Matrix or vector computation, e.g. matrix-matrix or matrix-vector multiplication, matrix factorization
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/11—Region-based segmentation
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Mathematical Physics (AREA)
- Data Mining & Analysis (AREA)
- Pure & Applied Mathematics (AREA)
- Computational Mathematics (AREA)
- Mathematical Analysis (AREA)
- Mathematical Optimization (AREA)
- General Engineering & Computer Science (AREA)
- Software Systems (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Databases & Information Systems (AREA)
- Evolutionary Computation (AREA)
- Algebra (AREA)
- Computing Systems (AREA)
- Artificial Intelligence (AREA)
- Life Sciences & Earth Sciences (AREA)
- Evolutionary Biology (AREA)
- Molecular Biology (AREA)
- General Health & Medical Sciences (AREA)
- Computational Linguistics (AREA)
- Biophysics (AREA)
- Biomedical Technology (AREA)
- Health & Medical Sciences (AREA)
- Bioinformatics & Computational Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Discrete Mathematics (AREA)
- Image Analysis (AREA)
- Image Processing (AREA)
Abstract
An image classification method based on frequency domain contrast learning comprises the following steps: s1: respectively carrying out two times of random data enhancement on the training set images, and carrying out two times of random data enhancement on the same image to obtain two different enhanced images; s2: performing discrete cosine transform on the image subjected to data enhancement to obtain an image transferred to a frequency domain; s3: the image transferred to the frequency domain passes through a depth network, and network parameters are learned through a comparison learning task to obtain the stable characteristics of the image; s5: and classifying the images in the test set by using the depth network with optimized parameters. The method can learn the stable characteristics of the same object under different backgrounds, and can better classify the non-independent images with the same distribution compared with the traditional classification method.
Description
Technical Field
The invention relates to the field of image classification, in particular to an image classification method based on frequency domain contrast learning.
Background
In recent years, with the continuous development of deep learning techniques, computers have achieved considerably high precision in various image classification tasks. However, the effectiveness of the mainstream image classification method is established on the premise that the training data set and the test set satisfy independent and same distribution. In practical application, training data and test data are often difficult to strictly satisfy the assumption of independent and same distribution, and in such a case, a model trained by a traditional method can perform well on a training data set, but an ideal classification effect cannot be achieved on the test data.
The classification methods proposed for the non-independent images with the same distribution are not many, and the currently proposed method is mainly based on a causal inference theory. The method inputs images into a depth model to extract features, takes each dimension of the features as an intervention variable in turn, and takes the features of other dimensions as confounding factors. The model is made to learn a set of sample weights, minimizing the association between each dimensional feature, thereby independently estimating the causal relationship of each dimensional feature to the classification result. However, there is not at all a complete lack of association between each dimension of an image feature, and different dimensions of features have different causal relationships to image classes. The method treats each dimension feature equally, and the classification effect is still not ideal enough.
The non-independent and identical distribution of the training data and the test data is mainly caused by different context information (including the background of the classification target, the texture of the classification target, the action of the classification target and the like) in the image, but the classification target has some characteristics which are not changed under different scenes. The stable characteristics of the learning target can effectively solve the classification problem of the non-independent same-distribution image. However, the diversity of the target stabilization features and the abstraction of the high-dimensional features extracted by the neural network pose challenges to the learning of the stabilization features.
Disclosure of Invention
Objects of the invention
In order to solve the technical problems in the background art, the invention provides an image classification method based on frequency domain contrast learning.
(II) technical scheme
In order to solve the above problems, the present invention provides an image classification method based on frequency domain contrast learning, which comprises the following steps:
s1: respectively carrying out random data enhancement twice on the training set image, wherein the data enhancement operation comprises cutting and size adjustment, horizontal turning, Gaussian blur, color dithering and gray level image conversion;
whether each data enhancement operation is executed is determined by the probability set in advance so as to carry out two times of random data enhancement on the same image to obtain two different enhanced images;
s2: performing discrete cosine transform on the image subjected to data enhancement to obtain an image transferred to a frequency domain;
wherein, the image is represented by RGB color table codes; extracting the frequency domain features of the image can be divided into the following two sub-steps:
s201: converting an image from an RGB color space to a YCbCr color space according to the following formula:
202: the picture converted to YCbCr color space is divided into 8 x 8 small blocks, and three channels are obtained for each block according to the formula F-AfATSolving a corresponding discrete cosine transform coefficient; the transformation matrix formula is as follows:
the original image is divided into 14 small blocks of 8 × 8, each small block contains 64 pixel points, and each pixel point has pixel values of 3 color channels, that is, 192 frequency domain coefficients can be obtained for each image small block, so that image frequency domain coefficients with the dimensions of (192, 14, 14) are obtained;
s3: the image transferred to the frequency domain passes through a depth network, and network parameters are learned through a comparison learning task to obtain the stable characteristics of the image; the method comprises the following substeps:
s301: respectively inputting the image frequency domain coefficients into a feature extraction layer of the depth network to obtain a feature h with the dimension of (N, 2048)iAnd hjThe network structure adopts a residual error neural network;
s302: h is to beiAnd hjInputting into a multi-layer perceptron to obtain an (N, 128) -dimensional feature z for comparison learning trainingiAnd zj;
S303: will ziAnd zjAnd (3) splicing according to the 0 th dimension to obtain the characteristics for calculating the comparative learning loss:
the loss of the comparative learning pre-training is calculated according to the following formula:
wherein σ is a positive number;
s304: performing parameter adjustment on the depth network in return by minimizing the loss of the contrast learning pre-training, and performing global parameter adjustment by adopting a back propagation algorithm until the loss of the contrast learning pre-training does not decrease any more, then converging the model, and ending the contrast learning pre-training step;
s4: predicting the classification result of the training set image by using the extracted features, further learning network parameters, and performing classification tasks;
wherein the 2018-dimensional features extracted in S301 are input into the full link layer and the softmax layerIn the classifier, a predicted classification result is obtained:n is the number of images in a batch, K is the number of categories of images;
and then calculating a cross entropy loss function by using the classification result:
and finally, carrying out global parameter adjustment through a back propagation algorithm, and optimizing the network parameters by taking the minimized cross entropy loss function as a target until the function value is not reduced any more.
S5: and classifying the images in the test set by using the depth network with optimized parameters:
performing random data enhancement on the test image according to S1, and converting the test image into a frequency domain according to S2; and finally, inputting the image converted into the frequency domain into a feature extraction layer of the depth network, and inputting the image into a classifier consisting of a full connection layer and a softmax layer to obtain a prediction result.
Preferably, in S2, the frequency domain image is input into a convolutional neural network to extract features.
Preferably, in S3, the model learns stable features through pre-training of the contrast learning task in combination with the frequency domain learning and contrast learning framework.
In the invention, images of each category are classified secondarily according to context information, and then relevant data sets are divided: training set: the image classification method comprises an image and a classification label corresponding to the image. And (3) test set: the image classification method comprises an image and a classification label corresponding to the image. But the context information of the images in the test set is different from that in the training set.
In the invention, random data enhancement is performed twice on each batch of pictures, and then the pictures are converted into the frequency domain to obtain two characteristics of the same image, and the training model is used for distinguishing whether the two characteristics come from the same image, so that the stable characteristics of the image are learned, and the classification effect of the non-independent images in the same distribution is improved.
The method can learn the stable characteristics of the same object under different backgrounds, and can better classify the non-independent images with the same distribution compared with the traditional classification method.
Drawings
FIG. 1 is a flowchart of a frequency domain contrast learning-based non-independent image classification method according to the present invention.
Fig. 2 is a model structure diagram of frequency domain contrast learning in the image classification method based on frequency domain contrast learning according to the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, the present invention will be described in further detail with reference to the accompanying drawings in conjunction with the following detailed description. It should be understood that the description is intended to be exemplary only, and is not intended to limit the scope of the present invention. Moreover, in the following description, descriptions of well-known structures and techniques are omitted so as to not unnecessarily obscure the concepts of the present invention.
As shown in fig. 1-2, the image classification method based on frequency domain contrast learning provided by the present invention includes the following steps:
s1: for training set image [ x1,x2,x3,……xN]Respectively carrying out random data enhancement twice, wherein the data enhancement operation comprises cutting and size adjustment, horizontal turning, Gaussian blur, color dithering and gray level image conversion;
whether each data enhancement operation is executed is determined by the probability set in advance so as to carry out two times of random data enhancement on the same image to obtain two different enhanced images;
s2: performing discrete cosine transform on the image subjected to data enhancement to obtain an image transferred to a frequency domain, namely:
[xi1,xi2,xi3,……xiN]and [ x ]j1,xj2,xj3,……xjN];
Wherein, the image is represented by RGB color table codes;
the enhanced image dimension is (3, 112, 112), where 3 represents that the image has 3 color channels, which are R, G respectively, and B, 112 is the image size set in advance;
extracting the frequency domain features of the image can be divided into the following two sub-steps:
s201: converting an image from an RGB color space to a YCbCr color space according to the following formula:
202: the picture converted to YCbCr color space is divided into 8 x 8 small blocks, and three channels are obtained for each block according to the formula F-AfATSolving a corresponding discrete cosine transform coefficient; the transformation matrix formula is as follows:
the original picture (3, 112, 112) can be divided into 14 8 × 8 small blocks, each small block contains 64 pixel points, and each pixel point has a pixel value of 3 color channels, that is, 192 frequency domain coefficients can be obtained for each image small block, so as to obtain an image frequency domain coefficient with a dimension of (192, 14, 14):
[x’i1,x’i2,x’i3,……x’iN]and [ x'j1,x’j2,x’j3,……x’jN];
S3: the image transferred to the frequency domain passes through a depth network, and network parameters are learned through a comparison learning task to obtain the stable characteristics of the image; the method comprises the following substeps:
s301: image frequency domain coefficient [ x'i1,x’i2,x’i3,……x’iN]And [ x'j1,x’j2,x’j3,……x’jN]Respectively inputting the feature extraction layers of the depth network to obtain a feature h with dimensions of (N, 2048)iAnd hjThe network structure thereof adoptsUsing a residual neural network;
s302: h is to beiAnd hjInputting into a multi-layer perceptron to obtain an (N, 128) -dimensional feature z for comparison learning trainingiAnd zj;
S303: will ziAnd zjAnd (3) splicing according to the 0 th dimension to obtain the characteristics for calculating the comparative learning loss:
the loss of the comparative learning pre-training is calculated according to the following formula:
wherein σ is a positive number;
s304: performing parameter adjustment on the depth network in return by minimizing the loss of the contrast learning pre-training, and performing global parameter adjustment by adopting a back propagation algorithm until the loss of the contrast learning pre-training does not decrease any more, then converging the model, and ending the contrast learning pre-training step;
s4: predicting the classification result of the training set image by using the extracted features, further learning network parameters, and performing classification tasks;
inputting the 2018 dimensional features extracted in the S301 into a classifier consisting of a full-link layer and a softmax layer to obtain a predicted classification result:n is the number of images in a batchMesh, K is the number of categories of the image;
and then calculating a cross entropy loss function by using the classification result:
and finally, carrying out global parameter adjustment through a back propagation algorithm, and optimizing the network parameters by taking the minimized cross entropy loss function as a target until the function value is not reduced any more.
S5: and classifying the images in the test set by using the depth network with optimized parameters:
performing random data enhancement on the test image according to S1, and converting the test image into a frequency domain according to S2; and finally, inputting the image converted into the frequency domain into a feature extraction layer of the depth network, and inputting the image into a classifier consisting of a full connection layer and a softmax layer to obtain a prediction result.
In the invention, images of each category are classified secondarily according to context information, and then relevant data sets are divided: training set: the image classification method comprises an image and a classification label corresponding to the image. And (3) test set: the image classification method comprises an image and a classification label corresponding to the image. But the context information of the images in the test set is different from that in the training set.
In the invention, random data enhancement is performed twice on each batch of pictures, and then the pictures are converted into the frequency domain to obtain two characteristics of the same image, and the training model is used for distinguishing whether the two characteristics come from the same image, so that the stable characteristics of the image are learned, and the classification effect of the non-independent images in the same distribution is improved.
In an alternative embodiment, in S2, the frequency domain image is input into a convolutional neural network to extract features.
In an alternative embodiment, in S3, the model is made to learn stable features through pre-training of the contrast learning task in conjunction with the frequency domain learning and contrast learning framework.
In conclusion, the method and the device can learn the stable characteristics of the same object under different backgrounds, and can better classify the non-independent images with the same distribution compared with the traditional classification method.
It is to be understood that the above-described embodiments of the present invention are merely illustrative of or explaining the principles of the invention and are not to be construed as limiting the invention. Therefore, any modification, equivalent replacement, improvement and the like made without departing from the spirit and scope of the present invention should be included in the protection scope of the present invention. Further, it is intended that the appended claims cover all such variations and modifications as fall within the scope and boundaries of the appended claims or the equivalents of such scope and boundaries.
Claims (3)
1. An image classification method based on frequency domain contrast learning is characterized by comprising the following steps:
s1: respectively carrying out random data enhancement twice on the training set image, wherein the data enhancement operation comprises cutting and size adjustment, horizontal turning, Gaussian blur, color dithering and gray level image conversion;
whether each data enhancement operation is executed is determined by the probability set in advance so as to carry out two times of random data enhancement on the same image to obtain two different enhanced images;
s2: performing discrete cosine transform on the image subjected to data enhancement to obtain an image transferred to a frequency domain;
wherein, the image is represented by RGB color table codes; extracting the frequency domain features of the image can be divided into the following two sub-steps:
s201: converting an image from an RGB color space to a YCbCr color space according to the following formula:
202: the picture converted to YCbCr color space is divided into 8 x 8 small blocks, and three channels are obtained for each block according to the formula F-AfATSolving a corresponding discrete cosine transform coefficient; the transformation matrix formula is as follows:
the original image is divided into 14 small blocks of 8 × 8, each small block contains 64 pixel points, and each pixel point has pixel values of 3 color channels, that is, 192 frequency domain coefficients can be obtained for each image small block, so that image frequency domain coefficients with the dimensions of (192, 14, 14) are obtained;
s3: the image transferred to the frequency domain passes through a depth network, and network parameters are learned through a comparison learning task to obtain the stable characteristics of the image; the method comprises the following substeps:
s301: respectively inputting the image frequency domain coefficients into a feature extraction layer of the depth network to obtain a feature h with the dimension of (N, 2048)iAnd hjThe network structure adopts a residual error neural network;
s302: h is to beiAnd hjInputting into a multi-layer perceptron to obtain an (N, 128) -dimensional feature z for comparison learning trainingiAnd zj;
S303: will ziAnd zjAnd (3) splicing according to the 0 th dimension to obtain the characteristics for calculating the comparative learning loss:
the loss of the comparative learning pre-training is calculated according to the following formula:
wherein σ is a positive number;
s304: performing parameter adjustment on the depth network in return by minimizing the loss of the contrast learning pre-training, and performing global parameter adjustment by adopting a back propagation algorithm until the loss of the contrast learning pre-training does not decrease any more, then converging the model, and ending the contrast learning pre-training step;
s4: predicting the classification result of the training set image by using the extracted features, further learning network parameters, and performing classification tasks;
inputting the 2018 dimensional features extracted in the S301 into a classifier consisting of a full-link layer and a softmax layer to obtain a predicted classification result:n is the number of images in a batch, K is the number of categories of images;
and then calculating a cross entropy loss function by using the classification result:
finally, global parameter adjustment is carried out through a back propagation algorithm, and a network parameter is optimized by taking a minimized cross entropy loss function as a target until the function value is not reduced any more;
s5: and classifying the images in the test set by using the depth network with optimized parameters:
performing random data enhancement on the test image according to S1, and converting the test image into a frequency domain according to S2; and finally, inputting the image converted into the frequency domain into a feature extraction layer of the depth network, and inputting the image into a classifier consisting of a full connection layer and a softmax layer to obtain a prediction result.
2. The image classification method based on frequency-domain contrast learning of claim 1, wherein in S2, the frequency-domain image is input into a convolutional neural network to extract features.
3. The image classification method based on frequency-domain contrast learning of claim 1, wherein in S3, in combination with the frequency-domain learning and contrast learning framework, the model is made to learn stable features through pre-training of the contrast learning task.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110164693.8A CN113077525A (en) | 2021-02-06 | 2021-02-06 | Image classification method based on frequency domain contrast learning |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110164693.8A CN113077525A (en) | 2021-02-06 | 2021-02-06 | Image classification method based on frequency domain contrast learning |
Publications (1)
Publication Number | Publication Date |
---|---|
CN113077525A true CN113077525A (en) | 2021-07-06 |
Family
ID=76609296
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110164693.8A Pending CN113077525A (en) | 2021-02-06 | 2021-02-06 | Image classification method based on frequency domain contrast learning |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN113077525A (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114357221A (en) * | 2022-03-15 | 2022-04-15 | 南京航空航天大学 | Self-supervision active learning method based on image classification |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20020186765A1 (en) * | 2001-06-05 | 2002-12-12 | Morley Steven A. | Selective chrominance decimation for digital images |
CN108009493A (en) * | 2017-11-30 | 2018-05-08 | 电子科技大学 | Face anti-fraud recognition methods based on action enhancing |
CN111127360A (en) * | 2019-12-20 | 2020-05-08 | 东南大学 | Gray level image transfer learning method based on automatic encoder |
CN111784633A (en) * | 2020-05-26 | 2020-10-16 | 西安理工大学 | Insulator defect automatic detection algorithm for power inspection video |
-
2021
- 2021-02-06 CN CN202110164693.8A patent/CN113077525A/en active Pending
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20020186765A1 (en) * | 2001-06-05 | 2002-12-12 | Morley Steven A. | Selective chrominance decimation for digital images |
CN108009493A (en) * | 2017-11-30 | 2018-05-08 | 电子科技大学 | Face anti-fraud recognition methods based on action enhancing |
CN111127360A (en) * | 2019-12-20 | 2020-05-08 | 东南大学 | Gray level image transfer learning method based on automatic encoder |
CN111784633A (en) * | 2020-05-26 | 2020-10-16 | 西安理工大学 | Insulator defect automatic detection algorithm for power inspection video |
Non-Patent Citations (2)
Title |
---|
HUAN SHAO,AND ETC: "Contrastive Learning in Frequency Domain for Non-I.I.D.Image Classification", 《SPRINGER LINK:INTERNATIONAL CONFERENCE ON MULTIMEDIA MODELING》 * |
张晶主编: "《多媒体信息与通信》", 30 November 2017, 西安:西安电子科技大学出版社 * |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114357221A (en) * | 2022-03-15 | 2022-04-15 | 南京航空航天大学 | Self-supervision active learning method based on image classification |
CN114357221B (en) * | 2022-03-15 | 2022-08-05 | 南京航空航天大学 | Self-supervision active learning method based on image classification |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN109949317B (en) | Semi-supervised image example segmentation method based on gradual confrontation learning | |
CN111046962B (en) | Sparse attention-based feature visualization method and system for convolutional neural network model | |
JP7381942B2 (en) | Control method, information processing device and control program | |
CN110717953B (en) | Coloring method and system for black-and-white pictures based on CNN-LSTM (computer-aided three-dimensional network-link) combination model | |
CN111986125A (en) | Method for multi-target task instance segmentation | |
CN111986126B (en) | Multi-target detection method based on improved VGG16 network | |
CN110929099B (en) | Short video frame semantic extraction method and system based on multi-task learning | |
CN111126115A (en) | Violence sorting behavior identification method and device | |
CN117523295B (en) | Passive domain adaptive image classification method based on class guide element learning | |
JP2009140369A (en) | Group learning device and group learning method, object detection device and object detection method, and computer program | |
CN116452862A (en) | Image classification method based on domain generalization learning | |
CN116563410A (en) | Electrical equipment electric spark image generation method based on two-stage generation countermeasure network | |
CN111583259A (en) | Document image quality evaluation method | |
CN113077525A (en) | Image classification method based on frequency domain contrast learning | |
CN117078656A (en) | Novel unsupervised image quality assessment method based on multi-mode prompt learning | |
CN112016592A (en) | Domain adaptive semantic segmentation method and device based on cross domain category perception | |
CN113807194B (en) | Enhanced power transmission line fault image recognition method | |
CN116129417A (en) | Digital instrument reading detection method based on low-quality image | |
CN115512207A (en) | Single-stage target detection method based on multipath feature fusion and high-order loss sensing sampling | |
CN115713464A (en) | Attention text super-resolution method based on text perception loss | |
CN115862015A (en) | Training method and device of character recognition system, and character recognition method and device | |
Yuan et al. | RM-IQA: A new no-reference image quality assessment framework based on range mapping method | |
CN112598043A (en) | Cooperative significance detection method based on weak supervised learning | |
Xu et al. | Drhnet: a deep residual network based on heterogeneous kernel for steganalysis | |
CN115063732B (en) | Action video classification method and system based on double-flow multi-resolution comprehensive modeling |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20210706 |