CN113537325B - Deep learning method for image classification based on extracted high-low layer feature logic - Google Patents
Deep learning method for image classification based on extracted high-low layer feature logic Download PDFInfo
- Publication number
- CN113537325B CN113537325B CN202110758190.3A CN202110758190A CN113537325B CN 113537325 B CN113537325 B CN 113537325B CN 202110758190 A CN202110758190 A CN 202110758190A CN 113537325 B CN113537325 B CN 113537325B
- Authority
- CN
- China
- Prior art keywords
- feature
- logic
- extraction unit
- layer
- feature extraction
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/084—Backpropagation, e.g. using gradient descent
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02D—CLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
- Y02D10/00—Energy efficient computing, e.g. low power processors, power management or thermal management
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Evolutionary Computation (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Software Systems (AREA)
- Mathematical Physics (AREA)
- Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- Computing Systems (AREA)
- Molecular Biology (AREA)
- General Health & Medical Sciences (AREA)
- Evolutionary Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses a deep learning method based on feature logic of an extraction high-low layer for image classification, which consists of a feature extraction network and a logic network, wherein the feature extraction network is formed by connecting a plurality of feature extraction units in series, the logic network is formed by connecting a plurality of logic extraction units in series, the feature extraction units are convolution layers or residual blocks, the logic extraction units are convolution dendritic modules formed by convolution layers and Hadamard products, the feature extraction network is single-input multi-output, is used for extracting image features of all layers from low to high, is input into images to be classified, is output into feature images of all layers, the logic network is multi-input single-output, is used for constructing a logic relation between the features of the high-low layer and classifying according to the logic relation, is input into feature images of all layers, and is output into classification results.
Description
Technical Field
The invention relates to the field of artificial intelligence, in particular to a deep learning method for classifying by extracting high-low layer feature logic applied to image signals.
Background
Classification is a fundamental problem in many fields such as fault diagnosis, automation, computer Vision (CV), and Natural Language Processing (NLP). Deep learning has been a useful tool to solve classification problems. Feature extraction has been a key to Convolutional Neural Networks (CNNs) mimicking human vision. The CNN performs feature extraction on the input image by convolution check. Due to the ability to translate, etc., and to stratify, CNNs can extract deeper image features for classification. Later, researchers improved deep CNNs with jump connections to yield residual neural networks res net. The model solves the problems of gradient disappearance, gradient dispersion and degradation of deep CNN, so that the neural network has stronger feature extraction capability. Biological dendrites in the brain have proven to have logical and/or/non-logical computing capabilities, enabling the calculation of logical relationships between input signals. An existing model that mimics biological dendritic functions is a dendritic network (DD) that can provide a measure for logical extraction, and as the number of layers increases, the DD can build more complex logical relationships. The convolution dendritic network (CDD) inherits the logic extraction capability of the DD and the feature extraction capability of the CNN, and makes up the defects that the DD cannot perform feature extraction and the CNN cannot perform logic extraction. CDD is therefore more concerned with the combination of features and relies on the logical relationships between features for classification. Many models in the field of target detection, such as Feature Pyramid Networks (FPN), densNet, YOLOv3, etc., improve the detection accuracy of small objects by fusing high-low layer features. In the image segmentation field, a large number of semantic structures are maintained by a high-level semantic segmentation result, but a small structure is seriously lost, a large number of details are maintained by a segmentation result of a low-level feature, but semantic category prediction is poor, so that a semantic segmentation model with better performance is obtained by fusing the high-level feature and the low-level feature and combining the advantages of the high-level feature and the low-level feature.
Many attempts have been made to emulate human visual perceptibility, but the following problems still remain. The traditional neural network model cannot do logical extraction between features. CDD inherits the advantages of DD and CNN, but the feature extraction layer of CDD is shallow, so that the feature extraction capability is weak. In addition, because the neural network has the hierarchical expression capability of the features, whether the high-level features or the low-level features contain a large amount of information of the images, the existing classification network only uses the high-level features to classify and can lose a large amount of information of the low-level features.
Disclosure of Invention
The invention aims to solve the technical problems that the traditional network cannot perform logic extraction between features, the feature extraction capability of a convolutional dendritic network (CDD) is weak, and the existing neural network cannot combine high-low-level features. The method is mainly applied to the field of image classification, can improve the classification precision of all traditional models which are used for classification only by extracting image features, and has the advantages of high classification precision, high convergence speed, good robustness and the like.
The technical scheme adopted for solving the technical problems is as follows: the deep learning method based on the feature logic of the high and low layers is characterized by comprising a feature extraction network and a logic network, wherein the feature extraction network is formed by connecting a plurality of feature extraction units in series, the logic network is formed by connecting a plurality of logic extraction units in series, the feature extraction units are convolution layers or residual blocks, the logic extraction units are convolution dendritic modules formed by convolution layers and Hadamard products, the connection relation among the units is that the input of the feature extraction unit of the layer is the output of the feature extraction unit of the upper layer, the input of the feature extraction unit of the layer is the output of the feature extraction unit of the upper layer and the output of the feature extraction unit of the layer, the feature extraction network is single-input multi-output, the function is to extract the image features of each layer from low to high, the input is the feature map of each layer, the logic network is multi-input single-output, the function is to construct the logic relation between the image features of the high and classify according to the logic relation, the input is the feature map of each layer, and the output is the classifying result, and the method specifically comprises the following steps:
step 1, an input image with an image tag is sent into a first layer convolution layer to perform dimension transformation and preliminary feature extraction to obtain a feature map 0;
in the method, in the process of the invention,is characteristic diagram 0, X in To input an image W first Weight matrix for first layer convolution layer, +.>Is a convolution;
step 2, sending the feature map 0 into a feature extraction unit 1 to obtain a feature map 1 after feature extraction;
in the method, in the process of the invention,is characteristic diagram 0, X F 1 Is characterized by FIG. 1, W F 1 A weight matrix W for the feature extraction unit 1 S 1 For the linear mapping weight matrix in the jump connection in the feature extraction unit 1, F (·) is a nonlinear activation function,/->Is a convolution;
step 3, simultaneously sending the feature map 0 and the feature map 1 to a logic extraction unit 1, and obtaining a feature logic map 1 after high-low layer feature logic combination;
in the method, in the process of the invention,logic diagram 1, for features>Is characteristic diagram 0, X F 1 For the features of FIG. 1, < >>Weight matrix for logic extraction unit 1, +.>Convolution, O is Hadamard product;
step 4, sending the feature map 1 into a feature extraction unit 2 to obtain a feature map 2 after feature extraction;
wherein X is F 1 As a feature of the figure 1 of the drawings,for the features of FIG. 2, < >>A weight matrix W for the feature extraction unit 2 S 2 For the linear mapping weight matrix in the jump connection in the feature extraction unit 2, F (·) is a nonlinear activation function,/->Is a convolution;
step 5, the feature logic diagram 1 and the feature diagram 2 are simultaneously sent to a logic extraction unit 2, and the feature logic diagram 2 is obtained after high-low-layer feature logic combination;
in the method, in the process of the invention,logic diagram 1, for features>Is characterized by logic diagram 2, X F 1 Is characterized by FIG. 1, X F 2 For the features of FIG. 2, < >>For the weight matrix of the logic extraction unit 2, +.>Convolution, O is Hadamard product;
and step 6, continuously extracting higher-level features by using a next-level feature extraction unit and logically combining the higher-level features and the lower-level features by using the next-level logic extraction unit until a highest-level feature logic diagram is obtained, wherein a recursive formula is expressed as follows:
where k is a layer flag (k=2, 3, …),is a characteristic diagram k-1,/or->For the feature map k->Is a characteristic logic diagram k-1,/o>Is a characteristic logic diagram k->Weight matrix for feature extraction unit k, W S k For the linear mapping weight matrix in the jump connection in the feature extraction unit k,/for the feature extraction unit k>For the weight matrix of the logic extraction unit 2, F (·) is a nonlinear activation function, ++>Is a convolution;
step 7, reforming the highest-level feature logic diagram into a one-dimensional tensor and sending the one-dimensional tensor into a classifier, and obtaining a classification result of the image by the classifier according to the highest-level feature logic;
wherein X is out In order to classify the result of the classification,is the feature logic diagram of the highest layer, W FC For the weight matrix of the full connection layer in the classifier, reshape (·) represents reforming variables into one-dimensional tensors;
step 8, comparing the classification result with the image label, calculating classification errors, and adjusting weights of the first-layer convolution layer, each feature extraction unit and the logic extraction unit by using an error back propagation algorithm to obtain an optimal image classification model after multiple iterations;
E=L(X lab ,X out )
wherein X is out For classification result, X lab For image labels, E is the classification error, W * For all weights in the model, W * ' is all weights of the model after adjustment of an error back propagation algorithm, L (·) is a loss function, and η is a learning rate;
and 9, classifying the input image without the image tag by using an optimal image classification model to obtain the category to which the input image belongs.
Further, expanding the formula in the step 3, so as to obtain a feature logic diagram 1 including all possible feature logic combinations between the feature diagram 0 and the feature diagram 1;
in the method, in the process of the invention,logic diagram 1, for features>Is characteristic diagram 0, X F 1 For the features of FIG. 1, < >>For logically extracting the weight of cell 1, f ij 0 Image features, f, of feature map 0 ij 1 For characterizing the image features of FIG. 1, AND logic f ij 0 ·f ij 1 Or logic f ij 0 +f ij 1 Non-logic-f ij * ,/>Is the weight of the logic extraction unit 1, where i, j is the matrix index (i=1, 2, …, n; j=1, 2, …, m), n is the length of the image, m is the width of the image, and represents any layer>For convolution, o is hadamard product, (padding=1, tap=1) indicates that one pixel is filled around the image and the convolution step is 1.
Further, by using the recurrence formula in step 6, feature logic combinations between all high-level and low-level features can be obtained, and then the weights before logic combination are adjusted by an error back propagation algorithm, so that the high-level and low-level feature logic relationship contributing to classification accuracy can be obtained.
The principle of the invention is as follows: traditional classification networks can extract more robust image features for classification, but ignore potential logical relationships between features. Meanwhile, the classification mode is only dependent on the high-level features, and a large number of detail features contained in the low-level features are ignored. Therefore, the invention combines the characteristic extraction module with better network robustness with the CDD module through a carefully designed architecture, extracts and combines the high-bottom image characteristics, and finally classifies the image characteristics by means of the logic relationship among the characteristics.
Compared with the prior art, the invention has the advantages that:
the advantages compared to conventional models (e.g., resNet, WRN, etc.) that are used for classification by extracting image features alone are: the invention considers the logic relation between the high-low-level characteristics, and obviously improves the classification precision and the convergence rate; the advantages compared to CDD networks that rely on logical relationships between features for classification are: the method has stronger feature extraction capability, and can extract higher-level image features, so that the model is more robust.
Drawings
Fig. 1 is a flow chart of the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, exemplary embodiments of the present invention will be described in further detail below with reference to the accompanying drawings of the specification. It should be understood that the specific embodiments described herein are for purposes of illustration only and are not intended to limit the scope of the invention. In addition, the technical features of the embodiments of the present invention described below may be combined with each other as long as they do not collide with each other.
In order to illustrate a deep learning method of extracting high-low level feature logic for classification according to the present invention, a complete image classification model is constructed by the present invention, and the image signals of two open source data sets Cifar10 and Cifar100 are classified and compared with two conventional models (res net, WRN) which rely on only high-level features for classification. And the model of the invention applied to ResNet is called ResNet-CDD, and the model applied to WRN is called WRN-CDD.
As shown in fig. 1, the present invention is composed of a feature extraction unit and a logic extraction unit. According to the size of the middle image of the Cifar10 and Cifar100 data, the input data is replaced by 3×32×32 tensors, the images are quantized and sent into a model, training is carried out on the model according to the steps 1 to 8, and the optimal classification model can be obtained after multiple iterations. The whole process of extracting the characteristics and obtaining the logic combination of the beneficial characteristics aims at reducing errors, and the iterative optimization is automatically carried out by taking an error back propagation algorithm as a method, so that manual participation is not needed. The model predictive classification process is steps 1 to 7. Because the weights in the model are already adjusted and optimized by a large amount of data in the training set, the error between the output of the image signal in the test set obtained by the model and the image label is also within the allowable range.
Tables 1,2, 3 and 4 show that in order to compare the present invention with two conventional models (ResNet, WRN) classified by only high-level features on two open source data sets, cifar10 and Cifar100, under the same framework, the model of the present invention applied to ResNet is called ResNet-CDD, and the model applied to WRN is called WRN-CDD.
From tables 1,2, 3 and 4, the following phenomena can be seen: on the Cifar10 data set with smaller classification difficulty, when the original model is smaller (the number of layers or the width or the parameter quantity is smaller, namely, the feature extraction capability is weaker), the precision of the original model is improved greatly, and when the model is gradually enlarged (namely, the feature extraction capability is gradually enhanced), the precision of the original model gradually reaches the precision level of the invention. On the Cifar100 data set with larger classification difficulty, the invention greatly improves the original model, and also presents a trend that the original model is improved from large to small as the original model is reduced from small to large (namely, the feature extraction capability of the original model is improved from large to strong).
TABLE 1 comparative experimental results of the invention (ResNet-CDD) with the prototype (ResNet) in the Cifar10 dataset
TABLE 2 comparative experimental results of the invention (ResNet-CDD) with the prototype (ResNet) in the Cifar100 dataset
TABLE 3 comparative experimental results of the invention (WRN-CDD) with the prototype model (WRN) in the Cifar10 dataset
TABLE 4 comparative experimental results of the invention (WRN-CDD) with the prototype model (WRN) in the Cifar100 dataset
First, the advantages of the invention, such as higher accuracy, faster convergence speed, and better robustness, can be seen from the foregoing. Secondly, the invention can greatly enhance the classification model with weaker feature extraction capability, which accords with the visual characteristics of human beings, and in practice, the human beings can realize high-precision classification only by extracting some low-level image features (such as outlines, textures, colors and the like) and then performing logical combination of brains, and the human beings are not limited to only logically combining the features of the same level, but combining all available high-level features. It should also be noted that if the identified object is relatively simple, only a few simple features are needed to distinguish, and that human vision does not need to be classified by logical combinations between features, which explains that the original model can achieve the same level of classification accuracy on the Cifar10 dataset as the present invention.
What is not described in detail in the present specification belongs to the prior art known to those skilled in the art.
The above examples are provided for the purpose of describing the present invention only and are not intended to limit the scope of the present invention. The scope of the invention is defined by the appended claims. Various equivalents and modifications that do not depart from the spirit and principles of the invention are intended to be included within the scope of the invention.
Claims (1)
1. The deep learning method based on the feature logic of the high and low layers is characterized by comprising a feature extraction network and a logic network, wherein the feature extraction network is formed by connecting a plurality of feature extraction units in series, the logic network is formed by connecting a plurality of logic extraction units in series, the feature extraction units are convolution layers or residual blocks, the logic extraction units are convolution dendritic modules formed by convolution layers and Hadamard products, the connection relation among the units is that the input of the feature extraction unit of the layer is the output of the feature extraction unit of the upper layer, the input of the feature extraction unit of the layer is the output of the feature extraction unit of the upper layer and the output of the feature extraction unit of the layer, the feature extraction network is single-input multi-output, the function is to extract the image features of each layer from low to high, the input is the feature map of each layer, the logic network is multi-input single-output, the function is to construct the logic relation between the image features of the high and classify according to the logic relation, the input is the feature map of each layer, and the output is the classifying result, and the method specifically comprises the following steps:
step 1, an input image with an image tag is sent into a first layer convolution layer to perform dimension transformation and preliminary feature extraction to obtain a feature map 0;
wherein X is F 0 Is characteristic diagram 0, X in To input an image W first Is the weight matrix of the first layer convolution layer,is a convolution;
step 2, sending the feature map 0 into a feature extraction unit 1 to obtain a feature map 1 after feature extraction;
wherein X is F 1 Is characterized by FIG. 1, W F 1 A weight matrix W for the feature extraction unit 1 S 1 For the linear mapping weight matrix in the jump connection in the feature extraction unit 1, F (·) is a nonlinear activation function,is a convolution;
step 3, simultaneously sending the feature map 0 and the feature map 1 to a logic extraction unit 1, and obtaining a feature map 1 after high-low layer feature logic combination, wherein the feature map 1 comprises all possible feature logic combinations between the feature map 0 and the feature map 1;
in the method, in the process of the invention,is characterized by logic diagram 1, X F 0 Is characteristic diagram 0, X F 1 For the features of FIG. 1, < >>For logically extracting the weight of cell 1, f ij 0 Image features, f, of feature map 0 ij 1 For characterizing the image features of FIG. 1, AND logic f ij 0 ·f ij 1 Or logic f ij 0 +f ij 1 Non-logic-f ij * ,/>Is the weight of the logic extraction unit 1, where i, j is the matrix index (i=1, 2, …, n; j=1, 2, …, m), n is the length of the image, m is the width of the image, and represents any layer>For convolution +.>Is Hadamard product;
step 4, sending the feature map 1 into a feature extraction unit 2 to obtain a feature map 2 after feature extraction;
wherein X is F 2 Is characterized by FIG. 2, W F 2 A weight matrix W for the feature extraction unit 2 S 2 A linear mapping weight matrix in jump connection in the feature extraction unit 2;
step 5, the feature logic diagram 1 and the feature diagram 2 are simultaneously sent to a logic extraction unit 2, and the feature logic diagram 2 is obtained after high-low-layer feature logic combination;
in the method, in the process of the invention,logic diagram 2, for feature>A weight matrix which is a logic extraction unit 2;
and step 6, continuously extracting higher-level features by using a next-level feature extraction unit and logically combining the higher-level features and the lower-level features by using the next-level logic extraction unit until a highest-level feature logic diagram is obtained, wherein the feature logic diagram comprises feature logic combinations among all the higher-level features and the lower-level features, and a recursive formula is expressed as follows:
where k is a layer mark, k=2, 3, …, X F k-1 For the characteristic diagram k-1, X F k As a characteristic map k of the graph,as a characteristic logic diagram k-1,for the characteristic logic diagram k, W F k Weight matrix for feature extraction unit k, W S k A linear mapping weight matrix in jump connection in the feature extraction unit k;
step 7, reforming the highest-level feature logic diagram into a one-dimensional tensor and sending the one-dimensional tensor into a classifier, and obtaining a classification result of the image by the classifier according to the highest-level feature logic;
wherein X is out In order to classify the result of the classification,is the feature logic diagram of the highest layer, W FC For the weight matrix of the full connection layer in the classifier, reshape (·) represents reforming variables into one-dimensional tensors;
step 8, comparing the classification result with the image label, calculating classification errors, and adjusting weights of a first-layer convolution layer, each feature extraction unit and a logic extraction unit by using an error back propagation algorithm, so as to obtain an optimal image classification model after multiple iterations;
E=L(X lab ,X out )
wherein X is out For classification result, X lab For image labels, E is the classification error, W * For all weights in the model, W * ' is all weights of the model after adjustment of an error back propagation algorithm, L (·) is a loss function, and η is a learning rate;
and 9, classifying the input image without the image tag by using an optimal image classification model to obtain the category to which the input image belongs.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110758190.3A CN113537325B (en) | 2021-07-05 | 2021-07-05 | Deep learning method for image classification based on extracted high-low layer feature logic |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110758190.3A CN113537325B (en) | 2021-07-05 | 2021-07-05 | Deep learning method for image classification based on extracted high-low layer feature logic |
Publications (2)
Publication Number | Publication Date |
---|---|
CN113537325A CN113537325A (en) | 2021-10-22 |
CN113537325B true CN113537325B (en) | 2023-07-11 |
Family
ID=78126781
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110758190.3A Active CN113537325B (en) | 2021-07-05 | 2021-07-05 | Deep learning method for image classification based on extracted high-low layer feature logic |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN113537325B (en) |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114886440B (en) * | 2022-07-13 | 2022-09-30 | 武汉工程大学 | Epileptic sample discharge classification model training and recognition method, system and electronic equipment |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2020216227A1 (en) * | 2019-04-24 | 2020-10-29 | 华为技术有限公司 | Image classification method and apparatus, and data processing method and apparatus |
WO2021057810A1 (en) * | 2019-09-29 | 2021-04-01 | 深圳数字生命研究院 | Data processing method, data training method, data identifying method and device, and storage medium |
CN112926691A (en) * | 2021-04-07 | 2021-06-08 | 北京航空航天大学 | Convolutional dendrite method for extracting feature logic for classification |
-
2021
- 2021-07-05 CN CN202110758190.3A patent/CN113537325B/en active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2020216227A1 (en) * | 2019-04-24 | 2020-10-29 | 华为技术有限公司 | Image classification method and apparatus, and data processing method and apparatus |
WO2021057810A1 (en) * | 2019-09-29 | 2021-04-01 | 深圳数字生命研究院 | Data processing method, data training method, data identifying method and device, and storage medium |
CN112926691A (en) * | 2021-04-07 | 2021-06-08 | 北京航空航天大学 | Convolutional dendrite method for extracting feature logic for classification |
Non-Patent Citations (1)
Title |
---|
基于深度卷积神经网络的图像分类算法;陈瑞瑞;;河南科技学院学报(自然科学版)(第04期);全文 * |
Also Published As
Publication number | Publication date |
---|---|
CN113537325A (en) | 2021-10-22 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN109492099B (en) | Cross-domain text emotion classification method based on domain impedance self-adaption | |
CN108664632B (en) | Text emotion classification algorithm based on convolutional neural network and attention mechanism | |
CN109711426B (en) | Pathological image classification device and method based on GAN and transfer learning | |
CN106650789B (en) | Image description generation method based on depth LSTM network | |
Dong et al. | Sparse fully convolutional network for face labeling | |
CN111368896A (en) | Hyperspectral remote sensing image classification method based on dense residual three-dimensional convolutional neural network | |
CN111507993A (en) | Image segmentation method and device based on generation countermeasure network and storage medium | |
CN109389171B (en) | Medical image classification method based on multi-granularity convolution noise reduction automatic encoder technology | |
CN112560432A (en) | Text emotion analysis method based on graph attention network | |
CN107247952B (en) | Deep supervision-based visual saliency detection method for cyclic convolution neural network | |
CN112819063B (en) | Image identification method based on improved Focal loss function | |
WO2023124342A1 (en) | Low-cost automatic neural architecture search method for image classification | |
Zhang et al. | Fine-grained vehicle recognition using lightweight convolutional neural network with combined learning strategy | |
CN113850311A (en) | Long-tail distribution image identification method based on grouping and diversity enhancement | |
CN113537325B (en) | Deep learning method for image classification based on extracted high-low layer feature logic | |
CN110111365B (en) | Training method and device based on deep learning and target tracking method and device | |
CN116258990A (en) | Cross-modal affinity-based small sample reference video target segmentation method | |
Lonij et al. | Open-world visual recognition using knowledge graphs | |
CN113378938B (en) | Edge transform graph neural network-based small sample image classification method and system | |
CN113378962B (en) | Garment attribute identification method and system based on graph attention network | |
CN112668633B (en) | Adaptive graph migration learning method based on fine granularity field | |
CN115761654B (en) | Vehicle re-identification method | |
CN116975578A (en) | Logic rule network model training method, device, equipment, program and medium | |
An | Xception network for weather image recognition based on transfer learning | |
CN115063374A (en) | Model training method, face image quality scoring method, electronic device and storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |