CN113962329A - Novel image recognition algorithm based on deep ensemble learning - Google Patents

Novel image recognition algorithm based on deep ensemble learning Download PDF

Info

Publication number
CN113962329A
CN113962329A CN202111351249.3A CN202111351249A CN113962329A CN 113962329 A CN113962329 A CN 113962329A CN 202111351249 A CN202111351249 A CN 202111351249A CN 113962329 A CN113962329 A CN 113962329A
Authority
CN
China
Prior art keywords
classification
prediction
model
models
image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202111351249.3A
Other languages
Chinese (zh)
Other versions
CN113962329B (en
Inventor
邓泽林
胡钰聪
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Changsha University of Science and Technology
Original Assignee
Changsha University of Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Changsha University of Science and Technology filed Critical Changsha University of Science and Technology
Priority to CN202111351249.3A priority Critical patent/CN113962329B/en
Publication of CN113962329A publication Critical patent/CN113962329A/en
Application granted granted Critical
Publication of CN113962329B publication Critical patent/CN113962329B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2415Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on parametric or probabilistic models, e.g. based on likelihood ratio or false acceptance rate versus a false rejection rate
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Computational Linguistics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Evolutionary Biology (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Probability & Statistics with Applications (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a new image recognition algorithm based on deep ensemble learning, which comprises an image data preprocessing module, a depth algorithm module, a deep learning algorithm integration module and a prediction output module, and is characterized in that the image data preprocessing module converts an input image into a model input matrix and expands data; the depth algorithm module is used for carrying out preference selection in a plurality of depth models, the selected models are used as a base learner for integrated learning, and the base learner is independently trained, so that the diversity among classifiers can be increased and the independence can be enhanced; the deep learning algorithm integration module carries out strategy integration on the multiple models, and maximizes the output of the models by combining the multiple output results of the multiple models; and the system prediction module outputs the judgment result to the integration module. The image recognition network based on ensemble learning relieves the ambiguity problem by fully utilizing the output result of the base learner, effectively improves the fault-tolerant capability of the model and obtains higher recognition accuracy.

Description

Novel image recognition algorithm based on deep ensemble learning
Technical Field
The invention relates to the field of image recognition, in particular to a novel image recognition method based on deep ensemble learning.
Background
The ensemble learning can efficiently solve the problem of practical application, and therefore, the ensemble learning attracts much attention in the field of machine learning. Initially, ensemble learning was aimed at improving the accuracy of automated decision-making systems, and this approach has been able to successfully solve a variety of machine learning problems today, and ensemble learning provides solutions to a variety of machine learning problems, primarily by combining multiple learners, whose models can solve many problems that a single model cannot solve.
Dasarathy et al, first proposed an integrated learning concept that achieves a range of enhanced recognition system performance by deploying a composite classifier system consisting of two or more component classifiers belonging to different classes. The deployment domain of these individual components (classifiers) is determined by the optimal partitioning of the problem space. The criteria for such an optimal division are in each case determined by the characteristics of the classifier component. The concepts, related methods, and possible benefits one may expect are designed by such a composite classifier system, exemplified in terms of partitioning the feature space to optimize the deployment of the composite system with linear and nearest neighbor classifiers as its components.
Hansen et al propose an integration model based on neural networks and optimize network parameters and architecture with cross validation, the integration model reduces residual error generalization errors by calling for the integration of similar networks, and has better generalization capability.
Kittler et al developed a general theoretical framework to combine classifiers using different pattern representations and showed that many existing schemes can be considered as special cases of composite classification, where all pattern representations are used jointly to make decisions. Experimental comparisons of various classifier combination schemes show that the combination rule developed under the most rigorous hypothesis-the sum rule-is superior to other classifier combination schemes. Analysis of the sensitivity of the various schemes to estimation errors shows that this finding is theoretically reasonable. .
Cheng et al studied a number of widely used integration methods, including unweighted averaging, majority voting, bayesian optimal classifiers, and (discrete) super learners, for image recognition tasks, and with deep neural networks as candidate algorithms. And several experiments are designed, and the candidate algorithms are the same network structure with different model check points in a single training process, the network with the same structure but randomly trained for multiple times, and the network with different structures. In addition, the over-confidence phenomenon of the neural network, and its influence on the integration method, were further studied. In all experiments, the super learner achieved the best performance in all integrated approaches.
Hafiz et al, proposes a simple, sequential, and efficient ensemble learning method using multiple deep networks to address the shortcomings of the traditional machine learning methods when complex data. The deep network used in the integration is ResNet 50. The model draws inspiration from a binary decision/classification tree.
In summary, although the image classification task achieves significant results in recognition rate through different integration methods, some problems still exist. The existing integrated classifiers only change the part of the model or the training part of the model, the output of the base classifier is not fully utilized, and the waste of characteristic information caused by neglecting other output results influences the final classification effect. Therefore, the invention provides a new classifier based on ensemble learning, and the classifier is applied to an image classification task to extract the high-level features of the image and improve the accuracy rate through the diversity of a plurality of models.
Disclosure of Invention
Compared with the traditional method, the image recognition system based on the ensemble learning can realize extremely high recognition accuracy on the image, increase the feature extraction efficiency, overcome the problems of low feature extraction efficiency and difficult classification in the traditional method and achieve higher accuracy. The problem that the traditional integrated classifier only changes the part of a model or a model training part and the output of a base classifier is not fully utilized is solved, a novel image classification integrated classifier based on integrated learning is provided, a plurality of output predictions of multiple models are combined, the features extracted by the base model can be utilized to the maximum extent, the model utilization rate is increased, the model robustness is increased, and the classification effect is improved by using a novel multi-model integration method through the network.
The solution proposed by the invention is to realize image recognition by adopting a network based on deep ensemble learning, which comprises the following steps: the device comprises an image data preprocessing module, a depth algorithm module, a depth learning algorithm integration module and a prediction output module.
The image data preprocessing module has two conditions of training and testing. The functions of the data preprocessing module comprise: image data is preprocessed and converted into an input form which can be accepted by a model image recognition module, image categories, namely real labels, need to be marked in a training mode, and a data set needed by a machine learning method is constructed.
The working method of image data preprocessing in the training mode comprises the following steps: and converting the image data into matrix data, and performing preprocessing such as uniform up-sampling, center cutting, rotation and the like on each sample image data to expand the data volume. After the training mode is finished, the parameters of the convolutional neural network can be updated, and the network capable of accurately identifying the specified target is obtained. And upsampling is unified to a fixed resolution, in the following embodiment, the upsampling results in an image resolution of 224 × 224, wherein inclusion-v 3 is 299 × 299.
The working method of image data preprocessing in the test mode comprises the following steps: the image data is converted into matrix data, each sample image data is subjected to unified upsampling without data expansion processing, and is uniformly upsampled to a fixed resolution, the resolution of the image obtained after upsampling in the following embodiment is 224 × 224, wherein the inclusion-v 3 is 299 × 299.
The deep algorithm module comprises the selection of base classifiers, based on the principle that the base classifiers can provide as much supplementary information as possible in classification, the test is carried out from a plurality of neural networks, and then a plurality of base classifiers with better classification efficiency are selected as the base classifiers. Several classifiers are selected that have different mechanisms in classification so that they can provide complementary information to each other. Furthermore, the single classifier chosen is complex and is known to achieve high performance in classification. These classifiers trained with the same training parameters will help increase diversity and enhance independence between classifiers.
The deep learning algorithm integration module comprises a proposed integration algorithm, in the model training process, an independent training mode is adopted, the model independence is enhanced, a large number of experiments show that many correct labels exist in the first k predicted values, and a plurality of models can be used for relieving the class ambiguity problem by considering that the first k outputs are the results of model feature extraction. Two results, namely top-1 and top-2 prediction, are output through each classifier, the top-1 prediction results of a plurality of models are placed into a top-1 prediction pool, the top-2 prediction results are placed into a top-2 prediction pool, when the top-1 prediction pool cannot obtain the maximum classification result, the top-1 prediction pool and the top-2 prediction pool are combined, the class with the most labels is used as a sample class to be output, if the maximum result cannot be obtained, each model is used as an independent 'evidence', confidence coefficient distribution of the evidence is fused, and the model maximum prediction value with the highest confidence coefficient corresponds to the label to be used as the final classification result.
And the system output module processes the output of the classifier and outputs a prediction result.
The invention has the beneficial effects that:
the invention provides a novel multi-model multi-output network based on ensemble learning based on the problem of image recognition, which can increase the utilization efficiency of features and improve the image recognition effect. The problem that the output result of a base learner is not fully utilized in the existing deep ensemble learning method can be solved, ambiguity problems are relieved by combining top-2 results of a plurality of models, and the fault tolerance of the models is effectively improved. And the identification capability of the model is improved by outputting results of multiple models and combining confidence coefficient distribution to obtain more useful features.
Drawings
FIG. 1 is a diagram of the base classifier EfficientNet model used in the present invention.
FIG. 2 is a diagram of a base classifier VGGNet model used in the present invention.
Fig. 3 is a diagram of a network model of the basis classifier inclusion-v 3 used in the present invention.
FIG. 4 is a diagram of multiple model independent training based on ensemble learning in an embodiment of the present invention.
FIG. 5 is a diagram of a model for fusing multiple output results of multiple models according to an embodiment of the present invention.
Detailed Description
The present invention is described in further detail below with reference to the attached drawings.
Examples
First, we selected a data set, which is a dermannist skin lesion data set, and which contains 7 classes, including 7007 training sets, 1003 verification sets, 2005 test sets, and 28 × 28 image sizes. The data set details are shown in the following table:
TABLE 1DermaMNIST data set introduction
Dataset Train Size Val Size Test Size Classes
DermaMNIST 7007 1003 2005 7
The next step is the selection of the base classifier. Based on the principle that the classifier can provide as much supplementary information as possible in classification, the classifier performs testing from a plurality of deep neural networks and then selects a plurality of base classifiers with better classification efficiency.
We selected neural network classifiers for testing to be:
(1) VGGNet, VGGNet is a deep convolutional neural network developed together by Google corporation, oxford university. It explores the relationship between the depth of the convolutional neural network and its performance, VGGNet obtains the champions of the ILSVRC 2014 game and the champions of the positioning projects by iteratively stacking a small convolutional kernel of 3x3 and a maximum pooling layer of 2x2, with an error rate of 7.5% on top 5.
(2) The EfficientNet is one of the most advanced neural network algorithms in the world at present, the algorithm has the quantization adjustment capacity of a complex network, and the optimal network parameters for specific requirements are obtained by comprehensively adjusting the depth (depth), the width (width) and the resolution (resolution) of an input picture, so that the network has the dual advantages of network size and identification accuracy. The inside of the EfficientNet model is realized by a plurality of MBConv convolution blocks, and the MBConv convolution blocks also use a structure similar to residual linking, so that the problem of model gradient disappearance is solved.
(3) ResNet, ResNet uses the identity mapping to directly transmit the output of the previous layer to the back layer, when the depth of the network is increased, the error cannot be increased, the deeper network cannot bring the rise of the error on the training set, and the problem of gradient disappearance is solved.
(4) The Incep-v 3 and the Incep-v 3 have strong image feature extraction and classification performance and are widely used image recognition models. It consists of symmetric and asymmetric building blocks, including convolution, average pooling, maximum pooling, dropout, and full-connected layers. Batch normalization is widely used throughout the model and applied to activation inputs. Two 3x3 convolution kernels are used for replacing a 5x5 convolution kernel, three 3x3 convolution kernels are used for replacing a 7x7 convolution kernel, the parameter number is reduced, calculation is accelerated, the nxn convolution kernels are further decomposed into 1xn and nx1 convolution kernels, the size of feature map is reduced, and the channel number is increased.
(5) The High Order RNN structure (HORNN) used in the DPN model to tie densnet and ResNet together demonstrates that densnet can extract new features from earlier levels, while ResNet is essentially a reuse of extracted features from earlier levels. By combining the advantages of the two structures, the DPN network can effectively improve the classification efficiency.
The model was trained using the environmental training data set shown in the table below.
Table 2 experimental environment table
Name(s) Configuration of
Operating system Ubuntu18.04
GPU NVIDIA GEFORCE RTX 2080Ti
CPU Inter xeon processor(skylake,IBRS),2
RAM 16GB
GPU correlation library CUDA10.2,CUDNN7.6
Deep learning framework Pytorch
The model training parameters are shown in the following table, and the Batchsize and the Epoch are properly adjusted according to the model parameter quantity to reduce the video memory occupied by the video card and the model training time.
TABLE 3 model initialization parameter Table
Figure BDA0003354464850000031
Figure BDA0003354464850000041
The test model is trained to obtain test results of five base classifiers, namely VGGNet, EfficientNet, ResNet, DPN92 and inclusion, and the test results are shown in the following table:
TABLE 4 DermaMNIST data set comparison of different basis classifier accuracy performance
Method Accuracy
VGGNet 76.41
EfficientNet 76.86
ResNet 73.22
DPN92 73.72
Inception-v3 76.86
Finally, according to the experimental effect, EfficientNet, VGGNet and inclusion-v 3 are selected as the basic networks for integration, and the network models are respectively shown in the following tables as the network structures shown in FIG. 1, FIG. 2 and FIG. 3:
table 5 selected base classifier network architecture
Figure BDA0003354464850000042
And in a final deep algorithm integration module of the model, an independent training mode is adopted, and basic classifiers are independently trained one by a basic network. Each neural network model outputs two prediction results, and the first deep neural network is assumed to generate h for the input image1(x) 1 column n-dimensional matrix, where n is the number of classes, and a second deep neural network generates h for the input image2(x) A third deep neural network generates h for the input image3(x) The maximum prediction labels of the three neural networks are used as a first classification pool, the second maximum prediction label is placed in a second classification pool, and the model is trained independently as shown in fig. 4. The best strategy for realizing the highest prediction performance by using the convolutional neural network is to independently train each base classifier, each model is independently trained to serve as a class of classified 'experts', each base learner gives out own classification prediction and is not interfered by other learners, the independence of the learners is ensured, and information acquisition redundancy caused by the interference of other learners is avoided. Each learner is an independent CNN architecture network with good prediction efficiency. Here we represent the base classifier as a shallow network and then integrate its output into a deeper network, generalizing the feature extraction capability.
And after the data passes through the independently trained base classifier, the prediction probability of each class of the target is output through a Soft-max function, the class where the maximum probability value is located is judged as the class of the target, namely Top-1, and the class where the second approximate probability value is located is judged as the second prediction class, namely Top-2. In practical application, the image data is input into the trained model, the system can determine and output the Top-1 and Top-2 categories of the target to be used as the input of the ensemble learning, and the multi-model output is combined with the model as shown in fig. 5. In this case, the model prediction is used instead of the softmax classification value, so that the maximum and minimum values generated by the Soft-max function are prevented from causing certain damage to other base classifier models. We express the first classification pool as a Majority voting (Majority voting) mathematical formula:
Figure BDA0003354464850000051
wherein, T represents T classifiers, N represents N categories, namely, the prediction result of the T classifiers to the category j is more than half of the total voting result, the prediction is the category j, otherwise, the prediction is rejected.
And if the first classification pool has a prediction rejection condition, namely a majority voting result is not generated, combining the first classification pool and the second classification pool, and continuously using a majority voting method to obtain the prediction. If the prediction is still rejected, the first classification pool and the second classification pool still do not produce most results after being combined. We treat each model as an independent "evidence", fuse confidence distributions of the evidence, and the mathematical formula is expressed as:
Figure BDA0003354464850000052
the above equation gives the probability that the predictions are correct in each of the n models, where p is the probability that the predictions are correct in a single model. The higher the confidence coefficient is, the more positive the output result of the model is, and the label corresponding to the maximum predicted value of the model with the highest confidence coefficient is taken as the final classification result.
The test model, the method of the invention, on the DermaMNIST public data set, compares with the traditional method and other convolution neural network methods respectively. Table 2 shows the comparison of the recognition accuracy obtained by different algorithms on the dermannist image dataset based on ensemble learning as proposed by the present invention. ResNet18, DenseNet, inclusion-v 3, EffectintNet, DPN92, MobileNet V2 and VGGNet11 are respectively single different convolutional neural network architectures, Ruta is an integrated learning method proposed by other people, and compared with the methods, the integrated learning method provided by the inventor obtains higher identification accuracy, and the accuracy is compared with the following table:
TABLE 6 comparison of Performance of different methods on DermaMNIST dataset
Method Accuracy
AutoKeras 74.9
Yang 76.8
ResNet18 73.22
Auto-sklearn 71.9
VGGNet11 76.41
Inception-v3 76.86
EfficientNet 76.86
MobileNetV2 75.81
DPN92 73.72
Ruta 77.40
Ours Method 77.55
The above embodiments describe in detail a specific embodiment of the new image recognition system based on ensemble learning, and the above embodiments only serve to help understanding the proposed method and core idea of the present invention, and according to the idea of the present invention, there may be some differences in the specific embodiment, and in summary, the content of the present specification should not be construed as limiting the present invention.

Claims (4)

1. A new ensemble learning based multiple output convolutional neural network for image recognition, comprising the steps of:
(1) selecting a data set for carrying out an image classification experiment from a plurality of image classification data sets;
(2) forward propagation: inputting the image into a classifier, and obtaining a plurality of model prediction results through a plurality of convolutional neural networks including residual error networks ResNet, VGGNet, EfficientNet, DPN92 and other deep learning networks;
(3) and (3) back propagation: firstly, parameters of the classifier are fixed, gradient descending is not carried out on the parameters, and meanwhile back propagation and parameter updating are carried out on a new classifier. Then, allowing the gradient of the classifier to descend, and performing back propagation and updating parameters;
(4) and integrating the models according to the proposed integration strategy, comparing the models with other latest single models and other integrated models in a model structure, testing the trained models by adopting Accuracy (Accuracy), and measuring classification results.
2. The ensemble learning-based multi-output convolutional neural network for image classification as claimed in claim 1,
in the step (2), based on a plurality of new classifiers of the deep neural network, each neural network model outputs two prediction results, and the first deep neural network is assumed to generate h for the input image1(x) 1 column n-dimensional matrix, where n is the number of classes, and a second deep neural network generates h for the input image2(x) A third deep neural network generates h for the input image3(x) The maximum prediction labels of the three neural networks are used as a first classification pool, a second maximum prediction label is placed in a second classification pool, and the first classification pool is expressed by a mathematical formula according to a Majority voting method (Majority voting):
Figure FDA0003354464840000011
wherein, T represents T classifiers, N represents N categories, namely, the prediction result of the T classifiers to the category j is more than half of the total voting result, the prediction is the category j, otherwise, the prediction is rejected.
And if the first classification pool has a prediction rejection condition, namely a majority voting result is not generated, combining the first classification pool and the second classification pool, and continuously using a majority voting method to obtain the prediction. If the prediction is still rejected, the first classification pool and the second classification pool still do not produce most results after being combined. We treat each model as an independent "evidence", fuse confidence distributions of the evidence, and the mathematical formula is expressed as:
Figure FDA0003354464840000012
the above equation gives the probability that the predictions are correct in each of the n models, where p is the probability that the predictions are correct in a single model. The higher the confidence coefficient is, the more positive the output result of the model is, and the label corresponding to the maximum predicted value of the model with the highest confidence coefficient is taken as the final classification result.
3. The ensemble learning-based multi-output convolutional neural network for image classification as claimed in claim 1,
the new classifier in the step (2) is a classifier based on ensemble learning, and the ensemble learning is a machine learning method which uses a series of base learners to learn and integrates all learning results by using a certain rule so as to obtain better learning effect than a single learner. The range of enhancing the performance of the recognition system can be achieved by utilizing a plurality of classifiers and utilizing the diversity and independence of the classifiers to perform better-level extraction on the image features and deploying a composite classifier system consisting of two or more component classifiers belonging to different categories.
4. The ensemble learning-based multi-output convolutional neural network for image classification as claimed in claim 1,
and (4) in the step (3), the new classifier utilizes a network with a plurality of residual modules in convolution, so that feature extraction can be ensured under the condition of not reducing features, and the problem of feature reduction caused by gradient disappearance due to too deep network of the classifier is prevented.
CN202111351249.3A 2021-11-15 2021-11-15 Novel image classification method based on deep ensemble learning Active CN113962329B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111351249.3A CN113962329B (en) 2021-11-15 2021-11-15 Novel image classification method based on deep ensemble learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111351249.3A CN113962329B (en) 2021-11-15 2021-11-15 Novel image classification method based on deep ensemble learning

Publications (2)

Publication Number Publication Date
CN113962329A true CN113962329A (en) 2022-01-21
CN113962329B CN113962329B (en) 2024-07-23

Family

ID=79470649

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111351249.3A Active CN113962329B (en) 2021-11-15 2021-11-15 Novel image classification method based on deep ensemble learning

Country Status (1)

Country Link
CN (1) CN113962329B (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114549906A (en) * 2022-02-28 2022-05-27 长沙理工大学 Improved image classification algorithm for step-by-step training of Top-k loss function
CN116580282A (en) * 2023-07-12 2023-08-11 四川大学华西医院 Neural network model-based pressure injury staged identification system and storage medium

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107564580A (en) * 2017-09-11 2018-01-09 合肥工业大学 Gastroscope visual aids processing system and method based on integrated study
CN109558898A (en) * 2018-11-09 2019-04-02 复旦大学 A kind of more options learning method of the high confidence level based on deep neural network
CN110210550A (en) * 2019-05-28 2019-09-06 东南大学 Image fine granularity recognition methods based on integrated study strategy
AU2020103901A4 (en) * 2020-12-04 2021-02-11 Chongqing Normal University Image Semantic Segmentation Method Based on Deep Full Convolutional Network and Conditional Random Field

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107564580A (en) * 2017-09-11 2018-01-09 合肥工业大学 Gastroscope visual aids processing system and method based on integrated study
CN109558898A (en) * 2018-11-09 2019-04-02 复旦大学 A kind of more options learning method of the high confidence level based on deep neural network
CN110210550A (en) * 2019-05-28 2019-09-06 东南大学 Image fine granularity recognition methods based on integrated study strategy
AU2020103901A4 (en) * 2020-12-04 2021-02-11 Chongqing Normal University Image Semantic Segmentation Method Based on Deep Full Convolutional Network and Conditional Random Field

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
周燕艳;范学满;海深;: "基于深度集成学习的手写数字识别", 陕西理工大学学报(自然科学版), no. 03, 20 June 2020 (2020-06-20) *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114549906A (en) * 2022-02-28 2022-05-27 长沙理工大学 Improved image classification algorithm for step-by-step training of Top-k loss function
CN116580282A (en) * 2023-07-12 2023-08-11 四川大学华西医院 Neural network model-based pressure injury staged identification system and storage medium

Also Published As

Publication number Publication date
CN113962329B (en) 2024-07-23

Similar Documents

Publication Publication Date Title
Bansal et al. Transfer learning for image classification using VGG19: Caltech-101 image data set
CN110689086B (en) Semi-supervised high-resolution remote sensing image scene classification method based on generating countermeasure network
Minaei-Bidgoli et al. Ensembles of partitions via data resampling
CN103605972B (en) Non-restricted environment face verification method based on block depth neural network
JP2022538866A (en) System and method for image preprocessing
CN107766894A (en) Remote sensing images spatial term method based on notice mechanism and deep learning
CN112733866A (en) Network construction method for improving text description correctness of controllable image
CN112232165B (en) Data processing method, device, computer and readable storage medium
Akhand et al. Convolutional neural network training incorporating rotation-based generated patterns and handwritten numeral recognition of major Indian scripts
CN113962329A (en) Novel image recognition algorithm based on deep ensemble learning
CN115222998B (en) Image classification method
Chen et al. A compact cnn-dblstm based character model for online handwritten chinese text recognition
CN105046323B (en) Regularization-based RBF network multi-label classification method
Cheng et al. A multiclassification method for iris data based on the hadamard error correction output code and a convolutional network
CN111371611B (en) Weighted network community discovery method and device based on deep learning
CN109344777A (en) The Optimum Classification method of target in hyperspectral remotely sensed image land use covering based on ELM
CN112164100A (en) Image registration method based on graph convolution neural network
CN111310820A (en) Foundation meteorological cloud chart classification method based on cross validation depth CNN feature integration
Zhang et al. Deep compression of probabilistic graphical networks
Shekar et al. Blockwise binary pattern: a robust and an efficient approach for offline signature verification
Li et al. A spectral clustering based filter-level pruning method for convolutional neural networks
Bansod et al. Analysis of convolution neural network architectures and their applications in industry 4.0
Wirayasa et al. Comparison of Convolutional Neural Networks Model Using Different Optimizers for Image Classification
Klein et al. Deep learning classification of photographic paper based on clustering by domain experts
CN113221676B (en) Target tracking method and device based on multidimensional features

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant