CN110705600A - Cross-correlation entropy based multi-depth learning model fusion method, terminal device and readable storage medium - Google Patents

Cross-correlation entropy based multi-depth learning model fusion method, terminal device and readable storage medium Download PDF

Info

Publication number
CN110705600A
CN110705600A CN201910843623.8A CN201910843623A CN110705600A CN 110705600 A CN110705600 A CN 110705600A CN 201910843623 A CN201910843623 A CN 201910843623A CN 110705600 A CN110705600 A CN 110705600A
Authority
CN
China
Prior art keywords
model
cross
classification
correlation entropy
image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201910843623.8A
Other languages
Chinese (zh)
Inventor
杜少毅
宋盼盼
国雷达
许美凤
崔文婷
万腾
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xian Jiaotong University
Original Assignee
Xian Jiaotong University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Xian Jiaotong University filed Critical Xian Jiaotong University
Priority to CN201910843623.8A priority Critical patent/CN110705600A/en
Publication of CN110705600A publication Critical patent/CN110705600A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Health & Medical Sciences (AREA)
  • Software Systems (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Mathematical Physics (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a multi-depth learning model fusion method based on cross-correlation entropy, terminal equipment and a readable storage medium, and belongs to the field of computer image processing. The invention trains and tests the collected data to obtain the image classification model. The invention can effectively finish the accurate classification of the images, and has good robustness and accuracy under the conditions of extremely similar image characteristics, much image noise and the like. The image classification is used as the basis of high-level visual tasks such as image detection, image segmentation, object tracking, behavior analysis and the like, and can be widely applied to a plurality of fields such as face recognition and intelligent video analysis in the security field, scene recognition in the traffic field, content-based image retrieval and image automatic classification in the internet field, image recognition in the medical field and the like.

Description

Cross-correlation entropy based multi-deep learning model fusion method, terminal device and readable storage medium
[ technical field ] A method for producing a semiconductor device
The invention belongs to the field of computer image processing, and relates to a cross-correlation entropy-based multi-deep learning model fusion method, terminal equipment and a readable storage medium.
[ background of the invention ]
Computer vision is one of the most popular research areas in the field of deep learning. Computer vision, as a cross-domain discipline, includes computer science (graphics, algorithms, theories, systems, architectures), mathematics (information retrieval, machine learning), engineering (robotics, speech, natural language processing, image processing), physics (optics), biology (neuroscience), psychology (cognitive science), and so on, which opens the way for the development of artificial intelligence. Image classification is the most fundamental task in computer vision, and the classification problem refers to giving a group of images of respective marked classes, predicting the classes of a new group of test images, measuring the accuracy of prediction, and realizing the minimum classification error. However, one of the most important disadvantages of the visual technology is that the requirements for image quality are relatively strict. The image information obtained by the vision technology has more noise under the conditions of point-of-view change, scale change, intra-class change, image deformation, image occlusion, background mottle and the like, and great challenge is brought to the precision of the classification algorithm.
[ summary of the invention ]
The present invention is directed to overcome the disadvantages of the prior art, and provides a method for fusing multiple deep learning models based on cross-correlation entropy, a terminal device and a readable storage medium
In order to achieve the purpose, the invention adopts the following technical scheme to realize the purpose:
a cross-correlation entropy based multi-deep learning model fusion method is suitable for being executed in computing equipment and comprises the following steps:
step 1: preprocessing the data set;
step 2: respectively training a plurality of single model classification convolution neural networks to enable each model to obtain an optimal result;
and step 3: and (2) using a multi-model integration method for the single model with the optimal result, and learning the contribution weight of each single-model classification convolutional neural network to the final result by using a multi-model fusion method based on a cross-correlation entropy criterion, wherein the objective function is as follows:
Figure BDA0002194490150000021
wherein w ═ w1,w2,…,wn]Representing the weights of the integrated model, b is the bias, m is the number of samples, σ is the kernel width, xiRepresenting the output of the ith sample in the plurality of single models, yiA true tag representing the ith sample;
using a gradient-rise method to obtain a parameter w which brings the objective function close to a maximumjAnd b;
and 4, step 4: and (4) training the target function in the step (3) to obtain an optimal image classification result of the multi-depth convolution neural network integrated model based on the cross-correlation entropy criterion.
The invention further improves the following steps:
the specific method of the step 2 is as follows:
step 2-1: training a plurality of deep convolution neural network models by using the same training set X according to the collected data, firstly initializing model variables w and b, and obtaining a predicted value by network forward propagation;
step 2-2: and (3) performing back propagation by using a loss function and a function optimization method, updating model weights w and b, and calculating the formula as follows:
Figure BDA0002194490150000022
Figure BDA0002194490150000023
where λ represents the learning rate, wjRepresents the weight of the jth parameter, k represents the kth iteration,
Figure BDA0002194490150000024
updating the k-th weight value of the jth model, taking J (w, b) as an objective function, and applying a gradient descent method to the parameter w in the objective functionjB, updating the weight;
step 2-3: and (5) repeating the step 2-1 to the step 2-2 until the training times are reached or the training target is obtained to obtain the optimal model.
The multiple deep convolutional neural network models are one or more of VGG, inclusion V4, Xception, ResNet or DensNet.
The specific method of the step 3 is as follows:
step 3-1: for a plurality of existing single models, learning the contribution weight w of each single model to the final result by using an MCC-based multi-model integration method1,w2,…,wn]N is the number of single models;
step 3-2: taking the correlation entropy criterion as an objective function, the calculation formula is as follows:
Figure BDA0002194490150000031
step 3-3: and (3) maximizing a target function defined by the cross-correlation entropy by using the corresponding relation obtained in the step (3-2), wherein the contribution of each single model to a final result is calculated according to the following formula:
Figure BDA0002194490150000032
Figure BDA0002194490150000033
the entropy criterion is subjected to gradient rise, and the objective function is close to the maximum value after multiple iterations to achieve the learning parameter wjAnd purpose of b;
step 3-4: judging whether the classification precision of the current model meets the requirement, if so, outputting a parameter wjB, if not, continuing to adjust the model parameters and returningStep 3-2, continuously performing iterative computation to obtain optimal parameters to obtain a final model
A memory device having stored therein a plurality of instructions adapted to be loaded and executed by a processor:
step 1: preprocessing the data set;
step 2: respectively training a plurality of single model classification convolution neural networks to enable each model to obtain an optimal result;
and step 3: and (2) using a multi-model integration method for the single model with the optimal result, and learning the contribution weight of each single-model classification convolutional neural network to the final result by using a multi-model fusion method based on a cross-correlation entropy criterion, wherein the objective function is as follows:
Figure BDA0002194490150000041
wherein w ═ w1,w2,…,wn]Representing the weights of the integrated model, b is the bias, m is the number of samples, σ is the kernel width, xiRepresenting the output of the ith sample in the plurality of single models, yiA true tag representing the ith sample;
using a gradient-rise method to obtain a parameter w which brings the objective function close to a maximumjAnd b;
and 4, step 4: and (4) training the target function in the step (3) to obtain an optimal image classification result of the multi-depth convolution neural network integrated model based on the cross-correlation entropy criterion.
A mobile terminal, comprising:
a processor adapted to implement instructions; and
a storage device adapted to store a plurality of instructions, the instructions adapted to be loaded and executed by a processor to:
step 1: preprocessing the data set;
step 2: respectively training a plurality of single model classification convolution neural networks to enable each model to obtain an optimal result;
and step 3: and (2) using a multi-model integration method for the single model with the optimal result, and learning the contribution weight of each single-model classification convolutional neural network to the final result by using a multi-model fusion method based on a cross-correlation entropy criterion, wherein the objective function is as follows:
Figure BDA0002194490150000051
wherein w ═ w1,w2,…,wn]Representing the weights of the integrated model, b is the bias, m is the number of samples, σ is the kernel width, xiRepresenting the output of the ith sample in the plurality of single models, yiA true tag representing the ith sample;
using a gradient-rise method to obtain a parameter w which brings the objective function close to a maximumjAnd b;
and 4, step 4: and (4) training the target function in the step (3) to obtain an optimal image classification result of the multi-depth convolution neural network integrated model based on the cross-correlation entropy criterion.
Compared with the prior art, the invention has the following beneficial effects:
the method comprises the steps of training a plurality of single model networks respectively by utilizing an acquired data set including scene recognition in the traffic field and skin disease classification in the medical field to enable the classification effect of each network to be optimal, then performing multi-model fusion by utilizing a maximum cross-correlation entropy criterion as a target function, and calculating the contribution weight of each single model network to a final result to obtain a final classification model.
The method utilizes a traffic scene data set acquired by unmanned vehicles from an artificial intelligent robot institute of the university of transport in Western Ann and pathological sections of the department of dermatology of a second subsidiary hospital of the university of transport in Western Ann as data sources, and finishes the classification work of the traffic scene and the pathological sections through the steps of data preprocessing, multiple single-model network training, multi-model integrated training based on maximum cross-correlation entropy and the like. Due to the fact that slicing properties of skin diseases have very high similarity, the judgment result of each single model is unstable and limited in precision, and therefore the method for model integration can well classify the slices in high precision. In addition, aiming at the problem that the slice has a lot of noise, the multi-model integration method based on the maximum cross-correlation entropy in the step 3.3) can be used for well reducing noise interference and improving the model accuracy. Thus, a function of robust and accurate classification of images can be realized.
[ description of the drawings ]
FIG. 1 is an overall structure of a convolutional neural network;
FIG. 2 is an overall framework of an integrated model based on a plurality of convolutional neural networks;
FIG. 3 shows the experimental results of different models.
[ detailed description ] embodiments
In order to make the technical solutions of the present invention better understood, the technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, not all of the embodiments, and are not intended to limit the scope of the present disclosure. Moreover, in the following description, descriptions of well-known structures and techniques are omitted so as to not unnecessarily obscure the concepts of the present disclosure. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
Various structural schematics according to the disclosed embodiments of the invention are shown in the drawings. The figures are not drawn to scale, wherein certain details are exaggerated and possibly omitted for clarity of presentation. The shapes of various regions, layers and their relative sizes and positional relationships shown in the drawings are merely exemplary, and deviations may occur in practice due to manufacturing tolerances or technical limitations, and a person skilled in the art may additionally design regions/layers having different shapes, sizes, relative positions, according to actual needs.
In the context of the present disclosure, when a layer/element is referred to as being "on" another layer/element, it can be directly on the other layer/element or intervening layers/elements may be present. In addition, if a layer/element is "on" another layer/element in one orientation, then that layer/element may be "under" the other layer/element when the orientation is reversed.
It should be noted that the terms "first," "second," and the like in the description and claims of the present invention and in the drawings described above are used for distinguishing between similar elements and not necessarily for describing a particular sequential or chronological order. It is to be understood that the data so used is interchangeable under appropriate circumstances such that the embodiments of the invention described herein are capable of operation in sequences other than those illustrated or described herein. Furthermore, the terms "comprises," "comprising," and "having," and any variations thereof, are intended to cover a non-exclusive inclusion, such that a process, method, system, article, or apparatus that comprises a list of steps or elements is not necessarily limited to those steps or elements expressly listed, but may include other steps or elements not expressly listed or inherent to such process, method, article, or apparatus.
The invention is described in further detail below with reference to the accompanying drawings:
referring to fig. 1, in order to accurately, stably and robustly classify images, the invention provides a cross-correlation entropy based multi-deep learning model fusion method and an application thereof in image classification, and the method specifically comprises three parts, namely data set collection and preprocessing, training of a plurality of single models, and multi-model fusion, as shown in fig. 1.
The method is specifically carried out according to the following steps:
step 1.1: in order to realize the function of classifying input images, data needs to be preprocessed, the preprocessing comprises preprocessing the data into a format which can be read by a neural network and normalizing, and the training model has better robustness through means of rotation, mirror image, cutting and the like.
Step 1.2: in order to evaluate and select the model, independent and identically distributed sampling needs to be performed on a data set, the data set is divided into a training set for training and learning of the model, a cross validation set is used for preventing overfitting of the model, and a test set is used for testing the discrimination capability of the model on the sample.
Step 2.1: respectively training the models in a plurality of deep convolutional neural networks, such as VGG, inclusion V4, Xconvergence, ResNet, DenNet and other networks by using the same training set X according to the collected data, so that each model obtains an optimal result, and storing model parameters w and b;
step 2.2: and (3) performing back propagation by using a loss function and a function optimization method, updating model weights w and b, and calculating the formula as follows:
Figure BDA0002194490150000071
Figure BDA0002194490150000081
where λ represents the learning rate, wjExpressing the weight of the jth parameter, J (w, b) is an objective function, and the gradient descent method is used for the parameter w in the objective functionjAnd b, updating the weight.
Step 2.3: and repeating the steps until the training times are reached or the training target is obtained to obtain the optimal model.
Because the accuracy of a single model is very limited, in order to solve the problem, the method uses a multi-model integration method, learns the contribution of each model to the final result, improves the sensitivity of the model, and provides the maximum mutual entropy criterion to improve the anti-noise capability of the model and acquire a more accurate prediction result aiming at more image noise points.
Step 3.1: for a plurality of existing single models, a multi-model integration method based on maximum cross-correlation entropy is used to learn the contribution weight w of each single model to the final result [ w ]1,w2,…,wn]And n is the number of single models trained.
Step 3.2: taking the correlation entropy criterion as an objective function, the calculation formula is as follows:
Figure BDA0002194490150000082
where there are m samples, w ═ w1,w2,…,wn]Weights, w, representing integration modelsjWeight, x, representing the jth modeliRepresenting the output of the ith sample in a plurality of models, yiThe authentic label representing the ith sample.
Step 3.3: after the step 3.2), maximizing the target function defined by the cross-correlation entropy by using the corresponding relation obtained by the step 3.2), wherein the contribution of each single model to the final result is calculated according to the following formula:
Figure BDA0002194490150000083
Figure BDA0002194490150000084
Figure BDA0002194490150000091
where σ denotes the kernel width and λ denotes the learning rate. The entropy criterion is subjected to gradient rise, and the objective function is close to the maximum value after multiple iterations to achieve the learning parameter wjAnd purpose of b.
Step 3.4: judging whether the classification precision of the current model meets the requirement, if so, outputting a parameter wjAnd b, if the model parameters do not meet the requirements, continuing to adjust the model parameters, returning to the step 3.2), and continuing iterative computation to obtain the optimal parameters.
Through the training, the multi-model fusion based on the cross-correlation entropy is obtained, and the accurate classification result of the input images can be obtained. The experiment on the medical skin data set is to classify pathological sections of basal cell carcinoma and seborrheic keratosis, the image characteristics of the pathological sections are very similar and have more noise, and the results of image classification are compared by applying a plurality of single model networks and the multi-deep learning model fusion method based on cross-correlation entropy provided by the invention to the image classification, and the method is shown in figure 3, wherein a basal cell carcinoma image is arranged above a middle line, and a seborrheic keratosis image is arranged below the middle line. The dashed box indicates a sample classification error. Wherein, (a) is an original pathological image, (b) is a ResNet classification result, (c) is an Xception classification result, (d) is a DensNet classification result, and (e) is an MCC-based integration method classification result.
The image classification method based on deep learning of the invention learns hierarchical feature description through a supervised or unsupervised mode, replaces the traditional work of manually designing or selecting image features, a convolutional neural network in a deep learning model has achieved amazing performance in the image field, the convolutional neural network utilizes image pixel information as input, retains all information of the input image to the maximum extent, and performs feature extraction through convolution operation, and the end-to-end learning method achieves very good effect and is widely applied.
The mobile terminal provided by the embodiment of the invention. The mobile terminal of this embodiment includes: a processor, a memory, and a computer program stored in the memory and executable on the processor. The processor, when executing the computer program, implements the steps in the above-described cross-correlation entropy based multi-deep learning model fusion method embodiment, such as the steps shown in fig. 1. Alternatively, the processor implements the functions of the modules/units in the above device embodiments when executing the computer program.
The computer program may be partitioned into one or more modules/units that are stored in the memory and executed by the processor to implement the invention.
The mobile terminal can be a desktop computer, a notebook, a palm computer, a cloud server and other computing equipment. The storage device may include, but is not limited to, a processor, a memory.
The processor may be a Central Processing Unit (CPU), other general purpose processor, a Digital Signal Processor (DSP), an Application Specific Integrated Circuit (ASIC), an off-the-shelf programmable gate array (FPGA) or other programmable logic device, discrete gate or transistor logic, discrete hardware components, etc.
The mobile terminal integrated module/unit, if implemented in the form of a software functional unit and sold or used as a separate product, may be stored in a computer-readable storage medium. Based on such understanding, all or part of the flow of the method according to the embodiments of the present invention may also be implemented by a computer program, which may be stored in a computer-readable storage medium, and when the computer program is executed by a processor, the steps of the method embodiments may be implemented. Wherein the computer program comprises computer program code, which may be in the form of source code, object code, an executable file or some intermediate form, etc. The computer-readable medium may include: any entity or device capable of carrying the computer program code, recording medium, usb disk, removable hard disk, magnetic disk, optical disk, computer memory, Read-only memory (ROM), Random Access Memory (RAM), electrical carrier wave signals, telecommunications signals, software distribution medium, etc. It should be noted that the computer readable medium may contain content that is subject to appropriate increase or decrease as required by legislation and patent practice in jurisdictions, for example, in some jurisdictions, computer readable media does not include electrical carrier signals and telecommunications signals as is required by legislation and patent practice.
The memory may be used to store the computer programs and/or modules, and the processor may implement various functions of the mobile terminal by executing or executing the computer programs and/or modules stored in the memory and calling data stored in the memory.
The above-mentioned contents are only for illustrating the technical idea of the present invention, and the protection scope of the present invention is not limited thereby, and any modification made on the basis of the technical idea of the present invention falls within the protection scope of the claims of the present invention.

Claims (6)

1. A cross-correlation entropy based multi-deep learning model fusion method suitable for being executed in a computing device is characterized by comprising the following steps of:
step 1: preprocessing the data set;
step 2: respectively training a plurality of single model classification convolution neural networks to enable each model to obtain an optimal result;
and step 3: and (2) using a multi-model integration method for the single model with the optimal result, and learning the contribution weight of each single-model classification convolutional neural network to the final result by using a multi-model fusion method based on a cross-correlation entropy criterion, wherein the objective function is as follows:
Figure FDA0002194490140000011
wherein w ═ w1,w2,…,wn]Representing the weights of the integrated model, b is the bias, m is the number of samples, σ is the kernel width, xiRepresenting the output of the ith sample in the plurality of single models, yiA true tag representing the ith sample;
using a gradient-rise method to obtain a parameter w which brings the objective function close to a maximumjAnd b;
and 4, step 4: and (4) training the target function in the step (3) to obtain an optimal image classification result of the multi-depth convolution neural network integrated model based on the cross-correlation entropy criterion.
2. The fusion method of the multi-deep learning model based on the cross-correlation entropy of claim 1, wherein the specific method of the step 2 is as follows:
step 2-1: training a plurality of deep convolution neural network models by using the same training set X according to the collected data, firstly initializing model variables w and b, and obtaining a predicted value by network forward propagation;
step 2-2: and (3) performing back propagation by using a loss function and a function optimization method, updating model weights w and b, and calculating the formula as follows:
Figure FDA0002194490140000012
Figure FDA0002194490140000021
where λ represents the learning rate, wjRepresents the weight of the jth parameter, k represents the kth iteration,
Figure FDA0002194490140000022
updating the k-th weight value of the jth model, taking J (w, b) as an objective function, and applying a gradient descent method to the parameter w in the objective functionjB, updating the weight;
step 2-3: and (5) repeating the step 2-1 to the step 2-2 until the training times are reached or the training target is obtained to obtain the optimal model.
3. The fusion method of the multiple deep learning models based on the cross-correlation entropy of claim 2, wherein the multiple deep convolutional neural network models are one or more of VGG, IncepotionV 4, Xception, ResNet or DensNet.
4. The fusion method of the multi-deep learning model based on the cross-correlation entropy of claim 1, wherein the specific method of the step 3 is as follows:
step 3-1: for a plurality of existing single models, learning the contribution weight w of each single model to the final result by using an MCC-based multi-model integration method1,w2,…,wn]N is the number of single models;
step 3-2: taking the correlation entropy criterion as an objective function, the calculation formula is as follows:
step 3-3: and (3) maximizing a target function defined by the cross-correlation entropy by using the corresponding relation obtained in the step (3-2), wherein the contribution of each single model to a final result is calculated according to the following formula:
Figure FDA0002194490140000024
Figure FDA0002194490140000025
Figure FDA0002194490140000031
the entropy criterion is subjected to gradient rise, and the objective function is close to the maximum value after multiple iterations to achieve the learning parameter wjAnd purpose of b;
step 3-4: judging whether the classification precision of the current model meets the requirement, if so, outputting a parameter wjAnd b, if the model parameters do not meet the requirements, continuing to adjust the model parameters, returning to the step 3-2, continuing iterative computation to obtain the optimal parameters, and obtaining the final model.
5. A memory device having stored therein a plurality of instructions adapted to be loaded and executed by a processor:
step 1: preprocessing the data set;
step 2: respectively training a plurality of single model classification convolution neural networks to enable each model to obtain an optimal result;
and step 3: and (2) using a multi-model integration method for the single model with the optimal result, and learning the contribution weight of each single-model classification convolutional neural network to the final result by using a multi-model fusion method based on a cross-correlation entropy criterion, wherein the objective function is as follows:
wherein w ═ w1,w2,…,wn]Representing the weights of the integrated model, b is the bias, m is the number of samples, σ is the kernel width, xiRepresents the ithOutput of samples in the plurality of single models, yiA true tag representing the ith sample;
using a gradient-rise method to obtain a parameter w which brings the objective function close to a maximumjAnd b;
and 4, step 4: and (4) training the target function in the step (3) to obtain an optimal image classification result of the multi-depth convolution neural network integrated model based on the cross-correlation entropy criterion.
6. A mobile terminal, comprising:
a processor adapted to implement instructions; and
a storage device adapted to store a plurality of instructions, the instructions adapted to be loaded and executed by a processor to:
step 1: preprocessing the data set;
step 2: respectively training a plurality of single model classification convolution neural networks to enable each model to obtain an optimal result;
and step 3: and (2) using a multi-model integration method for the single model with the optimal result, and learning the contribution weight of each single-model classification convolutional neural network to the final result by using a multi-model fusion method based on a cross-correlation entropy criterion, wherein the objective function is as follows:
Figure FDA0002194490140000041
wherein w ═ w1,w2,…,wn]Representing the weights of the integrated model, b is the bias, m is the number of samples, σ is the kernel width, xiRepresenting the output of the ith sample in the plurality of single models, yiA true tag representing the ith sample;
using a gradient-rise method to obtain a parameter w which brings the objective function close to a maximumjAnd b;
and 4, step 4: and (4) training the target function in the step (3) to obtain an optimal image classification result of the multi-depth convolution neural network integrated model based on the cross-correlation entropy criterion.
CN201910843623.8A 2019-09-06 2019-09-06 Cross-correlation entropy based multi-depth learning model fusion method, terminal device and readable storage medium Pending CN110705600A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910843623.8A CN110705600A (en) 2019-09-06 2019-09-06 Cross-correlation entropy based multi-depth learning model fusion method, terminal device and readable storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910843623.8A CN110705600A (en) 2019-09-06 2019-09-06 Cross-correlation entropy based multi-depth learning model fusion method, terminal device and readable storage medium

Publications (1)

Publication Number Publication Date
CN110705600A true CN110705600A (en) 2020-01-17

Family

ID=69194727

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910843623.8A Pending CN110705600A (en) 2019-09-06 2019-09-06 Cross-correlation entropy based multi-depth learning model fusion method, terminal device and readable storage medium

Country Status (1)

Country Link
CN (1) CN110705600A (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111523592A (en) * 2020-04-21 2020-08-11 易拍全球(北京)科贸有限公司 Historical relic artwork field image similarity measurement algorithm based on deep learning
CN112101172A (en) * 2020-09-08 2020-12-18 平安科技(深圳)有限公司 Weight grafting-based model fusion face recognition method and related equipment
CN112434732A (en) * 2020-11-17 2021-03-02 西安交通大学 Deep learning classification method based on feature screening
CN113496442A (en) * 2020-03-19 2021-10-12 荷盛崧钜智财顾问股份有限公司 Graph representation generation system, graph representation generation method and graph representation intelligent module thereof
CN113537150A (en) * 2021-08-11 2021-10-22 西安交通大学 Hyperspectral image target anomaly detection method, system, terminal and storage medium

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
苗胜军: "《复杂岩体边坡变形与失稳预测研究》", 31 August 2016, 北京冶金工业出版社 *
许美凤 等: "多卷积神经网络模型融合的皮肤病识别方法", 《西安交通大学学报》 *

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113496442A (en) * 2020-03-19 2021-10-12 荷盛崧钜智财顾问股份有限公司 Graph representation generation system, graph representation generation method and graph representation intelligent module thereof
CN111523592A (en) * 2020-04-21 2020-08-11 易拍全球(北京)科贸有限公司 Historical relic artwork field image similarity measurement algorithm based on deep learning
CN111523592B (en) * 2020-04-21 2023-05-09 易元数字(北京)科技集团有限公司 Image similarity measurement method based on deep learning in cultural relic artwork field
CN112101172A (en) * 2020-09-08 2020-12-18 平安科技(深圳)有限公司 Weight grafting-based model fusion face recognition method and related equipment
WO2021155713A1 (en) * 2020-09-08 2021-08-12 平安科技(深圳)有限公司 Weight grafting model fusion-based facial recognition method, and related device
CN112101172B (en) * 2020-09-08 2023-09-08 平安科技(深圳)有限公司 Weight grafting-based model fusion face recognition method and related equipment
CN112434732A (en) * 2020-11-17 2021-03-02 西安交通大学 Deep learning classification method based on feature screening
CN113537150A (en) * 2021-08-11 2021-10-22 西安交通大学 Hyperspectral image target anomaly detection method, system, terminal and storage medium
CN113537150B (en) * 2021-08-11 2024-04-02 西安交通大学 Hyperspectral image target anomaly detection method, system, terminal and storage medium

Similar Documents

Publication Publication Date Title
Albattah et al. A novel deep learning method for detection and classification of plant diseases
Tang et al. Deepchart: Combining deep convolutional networks and deep belief networks in chart classification
US11836603B2 (en) Neural network method and apparatus with parameter quantization
CN110705600A (en) Cross-correlation entropy based multi-depth learning model fusion method, terminal device and readable storage medium
CN111507378A (en) Method and apparatus for training image processing model
US8954365B2 (en) Density estimation and/or manifold learning
Ahmed et al. Multi-objects detection and segmentation for scene understanding based on Texton forest and kernel sliding perceptron
Ashok Kumar et al. Enhanced facial emotion recognition by optimal descriptor selection with neural network
Shiloh-Perl et al. Introduction to deep learning
Wu et al. Research on lightweight infrared pedestrian detection model algorithm for embedded Platform
Xiao et al. Apple ripeness identification from digital images using transformers
Goumiri et al. A new hybrid model of convolutional neural networks and hidden Markov chains for image classification
US20220188605A1 (en) Recurrent neural network architectures based on synaptic connectivity graphs
Brahim et al. RNN-and CNN-based weed detection for crop improvement: An overview
Demertzis et al. Large-scale geospatial data analysis: Geographic object-based scene classification in remote sensing images by GIS and deep residual learning
Jeyalakshmi et al. Accurate liver disease prediction system using convolutional neural network
Zhao et al. Spectral–spatial classification of hyperspectral images using trilateral filter and stacked sparse autoencoder
Lin et al. LA-Net: LSTM and attention based point cloud down-sampling and its application
CN116152587A (en) Training method of expression recognition model, facial expression recognition method and facial expression recognition device
Shang et al. Real-time Accurate Object Counting for Smart Farms
CN114140524A (en) Closed loop detection system and method for multi-scale feature fusion
Zhang et al. Research on multimedia image classification technology based on chaos optimization machine learning algorithm
Struckmeier et al. Autoencoding slow representations for semi-supervised data-efficient regression
Asiri Unmanned aerial vehicles assisted rice seedling detection using shark smell optimization with deep learning model
Sreenivasulu et al. Adaptive inception based on transfer learning for effective visual recognition

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20200117