CN111553475A - High-dimensional multi-mode evolution optimization method based on random embedding technology - Google Patents

High-dimensional multi-mode evolution optimization method based on random embedding technology Download PDF

Info

Publication number
CN111553475A
CN111553475A CN202010341693.6A CN202010341693A CN111553475A CN 111553475 A CN111553475 A CN 111553475A CN 202010341693 A CN202010341693 A CN 202010341693A CN 111553475 A CN111553475 A CN 111553475A
Authority
CN
China
Prior art keywords
dimensional
population
optimization
subtasks
matrix
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
CN202010341693.6A
Other languages
Chinese (zh)
Inventor
候亚庆
曹雨梦
杨鑫
张强
魏小鹏
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Dalian University of Technology
Original Assignee
Dalian University of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Dalian University of Technology filed Critical Dalian University of Technology
Priority to CN202010341693.6A priority Critical patent/CN111553475A/en
Publication of CN111553475A publication Critical patent/CN111553475A/en
Withdrawn legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2411Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on the proximity to a decision surface, e.g. support vector machines
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/004Artificial life, i.e. computing arrangements simulating life
    • G06N3/006Artificial life, i.e. computing arrangements simulating life based on simulated virtual individual or collective life forms, e.g. social simulations or particle swarm optimisation [PSO]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/12Computing arrangements based on biological models using genetic models
    • G06N3/126Evolutionary algorithms, e.g. genetic algorithms or genetic programming

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Data Mining & Analysis (AREA)
  • Biophysics (AREA)
  • Health & Medical Sciences (AREA)
  • General Physics & Mathematics (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Computation (AREA)
  • General Engineering & Computer Science (AREA)
  • Molecular Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Software Systems (AREA)
  • Biomedical Technology (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Evolutionary Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physiology (AREA)
  • Genetics & Genomics (AREA)
  • Image Analysis (AREA)

Abstract

The invention belongs to the crossing field of evolution calculation and machine learning, and relates to a high-dimensional multi-mode evolution optimization method based on a random embedding technology, which can be used for automatic parameter adjustment of a multi-classification Support Vector Machine (SVM) model. The invention is an improved optimization algorithm which is obtained based on a multi-time optimization parallel operation and is used for carrying out transfer learning among a plurality of subtasks by adopting a self-coding technology so as to improve the optimization efficiency, can quickly and effectively optimize and adjust a large number of parameters at the same time, and is very suitable for the super-parameter adjustment of a large-scale machine learning model. According to the method, a plurality of low-dimensional random embeddings are generated, each random embedding can be used as a substitute representation of an original high-dimensional task, uniqueness and correlation existing among subtasks are utilized, and beneficial transfer learning among subtasks is carried out, so that on one hand, the searching efficiency of the optimal solution is effectively improved, and meanwhile, the time consumption of multiple times of operation is avoided.

Description

High-dimensional multi-mode evolution optimization method based on random embedding technology
Technical Field
The invention belongs to the crossing field of evolution calculation and machine learning, and relates to a high-dimensional multi-mode evolution optimization method based on a random embedding technology, which can be used for automatic parameter adjustment of a multi-classification Support Vector Machine (SVM) model.
Background
In clinical work, medical images provide important auxiliary information for clinical decision making, but traditional image diagnosis is mainly based on subjective judgment of radiologists, diagnosis accuracy depends on personal experience of experts, and due to the fact that the level of doctors is limited, misdiagnosis is easily caused in many hospitals, and subsequent treatment is affected. If the computer-aided doctor can be used for detection, the diagnosis accuracy can be greatly improved, and the doctor is helped to make a subsequent treatment scheme.
In the aspect of computer-aided diagnosis, the current main method is to finely adjust a deep neural network of a natural picture and then use the fine adjustment for a classification task of medical images (Kermany, Goldbaum.et al in 2018). However, the features of the medical image are not rich in natural pictures, and the difference between different pictures is relatively small, so that the classification task of the medical image is more biased to fine-grained classification, a very deep neural network is not needed when a model is constructed, and conversely, if the network scale is enlarged to improve the classification accuracy rate by only deepening the neural network, overfitting is easily caused, and the classification effect is reduced. Compared with a deep neural network, the support vector machine is a traditional machine learning classification model, although the scale of processable data is small, the model classification speed is high, local optimization is not caused, and the support vector machine is combined with the neural network and is very suitable for being used as a medical image diagnosis task.
However, in a multi-disease diagnosis system constructed by using a support vector machine, many hyper-parameters need to be adjusted, and the problem of hyper-parameter tuning of a model is also a significant challenge in the field of machine learning. In the modeling process, not only how to build an optimal network structure needs to be considered, but also finding an optimal hyper-parameter combination is a very important task, and because parameter adjustment of the machine learning model has no fixed rule and mode, parameter adjustment can only be performed depending on human experience and random attempts, and manpower and time are consumed. With the continuous expansion of the super parameter dimension, the super parameter tuning becomes an NP difficult problem, and the optimal parameter combination is difficult to find only by manpower. In the development process of the machine learning algorithm, most of the time is used for constructing an optimal model, namely, an optimal hyper-parameter combination is searched, no matter manual parameter adjustment is carried out, or methods such as random search and grid search are used, a large amount of time is consumed, and the time cost is exponentially increased along with the amplification of the parameter number. Therefore, if automated hyper-parametric tuning can be achieved, modeling efficiency would be greatly improved.
The multi-classification support vector machine hyperparameter tuning essentially belongs to a combined optimization problem, and for the optimization problem, a widely applied method is an evolutionary algorithm. The evolutionary algorithm is a metaheuristic optimization algorithm based on population, and is inspired by Darwinian biological evolution theory. The genetic algorithm, the differential evolution and the like are easy to realize, have strong universality and are widely applied to solving the optimization problem. Although the evolutionary algorithm has wide potential applicability, previous researches show that the efficiency of the genetic algorithm is gradually reduced along with the increase of the problem dimension, so that if the existing genetic algorithm is expanded to the problem of high-dimensional hyperparameter optimization, the most direct solution idea is to reduce the dimension of the hyperparameter dimension through a dimension reduction technology and then optimize the hyperparameter dimension by applying the genetic algorithm. Among this class of methods, decomposition-based and embedding-based methods are currently the two most dominant technical approaches.
Among them, decomposition-based approaches have been successfully applied to separable problems, i.e., a high-dimensional search space can be decomposed into multiple low-dimensional sub-problems (Kandasamy, Schneider, and P' oczos 2015, Friesen and Domingos (2015)). However, due to the nature of the technology, the application of the technology to the hyper-parameter optimization problem of the machine learning model to be solved is not simple, because the hyper-parameters of the machine learning are usually not independent of each other, and often have complex dependency relationships, and the optimization needs to be carried out simultaneously.
On the other hand, considering that some decision variables do not have significant influence on the target problem in the optimization problem, the embedding method (random projection, random embedding, and so on Chen, Krause, and Castro (2012), Carpentier, Munos, and others 2012, (Wang et al 2013; Kaban, Bootkrajang, and Durrant2013, (Qian, Hu, and Yu 2016) can be used to perform efficient search on the high-dimensional space.
Research shows that for a high-dimensional model parameter optimization problem in machine learning, parameters of most dimensions do not significantly influence a target problem. I.e., these problems are all characterized by a low effective dimension, the mathematical description for solving such problems with random embedding is as follows: the target function f (X) X → RDEffective dimension de(de<D) If there is one dimension deThe linear subspace Ψ is such that
Figure BDA0002467649500000031
Then
f(x)=f(xe+x)=f(xe)
Wherein the content of the first and second substances,
Figure BDA0002467649500000032
the orthogonal complement of Ψ is represented. Ψ means the effective subspace of f, ΨRefers to the constant subspace of f. The equation represents the function f for the dimension deThe variation in the subspace Ψ (effective subspace) is very sensitive, but at the subspace ΨFunction value table on (constant subspace)Hardly changed. On the premise, the random embedding method can solve the optimization problem with high dimensionality. In particular, consider an effective dimension of deFunction f (X) X → R, random embedded matrix M ∈ RD×dEach element of the matrix is independently sampled in a normal distribution N (0,1), where D>d≥deThus for any x ∈ RDThere must be one solution y ∈ RdThus, therefore, it is
f(x)=f(My)
For any maximum x*∈RDMust exist in y*∈RdSo that f (My)*)=f(x*). Accordingly, the optimization problem f (x) is described as follows:
Figure BDA0002467649500000033
wherein g (y) ═ f (my). We can therefore optimize the low dimensional objective function g (y) and then use the matrix M to demap back into the high dimensional search space.
In order to make the evolutionary algorithm feasible, the boundary of the low-dimensional search space needs to be further defined. After mapping back to the high dimensional space, the point beyond the boundary is fixed on the boundary.
Figure BDA0002467649500000041
However, due to the randomness of the random embedding method, sometimes the optimal solution is not necessarily included in the compressed space. According to the method and assumption of random embedding, the probability that the optimal solution has γ (non-zero) does not fall within the boundaries of the generated low-dimensional subspace, and thus it is difficult to determine whether the randomly generated low-dimensional space is valid. Therefore, the random embedding method is not ideal for the machine learning model parameter adjustment problem.
Existing approaches overcome this problem by considering multiple random embeddings. Specifically, in the hyper-parametric optimization problem of the machine learning model, N different low-dimensional searched parameter spaces are independently sampled, and each parameter space is optimized to reduce the failure probability, that is, the probability that the optimal solution of the target problem is not contained in any one of the N low-dimensional spaces is γ (N), but because each search optimization is run from the beginning, no potential relation among the low-dimensional spaces is utilized, parallel optimization of the multiple independent spaces is very time-consuming, and especially under the condition that the computational resources are limited and the parallel optimization cannot be performed at high level, the defect is more significant. The biggest defect of the existing technical scheme is that the time consumption between sequence optimization is large due to the fact that the relevance between search results obtained by running of the existing technical scheme is not fully utilized after the existing technical scheme is run for multiple times.
Disclosure of Invention
In order to solve the problems, the high-dimensional multi-mode evolution optimization method provided by the invention is mainly based on a random embedding technology, joint evolution search is carried out on a plurality of low-dimensional spaces, gene migration is obtained by using a cross method, so that a target task can simultaneously obtain beneficial information of a plurality of low-dimensional subtasks, the relevance among the subtasks is fully utilized, the defect of optimization search caused by the randomness of random embedding is overcome, the search efficiency of optimal parameters is greatly improved, and the optimization process of the parameters is accelerated.
The purpose of the invention is realized by the following technical scheme:
a high-dimensional multi-mode evolution optimization method based on a random embedding technology is an improved optimization algorithm which is based on a multi-mode obtained by parallel operation of multiple times of optimization, namely, a self-coding technology is adopted among a plurality of subtasks to carry out transfer learning among the subtasks so as to improve the optimization efficiency, and comprises the following steps:
(1) the specific target value of the single-target optimization problem to be solved is represented by f (x), all parameters to be optimized are represented by a long vector x, and each dimension of x corresponds to a parameter to be optimized (namely, the optimization x finds the value of x for minimizing f (x)), and f (x) corresponding to each x can be calculated according to the specific single-target optimization problem.
(2) A classical genetic algorithm in an evolutionary algorithm is used as a basic optimization algorithm, and a random embedding thought is combined to optimize a target value; the parameters are set as follows:
setting the corresponding parameter values for the genetic algorithm: the number N of subtasks, the population size K of each subtask, the maximum iteration frequency generation or maximum evaluation frequency FEVs of the iteration stop condition, and the cross probability PCAnd the mutation probability PM
Setting corresponding parameter values for the random embedding idea: the dimensionality of the high-dimensional optimization problem, namely the number D of parameters to be adjusted, the dimensionality D of each subtask, and the effective dimensionality D of the whole problem set according to a random embedding theoryeAlgebraic interval G for occurrence of transfer learningw
(3) Initializing a population P of size K at low latitudes using the idea of random embedding, P ═ P1,p2,...,pK-wherein each individual randomly corresponds to a low-dimensional representation and each individual is uniformly randomly assigned to a respective subtask, thereby obtaining
Figure BDA0002467649500000051
Where N is the number of low-dimensional tasks generated.
(4) Initializing random embedding matrix { M1,M2,…MNTherein of
Figure BDA0002467649500000052
Wherein d ismultitaskIs the upper limit of the dimension of each low-dimensional subtask, so that the random embedding matrix { M ] is utilized by the random embedding method1,M2,…MNH, solve the high-dimensional optimization problem f (x), x ∈ RDThe random embedding idea is to replace the optimized original high-dimensional equation f (X) with the optimized low-dimensional equation g (y), wherein the DNA represents the gene material chromosome, namely the individual to be optimized, and finally the obtained low-dimensional DNA is substituted into a genetic algorithm to carry out genetic operation evolution, so that a plurality of low-dimensional search spaces y ∈ y are obtainedd_multitaskInstead of in the original high-dimensional space by parallel search optimizationThe optimization is carried out.
(5) The searching efficiency is improved by carrying out transfer learning on any two subtasks through a self-coding technology. Because different low-dimensional tasks have different perception views on the original high-dimensional task in the multi-mode optimization of the multi-task environment, the genetic material in one task cannot be directly used by other tasks, and the invention provides a self-coding technology for generating a corresponding mapping matrix W between any two subtasks. Assuming that the populations of the two subtasks are,
Figure BDA0002467649500000061
and
Figure BDA0002467649500000062
wherein Q separately optimizes two low-dimensional tasks Θi,ΘjA mapping matrix W between the twoijIs composed of (W)ijPi-Pj) Expressed in the least mean square of, e.g. formula
Figure BDA0002467649500000063
For this simple optimization problem, the least squares method can be used
Figure BDA0002467649500000064
Wherein the content of the first and second substances,
Figure BDA0002467649500000065
finger PiTranspose of (W)ijProviding a low-dimensional task ΘiAnd thetajThe connection mode between the two. Thus, PiThe solution obtained by population optimization may be obtained by multiplying by a matrix WijIs mapped to a population PjIn (1). If P isiAnd PjThe dimension of (2) is different, and the W matrix is generated by firstly carrying out zero filling operation on the shorter chromosome codes and then calculating. And obtaining a migration matrix between every two subtasks, so that the population can be subjected to migration learning.
(6) And generating a child population C1 by cross variation in the population of each subtask.
(7) And (4) applying the self-coding matrix W to perform migration learning among subtasks, and generating a child population C2 obtained by the migration learning for the population of each subtask.
(8) And merging the parent population P and the child populations C1 and C2, calculating the target value of each individual in the merged population, and selecting the K individuals with the best target values before by adopting an elite strategy according to the target values to update the population P.
(9) And (5) repeating the steps (5) to (8) until an iteration stop condition is reached, namely a set maximum evaluation time is reached, and finally obtaining a parameter value corresponding to each individual in the population P.
Compared with the prior art, the invention has the beneficial effects that: the invention is an effective high-dimensional optimization method, namely, a large number of parameters can be optimized and adjusted quickly and effectively at the same time, so the method is very suitable for the super-parameter adjustment of large-scale machine learning models, such as the super-parameter adjustment of neural network models, the super-parameter adjustment of SVM models and the like. Aiming at the optimization problem of high dimension and low effective dimension, the method provides an evolutionary multitask multi-mode optimization method based on a random embedding dimension reduction means. The method generates a plurality of low-dimensional random embeddings, and each random embeddings can be used as a substitute expression of an original (high-dimensional) task; and the uniqueness and the correlation existing among all the subtasks are utilized to carry out beneficial transfer learning among the subtasks, so that on one hand, the searching efficiency of the optimal solution is effectively improved, and meanwhile, the time consumption of multiple operations is avoided.
Drawings
Fig. 1 is a block diagram of a chest disease diagnosis system.
Fig. 2 is a block diagram of the feature extraction network VGG 16.
FIG. 3 is a flowchart of an SVM multi-classifier auto-parameter adjustment.
Fig. 4 is a schematic diagram of a random embedding process.
Detailed description of the preferred embodiments
The following further describes a specific embodiment of the present invention with reference to the drawings and technical solutions.
The method can be applied to the super-parameter adjusting process in a plurality of large-scale machine learning models, such as the super-parameter adjusting process of a large-scale neural network model or SVM model. The embodiment is particularly applied to a medical diagnosis system for diagnosing chest diseases, as shown in fig. 1. The chest disease diagnosis system adopts a multi-classification Support Vector Machine (SVM) as a model, carries out super-parameter adjustment on the SVM model by a high-dimensional multi-mode evolution optimization method, and trains the optimized SVM model to obtain the diagnosis system for chest disease diagnosis. The specific embodiments discussed are merely illustrative of implementations of the invention and do not limit the scope of the invention. The following describes embodiments of the present invention mainly for the problem of the chest disease diagnosis system, and specifically includes the following steps:
(1) breast medical image feature extraction module
The medical images of the chest in the data set are input into the existing pre-trained classical feature extraction neural network VGG16, and are used for extracting corresponding feature maps in the medical images and connecting the feature maps into a long feature vector. As shown in fig. 2: the structure of the pre-trained feature extraction network VGG16 is characterized in that a medical image is input into 2 3 x 3 convolutional layers (Cov) connected in series to perform primary feature extraction, then a pooling layer (Pool) is connected to compress an obtained feature map, the feature map is input into 3 series 3 x 3 convolutional layers (Cov) to perform final feature extraction after the feature map is repeated twice, finally the feature map is input into 3 series full-connection layers (Fc) to obtain a final long feature vector as an input vector of an SVM, and the feature extraction step of the medical image is completed.
(2) Basic setting is carried out on SVM models of multi-classification support vector machine
SVM is a classical classification model in machine learning, and the classification task refers to the training set with labels
Figure BDA0002467649500000081
A classifier is trained so that it can be applied to unlabeled data. The training objective of the SVM is to minimize the loss function:
Figure BDA0002467649500000082
u, 1, to getThe SVM model with better classification effect is obtained. Where w is the parameter matrix of the SVM model, λ is a constant coefficient, U is the number of samples, yiIs the classification label of the ith sample, xiIs the data of the ith sample, and b is the bias term of the SVM model.
For the diagnosis problem of chest diseases, as shown in table 1, the medical images of the chest need to be analyzed to determine whether the patient has the following 14 diseases: atelectasis, consolidation, infiltration, pneumothorax, edema, emphysema, fibrosis, effusion, pneumonia, pleural thickening, cardiac hypertrophy, nodules, mass, hernia, or a combination thereof. When the classification problem is solved by using the SVM model, the hyper-parameters of the model greatly influence the balance between two targets of errors and edge maximization of the SVM model on a training set. When the SVM model solves the multi-classification task, a common method is a one-to-one (OvO) strategy, and the multi-classification problem is converted into a plurality of two-classification problems. In a class H classification problem, training is performed according to a one-to-one (OvO) strategy
Figure BDA0002467649500000083
And two classifiers. Since 14 classes are trained in the chest disease diagnosis system, the SVM model used will be
Figure BDA0002467649500000091
The individual hyper-parameters need to be adjusted. And integrating 91 hyper-parameters to be optimized into a long vector x, wherein the value of each dimension in the vector x represents the value of one hyper-parameter, and the specific value range of the hyper-parameter adopts the commonly used SVM hyper-parameter range. And substituting each vector x, namely a group of hyper-parameters in a fixed value range, into a multi-classification Support Vector Machine (SVM) model for training to finally obtain a classification model with convergent training, and then representing the training effect of the group of hyper-parameters, namely an optimized target value, by testing the accuracy value tested on the classification model by a data set, so that the model with higher accuracy corresponds to a group of hyper-parameters with better optimization effect.
(3) Carrying out automatic parameter adjustment and optimization process on the SVM model (as shown in figure 3)
(3.1) basic setting of genetic Algorithm
The random embedding technology-based high-dimensional multi-mode evolution optimization method is applied to solving the problem of super-parameter adjustment of the multi-classification support vector machine SVM model, and a group of optimized super-parameters is obtained and substituted into the multi-classification support vector machine SVM model. All hyper-parameters to be optimized in the SVM model are integrated into a long vector x, and the accuracy of the SVM model corresponding to each x on a test set after the SVM model is finally trained to be converged is used as an optimization target value f (x). And (3) using a genetic algorithm GA as a basic evolutionary algorithm, and carrying out heuristic optimization based on different population quantities. Setting the number of the low-dimensional subtasks to be N-3, setting the population size of the whole problem to be K-30, setting the maximum iteration frequency generation or the maximum evaluation frequency FEVs to be FEs-300 under the iteration stop condition, and setting the cross probability P to be 300CProbability of mutation PMAnd the like. Setting corresponding parameter values for the random embedding idea: the dimension of the high-dimensional optimization problem, namely the number D of the hyper parameters to be adjusted is 91, the dimension D of each subtask is set to be 1/3 of the dimension D of the original problem, D is 30, and the effective dimension D of the whole problem is set according to the random embedding theorye10, the algebraic interval at which the transition occurs is set to Gw=1。
(3.2) multiple Low-dimensional subtask Generation modules
Initializing a series of random embedding matrixes (M) subject to normal distribution according to the idea of random embedding1,M2,…MNTherein of
Figure BDA0002467649500000101
Where D is the original high dimension, DmultitaskThe maximum value of the upper limit of the effective dimensions of the N subtasks, and the high-dimensional optimization problem g (y) ═ f (My) ═ f (x), x ∈ R is mapped by a matrixDConverting into a series of low-dimensional subproblems g (y), y ∈ yd_multitask
(3.3) multiple sub-population generating modules
Initializing a population P of size K at low latitudes, P ═ P1,p2,...,pKEach of which randomly corresponds to a low-dimensional expression and each of which is associated with a low-dimensional expressionThe volumes are uniformly and randomly distributed to the subtasks, thus obtaining
Figure BDA0002467649500000102
Where N is the number of low-dimensional subtasks generated.
(3.4) population evolution optimization module
Repeating the following operations until an iteration stop condition is reached, namely a set maximum evaluation number is reached: calculating and generating a migration matrix W matrix through a self-coding technology, wherein in each generation: firstly, evaluating each individual in the population, and calculating a target value f (x) of each individual in the population (namely, a training model module in the step (4)); then generating a child population C1 by cross variation in the population of each subtask; and then applying the matrix W obtained by self coding to execute the migration learning among the subtasks, and generating a child population C2 obtained by the migration learning for the population of each subtask. Calculating the target value (namely a training model module) of each individual in the offspring populations C1 and C2; finally, merging the parent population P and the child populations C1 and C2, and selecting the first K best individuals to update the population P by adopting an elite strategy according to the target value of each individual; substituting the finally obtained super-parameter value corresponding to each individual in the population P into the SVM model, wherein the SVM model is the finally obtained SVM model with higher precision and can be applied to medical diagnosis of chest diseases.
(4) Training model module
Substituting the individuals of which the target values are to be calculated, namely a group of fixed super-parameter values x, into the SVM model to be trained, and inputting the feature vectors extracted by the training data set through the feature extraction module into the SVM model for training until convergence; inputting the test data set into the trained SVM model for testing, and calculating the accuracy of the model on the test data set as the optimized target value f (x) of the individual x.

Claims (1)

1. A high-dimensional multi-mode evolution optimization method based on a random embedding technology is characterized by comprising the following steps:
(1) the specific target value of the single-target optimization problem to be solved is represented by f (x), all parameters to be optimized are represented by a long vector x, and each dimension of x corresponds to a parameter to be optimized
(2) A genetic algorithm is adopted as a basic optimization algorithm, and a random embedding thought is combined to optimize a target value; the parameters are set as follows:
setting the corresponding parameter values for the genetic algorithm: the number N of subtasks, the population size K of each subtask, the maximum iteration frequency generation or maximum evaluation frequency FEVs of the iteration stop condition, and the cross probability PCAnd the mutation probability PM
Setting corresponding parameter values for the random embedding idea: the dimensionality of the high-dimensional optimization problem, namely the number D of parameters to be adjusted, the dimensionality D of each subtask, and the effective dimensionality D of the whole problem set according to a random embedding theoryeAlgebraic interval G for occurrence of transfer learningw
(3) Initializing a population P of size K at low latitudes using the idea of random embedding, P ═ P1,p2,...,pK-wherein each individual randomly corresponds to a low-dimensional representation and each individual is uniformly randomly assigned to a respective subtask, thereby obtaining
Figure FDA0002467649490000011
Wherein N is the number of generated low-dimensional tasks;
(4) initializing random embedding matrix { M1,M2,…MNTherein of
Figure FDA0002467649490000012
Wherein d ismultitaskIs the upper limit of each low-dimensional subtask dimension, using a random embedding matrix { M }1,M2,…MNH, solve the high-dimensional optimization problem f (x), x ∈ RD(ii), translating into a low dimensional sub-problem g (y), g (y) f (my) f (x);
(5) a corresponding mapping matrix W is generated between any two subtasks through a self-coding technology, and migration learning is carried out to improve the search efficiency: assuming that the populations of the two subtasks are,
Figure FDA0002467649490000013
Figure FDA0002467649490000014
and
Figure FDA0002467649490000015
wherein Q separately optimizes two low-dimensional tasks Θi,ΘjA mapping matrix W between the twoijIs composed of (W)ijPi-Pj) Expressed in least mean square, i.e.
Figure FDA0002467649490000021
Then obtaining by least square method
Figure FDA0002467649490000022
Wherein the content of the first and second substances,
Figure FDA0002467649490000023
finger PiTranspose of (W)ijProviding a low-dimensional task ΘiAnd thetajThe connection mode between the two; thus, PiThe solution obtained by population optimization is multiplied by a matrix WijIs mapped to a population PjPerforming the following steps; if P isiAnd PjThe dimensions of the W matrix are different, and the W matrix is generated by firstly carrying out zero filling operation on the shorter chromosome codes and then calculating; obtaining a migration matrix between every two subtasks, namely performing migration learning on the population;
(6) generating a child population C1 by cross variation in the population of each subtask;
(7) performing migration learning among subtasks by applying a matrix W obtained by self-coding, and generating a child population C2 obtained by migration learning for the population of each subtask;
(8) merging the parent population P and the child populations C1 and C2, calculating the target value of each individual in the merged population, and selecting the K individuals with the best target values to update the population P according to the target values by adopting an elite strategy;
(9) and (5) repeating the steps (5) to (8) until an iteration stop condition is reached, namely a set maximum evaluation time is reached, and finally obtaining a parameter value corresponding to each individual in the population P.
CN202010341693.6A 2020-04-26 2020-04-26 High-dimensional multi-mode evolution optimization method based on random embedding technology Withdrawn CN111553475A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010341693.6A CN111553475A (en) 2020-04-26 2020-04-26 High-dimensional multi-mode evolution optimization method based on random embedding technology

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010341693.6A CN111553475A (en) 2020-04-26 2020-04-26 High-dimensional multi-mode evolution optimization method based on random embedding technology

Publications (1)

Publication Number Publication Date
CN111553475A true CN111553475A (en) 2020-08-18

Family

ID=72001194

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010341693.6A Withdrawn CN111553475A (en) 2020-04-26 2020-04-26 High-dimensional multi-mode evolution optimization method based on random embedding technology

Country Status (1)

Country Link
CN (1) CN111553475A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112508248A (en) * 2020-11-26 2021-03-16 山东大学 Integrated energy system day-ahead optimization method and system based on heterogeneous transfer learning
CN113589810A (en) * 2021-07-26 2021-11-02 南方科技大学 Intelligent dynamic autonomous obstacle avoidance movement method and device, server and storage medium

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112508248A (en) * 2020-11-26 2021-03-16 山东大学 Integrated energy system day-ahead optimization method and system based on heterogeneous transfer learning
CN113589810A (en) * 2021-07-26 2021-11-02 南方科技大学 Intelligent dynamic autonomous obstacle avoidance movement method and device, server and storage medium
CN113589810B (en) * 2021-07-26 2024-04-30 南方科技大学 Dynamic autonomous obstacle avoidance movement method and device for intelligent body, server and storage medium

Similar Documents

Publication Publication Date Title
Cai et al. Efficient architecture search by network transformation
Rani Parallel approach for diagnosis of breast cancer using neural network technique
Zhang et al. A novel fuzzy hybrid quantum artificial immune clustering algorithm based on cloud model
Muangkote et al. Rr-cr-IJADE: An efficient differential evolution algorithm for multilevel image thresholding
Zemmal et al. A new hybrid system combining active learning and particle swarm optimisation for medical data classification
CN108764280A (en) A kind of medical data processing method and system based on symptom vector
CN111553475A (en) High-dimensional multi-mode evolution optimization method based on random embedding technology
Chen et al. Binarized neural architecture search for efficient object recognition
Zheng et al. Dynamic distribution pruning for efficient network architecture search
CN113449802A (en) Graph classification method and device based on multi-granularity mutual information maximization
Liu et al. EACP: An effective automatic channel pruning for neural networks
Khan et al. Convolutional neural network based SARS-CoV-2 patients detection model using CT images
CN115393632A (en) Image classification method based on evolutionary multi-target neural network architecture structure
Weber et al. Automated labeling of electron microscopy images using deep learning
Zhan et al. Deep model compression via two-stage deep reinforcement learning
Khan et al. A multi-perspective revisit to the optimization methods of Neural Architecture Search and Hyper-parameter optimization for non-federated and federated learning environments
CN114093445A (en) Patient screening and marking method based on multi-label learning
CN108921853A (en) Image partition method based on super-pixel and clustering of immunity sparse spectrums
CN112668633A (en) Adaptive graph migration learning method based on fine granularity field
Jensen et al. Feature grouping-based fuzzy-rough feature selection
CN115249312A (en) Semi-supervised medical image classification method based on class decoupling distribution alignment
Hassan et al. A missing data imputation method based on salp swarm algorithm for diabetes disease
Soni et al. HMC: A hybrid reinforcement learning based model compression for healthcare applications
CN115063374A (en) Model training method, face image quality scoring method, electronic device and storage medium
Chandra et al. Bayesian multi-task learning for dynamic time series prediction

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
WW01 Invention patent application withdrawn after publication

Application publication date: 20200818

WW01 Invention patent application withdrawn after publication