CN109460862B - Method for solving multi-objective optimization problem based on MAB (multi-object-based) hyperheuristic algorithm - Google Patents

Method for solving multi-objective optimization problem based on MAB (multi-object-based) hyperheuristic algorithm Download PDF

Info

Publication number
CN109460862B
CN109460862B CN201811230929.8A CN201811230929A CN109460862B CN 109460862 B CN109460862 B CN 109460862B CN 201811230929 A CN201811230929 A CN 201811230929A CN 109460862 B CN109460862 B CN 109460862B
Authority
CN
China
Prior art keywords
value
algorithm
heuristic
operator
low
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201811230929.8A
Other languages
Chinese (zh)
Other versions
CN109460862A (en
Inventor
张淑艳
杨太龙
郭一
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhengzhou University
Original Assignee
Zhengzhou University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhengzhou University filed Critical Zhengzhou University
Priority to CN201811230929.8A priority Critical patent/CN109460862B/en
Publication of CN109460862A publication Critical patent/CN109460862A/en
Application granted granted Critical
Publication of CN109460862B publication Critical patent/CN109460862B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/04Forecasting or optimisation specially adapted for administrative or management purposes, e.g. linear programming or "cutting stock problem"
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/004Artificial life, i.e. computing arrangements simulating life
    • G06N3/006Artificial life, i.e. computing arrangements simulating life based on simulated virtual individual or collective life forms, e.g. social simulations or particle swarm optimisation [PSO]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/06Resources, workflows, human or project management; Enterprise or organisation planning; Enterprise or organisation modelling
    • G06Q10/063Operations research, analysis or management
    • G06Q10/0639Performance analysis of employees; Performance analysis of enterprise or organisation operations
    • G06Q10/06393Score-carding, benchmarking or key performance indicator [KPI] analysis

Abstract

The invention relates to the field of a heuristic algorithm, in particular to a method for solving a multi-target optimization problem based on a MAB heuristic algorithm, which takes an MAB strategy as a learning strategy, uses four performance evaluation mechanisms to evaluate the performance of each low-level heuristic operator, and better combines the advantages of each low-level heuristic operator through the learning and selecting mechanisms; the algorithm is used for carrying out experiments on a continuous multi-objective optimization problem set WFG, and good experimental results are obtained. The invention aims to solve the problems that the universality is poor during the design of the traditional heuristic algorithm and the effect of a single heuristic algorithm on some problem examples is poor, and provides a method for solving a multi-target optimization problem based on a super-heuristic algorithm of an MAB (multi-object-based optimization).

Description

Method for solving multi-objective optimization problem based on MAB (multi-object-based) hyperheuristic algorithm
Technical Field
The invention relates to the field of a super-heuristic algorithm, in particular to a method for solving a multi-objective optimization problem based on a MAB (multi-object-based) super-heuristic algorithm.
Background
The multi-objective optimization problem is used as a class of NP difficult-to-solve problems, and a large number of algorithms are used for solving the class of problems. Researchers often resort to heuristic algorithms tailored to the problem to obtain an acceptable solution in a reasonable amount of time. The heuristic algorithm has the advantages that domain knowledge can be conveniently integrated according to problems, and the disadvantage that the universality is poor, and a specialized algorithm needs to be designed aiming at the problems. The design of the heuristic algorithm has strong problem correlation, so that the workload of algorithm design is increased. In addition, a single heuristic may work best on some problem instances, but poorly on others, i.e., it may not be guaranteed that a single heuristic will achieve high quality results on all problem instances.
The problem can be solved by taking a Hyper-heuristic algorithm (HH) as a search method for selecting a heuristic operator or constructing the heuristic operator. A search problem and a set of heuristics associated with the problem (referred to as a set of low-level heuristics) are given. The hyper-heuristic algorithm is not used for directly searching the neighborhood space of the problem, but is used as a high-level heuristic strategy to increase the searching level to search the neighborhood space of the low-level heuristic operator. In the searching process, the hyper-heuristic algorithm selects and applies a proper low-level heuristic operator from the low-level heuristic operator set according to different solving states, and finally the selected low-level heuristic operator forms a low-level heuristic operator sequence. The method can combine the advantages of the low-level heuristic operators and avoid the disadvantages of the low-level heuristic operators to a certain extent. The solution of the hyper-heuristic algorithm on the single-target optimization problem has achieved a good effect, however, the application of the hyper-heuristic algorithm on the multi-target optimization problem is less.
A search of prior art documents revealed that Burke et al published in Springer (2005, pp: 129- "Metaheuristics:Progress as Real Problem Solvers"a hyperheuristic algorithm using reinforcement learning and tabu search as selection strategies is proposed for solving the multi-objective spatial allocation and temporal planning problem. An article "A new discrete rule based genetic algorithm for the multi-objective job shop problem" published by Vazzez-Rodriguez and Petrovic in Journal of Hearistics (2010, Vol.16, No.6, pp: 771-793) proposes a hyper-heuristic algorithm based on scheduling rules and genetic algorithms for solving the multi-objective job shop problem. However, the above-proposed hyper-heuristic algorithm only solves the non-continuity multi-objective optimization problem. Mashael Maashi et al, Expert Systems with Applications (2014, Vol.41, pp:4475-4493), and proposes a hyper-heuristic algorithm HH _ CF based on a selection function for solving the continuity multi-objective optimization problem, and achieves better effect. But the algorithm uses the running interval (CPU time) of a low-level heuristic operator as a parameter of a selection mechanism, so that the running effect of the algorithm on different machines is not stable. The article "A Hyper-Heuristic for the Multi-Objective Integration and Test Order Problem" published by Giovani Guiizzo et al in GECCO (2015, Madrid, Spain, pp: 1343) proposes two Hyper-Heuristic algorithms HITO-CF and HITO-MAB for solving the search-based Multi-Objective software engineering method Problem. HITO-CF and HITO-MAB use various crossover operators and mutation operators as low-level heuristic operators, and respectively use CF and MAB as selection strategies. In the selection process, the return value of the low-level heuristic operator is calculated by using a method of accumulation summation according to the dominance relation between the parent individuals and the child individuals. However, this method of return value calculation focuses on convergence of the solution set, and ignores the distributivity of the solution set.
Disclosure of Invention
The invention aims to overcome the problems that the universality is poor when the traditional heuristic algorithm is designed and the effect of a single heuristic algorithm on some problem examples is poor, and provides a method for solving a multi-target optimization problem based on a multi-objective heuristic algorithm of an MAB (multi-object-based) model, wherein the method takes an MAB strategy as a learning strategy, uses four performance evaluation mechanisms to evaluate the performance of each low-level heuristic operator, and better combines the advantages of each low-level heuristic operator through the learning and selecting mechanisms; the algorithm is used for carrying out experiments on a continuous multi-objective optimization problem set WFG, and good experimental results are obtained.
In order to achieve the purpose, the invention adopts the following technical scheme: a method for solving a multi-objective optimization problem based on a hyper-heuristic algorithm of MAB (multi-object-based B) comprises the following steps:
(1) giving the set H ═ H of the lower-level heuristic operator1,h2...hnumNum is the number of low-level heuristic operators, h1,h2...hnumOptimizing an algorithm or operator for multiple objectives;
(2) constructing a random initial population P of size N0Randomly generating N solutions in a decision space X;
(3) calculating the performance values of the heuristic operators of each lower layer on the four existing performance evaluation indexes; each low-level heuristic operator h1,h2...hnumFor the same initial population P0Respectively and independently run once, and each run is executed by hiNum, and calculating each heuristic operator h respectivelyiFour performance evaluation indices AE, RNI, SSC and UD; m is the number of times of executing the low-level heuristic operator in one iteration, and M is an integer in a range of (0, ∞);
1) AE: the evaluation algorithm obtains the calculated amount of the approximate solution set, the value range of the calculated amount is (0, infinity), and the smaller the AE value is, the better the performance is;
2) RNI: evaluating the proportion of the non-dominant solutions in the approximate solution set obtained by the algorithm, wherein the value range is (0, 1), the larger the RNI value is, the higher the proportion of the non-dominant solutions in the approximate solution set is, and if the RNI is 1, all individuals in the approximate solution set are non-dominant;
3) SSC: evaluating the size of a target space which can be covered by an approximate solution set obtained by an algorithm, wherein the value range is [0, ∞ ]; higher SSC values indicate greater coverage space like a solution set, which also indicates better performance in convergence and distribution;
4) UD: evaluating the distribution condition of the approximate solution set in the target space, wherein the value range is (0, 1), and the higher the UD value is, the better the distribution of the approximate solution set in the target space is, under the condition that the number of non-dominant solutions is the same;
(4) calculating the return value r of each low-level heuristic operator by using an operator performance evaluation mechanismiEach low-level heuristic operator evaluation mechanism comprises the following steps:
1) normalizing the performance data acquired by the heuristic operators at each low layer to map each performance evaluation index to the same value range [0, 1]]Inner, lower heuristic operator hiThe normalized values on the four performance evaluation indices were recorded as AEi_nor、RNIi_nor、SSCi_norAnd UDi_nor
2) According to the performance evaluation index value obtained after normalization, a formula is used for calculating the return value r of each low-level heuristic operatori
ri=(1-AEi_nor)+2RNIi_nor+SSCi_nor+UD i_nor
In formula (i): because of AEi_norThe smaller the value of (A) is, the better, and the larger the other three indexes are, the better, so that the return value r isiIn the calculation of (a) uses a low-level heuristic operator hiAE ofi_norReversal value, i.e. 1-AEi_norA value of (d); SSCi_norCan reflect the distribution and convergence of the population obtained after the algorithm execution, UDi_norThe distributivity of the population obtained after the algorithm is executed can be reflected; SSC in case of uniform population sizei_norAnd UDi_norThe larger the value of (A) is, the better; SSC pairs to avoid variation in population sizei_norAnd UDi_norInfluence of the value, therefore at riIn the calculation of (A) to SSCi_norAnd UDi_norRespectively matched with a non-dominant solution ratio RNIi_nor
(5) According to the reported value riCalculating the MAB value of each low-level heuristic operator, and the steps are as follows:
1) for a set of num lower-level heuristics H ═ H1,h2...hnumFor each hi(i ∈ { 1.,. num }) sets a performance tuple consisting of three variables
Figure GDA0002985843600000041
Wherein r isiIs a reported value,
Figure GDA0002985843600000042
Is the average return value, kiIs hiThe selected times, and all values are initialized to 0 in the algorithm initial stage;
2) update h using a formula-iPerformance tuple of
Figure GDA0002985843600000043
Figure GDA0002985843600000044
3) Heuristic operator h according to low layeriPerformance tuple V ofiComputing MAB of low-level heuristic operator at current stageiValue, formula (c) as follows:
Figure GDA0002985843600000045
wherein C is a control parameter, and controls the heuristic operator h of each lower layeriAverage performance of
Figure GDA0002985843600000046
A balance between the number of times an operator executes in the framework operation;
(6) selecting the lower-layer heuristic operator h with the maximum MAB valuetopI.e. top ═ argmax { MABiNum; if the algorithm with the maximum MAB value is multiple, one algorithm is randomly selected from the multiple algorithms;
(7) use of htopEvolved population PtExecute htopObtaining a new population P' by M generations, and calculating an application low-level heuristic operator htopObtaining performance values of the population P' on four performance evaluation indexes;
(8) updating population P with improved acceptance strategyt+1Execution h using SSC determinationtopWhether the later population has improvement or not, if not, the population P of the next iteration (t +1 generation)t+1By the t-th generation, i.e. Pt+1=Pt(ii) a If there is improvement, the population P of the t +1 th generationt+1Is replaced by the just acquired population P', i.e. Pt+1=P′;
(9) Judging whether the iteration times exceed a set maximum value W, if so, finishing the HH-MAB algorithm; if not, entering the next iteration and returning to the step (4); w is the maximum number of iterations of the algorithm, taking an integer in the range of (0, ∞).
The technical scheme of the invention has the following beneficial effects:
1. a novel MAB-based hyper-heuristic algorithm HH-MAB is provided for solving a multi-target optimization problem, and no low-level heuristic algorithm is learned in the aspects of calculated quantity, non-dominated solution-to-occupation ratio, solution set distribution and solution set convergence by using a MAB strategy in the hyper-heuristic algorithm at present.
2. A new return value algorithm is designed, and the performance of a low-level heuristic operator on the calculated amount, the non-dominated solution occupation ratio, the solution set distribution and the solution set convergence is comprehensively considered from multiple aspects by utilizing a method of normalizing and summing four performance evaluation indexes (AE, RNI, SSC and UD).
3. And normalizing each evaluation index value to enable the performance evaluation indexes with four different value intervals to be normalized in the same interval, so that the calculation of the return value has no preference for the four performance evaluation indexes, and a proper bottom layer heuristic operator is better selected at each decision point to improve the performance of the algorithm.
4. Compared with the existing hyper-heuristic methods (HH-RAND and HH-CF) for solving the multi-objective optimization problem, the HH-MAB selects the low-level heuristic operators according to the return values by using the MAB strategy, and the advantages of the low-level heuristic operators can be better combined by selecting the proper low-level heuristic operators at different solving stages, so that the algorithm performance is improved, and a better approximate solution set is obtained.
5. The HH-MAB adopts an improved acceptance strategy to update the population, and the method not only improves the performance of the algorithm, but also improves the stability of the algorithm.
Drawings
FIG. 1 is a flow chart of the HH-MAB algorithm framework of the present invention. FIG. 2 is a graph showing the average usage of the HH-RAND, HH-CF, and HH-MAB for three lower level heuristics (NSGA-II, SPEA2, MOGA) in the example, independently run 30 times on the WFG1-WFG9 test question.
FIG. 3 is a schematic diagram of the box plots on the evaluation index RNI of the HH-MAB and comparison algorithm of the example run independently 30 times on the WFG1-WFG9 test problems.
FIG. 4 is a schematic diagram of the box plot on the evaluation index SSC for 30 independent runs of the HH-MAB and comparison algorithm on WFG1-WFG9 test questions in the example.
FIG. 5 is a schematic diagram of the box map on the evaluation index UD of the HH-MAB and the comparison algorithm in the example, independently run 30 times on the WFG1-WFG9 test questions.
FIG. 6 is a schematic representation of the approximate Pareto front of the example HH-RAND, HH-CF, and HH-MAB, run independently 30 times, 50% on WFG1 test questions.
FIG. 7 is a schematic representation of the approximate Pareto front of the example HH-RAND, HH-CF, and HH-MAB, run independently 30 times, 50% on WFG2 test questions.
FIG. 8 is a graphical representation of the approximate Pareto front of the example HH-RAND, HH-CF, and HH-MAB, run independently 30 times, 50% on WFG3 test questions.
FIG. 9 is a schematic representation of the approximate Pareto front of the example HH-RAND, HH-CF, and HH-MAB, run independently 30 times, 50% on WFG4 test questions.
FIG. 10 is a schematic representation of the approximate Pareto front of the example HH-RAND, HH-CF, and HH-MAB running 30 times independently on WFG5 test questions, 50%.
FIG. 11 is a graphical representation of the approximate Pareto front of the example HH-RAND, HH-CF, and HH-MAB, run independently 30 times, 50% on WFG6 test questions.
FIG. 12 is a schematic representation of the approximate Pareto front of the example HH-RAND, HH-CF, and HH-MAB running 30 times independently on WFG7 test questions, 50%.
FIG. 13 is a schematic representation of the approximate Pareto front of the example HH-RAND, HH-CF, and HH-MAB running 30 times independently on WFG8 test questions, 50%.
FIG. 14 is a schematic representation of the approximate Pareto front of the example HH-RAND, HH-CF, and HH-MAB running 30 times independently on WFG9 test questions, 50%.
Fig. 15 is a schematic diagram of the original value and the normalized value of the performance data obtained by the lower-layer heuristic operator in step four 1).
Detailed Description
The invention will be further elucidated and described with reference to the drawings and specific embodiments.
For the multi-target optimization problem to be solved, firstly, a set H ═ H of the existing low-level heuristic operator aiming at the problem to be solved is given1,h2...hnum}. These low-level heuristics have strong problem dependencies. HH-MAB is used as a high-level heuristic strategy, and the space of a low-level heuristic operator is searched without directly solving the problem. The HH-MAB designs and uses an operator evaluation mechanism and a MAB selection mechanism to learn the performance of each low-level heuristic operator and uses an improved acceptance mechanism to decide whether to accept the generated new population.
Without loss of generality, assuming that a multi-objective optimization problem has n decision variables and m objective functions, the multi-objective optimization problem is defined as min F (x) ═ F1(x),f2(x),...,fm(x))TWherein x ═ x1,..,xn) Belongs to a decision vector (solution for short) with X as n dimension, wherein X is a decision space with n dimension and XiRepresents the ith decision component of x, i ∈ {1, 2.., n }; f (x) ═ f1(x),f2(x),...,fm(x))TAnd E is a target vector with m dimensions of Y, and Y is a target space with m dimensions of Y. f. of1(x),...,fm(x) M mutually conflicting objective functions; for the multi-objective optimization problem, if there are two solutions xAAnd xBWhen f is satisfied for all i ═ 1,2i(xA)≤fi(xB) And f is present on at least one targetj(xA)<fj(xB) Then call xADominating xB(ii) a If no other solution domination solution x exists, the solution x is called a Pareto optimal solution; a set formed by all Pareto optimal solutions is called a Pareto optimal solution set, and the mapping of the Pareto optimal solutions in a target space is called a Pareto front surface; for the multi-objective optimization problem, the objective of algorithm design is to obtain an approximate solution set that is uniformly distributed over the target space and converges towards the Pareto frontier.
As shown in fig. 1-15, the method of the present embodiment includes the following steps:
step one, a set H ═ H of a low-level heuristic operator is given1,h2...hnumNum is the number of the low-level heuristic operators, and in the invention, 3 low-level heuristic operators are selected, namely num is 3; h is1,h2,h3The method is an existing classical multi-objective optimization algorithm;
h 1:NSGAII(Deb,K.;Pratap,A.;Agarwal,S.;and Meyarivan,T.A fast and elitist multi-objective genetic algorithm:NSGA-II.IEEE Trans.on Evolutionary Computation,2002,6(2):182-197)
h2:SPEA2(Zitzler,E.;Laumanns,M.;and Thiele,L.SPEA2:Improving the Strength Pareto Evolutionary Algorithm.In:Proceedings of Evolutionary Methods for Design,Optimization and Control with Applications to Industrial Problems,Berlin,Germany:Springer-Verlag,2001.95-100)
h3:MOGA(Fonseca CM,Fleming PJ.Genetic algorithm for multiobjective optimization:Formulation,discussion and generation.In:Proceedings of the 5th Int’l Conf.on Genetic Algorithms,San Mateo:Morgan Kauffman Publishers,1993.416-423)。
the three algorithms have advantages and disadvantages and are more suitable for verifying the capability of HH-MAB in combining the advantages of low-level heuristic operators, the probability of crossover and mutation in each low-level heuristic operator is the same as that of a classical algorithm NSGA-II and most multi-objective evolutionary algorithms, the probability of crossover and mutation is respectively set to be 0.9 and 1/24, and the crossover and mutation distribution indexes are respectively 10 and 20.
Step two, constructing a random initial population P containing N solutions0(the starting population is designated P because it is the starting population, i.e., the 0 th generation population0(ii) a In this experiment, N is 100); in the decision space X, for the problem to be solved, each decision component X is setiThe value range is set as [ x ]il,xiu],i∈{1,2,...,n},xilAnd xiuRespectively representing the i-th decision component x of the solution x in the problem definitioniThe values of (1) are lower bound and upper bound, N values are randomly generated in the decision space XRandom initial population P of solutions0={x1,x2,...,xNThe method comprises the following steps: for each random solution xj(j ∈ {1, 2...., N }), the ith decision component x of whichj iIs its value range [ x ]il,xiu](i ∈ {1, 2.,. n }).
Calculating initial performance values of the heuristic operators at the lower layers on the four performance evaluation indexes so as to obtain the performance of a solution set through a multi-aspect measurement algorithm; each low-level heuristic operator h1,h2,h3Respectively for the same initial population P0Independently run once, and execute h every timeiFor M generations, i ∈ {1,2,3} (M is a user-defined parameter, which is the number of times of executing the low-level heuristic operator in one iteration, M is an integer in the range of (0, ∞), M is set to be 250 in this embodiment), and then each heuristic operator h is calculated separatelyiFour performance evaluation indexes (AE, RNI, SSC and UD), i ∈ {1,2,3 };
h1the four performance evaluation indexes of (1) are respectively recorded as AE1、RNI1、SSC1And UD1
h2The four performance evaluation indexes of (1) are respectively recorded as AE2、RNI2、SSC2And UD2
h3The four performance evaluation indexes of (1) are respectively recorded as AE3、RNI3、SSC3And UD3
1) Algorithm Effort (AE): the evaluation algorithm obtains the computation of the approximate solution set by dividing the fixed time step set by the user by the number of function evaluations performed within that step, over a range of (0, ∞), with smaller AE values indicating better Performance (Tan, K.C., Lee, T.H., & Khor, E.F. (2002)
2) Ratio of non-doped indicviduals (RNI): evaluating the proportion of the non-dominant solutions in the approximate solution set obtained by the algorithm, wherein the value is the number of the non-dominant solutions in the current population divided by the initial population size N, the value range is (0, 1), the larger the RNI value, the higher the proportion of the non-dominant solutions in the approximate solution set, and the better the Performance (Tan, K.C., Lee, T.H., & Khor, E.F. (2002). evolution algorithms for multi-object optimization: Performance assessment and compliance
3) Size of space converted or so-called S _ metric Hypervolume (SSC): the approximate solution set obtained by the evaluation algorithm is referenced to a reference point ref (ref)1,ref2,...,refm) The range of the target space covered by the space between the two is [0, ∞); reference point ref ═ ref1,ref2,...,refm) Assuming a solution in the target space dominated by all solutions in the set of approximate solutions, for minimizing the multi-objective optimization problem, generally refiA maximum value represented on the ith objective function; for the two target WFG problem, the common reference point is set to ref ═ 4, 4; higher SSC values indicate a more similar solution set with more coverage, indicating better performance in convergence and distribution (ziegler, e.,&Thiele,L.(1999).Multiobjective evolutionary algorithms:A comparative case study and the strength)
4) a uniformity distribution of a non-doped position (UD): evaluating the distribution of the non-dominant solution set in the target space, wherein the value range is (0, 1); in the case of the same number of non-dominant solutions, the higher the UD value indicates the better distribution of the approximate solution set in the target space (Srinivas, N., & Deb, K. (1994). Multi object optimization using non-dominant analysis in genetic algorithm, 2, 221-248.)
Fourthly, calculating the return value r of each low-level heuristic operator by using an operator performance evaluation mechanismi(ii) a Each low-level heuristic operator evaluation mechanism comprises the following steps:
1) normalizing the performance data acquired by the heuristic operators of each lower layer; the four performance evaluation indexes AE, RNI, SSC and UD are normalized from different value intervals to the same value range [0, 1%]Within; respectively obtaining the minimum value and the maximum value of the four performance evaluation indexes in the three low-level heuristic operators at the current stage, and respectively recording the minimum value and the maximum value as AEmin、AEmax、RNImin、RNImax、SSCmin、SSCmax、UDminAnd UDmaxWherein AEmin=argmin{AE1,AE2,AE3},AEmax=argmax{AE1,AE2,AE3},RNImin=argmin{RNI1,RNI2,RNI3},RNImax=argmax{RNI1,RNI2,RNI3},SSCmin=argmin{SSC1,SSC2,SSC3},SSCmax=argmax{SSC1,SSC2,SSC3},UDmin=argmin{UD1,UD2,UD3},UDmax=argmax{UD1,UD2,UD3}; if hiThe raw values before normalization of (i ∈ {1,2,3}) on the four performance evaluation indices are denoted as AEi、RNIi、SSCiAnd UDiThen h isiNormalized value AE on four Performance evaluation indicesi_nor,RNIi_nor,SSCi_norUD i_norThe calculation is as follows:
Figure GDA0002985843600000091
Figure GDA0002985843600000092
taking fig. 15 as an example, (a) is the original values of the three low-level heuristics on the four performance evaluation indexes before normalization, and (b) is the normalized values of the three low-level heuristics on the four performance evaluation indexes; taking AE as an example, AE can be known from (a)min=0.000108,AEmax0.000430; then h is1Normalization of value AE on AE1_nor=(0.000259-0.000108)/(0.000430-0.000108)=0.468944。
2) Calculating a heuristic index value of each low layer by using a formula (i) according to the normalized performance evaluation index value obtained in the step 1)Operator hiIs given a return value ri,i∈{1,2,3}:
ri=(1-AEi_nor)+2RNIi_nor+SSCi_nor+UD i_nor
Wherein AEi_norRepresenting the execution of a low-level heuristic operator hiCalculated amount of (SSC)i_norRepresenting execution of a low-level heuristic hiThe distribution and convergence of the later acquired population, UDi_norRepresenting execution of a low-level heuristic hiThe distribution of the population, RNI, is then obtainedi_norRepresenting low-level heuristic operator hiThe proportion of non-dominant solutions in the later population; due to RNIi_nor、SSCi_norAnd UDi_norIs that the larger the value the better the performance, and AEi_norThe smaller the value, the better the performance. Herein will be aEi_norFrom the minimization problem to the maximization problem, i.e. 1-AE is adopted in the formula (r)i_norA value of (d); however, in formula (I), UDi_norValue and SSCi_norThe value is influenced by the number of non-dominant solutions in the population; for example, UD when population individuals are few, even if the distribution of the current population is poori_norThe values may still be larger, even than UDs for populations that are better distributed but have more non-dominant solutions in the populationi_norThe value is high. In the execution process of the algorithm, the return value r is considered to be changed in the size of the populationiIn the calculation process we are SSCsi_norAnd UDi_norRespectively matched with an RNI with equal proportioni_nor
The formula is that four performance evaluation indexes are mixed, and after normalization, the value ranges of the performance evaluation indexes are the same, so that the heuristic operators h of the lower layers are all the sameiIs given a return value riThe four performance evaluation indexes are not preferred, and are comprehensively embodied; according to a formula, normalizing each low-level heuristic operator hiIs given a return value riAs shown in table 1; by h1For example, r1=(1-0.468944)+2×1.000000+1.000000+1.000000=4.531056。
TABLE 1 reporting values
h1 h2 h3
r 4.531056 4.406907 0.000000
Step five, according to the return value riCalculating the MAB value of each low-level heuristic operator, and the steps are as follows:
1) for the set of low-level heuristics H ═ H1,h2,h3H, for each low-level heuristic operator hiSetting a performance tuple consisting of three variables
Figure GDA0002985843600000101
Wherein r isiIs hiThe reported value obtained in step four,
Figure GDA0002985843600000102
Is from the beginning hiAverage value of the obtained reported values, kiIs hiThe number of times selected so far, i ∈ {1,2,3 }; three variables are initialized to r in the initial stage of the algorithmi=0,
Figure GDA0002985843600000103
ki=0;
2) Obtaining the return value r according to the step fouriUpdating the heuristic operator h of the lower layer by using a formula 2iCorresponding performance tuple
Figure GDA0002985843600000111
The last two variables in
Figure GDA0002985843600000112
And ki
Figure GDA0002985843600000113
3) Heuristic operator h according to low layeriPerformance tuple V ofiCalculating the heuristic operator h of the current stage low layeriMAB of (A)iValue (MAB)i>0) As shown in formula (c):
Figure GDA0002985843600000114
wherein C is a control parameter (set to 2 in the experiment) for controlling each low-level heuristic operator hiAverage return value of
Figure GDA0002985843600000115
And algorithm operation low-layer heuristic operator hiIs performed a number of times kiA balance is kept between;
step six, selecting a low-level heuristic operator h with the maximum MAB valuetopI.e. top ═ argmax { MAB1,MAB2,MAB3}; if a plurality of low-level heuristic operators simultaneously have the maximum MAB value, namely MABi=MABj>MABkWhere i, j, k ∈ {1,2,3}, then h is counted fromiAnd hjRandomly selects one as htop
Step seven, using a low-level heuristic operator htopEvolution of the Current population PtExecute htop250 generations altogetherObtaining a new population P'; calculating an application low-level heuristic operator h according to the performance evaluation indexes AE, RNI, SSC and UD introduced in the step threetopPerformance index value AE of obtained population P' on four performance evaluation indexestop、RNItop、SSCtopAnd UDtop
Step eight, adopting an improved acceptance strategy to update the population Pt+1(ii) a Since SSC is the only one of the four performance evaluation indexes we use to evaluate both the population convergence and the population distribution, the execution of h is determined using SSCtopWhether the later population has improved (i.e., whether the SSC value is large); if the SSC evaluation index value is not larger than the last iteration value, the population P of the next iteration (t +1 generation)t+1The t generation of the population PtReplacement, i.e. Pt+1=Pt(ii) a If the value becomes larger, the population P of the t +1 th generationt+1Is replaced by the just acquired population P', i.e. Pt+1=P′。
Step nine, judging whether the iteration number exceeds a set iteration maximum value W (W is a user-defined parameter and is the maximum iteration number of the algorithm, and an integer in a range of (0, ∞) is taken, wherein in the embodiment, W is set to be 25); if so, the HH-MAB algorithm ends; if not, entering the next iteration and returning to the step four.
FIGS. 6-14 are schematic representations of the HH-RAND, HH-CF, and HH-MAB approximate Pareto front for 30 independent runs on WFG1-WFG9 test questions, 50%. The closer the approximate Pareto front surface obtained by the three algorithms is to the real Pareto front surface (PF), the better the convergence of the algorithms is. As can be seen from FIGS. 6 to 14, the convergence properties of HH-MAB are more excellent than those of HH-RAND and HH-CF.
The above description is only an embodiment of the present invention. It should be noted that, in the present embodiment, an experiment is performed on two target WFG test sets, and the test set has 9 test cases, covers multiple aspects of Pareto front surface continuity \ discontinuity, Pareto front surface convexity \ concavity \ linearity, multiple modes \ single mode, fraud problem, and the like, and is a test set widely used for verifying the performance of the multi-target optimization algorithm. According to the characteristics of the hyper-heuristic algorithm, if the problem to be solved is changed, only the number and the types of the low-level heuristic operators need to be changed or replaced, and the high-level heuristic strategy does not need to be changed. Therefore, changing the number and the types of the heuristic operators of the lower layer still remains the protection scope of the patent.

Claims (1)

1. A method for solving a multi-objective optimization problem based on a MAB (multi-object-based) hyperheuristic algorithm is characterized by comprising the following steps: the method comprises the following steps:
(1) giving the set H ═ H of the lower-level heuristic operator1,h2...hnumNum is the number of low-level heuristic operators, h1,h2...hnumOptimizing an algorithm or operator for multiple objectives;
(2) constructing a random initial population P of size N0Randomly generating N solutions in a decision space X;
(3) calculating the performance values of the heuristic operators of each lower layer on the four existing performance evaluation indexes; each low-level heuristic operator h1,h2...hnumFor the same initial population P0Respectively and independently run once, and each run is executed by hiNum, and calculating each heuristic operator h respectivelyiFour performance evaluation indices AE, RNI, SSC and UD; m is the number of times of executing the low-level heuristic operator in one iteration, and M is an integer in a range of (0, ∞);
1) AE: the evaluation algorithm obtains the calculated amount of the approximate solution set, the value range of the calculated amount is (0, infinity), and the smaller the AE value is, the better the performance is;
2) RNI: evaluating the proportion of the non-dominant solutions in the approximate solution set obtained by the algorithm, wherein the value range is (0, 1), the larger the RNI value is, the higher the proportion of the non-dominant solutions in the approximate solution set is, and if the RNI is 1, all individuals in the approximate solution set are non-dominant;
3) SSC: evaluating the size of a target space which can be covered by an approximate solution set obtained by an algorithm, wherein the value range is [0, ∞ ]; higher SSC values indicate greater coverage space like a solution set, which also indicates better performance in convergence and distribution;
4) UD: evaluating the distribution condition of the approximate solution set in the target space, wherein the value range is (0, 1), and the higher the UD value is, the better the distribution of the approximate solution set in the target space is, under the condition that the number of non-dominant solutions is the same;
(4) calculating the return value r of each low-level heuristic operator by using an operator performance evaluation mechanismiEach low-level heuristic operator evaluation mechanism comprises the following steps:
1) normalizing the performance data acquired by the heuristic operators at each low layer to map each performance evaluation index to the same value range [0, 1]]Inner, lower heuristic operator hiThe normalized values on the four performance evaluation indices were recorded as AEi_nor、RNIi_nor、SSCi_norAnd UDi_nor
2) According to the performance evaluation index value obtained after normalization, a formula is used for calculating the return value r of each low-level heuristic operatori
ri=(1-AEi_nor)+2RNIi_nor+SSCi_nor+UDi_nor
In formula (i): because of AEi_norThe smaller the value of (A) is, the better, and the larger the other three indexes are, the better, so that the return value r isiIn the calculation of (a) uses a low-level heuristic operator hiAE ofi_norReversal value, i.e. 1-AEi_norA value of (d); SSCi_norCan reflect the distribution and convergence of the population obtained after the algorithm execution, UDi_norThe distributivity of the population obtained after the algorithm is executed can be reflected; SSC in case of uniform population sizei_norAnd UDi_norThe larger the value of (A) is, the better; SSC pairs to avoid variation in population sizei_norAnd UDi_norInfluence of the value, therefore at riIn the calculation of (A) to SSCi_norAnd UDi_norRespectively matched with a non-dominant solution ratio RNIi_nor
(5) According to the reported value riCalculating the MAB value of each low-level heuristic operator, and the steps are as follows:
1) for a set of num lower-level heuristics H ═ H1,h2...hnumFor each hi(i ∈ { 1.,. num }) setSetting a performance tuple consisting of three variables
Figure FDA0002985843590000021
Wherein r isiIs a reported value,
Figure FDA0002985843590000022
Is the average return value, kiIs hiThe selected times, and all values are initialized to 0 in the algorithm initial stage;
2) update h using a formula-iPerformance tuple of
Figure FDA0002985843590000023
Figure FDA0002985843590000024
3) Heuristic operator h according to low layeriPerformance tuple V ofiComputing MAB of low-level heuristic operator at current stageiValue, formula (c) as follows:
Figure FDA0002985843590000025
wherein C is a control parameter, and controls the heuristic operator h of each lower layeriAverage performance of
Figure FDA0002985843590000026
A balance between the number of times an operator executes in the framework operation;
(6) selecting the lower-layer heuristic operator h with the maximum MAB valuetopI.e. top ═ argmax { MABiNum; if the algorithm with the maximum MAB value is multiple, one algorithm is randomly selected from the multiple algorithms;
(7) use of htopEvolved population PtExecute htopObtaining a new population P' by M generations, and calculating an application low-level heuristic operator htopObtaining performance values of the population P' on four performance evaluation indexes;
(8) updating population P with improved acceptance strategyt+1Execution h using SSC determinationtopWhether the later population has improvement or not, if not, the population P of the next iteration (t +1 generation)t+1By the t-th generation, i.e. Pt+1=Pt(ii) a If there is improvement, the population P of the t +1 th generationt+1Is replaced by the just acquired population P', i.e. Pt+1=P′;
(9) Judging whether the iteration times exceed a set maximum value W, if so, finishing the HH-MAB algorithm; if not, entering the next iteration and returning to the step (4); w is the maximum number of iterations of the algorithm, taking an integer in the range of (0, ∞).
CN201811230929.8A 2018-10-22 2018-10-22 Method for solving multi-objective optimization problem based on MAB (multi-object-based) hyperheuristic algorithm Active CN109460862B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201811230929.8A CN109460862B (en) 2018-10-22 2018-10-22 Method for solving multi-objective optimization problem based on MAB (multi-object-based) hyperheuristic algorithm

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201811230929.8A CN109460862B (en) 2018-10-22 2018-10-22 Method for solving multi-objective optimization problem based on MAB (multi-object-based) hyperheuristic algorithm

Publications (2)

Publication Number Publication Date
CN109460862A CN109460862A (en) 2019-03-12
CN109460862B true CN109460862B (en) 2021-04-27

Family

ID=65608137

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201811230929.8A Active CN109460862B (en) 2018-10-22 2018-10-22 Method for solving multi-objective optimization problem based on MAB (multi-object-based) hyperheuristic algorithm

Country Status (1)

Country Link
CN (1) CN109460862B (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110826145B (en) * 2019-09-09 2020-08-28 西安工业大学 Automobile multi-parameter operation condition design method based on heuristic Markov chain evolution
CN111917529B (en) * 2020-07-15 2021-06-15 燕山大学 Underwater sound OFDM resource allocation method based on improved EXP3 algorithm
CN112631907B (en) * 2020-12-19 2024-04-02 北京化工大学 Super heuristic frame upper layer scheduling method based on utilization and exploration

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106156891A (en) * 2016-07-08 2016-11-23 华南师范大学 A kind of Scheduling method based on Pareto multiple target ant colony optimization algorithm
CN106384169A (en) * 2016-09-22 2017-02-08 合肥工业大学 Hyper-heuristic algorithm-based satellite task planning method
CN107705157A (en) * 2017-10-19 2018-02-16 大连理工大学 Automotive supplies Method for Sales Forecast method and system based on unified dynamic integrity model and meta-heuristic algorithm

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2010111312A2 (en) * 2009-03-23 2010-09-30 Riverbed Technology, Inc. Virtualized data storage system architecture
US8942659B2 (en) * 2011-09-08 2015-01-27 Drexel University Method for selecting state of a reconfigurable antenna in a communication system via machine learning

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106156891A (en) * 2016-07-08 2016-11-23 华南师范大学 A kind of Scheduling method based on Pareto multiple target ant colony optimization algorithm
CN106384169A (en) * 2016-09-22 2017-02-08 合肥工业大学 Hyper-heuristic algorithm-based satellite task planning method
CN107705157A (en) * 2017-10-19 2018-02-16 大连理工大学 Automotive supplies Method for Sales Forecast method and system based on unified dynamic integrity model and meta-heuristic algorithm

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
Alexandre Silvestre Ferreira;Richard Aderbal Gonçalves;Aurora Po.A Multi-Armed Bandit selection strategy for Hyper-heuristics.《2017 IEEE Congress on Evolutionary Computation (CEC)》.2017, *
Evaluating a Multi-objective Hyper-Heuristic for the Integration and Test Order Problem;Giovani Guizzo;Silvia R. Vergilio;Aurora T.R. Pozo;《2015 Brazilian Conference on Intelligent Systems (BRACIS)》;20151231;全文 *
Richard Aderbal Gonçalves etal..A New Hyper-Heuristic Based on a Restless Multi-armed Bandit for Multi-objective Optimization.《2017 Brazilian Conference on Intelligent Systems (BRACIS)》.2017, *
基于粒子群优化算法的类集成测试序列确定方法;张艳梅等;《计算机学报》;20180430;第41卷(第4期);全文 *

Also Published As

Publication number Publication date
CN109460862A (en) 2019-03-12

Similar Documents

Publication Publication Date Title
CN109460862B (en) Method for solving multi-objective optimization problem based on MAB (multi-object-based) hyperheuristic algorithm
Liang et al. A clustering-based differential evolution algorithm for solving multimodal multi-objective optimization problems
Zhu et al. An efficient evolutionary grey wolf optimizer for multi-objective flexible job shop scheduling problem with hierarchical job precedence constraints
CN112052071B (en) Cloud software service resource allocation method combining reinforcement learning and machine learning
Du et al. Managing noncooperative behaviors in large-scale group decision-making: Integration of independent and supervised consensus-reaching models
CN110389819B (en) Method and system for scheduling calculation intensive batch processing tasks
CN110138612A (en) A kind of cloud software service resource allocation methods based on QoS model self-correcting
Li et al. DLEA: A dynamic learning evolution algorithm for many-objective optimization
US20200292340A1 (en) Robot running path, generation method, computing device and storage medium
Xu et al. An efficient hybrid multi-objective particle swarm optimization with a multi-objective dichotomy line search
Li et al. A grid-based fitness strategy for evolutionary many-objective optimization
CN108153918B (en) Personalized movie recommendation method based on adaptive orthogonal intersection multi-objective optimization algorithm
CN111191343A (en) Multi-mode multi-target differential evolution algorithm based on random sequencing learning
CN115689069B (en) Power grid dispatching control method and system based on artificial intelligence
Jaszkiewicz et al. Multiobjective memetic algorithms
Liu et al. Multi-layer interaction preference based multi-objective evolutionary algorithm through decomposition
Horng et al. Bat algorithm assisted by ordinal optimization for solving discrete probabilistic bicriteria optimization problems
Li et al. A nondominated sorting genetic algorithm III with three crossover strategies for the combined heat and power dynamic economic emission dispatch with or without prohibited operating zones
Li et al. Achievement scalarizing function sorting for strength Pareto evolutionary algorithm in many-objective optimization
CN114065896A (en) Multi-target decomposition evolution algorithm based on neighborhood adjustment and angle selection strategy
Lasheng et al. Improved Initial clustering center selection algorithm for K-means
CN111311096A (en) QFD and KANO model-based multi-yield quality optimization method
CN115690476A (en) Automatic data clustering method based on improved harmony search algorithm
CN111813669B (en) Adaptive random test case generation method based on multi-target group intelligence
Cervone et al. Experimental validations of the learnable evolution model

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant