CN110533221A - Multipurpose Optimal Method based on production confrontation network - Google Patents

Multipurpose Optimal Method based on production confrontation network Download PDF

Info

Publication number
CN110533221A
CN110533221A CN201910688044.0A CN201910688044A CN110533221A CN 110533221 A CN110533221 A CN 110533221A CN 201910688044 A CN201910688044 A CN 201910688044A CN 110533221 A CN110533221 A CN 110533221A
Authority
CN
China
Prior art keywords
sample
network
group
variable
optimization
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201910688044.0A
Other languages
Chinese (zh)
Inventor
鲍亮
王方正
魏守鑫
方宝印
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xian University of Electronic Science and Technology
Original Assignee
Xian University of Electronic Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Xian University of Electronic Science and Technology filed Critical Xian University of Electronic Science and Technology
Priority to CN201910688044.0A priority Critical patent/CN110533221A/en
Publication of CN110533221A publication Critical patent/CN110533221A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/12Computing arrangements based on biological models using genetic models
    • G06N3/126Evolutionary algorithms, e.g. genetic algorithms or genetic programming
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/04Forecasting or optimisation specially adapted for administrative or management purposes, e.g. linear programming or "cutting stock problem"

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biophysics (AREA)
  • General Physics & Mathematics (AREA)
  • Software Systems (AREA)
  • Business, Economics & Management (AREA)
  • Mathematical Physics (AREA)
  • Artificial Intelligence (AREA)
  • Biomedical Technology (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Human Resources & Organizations (AREA)
  • Economics (AREA)
  • Strategic Management (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Development Economics (AREA)
  • Game Theory and Decision Science (AREA)
  • Genetics & Genomics (AREA)
  • Physiology (AREA)
  • Entrepreneurship & Innovation (AREA)
  • Marketing (AREA)
  • Operations Research (AREA)
  • Quality & Reliability (AREA)
  • Tourism & Hospitality (AREA)
  • General Business, Economics & Management (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

A kind of Multipurpose Optimal Method based on production confrontation network disclosed by the invention, solves the problems such as existing optimization algorithm time cost is high, and training difficulty is excessive, and training network is easy collapse.Its implementation: stochastical sampling obtains initial sample;Pareto solution therein is selected as training set;It randomly selects half in training set to be pre-processed as training sample, building production fights network, obtains generating sample after being iterated training;Judge whether to need to be trained optimization again according to evaluation number;The result for generating sample and other comparison algorithms through the invention is compared, evaluation algorithms superiority and inferiority.Present invention reduces time costs, improve network robustness and stability, and effect of optimization is obvious, can be used for the resource allocation of multiple targets, the production scheduling of multiple products and optimization of the multiple performances of various software systems etc..

Description

Multipurpose Optimal Method based on production confrontation network
Technical field
It is specifically a kind of to be based on production the invention belongs to field of computer technology, in particular to multi-target parameter optimizing Fight the Multipurpose Optimal Method of network, abbreviation MOGAN, can be used for using computer technology to the resource allocations of multiple targets, The optimization etc. of the production scheduling of multiple products and the multiple performances of various software systems.
Background technique
In industrial production and life, many problems are all made of the multiple targets for conflicting with each other and influencing.People's meeting Frequently encountering makes multiple targets in given area while optimization problem as optimal as possible, that is, multi-objective optimization question (multi-objectiveoptimization problem, MOP), existing optimization aim are more than one and need to more A target is handled simultaneously.Such as in the design of component, it is desirable to which big some stability of power is more by force, high-power at this time and high Stabilization is exactly a pair of conflicting target, and here it is a MOP problems;For a control system, it has various non- Functional attributes, such as runing time, handling capacity, processing number per second have handling capacity in a distributed manner for message system Kafka With two nonfunctional spaces of time delay, also referred to as target, in certain Kafka system, it is desirable to handling capacities a little and to postpone greatly It is a little bit smaller, the two targets be also it is conflicting, the task of multiple-objection optimization is exactly the configuration by adjusting its control system The noninferior solution of one group of the two target is found, this is also a MOP problem.There are many more other in industrial production and life MOP problem, be related to life and industrial every aspect, relevant method for solving to solve politics, finance, it is military, The programmed decision-making problem of the various aspects such as environment, the manufacturing, social security plays a crucial role, and is modern industry system The technical issues that need to address in engineering.
In single-object problem, usual optimal solution only one, and fairly simple and common mathematics side can be used Method finds out its optimal solution.However in multi-objective optimization question, mutually restricted between each target, an Objective may be made The improvement of energy is often to lose other target capabilities as cost, it is impossible to which there are one to be optimal all target capabilities all Solution, so for multi-objective optimization question, solution is usually set --- the Pareto disaggregation of a noninferior solution.
In multiple objective programming, due to the phenomenon that there are the conflicts between target with that can not compare, a solution is in some mesh It is best for putting on, may be poor in other targets.The solution insubjection solution (Non- of Pareto proposition multiple target Dominated set) concept, is defined as: assuming that it is any two solution S1 and S2 for all targets, S1 is superior to S2, then We claim S1 to dominate S2, if S1 is not dominated by other solutions, S1 is known as non-domination solution (insubjection solution), also referred to as Pareto Solution.
Multi-objective Evolutionary Algorithm (MOEA) is an analoglike biological evolution mechanism and the probability optimization of overall importance search that is formed Method, basic principle are described as follows: the population that multi-objective Evolutionary Algorithm generates at random from one group, by executing to population The evolutional operations such as selection, intersection and variation are evolved through excessive generation, individual fitness continuous improvement in population, to gradually force The Pareto optimal solution set of nearly multi-objective optimization question.Than more typical multi-objective Evolutionary Algorithm have NSGA2, PESA2 and SPEA2.For these three algorithms, advantage is more but its disadvantage is also obvious.
The advantages of NSGA2, is that operational efficiency height, disaggregation have good distributivity, has especially for low-dimensional optimization problem Preferable performance;Its shortcoming is that disaggregation process has defect in higher-dimension problem, the diversity of disaggregation is undesirable.
The advantages of PESA2, is that the convergence of its solution is fine, is easier close to optimal face, especially in higher-dimension problem feelings Under condition;But disadvantage is that selection operation can only once choose an individual, time loss is very big, and the multiplicity of class Property is bad.
The advantages of SPEA2, is that the good disaggregation of degree of distribution can be obtained, especially in the solution of higher-dimension problem, But its cluster process keeps diversity to take a long time, operational efficiency is not high.
Other than multi-objective Evolutionary Algorithm, there is also multi-objective particle swarm algorithm (MOPSO).Particle group optimizing (PSO) algorithm is a kind of simulation social action, the evolution technology based on swarm intelligence, with its unique search mechanism, remarkably Constringency performance, convenient computer realize, be widely used in engineering optimization field, multiple target PSO (MOPSO) calculate Method has been applied to different optimization fields, but there are computation complexities it is high, versatility is low, convergence is bad the disadvantages of.
Multiple-objection optimization is frequently encountered and to be solved the problems, such as in an engineering, although the above many methods are all attempted pair It is solved, but there are still variety of problems and deficiency.Generally speaking, some optimum results are not good enough, some training difficulty It is larger, need sample sizes it is too many, the optimization time also too long, spend cost it is too big.In addition to this each algorithm is also Other limitations, such as: NSGA2 disaggregation process in higher-dimension problem has defect, and the diversity of disaggregation is undesirable;PESA2 selection Operation can only once choose an individual, and time loss is very big, and the diversity of class is bad;SPEA2 cluster process is kept Diversity takes a long time, and operational efficiency is not high;And that there are computation complexities is high, versatility is low, convergence is bad etc. lacks by MOPSO Point.
Summary of the invention
It is an object of the invention in view of the above shortcomings of the prior art, propose that a kind of optimum results are more preferable, be easy to training, Multipurpose Optimal Method of the optimizing faster based on production confrontation network.
The present invention is a kind of Multipurpose Optimal Method based on production confrontation network, which is characterized in that including walking as follows It is rapid:
(1) initial sample is obtained:
(1a) selects optimization aim: for multi-objective optimization question present in a certain system, selection determination needs to optimize Multiple targets, it is assumed that have q optimization aim, there is p system variable to produce this q optimization aim in all system variables Raw to influence, each system variable has the data set and value range of oneself;
The maximum evaluation number of (1b) setting: each group of p system variable is defined as one group of independent variable, is obtained by one group of independent variable To the value of q optimization aim be defined as dependent variable, primary evaluation, which just refers to, obtains the mistake of corresponding dependent variable by one group of independent variable Journey;Evaluation number is set in optimization process as e, maximum evaluation number is E, and evaluation number e is initialized as zero;
(1c) determines initial sample size: determining that initializaing variable quantity is m group according to maximum evaluation number E, guarantees optimization Process can be completed in maximum evaluation number E;
(1d) stochastical sampling obtains initial sample: stochastical sampling is carried out to each system variable for being related to optimization aim, The initializaing variable of m group p dimension is obtained, evaluates the superiority and inferiority for measuring this m group initializaing variable with the operation method of multiple target place system, Each group of initializaing variable obtains the value of corresponding q optimization aim, and the initializaing variable of every group of stochastical sampling and the value of optimization aim are total With the initial sample of composition one group of p+q dimension;Initializaing variable sampling is evaluated and is combined with optimization aim once, and evaluation number e adds 1, Whole m group initializaing variable is traversed, the final number e that evaluates adds m, obtains the initial sample of m p+q dimension;
(2) production for constructing multiple-objection optimization fights network (MOGAN): it includes generating network G and differentiation network D, It generates network G and differentiates that network D is all made of three layers of full Connection Neural Network, the two is confronted with each other training, is promoted constantly excellent each other Change, is built into production confrontation network (MOGAN) of multiple-objection optimization;
(3) select training set from initial sample: all p+ are chosen in the definition solved according to Pareto from initial sample The Pareto solution of q dimension, then removes the value of q optimization aim in each group of Pareto solution, by the p of these removal optimization target values The Pareto solution of dimension is used as training set;
(4) production of the multiple-objection optimization of training building fights network (MOGAN):
(4a) determines training sample: from training set randomly choose half as training sample, then to training sample into The pretreatment of row data normalization, obtains pretreated training sample x;
(4b) inputs training sample into MOGAN: pretreated training sample x is input to production confrontation network Differentiate in network D;
(4c) generates sample: in the production confrontation network of multi-objective optimization question, using network G is generated, generating p dimension Generation sample z;
(4d) obtains differentiating result: will generate sample z and training sample x input and differentiates network D, output differentiates result;
(4e) training generates network G and differentiates network D: according to differentiating as a result, fixed generate network G, to differentiate network D into Row training, is continued to optimize, and is from training sample or to carry out self-generating net until differentiating that network D can be accurately judged to a sample The sample that network G is generated;Network D is differentiated according to differentiating as a result, fixing, and is trained, is continued to optimize, until sentencing to network G is generated Other network D cannot judge a sample be from training sample or come self-generating network G generation sample;
It obtains generating sample set after the training of (4f) successive ignition: executing step (4a)-(4e), complete once to generation The training of formula confrontation network;Judge whether to reach the number of iterations, if not reaching the number of iterations of design, repeats step (4a)-(4e), continues to train;If reaching the number of iterations of design, m is generated using network G is generated1Group p dimension is initial to be become Amount, and evaluate and measure this m1The superiority and inferiority of group sample, calculates the value of q dimension optimization aim, and evaluation number e adds m1, then m1Group p dimension Initializaing variable and q dimension optimization aim merge to obtain m1The generation sample set of group p+q dimension;
(5) judgement evaluation number: judging evaluation number, if evaluation number e reaches maximum evaluation number E at this time, Then the generation sample set that step (4f) is obtained executes step (9), further verifies effect of optimization as final result collection, Otherwise, step (6) are executed;
(6) it obtains intersecting result set: by m1The generation sample set and training sample x of group p+q dimension merge, after merging Sample set in select all p+q dimension Pareto solutions, removal Pareto solves the value of q optimization aim, obtains the initial change of p dimension The Pareto of amount is solved, and is carried out simulation binary system crossover operation to the Pareto solution of these p dimension initializaing variable, is obtained m2Group p dimension is just The intersection result set of beginning variable;
(7) variation result set is obtained: to m2The intersection result set of group p dimension initializaing variable carries out multinomial mutation operation, obtains To m3The variation result set of group p dimension initializaing variable;
(8) evaluation obtains new initial sample: the operation method of system measures this m to evaluate where multiple target3Group p dimension The superiority and inferiority of initializaing variable, each group of initializaing variable obtain the value of corresponding q optimization aim, every group of p initializaing variable and q optimization Mesh target value collectively constitutes the initial sample of one group of new p+q dimension, and the every evaluation of initializaing variable is simultaneously combined once with optimization aim, commented Valence number e adds 1, traverses whole m3The initializaing variable of group p dimension, the final number e that evaluates add m3, obtain m3Group p+q dimension it is new initial Sample;Carry out step (3)~step (5) again;
(9) effect of optimization is verified:
(9a) has optimization algorithm with other and optimizes to the selected p optimized variable of step (1), obtains corresponding Results of comparison collection;
(9b) optimizes the present invention by comparing final result collection and the results of comparison collection of other optimization algorithms of the invention Effect is verified.
Multiple-objection optimization is frequently encountered and to be solved the problems, such as in an engineering, and existing many methods all exist various Limitation and deficiency, some optimum results are not good enough, and some training difficulty are larger, the sample size of needs is too many, optimization also Time too long, spend cost it is too big.In addition to this, there are also other limitations for each algorithm, such as: NSGA2 is solved in higher-dimension problem Collection process has defect, and the diversity of disaggregation is undesirable;PESA2 selection operation can only once choose an individual, time loss It is very big, and the diversity of class is bad;SPEA2 cluster process keeps diversity to take a long time, and operational efficiency is not high;And MOPSO there are computation complexities it is high, versatility is low, convergence is bad the disadvantages of.
Compared with prior art, the present invention has the advantage that:
It is optimized using two network dual trainings, as a result well: by the present invention in that being based on two nets with one kind The mode of network dual training optimizes, and has broken the intrinsic thinking of original multi-objective optimization question solution, has used one Network carries out simulation and generates characteristic variable, another network judges performance quality, and alternating iteration carries out the side of the two processes Method optimizes, and as a result well, simultaneously because two networks use three layers of fully-connected network, is easy to trained.
Reduce the demand to training samples number: the present invention randomly chooses half sample characteristics by experiment sample every time Method, ensure that the diversity and randomness of training sample, it is also ensured that the quality of sample.Simultaneously as avoiding passing through Many experiments obtain the process of great amount of samples, have saved time cost to the maximum extent.
Sample size is effectively expanded, save time cost: invention introduces the intersection thoughts in genetic algorithm, pass through Simulation binary system is carried out to the sample of generation to intersect, and is expanded sample size, more effective search has been carried out to sample space, more The data distribution for having found Pareto solution fastly, has greatly saved time cost.
Avoid falling into local optimum: invention introduces the variation thought in genetic algorithm, by the sample to generation into The variation of row multinomial, effectively prevents the case where falling into local optimum, is easier to search out global optimum.
Detailed description of the invention
Fig. 1 is implementation flow chart of the invention;
Fig. 2 is the sub-process figure of production confrontation network internal logic in the present invention;
Fig. 3 is to differentiate network D in the present invention and generate the structure chart of network G;
Fig. 4 is the curve graph of the present invention and other optimization algorithms for the optimum results of ZDT1 function;
Fig. 5 is the curve graph of the present invention and other optimization algorithms for the optimum results of ZDT6 function.
Specific embodiment
With reference to the accompanying drawing with example to the detailed description of the invention
Embodiment 1
Multiple-objection optimization is the problem of being frequently encountered in industrial production, and many problems are all by conflicting with each other and influencing Multiple targets composition, the function for example, Kafka system two conflicting performance objective handling capacities and delay, in circuit design Rate and stability be also it is conflicting, these are all multi-objective optimization questions.Multi-objective optimization question is related to industrial Every aspect, relevant method for solving is to various aspects such as solution politics, finance, military affairs, environment, the manufacturing, social securities Programmed decision-making problem plays a crucial role, and is the technical issues that need to address in modern industry system engineering.Although existing There are many methods all to solve to multi-objective optimization question trial, but is having variety of problems and deficiency, some optimum results Not good enough, some training difficulty are larger, the sample size that needs is too many, the optimization time also too long, spend cost too big.This Invention expands research for this status, proposes a kind of Multipurpose Optimal Method based on production confrontation network.
The present invention be it is a kind of based on production confrontation network Multipurpose Optimal Method include the following steps: referring to Fig. 1
(1) initial sample is obtained:
(1a) selects optimization aim: for multi-objective optimization question present in a certain system, selection determination needs to optimize Multiple targets, it is assumed that have q optimization aim, there is p system variable to produce this q optimization aim in all system variables Raw to influence, each system variable has the data set and value range of oneself.
The maximum evaluation number of (1b) setting: each group of p system variable is defined as one group of independent variable, is obtained by one group of independent variable To the value of q optimization aim be defined as dependent variable, primary evaluation, which just refers to, obtains the mistake of corresponding dependent variable by one group of independent variable Journey;Evaluation number is set in optimization process as e, maximum evaluation number is E, and evaluation number e is initialized as zero.
(1c) determines initial sample size: determining that initializaing variable quantity is m group according to maximum evaluation number E, guarantees optimization Process can be completed in maximum evaluation number E.
(1d) stochastical sampling obtains initial sample: stochastical sampling is carried out to each system variable for being related to optimization aim, The initializaing variable of m group p dimension is obtained, evaluates the superiority and inferiority for measuring this m group initializaing variable with the operation method of multiple target place system, Each group of initializaing variable obtains the value of corresponding q optimization aim, and the initializaing variable of every group of stochastical sampling and the value of optimization aim are total With the initial sample of composition one group of p+q dimension;Initializaing variable sampling is evaluated and is combined with optimization aim once, and evaluation number e adds 1, Whole m group initializaing variable is traversed, the final number e that evaluates adds m, obtains the initial sample of m p+q dimension.
In numerous system variables, the present invention has been determined by p system related with q optimization aim and has become It measures, then the stochastical sampling in this p system variable, every group is intended to sample, and ties up initializaing variable with the p that these stochastical samplings form The value of the corresponding q optimization aim of affix behind forms a p+q and ties up initial sample, each p+q ties up initial sample The information of initializaing variable and corresponding optimization aim is contained, stochastical sampling has generality and result may be very well.
(2) production for constructing multiple-objection optimization fights network (MOGAN): it includes generating network G and differentiation network D, It generates network G and differentiates that network D is all made of three layers of full Connection Neural Network, the two is confronted with each other training, is promoted constantly excellent each other Change, is built into production confrontation network (MOGAN) of multiple-objection optimization.
In production confrontation network, training sample, which is input to, to be differentiated in network D, is trained, until differentiating network D energy Be accurately judged to a sample be from training sample or come self-generating network G generation sample;Random noise is input to life It in network G, is trained, until differentiating that network D can not judge that a sample is from training sample or to carry out self-generating net The sample that network G is generated, by the dual training of the two, last whole network output is the generation sample for generating network G and generating.
Unlike other optimization algorithms, present invention employs productions to fight network, based on zero in game theory And game idea, compared to more traditional model, there are two different networks for it --- and it generates network and differentiates network, rather than Single network, and training method, using dual training mode, this makes that difficulty is trained to reduce, these features to give birth to An accepted way of doing sth fights the multiple groups optimal solution for searching out multiple-objection optimization that network can be faster and better.
(3) select training set from initial sample: all p+ are chosen in the definition solved according to Pareto from initial sample The Pareto solution of q dimension, then removes the value of q optimization aim in each group of Pareto solution, by the p of these removal optimization target values The Pareto solution of dimension is used as training set.
(4) production of the multiple-objection optimization of training building fights network (MOGAN):
(4a) determines training sample: from training set randomly choose half as training sample, then to training sample into The pretreatment of row data normalization, obtains pretreated training sample x.The present invention only chooses half as training sample, compares The sample size that other optimization algorithms need is few.
(4b) inputs training sample into MOGAN: pretreated training sample x is input to production confrontation network Differentiate in network D.
(4c) generates sample: in the production confrontation network of multiple-objection optimization, using network G is generated, generating the life of p dimension At sample z.
(4d) obtains differentiating result: will generate sample z and training sample x input and differentiates network D, output differentiates result.
(4e) training generates network G and differentiates network D: according to differentiating as a result, fixed generate network G, to differentiate network D into Row training, is continued to optimize, and is from training sample or to carry out self-generating net until differentiating that network D can be accurately judged to a sample The sample that network G is generated;Network D is differentiated according to differentiating as a result, fixing, and is trained, is continued to optimize, until sentencing to network G is generated Other network D cannot judge a sample be from training sample or come self-generating network G generation sample.
It obtains generating sample set after the training of (4f) successive ignition: executing step (4a)-(4e), complete once to generation The training of formula confrontation network;Judge whether to reach the number of iterations, if not reaching the number of iterations of design, repeats step (4a)-(4e), continues to train;If reaching the number of iterations of design, m is generated using network G is generated1Group p dimension is initial to be become Amount, and evaluate and measure this m1The superiority and inferiority of group sample, calculates the value of q dimension optimization aim, and evaluation number e adds m1, then m1Group p dimension Initializaing variable and q dimension optimization aim merge to obtain m1The generation sample set of group p+q dimension.
(5) judgement evaluation number: judging evaluation number, if evaluation number e reaches maximum evaluation number E at this time, Then the generation sample set that step (4f) is obtained executes step (9), further verifies effect of optimization as final result collection, Otherwise, step (6) are executed.
(6) it obtains intersecting result set: by m1The generation sample set and training sample x of group p+q dimension merge, after merging Sample set in select all p+q dimension Pareto solutions, removal Pareto solves the value of q optimization aim, obtains the initial change of p dimension The Pareto of amount is solved, and is carried out simulation binary system crossover operation to the Pareto solution of these p dimension initializaing variable, is obtained m2Group p dimension is just The intersection result set of beginning variable.The present invention generates new sample, expands sample size, carry out to sample space by intersecting More effective search, has greatly saved time cost, reduces the cost of cost.
(7) variation result set is obtained: to m2The intersection result set of group p dimension initializaing variable carries out multinomial mutation operation, obtains To m3The variation result set of group p dimension initializaing variable.
(8) evaluation obtains new initial sample: the operation method of system measures this m to evaluate where multiple target3Group p dimension The superiority and inferiority of initializaing variable, each group of initializaing variable obtain the value of corresponding q optimization aim, every group of p initializaing variable and q optimization Mesh target value collectively constitutes the initial sample of one group of new p+q dimension, and the every evaluation of initializaing variable is simultaneously combined once with optimization aim, commented Valence number e adds 1, traverses whole m3The initializaing variable of group p dimension, the final number e that evaluates add m3, obtain m3Group p+q dimension it is new initial Sample;Step (3)~step (5) are carried out again, training set is selected again, carries out the training and judgement of a new round.
(9) effect of optimization is verified:
(9a) has optimization algorithm with other and optimizes to the selected p optimized variable of step (1), obtains corresponding Results of comparison collection;
(9b) optimizes the present invention by comparing final result collection and the results of comparison collection of other optimization algorithms of the invention Effect is verified.By the experimental results showed that, effect of optimization of the invention is got well than other optimization algorithms.
For the multi-objective optimization question frequently encountered in industrial production, The present invention gives a results more preferably, is easy to The technical solution of the Multipurpose Optimal Method of training, optimizing faster based on production confrontation network.In this process, first Pass through stochastical sampling and handled to have obtained training sample, then constructs production and fight network, with obtained training sample It is trained and is generated as a result, having obtained more samples by carrying out cross and variation to generation result, then again Secondary to be trained network, until having reached the evaluation number set, the sample generated at this time is exactly final result of the invention.
By the present invention in that being optimized with a kind of mode based on two network dual trainings, original multiple target is broken The intrinsic thinking of optimization problem solution carries out simulation using a network and generates characteristic variable, another network judgement property Can be fine or not, and alternating iteration carries out the method for the two processes to optimize, as a result well, simultaneously because two networks make With three layers of fully-connected network, it is easy to trained.
Embodiment 2
For Multipurpose Optimal Method based on production confrontation network with embodiment 1, step (2) is a kind of based on deep by designing The production of degree study fights network G AN model, contacts establishing between functional value and variable, and this method is enabled to be based on choosing The training sample taken, using the potential characteristic for generating model G and learning the solution of Pareto out, and using another differentiate network D into Row judgement, calculates error, continues to optimize result.Differentiating network D and generating the two networks of network G uses classical three layers complete Connect network structure, in which:
The production of building multiple-objection optimization of the invention fights network, wherein generating network G, is one and includes input Three layers of fully-connected network of layer, hidden layer and output layer, the input layer include 5 nodes, and each node is in [- 1,1] range Random number;The hidden layer has 128 nodes, and has weight relationship between each node and input layer, initialization weight be [- 1,1] random number in range;The output layer contains n node, and n is the quantity of optimization aim, and each node contains activation letter Number relu, wherein the value of n is the variable number of specific function;
Differentiation network D therein is three layers of fully-connected network comprising input layer, hidden layer and output layer, this is defeated Entering layer includes n node, and n is the quantity of optimization aim;The hidden layer has 128 nodes, and between each node and input layer There is weight relationship, initialization weight is also the random number in [- 1,1] range, and each node contains activation primitive sigmoid; The output layer contains 1 node, indicates the probability of input sample authenticity, and each node contains activation primitive tanh.
The present invention is optimized using the mode based on two network dual trainings, has broken original multi-objective optimization question The intrinsic thinking of solution carries out simulation using a network and generates characteristic variable, another network judges performance quality, and Alternating iteration carries out the method for the two processes to optimize, as a result well, simultaneously because two networks use three layers it is complete Network is connected, is easy to trained.
Embodiment 3
Multipurpose Optimal Method based on production confrontation network is with embodiment 1-2, to generation sample described in step (6) This collection merges with training sample and carries out simulation binary system crossover operation, obtains m2The intersection result set of group p dimension initializaing variable, tool Body is that two neighboring sample is successively carried out simulation binary system crossover operation in sequence in generating sample set, first to two samples This progress probabilistic determination intersect generating two new solutions, inherits aforementioned two if probability is less than the crossing-over rate of setting A part of feature of a solution, it is possible to very close Pareto solution, the purpose of this step of the invention is for enlarged sample Quantity preferably searches for entire sample space.Specific crossover process is carried out according to the formula that such as Imitating binary system intersects:
Wherein,It is two new samples after the two neighboring sample cross of jth group, x1j(t)x2j(t) it is The prechiasmal two neighboring sample of jth group, t indicate the t generation in the genetic algorithm that simulation binary system intersects, and j indicates to intersect behaviour The jth group of work, γjIt is the intersection of jth group,
ujFor random number, and uj∈ U (0,1), η are profile exponent, η > 0.
Invention introduces the intersection thoughts in genetic algorithm, pass through the sample after merging to generation sample set with training sample This collection carries out simulation binary system and intersects, and generates new sample, expands sample size, more effectively search to sample space Rope makes production confrontation network quickly have found the data distribution of Pareto solution, is conducive to generate better sample, greatly save About time cost.
Embodiment 4
Multipurpose Optimal Method based on production confrontation network is with embodiment 1-3, to cross knot described in step (7) Fruit collection carries out multinomial mutation operation, obtains m3The variation result set of group p dimension initializaing variable, specifically concentrates intersection result Each sample carries out the mutation operation in genetic algorithm, first carries out probabilistic determination to each sample, if probability is less than the change of setting Different rate then carries out the random variation of small part to sample, and specific change procedure is related with mutation operator, then generates p dimension just The variation sample of beginning variable realizes the mutation operator form of multinomial mutation operation are as follows:
v'k=vk+δ·(uk-lk), wherein
In formula, vkIndicate father's individual, v'kIndicate son individual, ukIndicate the upper bound of p system variable value range, lk Indicate the lower bound of p system variable value range, δ1=(vk-lk)/(uk-lk), δ1=(uk-vk)/(uk-lk), k indicates multinomial In kth generation in the genetic algorithm of formula variation, u is the random number in [0, a 1] section,ηmIt is profile exponent.
Invention introduces the variation thoughts in genetic algorithm, by carrying out multinomial variation to the sample of generation, effectively The case where falling into local optimum is avoided, is easier to search out global optimum.
The present invention has not only carried out cross and variation operation, also uses newly-generated sample as new initial sample, in this way It can quickly be scanned in sample space when being trained again, be conducive to find better result faster.
A more detailed example is given below, the present invention is further described.
Embodiment 5
Multipurpose Optimal Method based on production confrontation network is comprised the following steps that with embodiment 1-4 referring to Fig. 1
(1) initial sample is obtained:
(1a) selects optimization aim: for multi-objective optimization question present in a certain system, selection determination needs to optimize Multiple targets, it is assumed that have q optimization aim, there is p system variable to produce this q optimization aim in all system variables Raw to influence, each system variable has the data set and value range of oneself.
The maximum evaluation number of (1b) setting: each group of p system variable is defined as one group of independent variable, is obtained by one group of independent variable To the value of q optimization aim be defined as dependent variable, primary evaluation, which just refers to, obtains the mistake of corresponding dependent variable by one group of independent variable Journey;Evaluation number is set in optimization process as e, maximum evaluation number is E, and evaluation number e is initialized as zero;
(1c) determines initial sample size: determining that initializaing variable quantity is m group according to maximum evaluation number E, guarantees optimization Process can be completed in maximum evaluation number E;
(1d) stochastical sampling obtains initial sample: stochastical sampling is carried out to each system variable for being related to optimization aim, The initializaing variable of m group p dimension is obtained, evaluates the superiority and inferiority for measuring this m group initializaing variable with the operation method of multiple target place system, Each group of initializaing variable obtains the value of corresponding q optimization aim, and the initializaing variable of every group of stochastical sampling and the value of optimization aim are total With the initial sample of composition one group of p+q dimension;Initializaing variable sampling is evaluated and is combined with optimization aim once, and evaluation number e adds 1, Whole m group initializaing variable is traversed, the final number e that evaluates adds m, obtains the initial sample of m p+q dimension;
(2) production for constructing multiple-objection optimization fights network (MOGAN): it includes generating network G and differentiation network D, It generates network G and differentiates that network D is all made of three layers of full Connection Neural Network, the two is confronted with each other training, is promoted constantly excellent each other Change, is built into production confrontation network (MOGAN) of multiple-objection optimization;
Production based on deep learning fights network G AN model, contacts establishing between functional value and variable, so that this Method can learn the potential characteristic of Pareto solution out using model G is generated, and use another based on the training sample chosen A differentiation network D is judged, is calculated error, is continued to optimize result.
(3) select training set from initial sample: all p+ are chosen in the definition solved according to Pareto from initial sample The Pareto solution of q dimension, then removes the value of q optimization aim in each group of Pareto solution, by the p of these removal optimization target values The Pareto solution of dimension is used as training set;
(4) production of the multi-objective optimization question of training building fights network (MOGAN):
(4a) determines training sample: from training set randomly choose half as training sample, then to training sample into The pretreatment of row data normalization, obtains pretreated training sample x;
Before each iterative process, the half-sample in the training set that previous step selects, and random alignment will be randomly selected, As the training sample x of this time iterative process, be input in production confrontation network, with guarantee training sample diversity and Reliability.
Sample data is standardized, the value of each feature is made to meet the normal distribution of (0,1), generates energy The structural data of model training is enough carried out, so that production confrontation network is more stable, the Euclidean distance between feature is calculated More rationally.
(4b) inputs training sample into MOGAN: pretreated training sample x is input to production confrontation network Differentiate in network D;
(4c) generates sample: in the production confrontation network of multi-objective optimization question, using network G is generated, generating p dimension Generation sample z;
According to the random number that the input layer for generating network G is always in [- 1,1] range, by between hidden layer node Weight relationship the hidden layer node value of the network is calculated, which is transmitted to output layer, output layer again Nodal value calculated by relu function, obtain finally with the consistent generation sample z of training sample x form.
(4d) obtains differentiating result: will generate sample z and training sample x input and differentiates network D, output differentiates result;
The hidden layer node value of the network is calculated by the weight relationship with hidden layer, hidden layer node value is passed through Sigmoid function calculates, and is transmitted to output layer, and output layer nodal value finally passes through the calculating of tanh function, obtains differentiating that network D is closed In the differentiation probability of two groups of sample authenticities.
(4e) training generates network G and differentiates network D: according to differentiating as a result, fixed generate network G, to differentiate network D into Row training, is continued to optimize, and is from training sample or to carry out self-generating net until differentiating that network D can be accurately judged to a sample The sample that network G is generated;Network D is differentiated according to differentiating as a result, fixing, and is trained, is continued to optimize, until sentencing to network G is generated Other network D cannot judge a sample be from training sample or come self-generating network G generation sample;
The target formula of use is expressed as follows:
Wherein, the diversity factor of V expression generation sample and training sample, G expression generation network, D expression differentiation network, x~ pr(x) distribution about sample characteristics x is indicated, r indicates the number of parameters of sample, z~pn(z) point about sample characteristics z is indicated Cloth, n indicate the number of parameters of sample, its expectation is asked in E expression;
(4e1) is fixed to generate network G, is optimized by the loss formula of D to differentiation network D;
When to differentiating that network D is optimized, need to maximize the sum of the mean value of two probability, therefore according to the think of of deep learning Dimension obtains the loss function for differentiating network:
Two probability that step (4d) is acquired substitute into the loss function D_loss for differentiating network, by constantly minimizing this Loss function, optimization differentiate the weight between network D difference node layer.
(4e2) is fixed to differentiate network D, is optimized by the loss formula of G to network G is generated;
When optimizing to generation network G, need to minimize the mathematical expectation of probability for generating sample, therefore according to the think of of deep learning Dimension obtains the loss function for generating network:
Two probability that step (4d) is acquired substitute into the loss function G_loss for generating network, by constantly minimizing this Loss function, optimization generate the weight between network G difference node layer.
It obtains generating sample set after the training of (4f) successive ignition: executing step (4a)-(4e), complete once to generation The training of formula confrontation network.After primary training, judge whether to reach the number of iterations, if not reaching the number of iterations of design, Step (4a)-(4e) is repeated, continues to train;If reaching the number of iterations of design, m is generated using network G is generated1Group p Initializaing variable is tieed up, and evaluates and measures this m1The superiority and inferiority of group sample, calculates the value of q dimension optimization aim, and evaluation number e adds m1, then m1Group p dimension initializaing variable and q dimension optimization aim merge to obtain m1The generation sample set of group p+q dimension;
(5) judgement evaluation number: judging evaluation number, if evaluation number e reaches maximum evaluation number E at this time, Then the generation sample set that step (4f) is obtained executes step (9), further verifies effect of optimization as final result collection, Otherwise, step (6) are executed;
(6) it obtains intersecting result set: by m1The generation sample set and training sample x of group p+q dimension merge, after merging Sample set in select all p+q dimension Pareto solutions, removal Pareto solves the value of q optimization aim, obtains the initial change of p dimension The Pareto of amount is solved, and is carried out simulation binary system crossover operation to the Pareto solution of these p dimension initializaing variable, is obtained m2Group p dimension is just The intersection result set of beginning variable;
(7) variation result set is obtained: to m2The intersection result set of group p dimension initializaing variable carries out multinomial mutation operation, obtains To m3The variation result set of group p dimension initializaing variable;
(8) evaluation obtains new initial sample: the operation method of system measures this m to evaluate where multiple target3Group p dimension The superiority and inferiority of initializaing variable, each group of initializaing variable obtain the value of corresponding q optimization aim, every group of p initializaing variable and q optimization Mesh target value collectively constitutes the initial sample of one group of new p+q dimension, and the every evaluation of initializaing variable is simultaneously combined once with optimization aim, commented Valence number e adds 1, traverses whole m3The initializaing variable of group p dimension, the final number e that evaluates add m3, obtain m3Group p+q dimension it is new initial Sample;Step (3)~step (5) are carried out again, training set is selected again, carries out the training and judgement of a new round;
(9) effect of optimization is verified:
(9a) has optimization algorithm with other and optimizes to the selected p optimized variable of step (1), obtains corresponding Results of comparison collection;
(9b) optimizes the present invention by comparing final result collection and the results of comparison collection of other optimization algorithms of the invention Effect is verified.
By the present invention in that being optimized with a kind of mode based on two network dual trainings, original multiple target is broken The intrinsic thinking of optimization problem solution carries out simulation using a network and generates characteristic variable, another network judgement property Can be fine or not, and alternating iteration carries out the method for the two processes to optimize, as a result well, simultaneously because two networks make With three layers of fully-connected network, it is easy to trained.
Below with reference to specific application example, the present invention and its technical effect are explained again.
Embodiment 6
Multipurpose Optimal Method based on production confrontation network with embodiment 1-5,
Application example: optimizing kafka system using the present invention, obtains two optimization aims of the system --- and it is maximum The optimum results of handling capacity and minimum delay.
Kafka system be it is a kind of distributed, the message system based on publish/subscribe is write using Scala, it is with can Horizontal extension and high-throughput and be widely used, by different types of company, more families as a plurality of types of data pipes and Message system uses.
Step 1, initial sample is obtained by stochastical sampling;
(1a) selects optimization aim: the optimization aim selected in this example refers to for two performances of distributed system Kafka Mark --- handling capacity and delay, according to official document, determination has 11 system configuration variable x1, x2... ..., x11Shadow is generated to it It rings, i.e. system variable p=11, as shown in table 1:
The system variable of table 1kafka
Number Name variable Variable description Value (default)
1 num.network.threads Handle the Thread Count of network request Integer (3)
2 num.io.threads Handle the Thread Count of io Integer (8)
3 queued.max.requests Largest request quantity Integer (500)
4 num.replica.fetchers Thread Count for synchronization counterpart Integer (1)
5 socket.receive.buffer.bytes Socket data receiver buffer area byte number Integer (102400)
6 socket.send.buffer.bytes Socket data send buffer area byte number Integer (102400)
7 socket.request.max.bytes Socket requests maximum number of byte Integer (104857600)
8 buffer.memory The memory byte number of buffered message record Integer (33554432)
9 batch.size Batch processing byte number Integer (16384)
10 linger.ms Delay millisecond number when record is sent Integer (0)
11 compression.type Compression algorithm type Enumerate (none)
(1b) sets maximum evaluation number: maximum evaluation number E set in this example is 300, once evaluates and refers to change This 11 system configuration variable x of kafka1, x2... ..., x11, primary system is then run, handling capacity and delay are obtained;Setting is commented Valence number is e, is initialized as 0;
(1c) determines initial sample size: being 300 according to maximum evaluation number E, determines that the quantity of stochastical sampling is 100;
(1d) stochastical sampling obtains initial sample: to 11 independent variable x1, x2... ..., x30It is carried out between range of variables Stochastical sampling obtains the random initializaing variable of 100 group of 11 dimension, changes this 11 system configuration variable x of kafka1, x2... ..., x11, primary system is then run, obtains handling capacity and delay, each run evaluation number e adds 1, by x1, x2... ..., x11With Handling capacity, delay are merged into the initial sample of one 13 dimension, and the final number e that evaluates adds 100, obtains the initial sample of 100 group of 13 dimension This.
Obtain and the system variable of kafka and optimization aim of initial sample of the invention be it is closely related, walk herein In rapid, stochastical samplings are not carried out to more than all 100 a system variables, but according to actual test determined 11 with gulp down The amount of spitting and the relevant system variable of delay are sampled, and time cost is greatly saved in this.
Step 2, construct production confrontation network (MOGAN) of multi-objective optimization question: it includes generating network G and differentiation Network D generates network G and differentiates that network D is all made of three layers of full Connection Neural Network, and the two is confronted with each other training, promotes each other It continues to optimize, is built into production confrontation network (MOGAN) of multi-objective optimization question.
The multi-objective optimization question of system software performance belongs to multi-data processing, therefore optimizes Shi Yaokao to performance Consider influencing each other between each feature.This example is by designing a kind of production confrontation network G AN based on deep learning Model is contacted establishing between the performance and feature of Kafka, allows the invention to utilize based on the training sample chosen The potential characteristic that model G learns the configuration of good performance of Kafka out is generated, and differentiates that network D is judged using another, is counted Error is calculated, result is continued to optimize.The model does not use the previous performance that carries out and optimizes weight size between necessarily searching different characteristic The thinking of relationship, and the relationship of performance and feature is sought using the fitness of network, it continues to optimize, and directly obtain matching for optimization Set parameter.The result shows that the present invention can be obtained by exploring relationship of the different characteristic inside its configuration space so that performance Preferable feature configuration.
As shown in Fig. 3, production of the invention fights network, comprising: and differentiate network model D and generates network model G, The two networks use three layers of classical fully-connected network structure, in which:
Generation network model G of the invention is one and includes input layer, hidden layer and output layer as shown in Fig. 3 (b) Three layers of fully-connected network, the input layer include 5 nodes, each node be [- 1,1] range in random number;The hidden layer There are 128 nodes, and have weight relationship between each node and input layer, initialization weight is random in [- 1,1] range Number;The output layer contains n node, and each node contains activation primitive relu, and wherein the value of n is the variable of specific function Number, in this example, the variable number n of kafka are 11.
Differentiation network model D of the invention is one and includes input layer, hidden layer and output layer as shown in Fig. 3 (a) Three layers of fully-connected network, the input layer include n node, i.e. 11 nodes;The hidden layer has 128 nodes, and each section Have a weight relationship between point and input layer, initialization weight is also the random number in [- 1,1] range, and each node contain it is sharp Function sigmoid living;The output layer contains 1 node, indicates the probability of input sample authenticity, and each node contains activation Function tanh.
Step 3, training set is selected from initial sample, as training set;
In view of also having the configuration so that better performances around the configuration feature of better performances, it should choose performance most Good sample training, this example according to actual needs, are selected in sample and are solved about handling capacity and the Pareto of delay, from these 32 (taking whole if less than 32) are chosen in Pareto solution as training set is iterated training.
Step 4, the production of the multi-objective optimization question of training building fights network (MOGAN);
Referring to Fig. 2, this step is implemented as follows:
(4a) randomly chooses half as training sample from training set, and it is pre- then to carry out data normalization to training sample Processing, obtains pretreated training sample x.
In each iterative process, the half-sample randomly selected in the training set of step 3 selection carries out random alignment, as This time training sample x of iterative process, training sample x is 16 in this example.It is input in production confrontation network, to guarantee The diversity and reliability of training sample.Why half is selected, is because if selection is very little, the potential characteristic of sample is not easy to learn It practises;If selection is too many, and not can guarantee sample diversity.
The system variable of selection is traversed, first judges whether to be enumerated variable, then directly inputs life if not enumerated variable An accepted way of doing sth fights network;If it is enumerated variable, then need to use N shapes to enumerated variable progress one-hot coding processing is belonged to The N number of state of state register pair is encoded, and each state has other independent register-bits, and there was only one when any Effectively, using a this efficient coding is that classified variable is indicated as binary vector for position;It, can general piece by one-hot coding The value for lifting variable expands to theorem in Euclid space, some value of enumerated variable just corresponds to some point of theorem in Euclid space, while can be with Enumerated variable is discretized into the combination of multiple variables, is directly handled with being generated formula confrontation network, so that the Europe between variable It is more reasonable that formula distance calculates.
Variable-value is standardized, the value of each variable is made to meet the normal distribution of (0,1), generates energy The structural data of model training is enough carried out, so that production confrontation network is more stable, the Euclidean distance between variable is calculated More rationally.Why do so, is the distance between variable because in most of machine learning or deep learning algorithm It calculates or the calculating of similarity is very important, and this example is all in theorem in Euclid space for the calculating of distance or similarity Similarity calculation is carried out, and production fights network as a kind of deep learning algorithm, the pretreatment for needing to be normalized mentions Rise algorithm stability and robustness.
Pretreated training sample x is input to production confrontation network by (4b);
(4c), which is used, generates network G, generation and the consistent generation sample z of training sample x dimension,
According to the random number that the input layer for generating network G is always in [- 1,1] range, by between hidden layer node Weight relationship the hidden layer node value of the network is calculated, which is transmitted to output layer, output layer again Nodal value calculated by relu function, obtain finally it is consistent with training sample x form 11 dimension generation sample z.
(4d) will generate the training sample x that sample z and step (4b) are selected and input differentiation network D respectively, by with it is hidden The hidden layer node value of the network is calculated in weight relationship containing layer, and hidden layer node value is calculated by sigmoid function, passes It is delivered to output layer, output layer nodal value finally passes through the calculating of relu function, obtains differentiating network D about two groups of sample authenticities Differentiate probability.
(4e) optimizes production confrontation network according to target formula:
The target formula is expressed as follows:
Wherein, V indicates to generate the diversity factor of sample z and training sample x, and G indicates to generate network, and D indicates to differentiate network, x ~pr(x) data distribution about training sample x is indicated, r indicates the variable quantity of training sample, z~pn(z) it indicates about life At the data distribution of sample z, n indicates to generate the variable quantity of sample,The equal of training sample x data distribution is sought in expression Value,The mean value of sample z data distribution is sought survival into expression.
(4e1) optimizes differentiation network D:
It can be seen that by above-mentioned target formula, when to differentiating that network D is optimized, need to maximize the mean value of two probability The sum of, therefore according to the thinking of deep learning, obtain the loss function for differentiating network D:
Two probability that step (4c) is acquired substitute into the loss function D_loss for differentiating network, by constantly changing difference Weight between node layer, i.e. w in Fig. 3 (a), constantly minimizes the loss function, is optimized with this and differentiates network D.
(4e2) is optimized to network G is generated:
It can be seen that by above-mentioned target formula, when optimizing to generation network G, need to minimize the probability for generating sample Mean value, therefore according to the thinking of deep learning, obtain the loss function for generating network G:
Two probability that step (4d) is acquired substitute into the loss function G_loss for generating network, by constantly changing difference Weight between node layer, i.e. w in Fig. 3 (b), constantly minimizes the loss function, is optimized with this and generates network D.
By the two processes, the ability for generating network generation authentic specimen can be promoted, and differentiate that network judgement sample is true The ability of property can also be promoted.
(4f) repeats (4a) and arrives (4e), until reaching the number of iterations for meeting and setting, according to actual needs, saves last The generation sample z' that network finally generates is generated several times, and as final optimization pass as a result, in this example, z' expression generates 100 The initializaing variable of 11 dimension of group, then changes this 11 system configuration variable x of kafka1, x2... ..., x11, primary system is run, is obtained To handling capacity and delay, merge to obtain the generation sample set of 100 group of 13 dimension with initializaing variable.
The number of iterations of the present invention rule of thumb value, this example value 300000 times.
Step 5, judge to evaluate number;
In above-mentioned steps, it is once to comment that 11 dimension system variables of each pair of kafka system, which are once changed and run, Valence, using the generation sample set of step 4 as final result collection, carries out step 9, further if evaluation number reaches limitation at this time Optimum results are verified, otherwise, carry out step 6, training optimization production fights network again.
Step 6, binary system crossover operation is simulated;
Generation sample set and training sample x that 100 group 13 in step 4 is tieed up are merged, from the sample set after merging 13 all dimension Pareto solutions are selected, removal Pareto solves the value of 2 optimization aims, obtains the Pareto of 11 dimension initializaing variables Solution carries out simulation binary system crossover operation to the Pareto solution of these 11 dimension initializaing variables, obtains m211 dimension initializaing variable of group Intersect result set.m2For the quantity for intersecting result set, in operating process because the reason of crossing-over rate, some samples intersect, Some is not intersected, m2Quantity be not fixed.
Intersection is the step key operation in genetic algorithm, simulates producing offspring in nature, refers to two phases The chromosome mutually matched is exchanged with each other its portion gene in some way, to form two new individuals.Refer in the present invention Be two generate solution between data cross, generate new solution, a part of feature of both of the aforesaid solution inherited, so having It may closely Pareto solve, the purpose of this step is preferably to search for entire sample space for enlarged sample quantity.
In this example, the system variable tieed up to two group 11 carries out data cross, generates two groups of 11 new dimension system variables, A part of feature for inheriting aforementioned two groups of system variables, it is possible to the very close Pareto about handling capacity and delay Solution.
Step 7, multinomial makes a variation;
To m in step 62The intersection result set of 11 dimension initializaing variable of group carries out multinomial mutation operation, obtains m311 dimension of group is just The variation result set of beginning variable.The reason of in operating process because of aberration rate, some sample variations, some does not make a variation, so m3 Quantity be not fixed.
Variation is another step key operation of genetic algorithm, simulates the gene mutation in nature, i.e. child replicates father Possible (probability of very little) generates certain copy errors when female gene, and variation generates new chromosome, shows new character.In Middle finger of the present invention solves the random variation of progress a part to generating, and generates new solution, the purpose of this step is in order to avoid falling into Locally optimal solution is easier to search out globally optimal solution.
In this example, the data for exactly tieing up system variable to a certain group 11 are changed at random.
Step 8, evaluation obtains new initial sample;
For m in step 73The variation result set of 11 dimension initializaing variable of group, obtains this m by operating system3Group initializaing variable Handling capacity and delay, each run evaluation number e all add 1, by x1, x2... ..., x11One is merged into handling capacity, delay The initial sample of 13 dimensions, the final number e that evaluates add m3, obtain m3The initial sample of 13 dimension of group, carries out step 2~step 5 again, Training optimization production fights network.
Step 9, effect of optimization is verified;
According to production fight network generate final result collection, on kafka again operation to be tested, obtain as The HV value of the following table 2.HV is hypervolume index, and what it was measured is the non-dominant disaggregation and reference obtained by multi-objective optimization algorithm The volume in the dimension region in the object space that point surrounds, the value of HV is bigger, illustrates that optimum results are better.
Table 2: the HV value of the present invention and other optimization algorithms
Experimental data Different Optimization method The present invention NSGAII IBEA MOEA/D MOPSO MOEA/D-EGO K-RVEA PAREGO CSEA
First time data 0.9742 0.9547 0.9559 0.9578 0.9566 0.9657 0.9584 0.9665 0.9676
Second of data 0.9705 0.9658 0.9659 0.9653 0.9642 0.9682 0.9587 0.9677 0.9665
As seen from Table 2, the MOGAN algorithm in the present invention is bigger compared to other algorithms HV value, this illustrates that the present invention is better than Other existing optimization algorithms also demonstrate the Multipurpose Optimal Method for fighting network the present invention is based on production and are directed to kafka The validity and reasonability of performance optimization problem.
Embodiment 7
With embodiment 1-6, this example is one and is applied to solve Multipurpose Optimal Method based on production confrontation network The application example of the multi-objective optimization question of ZDT1.
Application example: ZDT1 minimum function value f1 and f2 are found using the present invention
ZDT is a kind of test function, wherein contain multi-objective optimization question, share ZDT1, ZDT2 ..., ZDT6 six Function describes entire optimization process and as a result, the formula of ZDT1 is as follows in this example by taking ZDT1 as an example:
f1=x1
0≤xi≤ 1, i=1 ..., n
There are the system variable between 30 0-1, the system variable x in this example in this example1, x2... ..., x30It indicates, In In this example, x1, x2... ..., x30Namely independent variable has an impact optimization aim, i.e. p=30, and optimization aim is ZDT1's Two functional values f1 and f2, that is, dependent variable, i.e. q=2.Optimization aim of the invention be f1 and f2 are minimized, but It is that cannot be directly changed f1 and f2, f1 and f2 can only be changed by changing independent variable x, realize the minimum to f1 and f2.This Equally set maximum evaluation number E in example is 300, determines the sample that initial sample is tieed up in this example for 100 group 32.
By the identical step of embodiment 6, after carrying out constantly training optimization to production confrontation network, available fixation The generation sample (generating sample only includes 30 independents variable, does not include f1 and f2) of 30 dimensions of number, by the formula of ZDT1 into The f1 and f2 for generating sample is calculated in row, draws the function curve of optimization sample f1 and f2, compares of the invention with other The functional value size of algorithm compares, as a result if Fig. 4, Fig. 4 are the present invention and other optimization algorithms for ZDT1 function The curve graph of optimum results, the abscissa of Fig. 4 are the values of f1, and ordinate is the value of f2, share MOGAN and k-rvea of the present invention, Moead, nsga, mopso and ibea etc. totally 6 optimization algorithms as a result, wherein curve of the invention is no any symbol Solid line.
Since the target in this example is to keep f1 and f2 as small as possible, and from fig. 4, it can be seen that under ZDT1 function, this hair Other comparison algorithms of 2.3-4.3 range internal ratio want low to curve representated by MOGAN in bright within the scope of 0-0.5, on f2 on f1 Very much, this illustrates that MOGAN method of the present invention is better than other existing optimization algorithms, also demonstrates that the present invention is based on production confrontation The validity and reasonability of the Multipurpose Optimal Method of network.
Embodiment 8
With embodiment 1-7, this example is one and is applied to solve Multipurpose Optimal Method based on production confrontation network The application example of the multi-objective optimization question of ZDT6.
Application example: ZDT6 minimum function value f1 and f2 are found using the present invention
ZDT is the common test function of multi-objective optimization question, share ZDT1, ZDT2 ..., six functions of ZDT6, In In this example by taking ZD6 as an example, entire optimization process is described and as a result, the formula of ZDT6 is as follows:
0≤xi≤ 1, i=1 ..., n
There are the system variable between 30 0-1, the system variable x in this example in this example1, x2... ..., x30It indicates, In In this example, x1, x2... ..., x30Namely independent variable has an impact optimization aim, i.e. p=30, and optimization aim is ZDT6's Two functional values f1 and f2, that is, dependent variable, i.e. q=2.Optimization aim of the invention be f1 and f2 are minimized, but It is that cannot be directly changed f1 and f2, f1 and f2 can only be changed by changing independent variable x, realize the minimum to f1 and f2.This Equally set maximum evaluation number E in example is 300, determines the sample that initial sample is tieed up in this example for 100 group 32.
By the identical step of embodiment 6, after carrying out constantly training optimization to production confrontation network, available fixation The generation sample (generating sample only includes 30 independents variable, does not include f1 and f2) of 30 dimensions of number, by the formula of ZDT6 into The f1 and f2 for generating sample is calculated in row, draws the function curve of optimization sample f1 and f2, compares of the invention with other The functional value size of algorithm compares, as a result if Fig. 5, Fig. 5 are the present invention and other optimization algorithms for ZDT1 function The curve graph of optimum results, the abscissa of Fig. 5 are the values of f1, and ordinate is the value of f2, share MOGAN and k-rvea of the present invention, Moead, nsga, mopso and ibea etc. totally 6 optimization algorithms as a result, wherein curve of the invention is no any symbol Solid line.
Since the target in this example is to keep f1 and f2 as small as possible, and from fig. 5, it can be seen that under ZDT6 function, this hair Other comparison algorithms of curve ratio representated by bright MOGAN want low, this illustrates that MOGAN method of the present invention is better than other existing optimizations Algorithm, also demonstrate the present invention is based on production confrontation network Multipurpose Optimal Method validity and reasonability.
Embodiment 9
Multipurpose Optimal Method based on production confrontation network with embodiment 1-8,
Application example: the maximum detection rates of image detecting method and maximum detection accuracy are optimized using the present invention
Specific implementation step is same as Example 6, and in image detection, improving detection rates and improving detection accuracy is two A conflicting target, i.e. optimization aim q=2, when improving detection rates, detection accuracy will be reduced, and improve detection essence When spending, detection rates will be reduced.After optimization of the invention, it is last it is available about recognition rate and accuracy of identification this Multiple optimal solutions of two targets therefrom select a detection rates and all very high combination of detection accuracy, can be used for specific Practical application in.
In brief, a kind of Multipurpose Optimal Method (MOGAN) based on production confrontation network disclosed by the invention, it is main It is high to solve prior art time cost, training difficulty is excessive, and training network is easy the problems such as collapsing.Its implementation: (1) it obtains Take initial sample;(2) production for constructing multiple-objection optimization fights network;(3) training set is selected from initial sample;(4) it instructs The production for practicing the multiple-objection optimization of building fights network;(5) judgement evaluation number;(6) it obtains intersecting result set;(7) it obtains Make a variation result set;(8) evaluation obtains new initial sample;(9) effect of optimization is verified.Present invention reduces time costs, improve Network robustness and stability, effect of optimization is obvious, can be used for the resource allocation of multiple targets, the production scheduling of multiple products With the optimization etc. of the multiple performances of various software systems.

Claims (4)

1. a kind of Multipurpose Optimal Method based on production confrontation network, which comprises the steps of:
(1) initial sample is obtained:
(1a) selects optimization aim: for multi-objective optimization question present in a certain system, selecting determining needs to optimize more A target, it is assumed that have q optimization aim, there is p system variable to generate shadow to this q optimization aim in all system variables It rings, each system variable has the data set and value range of oneself;
The maximum evaluation number of (1b) setting: each group of p system variable is defined as one group of independent variable, is obtained by one group of independent variable The value of q optimization aim is defined as dependent variable, and primary evaluation, which just refers to the process of, obtains corresponding dependent variable by one group of independent variable;It is excellent Evaluation number is set during changing as e, maximum evaluation number is E, and evaluation number e is initialized as zero;
(1c) determines initial sample size: determining that initializaing variable quantity is m group according to maximum evaluation number E, guarantees optimization process It can be completed in maximum evaluation number E;
(1d) stochastical sampling obtains initial sample: carrying out stochastical sampling to each system variable for being related to optimization aim, obtains m The initializaing variable of group p dimension evaluates the superiority and inferiority for measuring this m group initializaing variable with the operation method of multiple target place system, each Group initializaing variable obtains the value of corresponding q optimization aim, common group of value of the initializaing variable of every group of stochastical sampling and optimization aim The initial sample tieed up at one group of p+q;Initializaing variable sampling is evaluated and is combined with optimization aim once, and evaluation number e adds 1, traversal Whole m group initializaing variables, the final number e that evaluates add m, obtain the initial sample of m p+q dimension;
(2) construct multiple-objection optimization production fight network: it include generate network G and differentiate network D, generate network G with Differentiate that network D is all made of three layers of full Connection Neural Network, the two is confronted with each other training, and promotion is continued to optimize each other, is built into more The production of objective optimization fights network;
(3) select training set from initial sample: all p+q dimensions are chosen in the definition solved according to Pareto from initial sample Pareto solution, then remove the value of q optimization aim in each group of Pareto solution, by these removal optimization target values p dimension Pareto solution be used as training set;
(4) production of the multiple-objection optimization of training building fights network:
(4a) determines training sample: randomly choosing half from training set as training sample, then counts to training sample It is pre-processed according to standardization, obtains pretreated training sample x;
(4b) inputs training sample into MOGAN: pretreated training sample x being input to the differentiation of production confrontation network In network D;
(4c) generates sample: in the production confrontation network of multiple-objection optimization, using network G is generated, generating the generation sample of p dimension This z;
(4d) obtains differentiating result: will generate sample z and training sample x input and differentiates network D, output differentiates result;
(4e) training generates network G and differentiates network D: according to differentiation as a result, fixed generate network G, instructing to differentiation network D Practice, continue to optimize, is from training sample or to carry out self-generating network G until differentiating that network D can be accurately judged to a sample The sample of generation;Network D is differentiated according to differentiating as a result, fixing, and is trained, is continued to optimize again to network G is generated, until differentiating Network D cannot judge a sample be from training sample or come self-generating network G generation sample;
It obtains generating sample set after the training of (4f) successive ignition: executing step (4a)-(4e), complete once to production pair The training of anti-network;Judge whether to reach the number of iterations, if not reaching the number of iterations of design, repeat step (4a)- (4e) continues to train;If reaching the number of iterations of design, m is generated using network G is generated1Group p ties up initializaing variable, and This m is measured in evaluation1The superiority and inferiority of group sample, calculates the value of q dimension optimization aim, and evaluation number e adds m1, then m1Group p dimension is initial to be become Amount and q dimension optimization aim merge to obtain m1The generation sample set of group p+q dimension;
(5) judgement evaluation number: judging evaluation number, if evaluation number e reaches maximum evaluation number E at this time, will walk Suddenly then the generation sample set that (4f) is obtained executes step (9), further verifies effect of optimization as final result collection, otherwise, It executes step (6);
(6) it obtains intersecting result set: by m1The generation sample set and training sample x of group p+q dimension merge, from the sample after merging This concentration selects all p+q dimension Pareto solutions, and removal Pareto solves the value of q optimization aim, obtains p dimension initializaing variable Pareto solution carries out simulation binary system crossover operation to the Pareto solution of these p dimension initializaing variable, obtains m2Group p dimension is initial to be become The intersection result set of amount;
(7) variation result set is obtained: to m2The intersection result set of group p dimension initializaing variable carries out multinomial mutation operation, obtains m3Group The variation result set of p dimension initializaing variable;
(8) evaluation obtains new initial sample: the operation method of system measures this m to evaluate where multiple target3Group p dimension is initial The superiority and inferiority of variable, each group of initializaing variable obtain the value of corresponding q optimization aim, every group of p initializaing variable and q optimization aim Value collectively constitute the initial sample of one group of new p+q dimension, the every evaluation of initializaing variable simultaneously combine once with optimization aim, is evaluated secondary Number e adds 1, traverses whole m3The initializaing variable of group p dimension, the final number e that evaluates add m3, obtain m3The new initial sample of group p+q dimension This;Carry out step (3)~step (5) again;
(9) effect of optimization is verified:
(9a) has optimization algorithm with other and optimizes to the selected p optimized variable of step (1), obtains corresponding control Result set;
(9b) passes through the results of comparison collection for comparing final result collection and other optimization algorithms of the invention, to effect of optimization of the present invention It is verified.
2. the Multipurpose Optimal Method according to claim 1 based on production confrontation network, which is characterized in that step (2) production of the building multiple-objection optimization described in fights network, wherein generating network G, is one and includes input layer, hidden layer With three layers of fully-connected network of output layer, which includes 5 nodes, and each node is the random number in [- 1,1] range; The hidden layer has 128 nodes, and has weight relationship between each node and input layer, and initialization weight is in [- 1,1] range Random number;The output layer contains p node, and p is the quantity of system variable, and each node contains activation primitive relu;
Differentiation network D therein is three layers of fully-connected network comprising input layer, hidden layer and output layer, the input layer Comprising p node, p is the quantity of system variable;The hidden layer has 128 nodes, and has the right between each node and input layer Series of fortified passes system, initialization weight is also the random number in [- 1,1] range, and each node contains activation primitive sigmoid;This is defeated Layer contains 1 node out, indicates the probability for differentiating network D input sample authenticity, and each node contains activation primitive tanh.
3. the Multipurpose Optimal Method according to claim 1 based on production confrontation network, which is characterized in that step (6) simulation binary system crossover operation is carried out to generation sample set described in, obtains m2The intersection result of group p dimension initializaing variable Two neighboring sample is successively carried out simulation binary system crossover operation in sequence specifically in generating sample set by collection, first right Two samples carry out probabilistic determination, if probability is less than the crossing-over rate of setting, intersect generating two new solutions, specific to intersect Process is carried out according to the formula that such as Imitating binary system intersects:
Wherein,It is two new samples after the two neighboring sample cross of jth group, x1j(t) x2jIt (t) is to intersect The preceding two neighboring sample of jth group, t indicate the t generation in the genetic algorithm that simulation binary system intersects, and j indicates crossover operation Jth group, γjIt is the crossing-over rate of jth group,
ujFor random number, and uj∈ U (0,1), η are profile exponent, η > 0.
4. the Multipurpose Optimal Method according to claim 1 based on production confrontation network, which is characterized in that step (7) multinomial mutation operation is carried out to generation sample set described in, obtains m3The variation result set of group p dimension initializaing variable, tool Body is the mutation operation carried out in genetic algorithm to each generation sample generated in sample set, first carries out probability to each sample Judgement carries out the random variation of small part, specific change procedure and variation to sample if probability is less than the aberration rate of setting Operator is related, then generates the variation sample of p dimension initializaing variable, realizes the mutation operator form of multinomial mutation operation are as follows:
v'k=vk+δ·(uk-lk), wherein
In formula, vkIndicate father's individual, v'kIndicate son individual, ukIndicate the upper bound of p system variable value range, lkIndicate p The lower bound of a system variable value range, δ1=(vk-lk)/(uk-lk), δ1=(uk-vk)/(uk-lk), the variation of k representative polynomial Genetic algorithm in kth generation, u is the random number in [0, a 1] section,ηmIt is profile exponent.
CN201910688044.0A 2019-07-29 2019-07-29 Multipurpose Optimal Method based on production confrontation network Pending CN110533221A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910688044.0A CN110533221A (en) 2019-07-29 2019-07-29 Multipurpose Optimal Method based on production confrontation network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910688044.0A CN110533221A (en) 2019-07-29 2019-07-29 Multipurpose Optimal Method based on production confrontation network

Publications (1)

Publication Number Publication Date
CN110533221A true CN110533221A (en) 2019-12-03

Family

ID=68661960

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910688044.0A Pending CN110533221A (en) 2019-07-29 2019-07-29 Multipurpose Optimal Method based on production confrontation network

Country Status (1)

Country Link
CN (1) CN110533221A (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111481935A (en) * 2020-04-08 2020-08-04 网易(杭州)网络有限公司 Configuration method, device, equipment and medium for AI models of games with different styles
CN112488486A (en) * 2020-11-25 2021-03-12 吉林大学 Multi-criterion decision method based on zero sum game
CN112508093A (en) * 2020-12-03 2021-03-16 北京百度网讯科技有限公司 Self-training method and device, electronic equipment and readable storage medium
WO2021128805A1 (en) * 2019-12-24 2021-07-01 浙江大学 Wireless network resource allocation method employing generative adversarial reinforcement learning
CN113434459A (en) * 2021-06-30 2021-09-24 电子科技大学 Network-on-chip task mapping method based on generation of countermeasure network

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2021128805A1 (en) * 2019-12-24 2021-07-01 浙江大学 Wireless network resource allocation method employing generative adversarial reinforcement learning
CN111481935A (en) * 2020-04-08 2020-08-04 网易(杭州)网络有限公司 Configuration method, device, equipment and medium for AI models of games with different styles
CN111481935B (en) * 2020-04-08 2023-04-18 网易(杭州)网络有限公司 Configuration method, device, equipment and medium for AI models of games with different styles
CN112488486A (en) * 2020-11-25 2021-03-12 吉林大学 Multi-criterion decision method based on zero sum game
CN112488486B (en) * 2020-11-25 2022-04-15 吉林大学 Multi-criterion decision method based on zero sum game
CN112508093A (en) * 2020-12-03 2021-03-16 北京百度网讯科技有限公司 Self-training method and device, electronic equipment and readable storage medium
CN112508093B (en) * 2020-12-03 2022-01-28 北京百度网讯科技有限公司 Self-training method and device, electronic equipment and readable storage medium
CN113434459A (en) * 2021-06-30 2021-09-24 电子科技大学 Network-on-chip task mapping method based on generation of countermeasure network

Similar Documents

Publication Publication Date Title
CN110533221A (en) Multipurpose Optimal Method based on production confrontation network
Cantú-Paz et al. An empirical comparison of combinations of evolutionary algorithms and neural networks for classification problems
Derrac et al. IFS-CoCo: Instance and feature selection based on cooperative coevolution with nearest neighbor rule
Tian et al. A fast clustering based evolutionary algorithm for super-large-scale sparse multi-objective optimization
CN115481727A (en) Intention recognition neural network generation and optimization method based on evolutionary computation
Hu Flow-based tolerance rough sets for pattern classification
Pangilinan et al. Pareto-optimality of oblique decision trees from evolutionary algorithms
De Matteis et al. A new approach to fuzzy random forest generation
Yu et al. Unsupervised euclidean distance attack on network embedding
Farooq Genetic algorithm technique in hybrid intelligent systems for pattern recognition
Parvin et al. A scalable method for improving the performance of classifiers in multiclass applications by pairwise classifiers and GA
Siwik et al. Elitist evolutionary multi-agent system in solving noisy multi-objective optimization problems
CN112183598A (en) Feature selection method based on genetic algorithm
Hong et al. Improving performance insensitivity of large-scale multiobjective optimization via Monte Carlo tree search
CN113704570B (en) Large-scale complex network community detection method based on self-supervision learning type evolution
CN114819181A (en) Multi-target federal learning evolution method based on improved NSGA-III
Hu et al. Apenas: An asynchronous parallel evolution based multi-objective neural architecture search
Davarynejad et al. Accelerating convergence towards the optimal pareto front
Moyano et al. Tree-shaped ensemble of multi-label classifiers using grammar-guided genetic programming
Van Truong et al. A Coevolutionary approach for classification problems: Preliminary results
Khaparde et al. A new distributed differential evolution algorithm
Joldos et al. A parallel evolutionary approach to community detection in complex networks
Antonelli et al. Multi-objective evolutionary design of fuzzy rule-based systems
Harrison Network similarity measures and automatic construction of graph models using genetic programming
Berlanga et al. A novel genetic cooperative-competitive fuzzy rule based learning method using genetic programming for high dimensional problems

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
WD01 Invention patent application deemed withdrawn after publication
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20191203