CN108053068A - The method that mankind attacker cooperation behavior modeled and formulated corresponding Defending Policy - Google Patents
The method that mankind attacker cooperation behavior modeled and formulated corresponding Defending Policy Download PDFInfo
- Publication number
- CN108053068A CN108053068A CN201711324273.1A CN201711324273A CN108053068A CN 108053068 A CN108053068 A CN 108053068A CN 201711324273 A CN201711324273 A CN 201711324273A CN 108053068 A CN108053068 A CN 108053068A
- Authority
- CN
- China
- Prior art keywords
- mrow
- msub
- msubsup
- munder
- attacker
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 37
- 238000005457 optimization Methods 0.000 claims abstract description 21
- 230000003044 adaptive effect Effects 0.000 claims abstract description 5
- 230000006399 behavior Effects 0.000 claims description 19
- 238000012886 linear function Methods 0.000 claims description 18
- 230000006870 function Effects 0.000 claims description 10
- 238000009472 formulation Methods 0.000 claims description 9
- 239000000203 mixture Substances 0.000 claims description 9
- 230000007246 mechanism Effects 0.000 claims description 7
- 230000003247 decreasing effect Effects 0.000 claims description 4
- 230000004048 modification Effects 0.000 claims description 4
- 238000012986 modification Methods 0.000 claims description 4
- 230000011218 segmentation Effects 0.000 claims description 4
- 102000002423 Octamer Transcription Factor-6 Human genes 0.000 claims description 3
- 108010068113 Octamer Transcription Factor-6 Proteins 0.000 claims description 3
- 239000005018 casein Substances 0.000 claims description 3
- BECPQYXYKAMYBN-UHFFFAOYSA-N casein, tech. Chemical compound NCCCCC(C(O)=O)N=C(O)C(CC(O)=O)N=C(O)C(CCC(O)=N)N=C(O)C(CC(C)C)N=C(O)C(CCC(O)=O)N=C(O)C(CC(O)=O)N=C(O)C(CCC(O)=O)N=C(O)C(C(C)O)N=C(O)C(CCC(O)=N)N=C(O)C(CCC(O)=N)N=C(O)C(CCC(O)=N)N=C(O)C(CCC(O)=O)N=C(O)C(CCC(O)=O)N=C(O)C(COP(O)(O)=O)N=C(O)C(CCC(O)=N)N=C(O)C(N)CC1=CC=CC=C1 BECPQYXYKAMYBN-UHFFFAOYSA-N 0.000 claims description 3
- 235000021240 caseins Nutrition 0.000 claims description 3
- 238000000354 decomposition reaction Methods 0.000 claims description 3
- 230000000007 visual effect Effects 0.000 claims description 3
- 230000001419 dependent effect Effects 0.000 claims description 2
- 238000002474 experimental method Methods 0.000 abstract description 10
- 238000012360 testing method Methods 0.000 abstract description 6
- FNMKZDDKPDBYJM-UHFFFAOYSA-N 3-(1,3-benzodioxol-5-yl)-7-(3-methylbut-2-enoxy)chromen-4-one Chemical compound C1=C2OCOC2=CC(C2=COC=3C(C2=O)=CC=C(C=3)OCC=C(C)C)=C1 FNMKZDDKPDBYJM-UHFFFAOYSA-N 0.000 description 7
- 230000008901 benefit Effects 0.000 description 3
- 238000010586 diagram Methods 0.000 description 3
- 238000007476 Maximum Likelihood Methods 0.000 description 2
- 230000008859 change Effects 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 241000208340 Araliaceae Species 0.000 description 1
- 241000282327 Felis silvestris Species 0.000 description 1
- 235000005035 Panax pseudoginseng ssp. pseudoginseng Nutrition 0.000 description 1
- 235000003140 Panax quinquefolius Nutrition 0.000 description 1
- 230000009471 action Effects 0.000 description 1
- 230000006978 adaptation Effects 0.000 description 1
- JLQUFIHWVLZVTJ-UHFFFAOYSA-N carbosulfan Chemical compound CCCCN(CCCC)SN(C)C(=O)OC1=CC=CC2=C1OC(C)(C)C2 JLQUFIHWVLZVTJ-UHFFFAOYSA-N 0.000 description 1
- 230000007812 deficiency Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 238000011156 evaluation Methods 0.000 description 1
- 238000010304 firing Methods 0.000 description 1
- 235000008434 ginseng Nutrition 0.000 description 1
- 230000008520 organization Effects 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 230000010415 tropism Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q10/00—Administration; Management
- G06Q10/04—Forecasting or optimisation specially adapted for administrative or management purposes, e.g. linear programming or "cutting stock problem"
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q10/00—Administration; Management
- G06Q10/06—Resources, workflows, human or project management; Enterprise or organisation planning; Enterprise or organisation modelling
- G06Q10/063—Operations research, analysis or management
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q50/00—Systems or methods specially adapted for specific business sectors, e.g. utilities or tourism
- G06Q50/10—Services
- G06Q50/26—Government or public services
- G06Q50/265—Personal security, identity or safety
Landscapes
- Business, Economics & Management (AREA)
- Engineering & Computer Science (AREA)
- Human Resources & Organizations (AREA)
- Strategic Management (AREA)
- Economics (AREA)
- Tourism & Hospitality (AREA)
- Entrepreneurship & Innovation (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Marketing (AREA)
- General Physics & Mathematics (AREA)
- General Business, Economics & Management (AREA)
- Development Economics (AREA)
- Educational Administration (AREA)
- Quality & Reliability (AREA)
- Operations Research (AREA)
- Game Theory and Decision Science (AREA)
- Computer Security & Cryptography (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Primary Health Care (AREA)
- Medicines Containing Antibodies Or Antigens For Use As Internal Diagnostic Agents (AREA)
Abstract
The present invention discloses the mankind attacker cooperation behavior modeling in a kind of security protection and the method for formulating corresponding Defending Policy, first, an adaptive human behavior model is proposed, for optimizing distribution of the tool there are two limited defence resource in the repetition SSG for the attacker that may cooperate;Secondly, devise an efficient algorithm and approximate solve the Nonlinear Nonconvex optimization problem obtained by model, final output Defending Policy;Finally, true man's simulated experiment has been carried out to test the validity of model of the present invention and method.
Description
Technical field
The present invention relates to the mankind attacker cooperation behavior modeling in a kind of security protection and formulate corresponding Defending Policy
Method belongs to technical field of safety protection.
Background technology
The safe games of Stackelberg (SSG) are applied in a variety of safeguard protection scenes in real life, such as beautiful
State's Coast Guard, federal air marshal, Los Angeles Airport etc..
Instantly it is to patrol to needing the most common defence method that target to be protected is taken in the case where defence resource is limited
It patrols.Patrolman only has limited defence resource (such as human resources), so they cannot be a piece of very big in guard of same time
Area every nook and cranny.In addition, attacker can be with the patrol rule of observation patrol person, and corresponding attack action is designed to keep away
Exempt to be grabbed.Therefore, it is significant for defence organization's optimization Defending Policy.
In traditional single-wheel SSG, defender (leaders) arranges patrol strategy first, then attacker (follower) root
Target attack is selected to the observation of Defending Policy according to it.
The main problem of traditional single-wheel SSG is the incomplete reasonability for not accounting for mankind attacker, and without abundant
Utilize the passing data-optimized Defending Policy of attack.
The content of the invention
Goal of the invention:For problems of the prior art and deficiency, the present invention provides the people in a kind of security protection
The method that class attacker cooperation behavior modeled and formulated corresponding Defending Policy, first, it is proposed that an adaptive human behavior
Model, for optimizing distribution of the tool there are two limited defence resource in the repetition SSG for the attacker that may cooperate;Secondly, design
One efficient algorithm approximate solve the Nonlinear Nonconvex optimization problem obtained by model, final output Defending Policy;
Finally, true man's simulated experiment has been carried out to test the validity of model of the present invention and method.
Technical solution:A kind of side of the modeling of mankind attacker cooperation behavior and the corresponding Defending Policy of formulation in security protection
Method, including following content:
(1) cooperative mechanism in SSG is repeated
Repeatedly defender and two attackers, the two attackers can choose whether to cooperate there are one cooperative games;It is anti-
The person of keeping periodically arranges new Defending Policy as the leader in game, and attacker is as follower, according to it to defender
Attack is made in the observation of the Defending Policy of arrangement;This game is formed more by taking turns, thus defender can periodically according to front-wheel
The data that are collected into secondary change Defending Policy;Attacker independently attacks or cooperation attack;In the case where independently attacking,
Whether the income that attacker attacks every time is only dependent upon the attacker this attack and succeeds.And in the case where cooperation is attacked, it attacks
The income that the person of hitting attacks every time by be two people's income sums half;An extra returns ∈ is introduced simultaneously, i.e., when an attack
When person has successfully carried out once cooperation attack, he will obtain extra returns ∈;
Attackers need to make two main selections in game:1) selection attack target and 2) choose whether with separately
One attacker cooperates.Only when two attackers select cooperation in certain attack, a cooperative relationship just calculates foundation.
The defence resource summation M possessed by defender defends a target collection T={ 1 ..., | T | }, and T is by T1With
T2Two are not overlapped subset and are combined into;T1Represent the target collection that first attacker may attack, T2Represent second attacker
The target collection that may be attacked, and T1=T-T2;Vector x is defended, per one-dimensional element xiDefender is represented in each target
The defended probability of the defence resource namely target i of input.
Income of one attacker in target of attack i depend on 1) this target it is whether defended and 2) two attack
Whether person cooperates.When an attacker selects independent attack, the income that he attacks not defended target i isAttack quilt
The income of the target i of defence isRelatively, defender is in the income of the previous caseIn the income of latter situation
ForThe total revenue of defender for it in two incomes by target of attack and.AndIf two
A attacker agrees to cooperate, they will obtain extra returns ∈ when attacking not defended target.
(2) incomplete reasonability and adaptivity
In some illegal activities, attacker is typically considered Imperfect Rationality.For the mankind's in SSG is repeated
Incomplete reasonability proposes a model.
Tendentiousness of the attacker for some target is defined first.
Define 1:One attacker is in R wheels to the tendentiousness of target iIt is defined by following equation:
WhereinIt isWithDifference,Represent the single attack that attacker r wheels carry out at target i
Average yield,Represent the single attack average yield for all attacks that attacker carries out in r wheels, QrFor a vector,
Its i-th dimension isC be a constant, Var (Qr) represent { QrVariance.Therefore in fixation { QrDistribution in the case of, increase
AddAlso can accordingly increase.First equation explanationForAverage (r=1 ..., R).
When attacker tends to target of attack iTo be just, on the contrary will be negative, and attacker be not revealed inclining for target
Tropism is set to 0.
Adaptive subjectivity revenue function:
D=1/ (N are setr- r), wherein vector ω=(ω1, ω2, ω3) for parameter to be learned, NrAlways to take turns number, r is represented
For learning the round of ω.
(3) optimization problem provides
With reference to human behavior model and from the parameter that attack data learning obtains, established by equation 1 to equation 19
Generate the optimization problem of optimal Defending Policy.
D=d1+d2 (2)
α1, α2, β ∈ { 0,1 } (3)
β≤α1 (17)
β≤α2 (18)
α1+α2≤β+1 (19)
In all equatioies, Z represents a larger constant, and c represents cooperation, and nc represents uncooperative.
Formula 1 gives the expected revenus of the target of optimization problem, i.e. defender.From 2 visual target of formula by d1And d2Two parts
It forms.We define three binary variables, wherein α in formula 31And α2Attacker is represented each for the selection of cooperation, β generations
The final cooperation state of table.Only when two attackers agree to cooperation, β 1, β is 0 in all other cases.This
It can be ensured by the constraints in formula 17 and 19.Formula 4 and 5 ensure that the every one-dimensional all between [0,1] of defence vector x, and
The total resources of defender are not more than M.In formula 6 to 11, based on α1、α2Two parts of defender's expected revenus are defined with β ---
d1And d2。
With d1Exemplified by, if (i.e. β=0) is not reached in cooperation, for d1Constraints will be determined by formula 6 and 7.On the contrary,
If (i.e. β=1), d are reached in cooperation1Constraints will be determined by formula 8.Similarly, formula 9 to 11 is constituted to d2Constraint.
Formula 12 to 14 respectively define attacker cooperate with it is uncooperative when expected revenus.Here two attackers are distinguished
Modeling, so they have different expected revenusesWithBy formula 15 and 16, α ensure that1And α2It is to be received based on expectation
What benefit determined.As previously mentioned, three constraintss in formula 17 to 19 are for ensuring that correctly running for cooperative mechanism, i.e., only
There is the cooperation when two attackers select cooperation just to reach.
(4) optimal policy is calculated
Because above-mentioned optimization problem is complicated, this optimization problem is decomposed into 4 subproblems.By solving subproblem, and
Globally optimal solution is selected in the optimal solution of subproblem, just can solve former problem, obtains defence vector x.
PROBLEM DECOMPOSITION
The angle of cooperation whether is selected from attacker, may be occurred there are four types of situation altogether.Table 3 summarizes defender at this
Expected revenus (while being also the object function of subproblem) D in the case of four kinds1、D2、D3And D4。
Table 3
Expected revenus in given table 3, just can correspondingly establish 4 subproblems.Without loss of generality, one of them is only provided
The definition (formula 20 to 24) of subproblem.In this subproblem, first attacker selects cooperation and second attacker selects
It is uncooperative.The mark used in subproblemWithCome from the definition in previous formula 12 to 14.
Based on approximate method
Four subproblems suffer from the simple form of comparison, but the globally optimal solution for finding subproblem is still difficult
, therefore piecewise linear function is introduced, constraints is loosened, it is secondary that subproblem is converted into a MIXED INTEGER quadratic constraints
Planning problem (MIQCQP) solves.Such issues that can be with any MIQCQP solutions come numerical solution.
Some variable replacements are done first.Order So D4A quadric form can be rewritten as, sees formula 26.Because introduce four
A new variable to s and z according to definition, it is necessary to increase by four constraintss.Here as example is only provided for spPact
Beam, others constraint can be provided similarly.
Formula 25 is a nonlinear restriction, therefore it is replaced with two groups and slightly loosened to obtain quadratic constraints condition.Now
It illustrates how to find such two set of segmentation linear functions.
OrderIt needs to find two groups of suitably piecewise linear functionsWithSo thatHere f (xi) on [0,1] it is a monotone decreasing convex function, because
For negative.
Use xikDefinitionWith
WhereinIt representsThe slope of middle kth section broken line,It representsThe slope of middle kth section broken line.So
In order to defineWithOnly it needs to be determined thatWith
Theorem 1:According to piecewise linear functionWithMeet
After the piecewise linear function met the requirements is determined, the constraints in formula 25 can be loosened.Specifically, will
Formula 25 is replaced with following two groups of inequality.
This two groups of inequality are denoted as Cons (sp).By sr、zpAnd zrThe similar other inequality introduced are denoted as respectively
Cons(sr)、Cons(zp) and Cons (zr)。
Based on the modification above to subproblem, corresponding MIQCQP is established:
Cons(sp), Cons (sr), Cons (zp), Cons (zr) (30)
Subproblem is rewritten with the variable newly introduced in formula 26 to 29.Formula 30 represents the quadratic constraints of generation.31 He of formula
32 be to xtkConstraint.
MIQCQP there are many ∈-global optimization solution, come by the SCIP methods proposed using Vigerske S.et al.
Solve MIQCQP.
Theorem 2:NoteForWithThe distance between, then limK→+∞
Dis=0.
Description of the drawings
Fig. 1 be as K=4,WithComparison diagram;
Fig. 2 be as K=10,WithComparison diagram;
Fig. 3 is the experiment interface of participant in experiment;
Fig. 4 is the earnings structure S1 and S2 used in two groups of experiments;
When Fig. 5 is K=10, using earnings structure S1 manoeuvre second wheel in " Maximin " method generate strategy and
The strategy of model generation proposed by the present invention, wherein (a) is the strategy of " Maximin " method generation, (b) proposes for the present invention
The strategy of model generation;
Fig. 6 is the loss comparison diagram that defender often takes turns;
Fig. 7 is the ratio chart that cooperation attack accounts for general offensive number.
Specific embodiment
With reference to specific embodiment, the present invention is furture elucidated, it should be understood that these embodiments are merely to illustrate the present invention
Rather than limit the scope of the invention, after the present invention has been read, those skilled in the art are to the various equivalences of the present invention
The modification of form falls within the application scope as defined in the appended claims.
A kind of method of the modeling of mankind attacker cooperation behavior and the corresponding Defending Policy of formulation in security protection, including such as
Lower content:
(1) cooperative mechanism in SSG is repeated
Repeatedly defender and two attackers, the two attackers can choose whether to cooperate there are one cooperative games;One
A defender or attacker can represent a tissue and not necessarily only represent personal;Defender as the leader in game,
Periodically arrange new Defending Policy, and attacker does the observation of the Defending Policy of defender's arrangement according to it as follower
Go out attack;This game is formed more by taking turns, so defender can periodically change according to the data that be collected into round before
Defending Policy;Attacker independently attacks or cooperation attack;In the case where independently attacking, the income that attacker attacks every time is only
Depending on the attacker, whether this attack succeeds.And in the case where cooperation is attacked, the income that attacker attacks every time will be
The half of two people's income sums;An extra returns ∈ is introduced simultaneously, i.e., when an attacker is successfully once cooperated
During attack, he will obtain extra returns ∈;
Generally speaking, attackers need to make two main selections in game:1) target of selection attack and 2) choosing
It selects and whether cooperates with another attacker.Only when two attackers select cooperation in certain attack, once cooperation is closed
System just calculates and establishes.
Table 1
In order to further describe this game, the mark being introduced into table 1.One target collection T=is defendd by M
{ 1 ..., | T | }, T is by T1And T2Two are not overlapped subset and are combined into.T1Represent the object set that first attacker may attack
It closes, T2Represent the target collection that second attacker may attack, and T1=T-T2.The two subsets are non-overlapped, if becauseAttacker may attack identical target, this probability that them will be caused all to be arrested in once attacking
Rise.Vector x is defended, per one-dimensional element xiRepresent defence resource namely the target i that defender is put into each target
Defended probability.
Table 2
Income of one attacker in target of attack i depend on 1) this target it is whether defended and 2) two attack
Whether person cooperates.When an attacker selects independent attack, the income that he attacks not defended target i isAttack quilt
The income of the target i of defence isRelatively, defender is in the income of the previous caseIn the income of latter situation
ForThe total revenue of defender for it in two incomes by target of attack and.Because income and mistake that attacker successfully obtains
Loss and income of the loss being subject to respectively from defender are lost, so it is contemplated that zero-sum game.Then haveWith
AndIf two attackers agree to cooperate, they will obtain extra returns when attacking not defended target
∈.The income of defender and each attacker are shown in Table 2 under cooperation.In table 2, i ∈ T1Represent first attacker's selection
Target of attack, j ∈ T2Represent the target of attack of second attacker's selection." A1 " represents first attacker, and " A2 " is represented
Second attacker." S " is represented successfully, and " F " represents failure.Such as " A1S, A2F " represent first attacker success and second
A attacker's failure.
(2) incomplete reasonability and adaptivity
In some illegal activities, attacker is typically considered Imperfect Rationality.For the mankind's in SSG is repeated
Incomplete reasonability proposes a model.
Tendentiousness of the attacker for some target is defined first.
Define 1:One attacker is in R wheels to the tendentiousness of target iIt is defined by following equation:
WhereinIt isWithDifference,Represent the single attack that attacker r wheels carry out at target i
Average yield,Represent the single attack average yield for all attacks that attacker carries out in r wheels, QrFor a vector,
Its i-th dimension isC be a constant, Var (Qr) represent { QrVariance.Therefore in fixation { QrDistribution in the case of, increase
AddAlso can accordingly increase.First equation explanationForAverage (r=1 ..., R).
Generally speaking, when the average single attack income that attacker is obtained in some target i in passing round is more higher than institute
There is the overall average single attack income of target,It will be higher.This is also consistent with intuition --- and attacker should be more likely to attack
It hits him and has obtained the target of higher average single income.
It should be noted that it is likely to occur the feelings that some target is not all attacked in a certain even all rounds of wheel
Condition, the information of these targets need to be disclosed.Because when attacker tends to target of attack iWill be just, otherwise will be it is negative,
So attacker is set to 0 by us in the tendentiousness for not being revealed target.
There is above tendentious definition, we have proposed adaptive subjective revenue functions:
ASU(xi) in there are a kind of adaptation mechanism, it by introduce d andTo realize.Here d is one to be directed to
The parameter that target information is insufficient and introduces.D=1/ (N are setr- r), wherein vector ω=(ω1, ω2, ω3) it is ginseng to be learned
Number, NrAlways to take turns number, r represents the round for being used for learning ω.So will increase with the increase d of round, represent attacker and obtain
Obtained more information.
In xiWeight before multiplyButWithMultiply beforeThis is because
Attacker will be according to tendentiousness in modelIts subjective revenue function is adjusted, it is necessary to ensure after parameter substitution learn,
Work as tendentiousnessDuring increase, ASU (xi) also corresponding increase.In an experiment, all ω learnt1And ω30 is all no more than,
All ω20 is all greater than, so being made that such adjustment to weight.
Finally, the model of the present invention is summarized in the case where cooperation repeats SSG.In game, two attackers can select be
No cooperation.At each occurrence, they have different attack probability at target i, are denoted asWithout loss of generality, only provide belowCalculating side
Method, other three probability can also be calculated similarly.
Above formula subjectivity revenue functionIt is given by:
Wherein parameterIt is with non-conjunction of the maximum Likelihood (MLE) from first attacker
Make what attack data learning arrived.For ease of use, we willIt is rewritten as xiLinear expression.
(3) optimization problem provides
With reference to human behavior model and from the parameter that attack data learning obtains, we arrive equation 19 by equation 1
Establish the optimization problem for generating optimal Defending Policy.
D=d1+d2 (2)
α1, α2, β ∈ { 0,1 } (3)
β≤α1 (17)
β≤α2 (18)
α1+α2≤β+1 (19)
In all equatioies, it is (desirable that Z represents one big constantC generations
Table cooperation, nc represent uncooperative.
Formula 1 gives the expected revenus of the target of optimization problem, i.e. defender.From 2 visual target of formula by d1And d2Two parts
It forms.We define the integer variable that three values are only 0 or 1, wherein α in formula 31And α2It is each right to represent attacker
In the selection of cooperation, β represents final cooperation state.Only when two attackers agree to cooperation, β 1, at any other
In the case of β all be 0.This can be ensured by the constraints in formula 17 and 19.Formula 4 and 5 ensure that defence vector x per it is one-dimensional all
Between [0,1], and the total resources of defender are not more than M.
In formula 6 to 11, based on α1、α2Two parts of defender's expected revenus are defined with β --- d1And d2.With d1For
Example, if (i.e. β=0) is not reached in cooperation, for d1Constraints will be determined by formula 6 and 7.On the contrary, if cooperation is reached (i.e.
β=1), d1Constraints will be determined by formula 8.Similarly, formula 9 to 11 is constituted to d2Constraint.
Formula 12 to 14 respectively define attacker cooperate with it is uncooperative when expected revenus.Here two attackers are distinguished
Modeling, so they have different expected revenusesWithBy formula 15 and 16, α ensure that1And α2It is to be received based on expectation
What benefit determined.As previously mentioned, three constraintss in formula 17 to 19 are for ensuring that correctly running for cooperative mechanism, i.e., only
There is the cooperation when two attackers select cooperation just to reach.
(4) optimal policy is calculated
Because above-mentioned optimization problem is complicated, this optimization problem is decomposed into 4 subproblems.By solving subproblem, and
Globally optimal solution is selected in the optimal solution of subproblem, just can solve former problem, obtains defence vector x.As described in table 1, x is represented
(i.e. M is defended resource preventing every time to the specific defence scheme of each target any time defended probability, x and defender
Which position should be appeared in keeping) it is of equal value, therefore be just equivalent to obtain Defending Policy we can say that having solved x.
PROBLEM DECOMPOSITION
The angle of cooperation whether is selected from attacker, may be occurred there are four types of situation altogether.Table 3 summarizes defender at this
Expected revenus (while being also the object function of subproblem) D in the case of four kinds1、D2、D3And D4。
Table 3
Expected revenus in given table 3, just can correspondingly establish 4 subproblems.Without loss of generality, one of them is only provided
The definition (formula 20 to 24) of subproblem.In this subproblem, first attacker selects cooperation and second attacker selects
It is uncooperative.The mark used in subproblemWithCome from the definition in previous formula 12 to 14.
Based on approximate method
Four subproblems suffer from the simple form of comparison, but the globally optimal solution for finding subproblem is still difficult
, therefore piecewise linear function is introduced, constraints is loosened, it is secondary that subproblem is converted into a MIXED INTEGER quadratic constraints
Planning problem (MIQCQP) solves.Such issues that can be with any MIQCQP solutions come numerical solution.
Some variable replacements are done first.Order So D4A quadric form can be rewritten as, sees formula 26.Because introduce four
A new variable to s and z according to definition, it is necessary to increase by four constraintss.Here as example is only provided for spPact
Beam, others constraint can be provided similarly.
Formula 25 is a nonlinear restriction, therefore it is replaced with two groups and slightly loosened to obtain quadratic constraints condition.Now
It illustrates how to find such two set of segmentation linear functions.
OrderIt needs to find two groups of suitably piecewise linear functionsWithSo thatHere f (xi) on [0,1] it is a monotone decreasing convex function, because
For negative.Based on descending and convexity, we are illustrated in Fig. 1 as K=10, f (xi) and piecewise linear function
WithFunctional image on [0,1].
It will be seen that [0,1] is equally divided into K (K=4 here) part from Fig. 1, per portion by a new variable
xik(k=1,2 ..., K) is represented.Obviously,And Use xikDefinitionWith
WhereinIt representsThe slope of middle kth section broken line,It representsThe slope of middle kth section broken line.So
In order to defineWithOnly it needs to be determined thatWithIt will calculate WithStep
Suddenly it is summarized as algorithm 1.
Above-mentioned algorithm 1 is specifically described as:
1) it is directly substituted into firstIt calculatesValue.
2) slope of kth section (k=1 ..., K) is then calculated.
3) Δ is madeI, kForOnMaximum, calculate ΔI, kThe value and basis of (k=1 ..., K)
ΔI, kIt calculates
4) Δ is takenI, 0=ΔI, 1,
5) basisIt calculates
6) exportWith
Theorem 1:The piecewise linear function generated according to algorithm 1WithMeet
It proves:BecauseIt is [0, a 1] Convex Functions, it is clear that
Below based onIt provesGiven k ∈ 1,
2 ..., K }, orderHave
AndNotice g (xi) andAll it isOn linear function.Therefore,
On the other hand, becauseIt can be seen that
Thus obtainSo
After the piecewise linear function met the requirements is determined with algorithm 1, the constraints in formula 25 can be loosened.Tool
Body, the following two groups of inequality of formula 25 are replaced.
This two groups of inequality are denoted as Cons (sp).By sr、zpAnd zrThe similar other inequality introduced are denoted as respectively
Cons(sr)、Cons(zp) and Cons (zr)。
Based on the modification above to subproblem, corresponding MIQCQP is established:
Cons(sp), Cons (sr), Cons (zp), Cons (zr) (30)
Subproblem is rewritten with the variable newly introduced in formula 26 to 29.Formula 30 represent the method for the present invention generation it is secondary about
Beam.Formula 31 and 32 is to xtkConstraint.
MIQCQP there are many ∈-global optimization solution, come by the SCIP methods proposed using Vigerske S.et al.
Solve MIQCQP.
Theorem 2:NoteForWithThe distance between, then limK→+∞
Dis=0.
It proves:It is based onWithDefinition andIt is clear toOrder
Thenφ (x) existsOn maximum be
NoteThen
NoteIt obtains
ΔI, 1=wlnw-w+1
Because
It obtains
Theorem 2 shows that the method for the present invention is capable of providing and constraints arbitrary extent is loosened, this is because two set of segmentation
The distance between linear function can be arbitrarily small.Fig. 2 is illustrated as K=10,、f(xi) andWhat can be reached connects
Short range degree.Therefore, after a suitably K has been selected, the method for the present invention can obtain the approximate solution of four subproblems.
(5) test
In order to estimate the parameter in cooperation human behavior model and test its validity, tested, and analyze reality
Test result.
Test general view
In an experiment, participant needs to play the part of the attacker for being carried out wildcat operation by Conservation areas at one piece, carries out manoeuvre.
The panel of manoeuvre such as Fig. 3.
Protected area has been divided into 2 pieces of regions, and each region is divided into as 9 fritters, and each fritter represents a target.
Manoeuvre needs 2 participants to simultaneously participate in, they are referred to as Alice and Bob.Alice and Bob can be with before manoeuvre starts
Observe 18 all targets, income at each target when the defence probability of defender, attacker's independence success attack and
Punishment during independent attack failure can all show them.As auxiliary, a temperature figure is provided to aid in participant's observation every
The defence intensity of defender at a target.The color of target is deeper in figure, its defended probability is higher.In addition, also it is two
A participant respectively provides a form, when form summarizes cooperation attack can obtainable income and its generation probability.Alice
The target of half can only be all attacked with Bob, i.e. Alice can only attack 9, left side target, and Bob can only attack 9 targets in the right.
Two participants need 1) from oneself one target of selection and can 2) to decide whether and another one participant closes in firing area
Make.After two participants press attack (Attack) key, this attack terminates, and the result of attack can be illustrated in manoeuvre panel
Portion, two respective remaining sums of participant can also be updated according to result.
Experiment setting
Participants, which need to play, amounts to 5 wheel manoeuvres, often takes turns manoeuvre and is made of several offices.In the first round, due to not yet
Any attack data are collected into, we use Gholami, the Defending Policy for the method generation first round that S.et al. are proposed, with
Under we term it " Maximin " methods.For the validity of evaluation method, the strategy for respectively generating " Maximin " method
It is applied to the strategy of model proposed by the present invention generation in 5 wheel manoeuvres.Two groups of manoeuvres are provided with altogether, they have different
Earnings structure S1 and S2, every group of manoeuvre tool correspond to the strategy that two methods generate respectively there are two 5 wheel manoeuvres.
The two kinds of earnings structure such as Fig. 4 (receipts of independence success attack at each target of digitized representation used in an experiment
Benefit).Punishment at each target is set to -1.The extra returns ∈ of co-participant is set to 1.Total defence resource of defender
M is set to 9, it means that the same time only has 9 targets can be defended.
Interpretation of result
In cooperation SSG, our main target is the cooperation destroyed between attacker.When Fig. 5 illustrates K=10, adopt
It is generated with the tactful and proposed by the present invention model that " Maximin " method generates in the second wheel of the manoeuvre of earnings structure S1
Strategy (percentage that probability is defended at each target of digitized representation).As seen from Figure 5, the defence intensity that two attackers face has
Significant difference.
For the incomplete reasonability of mankind attacker, we have collected data and are drilled according to often wheel (except last wheel)
It practises data and has estimated model parameter.Here we listed in table 4 and table 5 from attack behavior learning to parameter.
Table 4
Table 5
In order to show validity of the model in terms of defender's average loss is reduced of the present invention, carry out in figure 6 pair
Than.As seen from Figure 6, except the first round, the strategy that our model obtains in two kinds of earnings structures can provide in each round
The significantly low defender loss than " Maximin " method.In addition, model of the invention is in the second wheel horse back initially applied
With regard to defender's loss can be reduced.
Finally calculate cooperative rate, i.e., cooperation attack proportion in one wheel.Fig. 7 shows the experiment on two kinds of earnings structures
As a result.When using the model of the present invention, the cooperative rate of the second wheel person of beginning participating in is decreased obviously and is kept low.With
Upper result is consistent with the result in Fig. 6, and further illustrates the method for the present invention better than " Maximin " method.
Claims (7)
1. a kind of method of the modeling of mankind attacker cooperation behavior and the corresponding Defending Policy of formulation in security protection, feature exist
In including following content:
(1) cooperative mechanism in SSG is repeated
Game is formed more by taking turns, and attackers need to make two selections in game:1) target of selection attack and 2) selection
Whether cooperate with another attacker.Only when two attackers select cooperation in certain attack, a cooperative relationship
It just calculates and establishes;In the case where cooperation is attacked, income that attacker attacks every time by be two people's income sums half;It introduces simultaneously
One extra returns ∈, i.e., when an attacker has successfully carried out once cooperation attack, he will obtain extra returns ∈;
(2) incomplete reasonability and adaptivity
Incomplete reasonability for the mankind in SSG is repeated proposes a model;
Tendentiousness of the attacker for some target is defined first.
Define 1:One attacker is in R wheels to the tendentiousness of target iIt is defined by following equation:
<mrow>
<msubsup>
<mi>I</mi>
<mi>i</mi>
<mi>R</mi>
</msubsup>
<mo>=</mo>
<mfrac>
<mrow>
<msubsup>
<mo>&Sigma;</mo>
<mrow>
<mi>r</mi>
<mo>=</mo>
<mn>1</mn>
</mrow>
<mi>R</mi>
</msubsup>
<msubsup>
<mi>N</mi>
<mi>i</mi>
<mi>r</mi>
</msubsup>
</mrow>
<mi>R</mi>
</mfrac>
</mrow>
<mrow>
<msubsup>
<mi>N</mi>
<mi>i</mi>
<mi>r</mi>
</msubsup>
<mo>=</mo>
<mfrac>
<msubsup>
<mi>Q</mi>
<mi>i</mi>
<mi>r</mi>
</msubsup>
<mrow>
<mi>C</mi>
<mo>*</mo>
<mi>V</mi>
<mi>a</mi>
<mi>r</mi>
<mrow>
<mo>(</mo>
<msup>
<mi>Q</mi>
<mi>r</mi>
</msup>
<mo>)</mo>
</mrow>
</mrow>
</mfrac>
</mrow>
<mrow>
<msubsup>
<mi>Q</mi>
<mi>i</mi>
<mi>r</mi>
</msubsup>
<mo>=</mo>
<msubsup>
<mi>avg</mi>
<mi>i</mi>
<mi>r</mi>
</msubsup>
<mo>-</mo>
<msubsup>
<mi>avg</mi>
<mrow>
<mi>a</mi>
<mi>l</mi>
<mi>l</mi>
</mrow>
<mi>r</mi>
</msubsup>
</mrow>
WhereinIt isWithDifference,Represent being averaged for the single attack that attacker r wheels carry out at target i
Income,Represent the single attack average yield for all attacks that attacker carries out in r wheels, QrFor a vector,
I is tieed upC be a constant, Var (Qr) represent { QrVariance;
When attacker tends to target of attack iTo be just, on the contrary will be negative, and attacker be not revealed the tendentiousness of target
It is set to 0.
Adaptive subjectivity revenue function:
<mrow>
<msup>
<mi>ASU</mi>
<mi>R</mi>
</msup>
<mrow>
<mo>(</mo>
<msub>
<mi>x</mi>
<mi>i</mi>
</msub>
<mo>)</mo>
</mrow>
<mo>=</mo>
<mrow>
<mo>(</mo>
<mn>1</mn>
<mo>-</mo>
<mi>d</mi>
<mo>*</mo>
<msubsup>
<mi>I</mi>
<mi>i</mi>
<mi>R</mi>
</msubsup>
<mo>)</mo>
</mrow>
<msub>
<mi>w</mi>
<mn>1</mn>
</msub>
<msub>
<mi>x</mi>
<mi>i</mi>
</msub>
<mo>+</mo>
<mrow>
<mo>(</mo>
<mn>1</mn>
<mo>+</mo>
<mi>d</mi>
<mo>*</mo>
<msubsup>
<mi>I</mi>
<mi>i</mi>
<mi>R</mi>
</msubsup>
<mo>)</mo>
</mrow>
<msub>
<mi>w</mi>
<mn>2</mn>
</msub>
<msubsup>
<mi>R</mi>
<mi>i</mi>
<mi>a</mi>
</msubsup>
<mo>+</mo>
<mrow>
<mo>(</mo>
<mn>1</mn>
<mo>+</mo>
<mi>d</mi>
<mo>*</mo>
<msubsup>
<mi>I</mi>
<mi>i</mi>
<mi>R</mi>
</msubsup>
<mo>)</mo>
</mrow>
<msub>
<mi>w</mi>
<mn>3</mn>
</msub>
<msubsup>
<mi>P</mi>
<mi>i</mi>
<mi>a</mi>
</msubsup>
</mrow>
D=1/ (N are setr- r), wherein vector ω=(ω1,ω2,ω3) for parameter to be learned, NrAlways to take turns number, r representatives are used for
Learn the round of ω;
(3) optimization problem provides
With reference to human behavior model and from the parameter that attack data learning obtains, the optimization for generating optimal Defending Policy is established
Problem;
(4) optimal policy is calculated
Optimization problem is decomposed into 4 subproblems.By solving subproblem, and global optimum is selected in the optimal solution of subproblem
Solution, just can solve former problem, obtain defence vector x.
2. the modeling of mankind attacker cooperation behavior and the corresponding Defending Policy of formulation in a kind of security protection as described in power requires 1
Method, which is characterized in that repeat cooperative game there are one defender and two attackers, the two attackers can select be
No cooperation;Defender periodically arranges new Defending Policy as the leader in game, and attacker is as follower, according to
It makes attack to the observation of the Defending Policy of defender's arrangement;This game is formed more by taking turns, so defender can be regular
Defending Policy is changed according to the data being collected into round before;Attacker independently attacks or cooperation attack;Independently attacking
In the case of hitting, whether income that attacker attacks every time is only dependent upon the attacker this attack and succeeds.
3. the modeling of mankind attacker cooperation behavior and the corresponding Defending Policy of formulation in a kind of security protection as described in power requires 1
Method, which is characterized in that the defence resource summation M possessed by defender come defend a target collection T=1 ..., | T
|, T is by T1And T2Two are not overlapped subset and are combined into;T1Represent the target collection that first attacker may attack, T2Represent the
The target collection that two attackers may attack, and T1=T-T2;Vector x is defended, per one-dimensional element xiDefender is represented to exist
The defence resource of each target input namely the defended probability of target i.
4. the modeling of mankind attacker cooperation behavior and the corresponding Defending Policy of formulation in a kind of security protection as described in power requires 1
Method, which is characterized in that income of the attacker in target of attack i depend on 1) this target it is whether defended and
2) whether two attackers cooperate;When an attacker selects independent attack, the income that he attacks not defended target i isThe income for attacking defended target i isRelatively, defender is in the income of the previous caseIn latter feelings
The income of condition isThe total revenue of defender for it in two incomes by target of attack and;AndIf two attackers agree to cooperate, they will obtain extra returns ∈ when attacking not defended target.
5. the modeling of mankind attacker cooperation behavior and the corresponding Defending Policy of formulation in a kind of security protection as described in power requires 1
Method, which is characterized in that with reference to human behavior model and from the obtained parameter of attack data learning, arrived by equation 1
Equation 19 establishes the optimization problem for generating optimal Defending Policy.
<mrow>
<munder>
<mi>max</mi>
<mi>x</mi>
</munder>
<mi>D</mi>
<mo>-</mo>
<mo>-</mo>
<mo>-</mo>
<mrow>
<mo>(</mo>
<mn>1</mn>
<mo>)</mo>
</mrow>
</mrow>
D=d1+d2 (2)
α1, α2, β ∈ { 0,1 } (3)
<mrow>
<munder>
<mo>&Sigma;</mo>
<mrow>
<mi>i</mi>
<mo>&Element;</mo>
<msub>
<mi>T</mi>
<mn>1</mn>
</msub>
<mo>&cup;</mo>
<msub>
<mi>T</mi>
<mn>2</mn>
</msub>
</mrow>
</munder>
<msub>
<mi>x</mi>
<mi>i</mi>
</msub>
<mo>&le;</mo>
<mi>M</mi>
<mo>-</mo>
<mo>-</mo>
<mo>-</mo>
<mrow>
<mo>(</mo>
<mn>4</mn>
<mo>)</mo>
</mrow>
</mrow>
<mrow>
<msub>
<mi>x</mi>
<mi>i</mi>
</msub>
<mo>&Element;</mo>
<mo>&lsqb;</mo>
<mn>0</mn>
<mo>,</mo>
<mn>1</mn>
<mo>&rsqb;</mo>
<mo>,</mo>
<mo>&ForAll;</mo>
<mi>i</mi>
<mo>&Element;</mo>
<msub>
<mi>T</mi>
<mn>1</mn>
</msub>
<mo>&cup;</mo>
<msub>
<mi>T</mi>
<mn>2</mn>
</msub>
<mo>-</mo>
<mo>-</mo>
<mo>-</mo>
<mrow>
<mo>(</mo>
<mn>5</mn>
<mo>)</mo>
</mrow>
</mrow>
<mrow>
<msub>
<mi>d</mi>
<mn>1</mn>
</msub>
<mo>-</mo>
<munder>
<mo>&Sigma;</mo>
<mrow>
<mi>i</mi>
<mo>&Element;</mo>
<msub>
<mi>T</mi>
<mn>1</mn>
</msub>
</mrow>
</munder>
<msubsup>
<mi>q</mi>
<mi>i</mi>
<mrow>
<mi>n</mi>
<mi>c</mi>
</mrow>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>)</mo>
</mrow>
<mo>&lsqb;</mo>
<msubsup>
<mi>R</mi>
<mi>i</mi>
<mi>d</mi>
</msubsup>
<msub>
<mi>x</mi>
<mi>i</mi>
</msub>
<mo>+</mo>
<msubsup>
<mi>P</mi>
<mi>i</mi>
<mi>d</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mn>1</mn>
<mo>-</mo>
<msub>
<mi>x</mi>
<mi>i</mi>
</msub>
<mo>)</mo>
</mrow>
<mo>&rsqb;</mo>
<mo>&le;</mo>
<msub>
<mi>&alpha;</mi>
<mn>1</mn>
</msub>
<mi>Z</mi>
<mo>-</mo>
<mo>-</mo>
<mo>-</mo>
<mrow>
<mo>(</mo>
<mn>6</mn>
<mo>)</mo>
</mrow>
</mrow>
<mrow>
<msub>
<mi>d</mi>
<mn>1</mn>
</msub>
<mo>-</mo>
<munder>
<mo>&Sigma;</mo>
<mrow>
<mi>i</mi>
<mo>&Element;</mo>
<msub>
<mi>T</mi>
<mn>1</mn>
</msub>
</mrow>
</munder>
<msubsup>
<mi>q</mi>
<mi>i</mi>
<mi>c</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>)</mo>
</mrow>
<mo>&lsqb;</mo>
<msubsup>
<mi>R</mi>
<mi>i</mi>
<mi>d</mi>
</msubsup>
<msub>
<mi>x</mi>
<mi>i</mi>
</msub>
<mo>+</mo>
<msubsup>
<mi>P</mi>
<mi>i</mi>
<mi>d</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mn>1</mn>
<mo>-</mo>
<msub>
<mi>x</mi>
<mi>i</mi>
</msub>
<mo>)</mo>
</mrow>
<mo>&rsqb;</mo>
<mo>&le;</mo>
<mrow>
<mo>(</mo>
<mn>1</mn>
<mo>-</mo>
<msub>
<mi>&alpha;</mi>
<mn>1</mn>
</msub>
<mo>)</mo>
</mrow>
<mi>Z</mi>
<mo>+</mo>
<mi>&beta;</mi>
<mi>Z</mi>
<mo>-</mo>
<mo>-</mo>
<mo>-</mo>
<mrow>
<mo>(</mo>
<mn>7</mn>
<mo>)</mo>
</mrow>
</mrow>
<mrow>
<msub>
<mi>d</mi>
<mn>1</mn>
</msub>
<mo>-</mo>
<munder>
<mo>&Sigma;</mo>
<mrow>
<mi>i</mi>
<mo>&Element;</mo>
<msub>
<mi>T</mi>
<mn>1</mn>
</msub>
</mrow>
</munder>
<msubsup>
<mi>q</mi>
<mi>i</mi>
<mi>c</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>)</mo>
</mrow>
<mo>&lsqb;</mo>
<msubsup>
<mi>R</mi>
<mi>i</mi>
<mi>d</mi>
</msubsup>
<msub>
<mi>x</mi>
<mi>i</mi>
</msub>
<mo>+</mo>
<mrow>
<mo>(</mo>
<msubsup>
<mi>P</mi>
<mi>i</mi>
<mi>d</mi>
</msubsup>
<mo>-</mo>
<mo>&Element;</mo>
<mo>)</mo>
</mrow>
<mrow>
<mo>(</mo>
<mn>1</mn>
<mo>-</mo>
<msub>
<mi>x</mi>
<mi>i</mi>
</msub>
<mo>)</mo>
</mrow>
<mo>&rsqb;</mo>
<mo>&le;</mo>
<mrow>
<mo>(</mo>
<mn>1</mn>
<mo>-</mo>
<msub>
<mi>&alpha;</mi>
<mn>1</mn>
</msub>
<mo>)</mo>
</mrow>
<mi>Z</mi>
<mo>+</mo>
<mrow>
<mo>(</mo>
<mn>1</mn>
<mo>-</mo>
<mi>&beta;</mi>
<mo>)</mo>
</mrow>
<mi>Z</mi>
<mo>-</mo>
<mo>-</mo>
<mo>-</mo>
<mrow>
<mo>(</mo>
<mn>8</mn>
<mo>)</mo>
</mrow>
</mrow>
<mrow>
<msub>
<mi>d</mi>
<mn>2</mn>
</msub>
<mo>-</mo>
<munder>
<mo>&Sigma;</mo>
<mrow>
<mi>j</mi>
<mo>&Element;</mo>
<msub>
<mi>T</mi>
<mn>2</mn>
</msub>
</mrow>
</munder>
<msubsup>
<mi>q</mi>
<mi>j</mi>
<mrow>
<mi>n</mi>
<mi>c</mi>
</mrow>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>)</mo>
</mrow>
<mo>&lsqb;</mo>
<msubsup>
<mi>R</mi>
<mi>j</mi>
<mi>d</mi>
</msubsup>
<msub>
<mi>x</mi>
<mi>j</mi>
</msub>
<mo>+</mo>
<msubsup>
<mi>P</mi>
<mi>j</mi>
<mi>d</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mn>1</mn>
<mo>-</mo>
<msub>
<mi>x</mi>
<mi>j</mi>
</msub>
<mo>)</mo>
</mrow>
<mo>&rsqb;</mo>
<mo>&le;</mo>
<msub>
<mi>&alpha;</mi>
<mn>2</mn>
</msub>
<mi>Z</mi>
<mo>-</mo>
<mo>-</mo>
<mo>-</mo>
<mrow>
<mo>(</mo>
<mn>9</mn>
<mo>)</mo>
</mrow>
</mrow>
<mrow>
<msub>
<mi>d</mi>
<mn>2</mn>
</msub>
<mo>-</mo>
<munder>
<mo>&Sigma;</mo>
<mrow>
<mi>j</mi>
<mo>&Element;</mo>
<msub>
<mi>T</mi>
<mn>2</mn>
</msub>
</mrow>
</munder>
<msubsup>
<mi>q</mi>
<mi>j</mi>
<mi>c</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>)</mo>
</mrow>
<mo>&lsqb;</mo>
<msubsup>
<mi>R</mi>
<mi>j</mi>
<mi>d</mi>
</msubsup>
<msub>
<mi>x</mi>
<mi>j</mi>
</msub>
<mo>+</mo>
<msubsup>
<mi>P</mi>
<mi>j</mi>
<mi>d</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mn>1</mn>
<mo>-</mo>
<msub>
<mi>x</mi>
<mi>j</mi>
</msub>
<mo>)</mo>
</mrow>
<mo>&rsqb;</mo>
<mo>&le;</mo>
<mrow>
<mo>(</mo>
<mn>1</mn>
<mo>-</mo>
<msub>
<mi>&alpha;</mi>
<mn>2</mn>
</msub>
<mo>)</mo>
</mrow>
<mi>Z</mi>
<mo>+</mo>
<mi>&beta;</mi>
<mi>Z</mi>
<mo>-</mo>
<mo>-</mo>
<mo>-</mo>
<mrow>
<mo>(</mo>
<mn>10</mn>
<mo>)</mo>
</mrow>
</mrow>
<mrow>
<msub>
<mi>d</mi>
<mn>2</mn>
</msub>
<mo>-</mo>
<munder>
<mo>&Sigma;</mo>
<mrow>
<mi>j</mi>
<mo>&Element;</mo>
<msub>
<mi>T</mi>
<mn>2</mn>
</msub>
</mrow>
</munder>
<msubsup>
<mi>q</mi>
<mi>j</mi>
<mi>c</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>)</mo>
</mrow>
<mo>&lsqb;</mo>
<msubsup>
<mi>R</mi>
<mi>j</mi>
<mi>d</mi>
</msubsup>
<msub>
<mi>x</mi>
<mi>j</mi>
</msub>
<mo>+</mo>
<mrow>
<mo>(</mo>
<msubsup>
<mi>P</mi>
<mi>j</mi>
<mi>d</mi>
</msubsup>
<mo>-</mo>
<mo>&Element;</mo>
<mo>)</mo>
</mrow>
<mrow>
<mo>(</mo>
<mn>1</mn>
<mo>-</mo>
<msub>
<mi>x</mi>
<mi>j</mi>
</msub>
<mo>)</mo>
</mrow>
<mo>&rsqb;</mo>
<mo>&le;</mo>
<mrow>
<mo>(</mo>
<mn>1</mn>
<mo>-</mo>
<msub>
<mi>&alpha;</mi>
<mn>2</mn>
</msub>
<mo>)</mo>
</mrow>
<mi>Z</mi>
<mo>+</mo>
<mrow>
<mo>(</mo>
<mn>1</mn>
<mo>-</mo>
<mi>&beta;</mi>
<mo>)</mo>
</mrow>
<mi>Z</mi>
<mo>-</mo>
<mo>-</mo>
<mo>-</mo>
<mrow>
<mo>(</mo>
<mn>11</mn>
<mo>)</mo>
</mrow>
</mrow>
<mrow>
<mtable>
<mtr>
<mtd>
<mrow>
<msubsup>
<mi>U</mi>
<mi>a</mi>
<mi>c</mi>
</msubsup>
<mo>=</mo>
<mfrac>
<mn>1</mn>
<mn>2</mn>
</mfrac>
<mo>{</mo>
<munder>
<mo>&Sigma;</mo>
<mrow>
<mi>i</mi>
<mo>&Element;</mo>
<msub>
<mi>T</mi>
<mn>1</mn>
</msub>
</mrow>
</munder>
<msubsup>
<mi>q</mi>
<mi>i</mi>
<mi>c</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>)</mo>
</mrow>
<mo>&lsqb;</mo>
<mrow>
<mo>(</mo>
<msubsup>
<mi>R</mi>
<mi>i</mi>
<mi>a</mi>
</msubsup>
<mo>+</mo>
<mo>&Element;</mo>
<mo>)</mo>
</mrow>
<mrow>
<mo>(</mo>
<mn>1</mn>
<mo>-</mo>
<msub>
<mi>x</mi>
<mi>i</mi>
</msub>
<mo>)</mo>
</mrow>
<mo>+</mo>
<msubsup>
<mi>P</mi>
<mi>i</mi>
<mi>a</mi>
</msubsup>
<msub>
<mi>x</mi>
<mi>i</mi>
</msub>
<mo>&rsqb;</mo>
<mo>+</mo>
</mrow>
</mtd>
</mtr>
<mtr>
<mtd>
<mrow>
<munder>
<mo>&Sigma;</mo>
<mrow>
<mi>j</mi>
<mo>&Element;</mo>
<msub>
<mi>T</mi>
<mn>2</mn>
</msub>
</mrow>
</munder>
<msubsup>
<mi>q</mi>
<mi>j</mi>
<mi>c</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>)</mo>
</mrow>
<mo>&lsqb;</mo>
<mrow>
<mo>(</mo>
<msubsup>
<mi>R</mi>
<mi>j</mi>
<mi>a</mi>
</msubsup>
<mo>+</mo>
<mo>&Element;</mo>
<mo>)</mo>
</mrow>
<mrow>
<mo>(</mo>
<mn>1</mn>
<mo>-</mo>
<msub>
<mi>x</mi>
<mi>j</mi>
</msub>
<mo>)</mo>
</mrow>
<mo>+</mo>
<msubsup>
<mi>P</mi>
<mi>j</mi>
<mi>a</mi>
</msubsup>
<msub>
<mi>x</mi>
<mi>j</mi>
</msub>
<mo>&rsqb;</mo>
<mo>}</mo>
</mrow>
</mtd>
</mtr>
</mtable>
<mo>-</mo>
<mo>-</mo>
<mo>-</mo>
<mrow>
<mo>(</mo>
<mn>12</mn>
<mo>)</mo>
</mrow>
</mrow>
<mrow>
<msubsup>
<mi>U</mi>
<mrow>
<mi>a</mi>
<mn>1</mn>
</mrow>
<mrow>
<mi>n</mi>
<mi>c</mi>
</mrow>
</msubsup>
<mo>=</mo>
<munder>
<mo>&Sigma;</mo>
<mrow>
<mi>i</mi>
<mo>&Element;</mo>
<msub>
<mi>T</mi>
<mn>1</mn>
</msub>
</mrow>
</munder>
<msubsup>
<mi>q</mi>
<mi>i</mi>
<mrow>
<mi>n</mi>
<mi>c</mi>
</mrow>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>)</mo>
</mrow>
<mo>&lsqb;</mo>
<msubsup>
<mi>R</mi>
<mi>i</mi>
<mi>a</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mn>1</mn>
<mo>-</mo>
<msub>
<mi>x</mi>
<mi>i</mi>
</msub>
<mo>)</mo>
</mrow>
<mo>+</mo>
<msubsup>
<mi>P</mi>
<mi>i</mi>
<mi>a</mi>
</msubsup>
<msub>
<mi>x</mi>
<mi>i</mi>
</msub>
<mo>&rsqb;</mo>
<mo>-</mo>
<mo>-</mo>
<mo>-</mo>
<mrow>
<mo>(</mo>
<mn>13</mn>
<mo>)</mo>
</mrow>
</mrow>
<mrow>
<msubsup>
<mi>U</mi>
<mrow>
<mi>a</mi>
<mn>2</mn>
</mrow>
<mrow>
<mi>n</mi>
<mi>c</mi>
</mrow>
</msubsup>
<mo>=</mo>
<munder>
<mo>&Sigma;</mo>
<mrow>
<mi>j</mi>
<mo>&Element;</mo>
<msub>
<mi>T</mi>
<mn>2</mn>
</msub>
</mrow>
</munder>
<msubsup>
<mi>q</mi>
<mi>j</mi>
<mrow>
<mi>n</mi>
<mi>c</mi>
</mrow>
</msubsup>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>)</mo>
</mrow>
<mo>&lsqb;</mo>
<msubsup>
<mi>R</mi>
<mi>j</mi>
<mi>a</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mn>1</mn>
<mo>-</mo>
<msub>
<mi>x</mi>
<mi>j</mi>
</msub>
<mo>)</mo>
</mrow>
<mo>+</mo>
<msubsup>
<mi>P</mi>
<mi>j</mi>
<mi>a</mi>
</msubsup>
<msub>
<mi>x</mi>
<mi>j</mi>
</msub>
<mo>&rsqb;</mo>
<mo>-</mo>
<mo>-</mo>
<mo>-</mo>
<mrow>
<mo>(</mo>
<mn>14</mn>
<mo>)</mo>
</mrow>
</mrow>
<mrow>
<mo>-</mo>
<msub>
<mi>&alpha;</mi>
<mn>1</mn>
</msub>
<mi>Z</mi>
<mo>&le;</mo>
<msubsup>
<mi>U</mi>
<mrow>
<mi>a</mi>
<mn>1</mn>
</mrow>
<mrow>
<mi>n</mi>
<mi>c</mi>
</mrow>
</msubsup>
<mo>-</mo>
<msubsup>
<mi>U</mi>
<mi>a</mi>
<mi>c</mi>
</msubsup>
<mo>&le;</mo>
<mrow>
<mo>(</mo>
<mn>1</mn>
<mo>-</mo>
<msub>
<mi>&alpha;</mi>
<mn>1</mn>
</msub>
<mo>)</mo>
</mrow>
<mi>Z</mi>
<mo>-</mo>
<mo>-</mo>
<mo>-</mo>
<mrow>
<mo>(</mo>
<mn>15</mn>
<mo>)</mo>
</mrow>
</mrow>
<mrow>
<mo>-</mo>
<msub>
<mi>&alpha;</mi>
<mn>2</mn>
</msub>
<mi>Z</mi>
<mo>&le;</mo>
<msubsup>
<mi>U</mi>
<mrow>
<mi>a</mi>
<mn>2</mn>
</mrow>
<mrow>
<mi>n</mi>
<mi>c</mi>
</mrow>
</msubsup>
<mo>-</mo>
<msubsup>
<mi>U</mi>
<mi>a</mi>
<mi>c</mi>
</msubsup>
<mo>&le;</mo>
<mrow>
<mo>(</mo>
<mn>1</mn>
<mo>-</mo>
<msub>
<mi>&alpha;</mi>
<mn>2</mn>
</msub>
<mo>)</mo>
</mrow>
<mi>Z</mi>
<mo>-</mo>
<mo>-</mo>
<mo>-</mo>
<mrow>
<mo>(</mo>
<mn>16</mn>
<mo>)</mo>
</mrow>
</mrow>
β≤α1 (17)
β≤α2 (18)
α1+α2≤β+1 (19)
In all equatioies, Z represents a larger constant, and c represents cooperation, and nc represents uncooperative;
Formula 1 gives the expected revenus of the target of optimization problem, i.e. defender;From 2 visual target of formula by d1And d2Two parts structure
Into;Three binary variables defined in formula 3, wherein α1And α2Attacker is represented each for the selection of cooperation, and β is represented most
Whole cooperation state;Only when two attackers agree to cooperation, β 1, β is 0 in all other cases;This is by formula 17
Ensure with the constraints in 19;Formula 4 and 5 ensure that the every one-dimensional all between [0,1] of defence vector x, and defender's is total
Resource is not more than M;
In formula 6 to 11, based on α1、α2Two parts of defender's expected revenus are defined with β --- d1And d2;
Formula 12 to 14 respectively define attacker cooperate with it is uncooperative when expected revenus;Here two attackers are built respectively
Mould, so they have different expected revenusesWithBy formula 15 and 16, α ensure that1And α2It is based on expected revenus
It determines;As previously mentioned, three constraintss in formula 17 to 19 are for ensuring that correctly running for cooperative mechanism, that is, only have
Cooperation is just reached when two attackers select cooperation.
6. the modeling of mankind attacker cooperation behavior and the corresponding Defending Policy of formulation in a kind of security protection as described in power requires 1
Method, which is characterized in that optimization problem is decomposed into 4 subproblems;By solving subproblem, and in the optimal solution of subproblem
Middle selection globally optimal solution just can solve former problem, obtain defence vector x.
PROBLEM DECOMPOSITION
The angle of cooperation whether is selected from attacker, may be occurred there are four types of situation altogether;Following table summarize defender this four
Expected revenus D in the case of kind1、D2、D3And D4;
Table 3
Expected revenus in given table, just can correspondingly establish 4 subproblems.
7. the modeling of mankind attacker cooperation behavior and the corresponding Defending Policy of formulation in a kind of security protection as described in power requires 6
Method, which is characterized in that find the globally optimal solution of subproblem based on approximate method, introduce piecewise linear function, will about
Beam condition is loosened, and subproblem is converted into a MIXED INTEGER quadratically constrained quadratic programming problem (MIQCQP) to solve;
Variable replacement is done first;Order That
D4A quadric form can be rewritten as, sees formula 26;Because four new variables are introduced, it is necessary to according to definition to s and z
Increase by four constraintss;Here only provide for spConstraint, others constraint can similarly provide;
<mrow>
<msub>
<mi>s</mi>
<mi>p</mi>
</msub>
<munder>
<mo>&Sigma;</mo>
<mrow>
<mi>i</mi>
<mo>&Element;</mo>
<msub>
<mi>T</mi>
<mn>1</mn>
</msub>
</mrow>
</munder>
<msup>
<mi>e</mi>
<mrow>
<msubsup>
<mi>ASU</mi>
<mn>1</mn>
<mi>c</mi>
</msubsup>
<mrow>
<mo>(</mo>
<msub>
<mi>x</mi>
<mi>i</mi>
</msub>
<mo>)</mo>
</mrow>
</mrow>
</msup>
<mo>-</mo>
<msup>
<mi>e</mi>
<mrow>
<msubsup>
<mi>ASU</mi>
<mn>1</mn>
<mi>c</mi>
</msubsup>
<mrow>
<mo>(</mo>
<msub>
<mi>x</mi>
<mi>p</mi>
</msub>
<mo>)</mo>
</mrow>
</mrow>
</msup>
<mo>=</mo>
<mn>0</mn>
<mo>,</mo>
<mo>&ForAll;</mo>
<mi>p</mi>
<mo>&Element;</mo>
<msub>
<mi>T</mi>
<mn>1</mn>
</msub>
<mo>-</mo>
<mo>-</mo>
<mo>-</mo>
<mrow>
<mo>(</mo>
<mn>25</mn>
<mo>)</mo>
</mrow>
</mrow>
Formula 25 is a nonlinear restriction, therefore it is replaced with two groups and slightly loosened to obtain quadratic constraints condition;Present explanation
How such two set of segmentation linear functions are found.
OrderIt needs to find two groups of suitably piecewise linear functionsWithSo thatHere f (xi) on [0,1] it is a monotone decreasing convex function, because
For negative;
AndUse xikDefinitionWith
<mrow>
<msubsup>
<mi>l</mi>
<mi>i</mi>
<mn>1</mn>
</msubsup>
<mrow>
<mo>(</mo>
<msub>
<mi>x</mi>
<mi>i</mi>
</msub>
<mo>)</mo>
</mrow>
<mo>=</mo>
<mn>1</mn>
<mo>+</mo>
<munderover>
<mo>&Sigma;</mo>
<mrow>
<mi>k</mi>
<mo>=</mo>
<mn>1</mn>
</mrow>
<mi>K</mi>
</munderover>
<msubsup>
<mi>v</mi>
<mrow>
<mi>i</mi>
<mi>k</mi>
</mrow>
<mrow>
<mi>u</mi>
<mi>p</mi>
</mrow>
</msubsup>
<msub>
<mi>x</mi>
<mrow>
<mi>i</mi>
<mi>k</mi>
</mrow>
</msub>
</mrow>
<mrow>
<msubsup>
<mi>l</mi>
<mi>i</mi>
<mn>2</mn>
</msubsup>
<mrow>
<mo>(</mo>
<msub>
<mi>x</mi>
<mi>i</mi>
</msub>
<mo>)</mo>
</mrow>
<mo>=</mo>
<msubsup>
<mi>l</mi>
<mi>i</mi>
<mn>2</mn>
</msubsup>
<mrow>
<mo>(</mo>
<mn>0</mn>
<mo>)</mo>
</mrow>
<mo>+</mo>
<munderover>
<mo>&Sigma;</mo>
<mrow>
<mi>k</mi>
<mo>=</mo>
<mn>1</mn>
</mrow>
<mi>K</mi>
</munderover>
<msubsup>
<mi>v</mi>
<mrow>
<mi>i</mi>
<mi>k</mi>
</mrow>
<mrow>
<mi>l</mi>
<mi>o</mi>
<mi>w</mi>
</mrow>
</msubsup>
<msub>
<mi>x</mi>
<mrow>
<mi>i</mi>
<mi>k</mi>
</mrow>
</msub>
</mrow>
WhereinIt representsThe slope of middle kth section broken line,It representsThe slope of middle kth section broken line;So in order to
DefinitionWithOnly it needs to be determined thatWith
Theorem 1:According to piecewise linear functionWithMeetIt is full determining
After the piecewise linear function required enough, the constraints in formula 25 is loosened;Specifically, formula 25 is replaced with following two groups of inequality
It changes;
<mrow>
<mi>C</mi>
<mi>o</mi>
<mi>n</mi>
<mi>s</mi>
<mrow>
<mo>(</mo>
<msub>
<mi>s</mi>
<mi>p</mi>
</msub>
<mo>)</mo>
</mrow>
<mo>:</mo>
<mfenced open = "{" close = "">
<mtable>
<mtr>
<mtd>
<mrow>
<msub>
<mi>s</mi>
<mi>p</mi>
</msub>
<munder>
<mo>&Sigma;</mo>
<mrow>
<mi>i</mi>
<mo>&Element;</mo>
<msub>
<mi>T</mi>
<mn>1</mn>
</msub>
</mrow>
</munder>
<msup>
<mi>e</mi>
<msubsup>
<mi>b</mi>
<mrow>
<mn>1</mn>
<mi>t</mi>
</mrow>
<mi>c</mi>
</msubsup>
</msup>
<msubsup>
<mi>l</mi>
<mi>i</mi>
<mn>1</mn>
</msubsup>
<mrow>
<mo>(</mo>
<msub>
<mi>x</mi>
<mi>i</mi>
</msub>
<mo>)</mo>
</mrow>
<mo>-</mo>
<msup>
<mi>e</mi>
<msubsup>
<mi>b</mi>
<mrow>
<mn>1</mn>
<mi>p</mi>
</mrow>
<mi>c</mi>
</msubsup>
</msup>
<msubsup>
<mi>l</mi>
<mi>p</mi>
<mn>2</mn>
</msubsup>
<mrow>
<mo>(</mo>
<msub>
<mi>x</mi>
<mi>p</mi>
</msub>
<mo>)</mo>
</mrow>
<mo>&GreaterEqual;</mo>
<mn>0</mn>
<mo>,</mo>
</mrow>
</mtd>
<mtd>
<mrow>
<mo>&ForAll;</mo>
<mi>p</mi>
<mo>&Element;</mo>
<msub>
<mi>T</mi>
<mn>1</mn>
</msub>
</mrow>
</mtd>
</mtr>
<mtr>
<mtd>
<mrow>
<msub>
<mi>s</mi>
<mi>p</mi>
</msub>
<munder>
<mo>&Sigma;</mo>
<mrow>
<mi>i</mi>
<mo>&Element;</mo>
<msub>
<mi>T</mi>
<mn>1</mn>
</msub>
</mrow>
</munder>
<msup>
<mi>e</mi>
<msubsup>
<mi>b</mi>
<mrow>
<mn>1</mn>
<mi>t</mi>
</mrow>
<mi>c</mi>
</msubsup>
</msup>
<msubsup>
<mi>l</mi>
<mi>i</mi>
<mn>2</mn>
</msubsup>
<mrow>
<mo>(</mo>
<msub>
<mi>x</mi>
<mi>i</mi>
</msub>
<mo>)</mo>
</mrow>
<mo>-</mo>
<msup>
<mi>e</mi>
<msubsup>
<mi>b</mi>
<mrow>
<mn>1</mn>
<mi>p</mi>
</mrow>
<mi>c</mi>
</msubsup>
</msup>
<msubsup>
<mi>l</mi>
<mi>p</mi>
<mn>1</mn>
</msubsup>
<mrow>
<mo>(</mo>
<msub>
<mi>x</mi>
<mi>p</mi>
</msub>
<mo>)</mo>
</mrow>
<mo>&le;</mo>
<mn>0</mn>
<mo>,</mo>
</mrow>
</mtd>
<mtd>
<mrow>
<mo>&ForAll;</mo>
<mi>p</mi>
<mo>&Element;</mo>
<msub>
<mi>T</mi>
<mn>1</mn>
</msub>
</mrow>
</mtd>
</mtr>
</mtable>
</mfenced>
</mrow>
This two groups of inequality are denoted as Cons (sp);By sr、zpAnd zrThe similar other inequality introduced are denoted as Cons respectively
(sr)、Cons(zp) and Cons (zr);
Based on the modification above to subproblem, corresponding MIQCQP is established:
<mrow>
<munder>
<mi>max</mi>
<mrow>
<mi>x</mi>
<mo>,</mo>
<mi>s</mi>
<mo>,</mo>
<mi>z</mi>
</mrow>
</munder>
<munder>
<mo>&Sigma;</mo>
<mrow>
<mi>i</mi>
<mo>&Element;</mo>
<msub>
<mi>T</mi>
<mn>1</mn>
</msub>
</mrow>
</munder>
<msub>
<mi>s</mi>
<mi>i</mi>
</msub>
<mo>&lsqb;</mo>
<msubsup>
<mi>R</mi>
<mi>i</mi>
<mi>d</mi>
</msubsup>
<msub>
<mi>x</mi>
<mi>i</mi>
</msub>
<mo>+</mo>
<msubsup>
<mi>P</mi>
<mi>i</mi>
<mi>d</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mn>1</mn>
<mo>-</mo>
<msub>
<mi>x</mi>
<mi>i</mi>
</msub>
<mo>)</mo>
</mrow>
<mo>&rsqb;</mo>
<mo>+</mo>
<munder>
<mo>&Sigma;</mo>
<mrow>
<mi>j</mi>
<mo>&Element;</mo>
<msub>
<mi>T</mi>
<mn>2</mn>
</msub>
</mrow>
</munder>
<msub>
<mi>z</mi>
<mi>j</mi>
</msub>
<mo>&lsqb;</mo>
<msubsup>
<mi>R</mi>
<mi>j</mi>
<mi>d</mi>
</msubsup>
<msub>
<mi>x</mi>
<mi>j</mi>
</msub>
<mo>+</mo>
<msubsup>
<mi>P</mi>
<mi>j</mi>
<mi>d</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mn>1</mn>
<mo>-</mo>
<msub>
<mi>x</mi>
<mi>j</mi>
</msub>
<mo>)</mo>
</mrow>
<mo>&rsqb;</mo>
<mo>-</mo>
<mo>-</mo>
<mo>-</mo>
<mrow>
<mo>(</mo>
<mn>26</mn>
<mo>)</mo>
</mrow>
</mrow>
<mrow>
<munder>
<mo>&Sigma;</mo>
<mrow>
<mi>i</mi>
<mo>&Element;</mo>
<msub>
<mi>T</mi>
<mn>1</mn>
</msub>
<mo>&cup;</mo>
<msub>
<mi>T</mi>
<mn>2</mn>
</msub>
</mrow>
</munder>
<msub>
<mi>x</mi>
<mi>i</mi>
</msub>
<mo>&le;</mo>
<mi>M</mi>
<mo>-</mo>
<mo>-</mo>
<mo>-</mo>
<mrow>
<mo>(</mo>
<mn>27</mn>
<mo>)</mo>
</mrow>
</mrow>
<mrow>
<mtable>
<mtr>
<mtd>
<mrow>
<munder>
<mo>&Sigma;</mo>
<mrow>
<mi>i</mi>
<mo>&Element;</mo>
<msub>
<mi>T</mi>
<mn>1</mn>
</msub>
</mrow>
</munder>
<msub>
<mi>z</mi>
<mi>i</mi>
</msub>
<mo>&lsqb;</mo>
<msubsup>
<mi>R</mi>
<mi>i</mi>
<mi>a</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mn>1</mn>
<mo>-</mo>
<msub>
<mi>x</mi>
<mi>i</mi>
</msub>
<mo>)</mo>
</mrow>
<mo>+</mo>
<msubsup>
<mi>P</mi>
<mi>i</mi>
<mi>a</mi>
</msubsup>
<msub>
<mi>x</mi>
<mi>i</mi>
</msub>
<mo>&rsqb;</mo>
<mo>&le;</mo>
<mfrac>
<mn>1</mn>
<mn>2</mn>
</mfrac>
<mo>{</mo>
<munder>
<mo>&Sigma;</mo>
<mrow>
<mi>i</mi>
<mo>&Element;</mo>
<msub>
<mi>T</mi>
<mn>1</mn>
</msub>
</mrow>
</munder>
<msub>
<mi>s</mi>
<mi>i</mi>
</msub>
<mo>&lsqb;</mo>
<mrow>
<mo>(</mo>
<msubsup>
<mi>R</mi>
<mi>i</mi>
<mi>a</mi>
</msubsup>
<mo>+</mo>
<mo>&Element;</mo>
<mo>)</mo>
</mrow>
<mrow>
<mo>(</mo>
<mn>1</mn>
<mo>-</mo>
<msub>
<mi>x</mi>
<mi>i</mi>
</msub>
<mo>)</mo>
</mrow>
<mo>+</mo>
<msubsup>
<mi>P</mi>
<mi>i</mi>
<mi>a</mi>
</msubsup>
<msub>
<mi>x</mi>
<mi>i</mi>
</msub>
<mo>&rsqb;</mo>
</mrow>
</mtd>
</mtr>
<mtr>
<mtd>
<mrow>
<mo>+</mo>
<munder>
<mo>&Sigma;</mo>
<mrow>
<mi>j</mi>
<mo>&Element;</mo>
<msub>
<mi>T</mi>
<mn>2</mn>
</msub>
</mrow>
</munder>
<msub>
<mi>s</mi>
<mi>j</mi>
</msub>
<mo>&lsqb;</mo>
<mrow>
<mo>(</mo>
<msubsup>
<mi>R</mi>
<mi>j</mi>
<mi>a</mi>
</msubsup>
<mo>+</mo>
<mo>&Element;</mo>
<mo>)</mo>
</mrow>
<mrow>
<mo>(</mo>
<mn>1</mn>
<mo>-</mo>
<msub>
<mi>x</mi>
<mi>j</mi>
</msub>
<mo>)</mo>
</mrow>
<mo>+</mo>
<msubsup>
<mi>P</mi>
<mi>j</mi>
<mi>a</mi>
</msubsup>
<msub>
<mi>x</mi>
<mi>j</mi>
</msub>
<mo>&rsqb;</mo>
<mo>}</mo>
</mrow>
</mtd>
</mtr>
</mtable>
<mo>-</mo>
<mo>-</mo>
<mo>-</mo>
<mrow>
<mo>(</mo>
<mn>28</mn>
<mo>)</mo>
</mrow>
</mrow>
<mrow>
<mtable>
<mtr>
<mtd>
<mrow>
<munder>
<mo>&Sigma;</mo>
<mrow>
<mi>j</mi>
<mo>&Element;</mo>
<msub>
<mi>T</mi>
<mn>2</mn>
</msub>
</mrow>
</munder>
<msub>
<mi>z</mi>
<mi>j</mi>
</msub>
<mo>&lsqb;</mo>
<msubsup>
<mi>R</mi>
<mi>j</mi>
<mi>a</mi>
</msubsup>
<mrow>
<mo>(</mo>
<mn>1</mn>
<mo>-</mo>
<msub>
<mi>x</mi>
<mi>j</mi>
</msub>
<mo>)</mo>
</mrow>
<mo>+</mo>
<msubsup>
<mi>P</mi>
<mi>j</mi>
<mi>a</mi>
</msubsup>
<msub>
<mi>x</mi>
<mi>j</mi>
</msub>
<mo>&rsqb;</mo>
<mo>&GreaterEqual;</mo>
<mfrac>
<mn>1</mn>
<mn>2</mn>
</mfrac>
<mo>{</mo>
<munder>
<mo>&Sigma;</mo>
<mrow>
<mi>i</mi>
<mo>&Element;</mo>
<msub>
<mi>T</mi>
<mn>1</mn>
</msub>
</mrow>
</munder>
<msub>
<mi>s</mi>
<mi>i</mi>
</msub>
<mo>&lsqb;</mo>
<mrow>
<mo>(</mo>
<msubsup>
<mi>R</mi>
<mi>i</mi>
<mi>a</mi>
</msubsup>
<mo>+</mo>
<mo>&Element;</mo>
<mo>)</mo>
</mrow>
<mrow>
<mo>(</mo>
<mn>1</mn>
<mo>-</mo>
<msub>
<mi>x</mi>
<mi>i</mi>
</msub>
<mo>)</mo>
</mrow>
<mo>+</mo>
<msubsup>
<mi>P</mi>
<mi>i</mi>
<mi>a</mi>
</msubsup>
<msub>
<mi>x</mi>
<mi>i</mi>
</msub>
<mo>&rsqb;</mo>
</mrow>
</mtd>
</mtr>
<mtr>
<mtd>
<mrow>
<mo>+</mo>
<munder>
<mo>&Sigma;</mo>
<mrow>
<mi>j</mi>
<mo>&Element;</mo>
<msub>
<mi>T</mi>
<mn>2</mn>
</msub>
</mrow>
</munder>
<msub>
<mi>s</mi>
<mi>j</mi>
</msub>
<mo>&lsqb;</mo>
<mrow>
<mo>(</mo>
<msubsup>
<mi>R</mi>
<mi>j</mi>
<mi>a</mi>
</msubsup>
<mo>+</mo>
<mo>&Element;</mo>
<mo>)</mo>
</mrow>
<mrow>
<mo>(</mo>
<mn>1</mn>
<mo>-</mo>
<msub>
<mi>x</mi>
<mi>j</mi>
</msub>
<mo>)</mo>
</mrow>
<mo>+</mo>
<msubsup>
<mi>P</mi>
<mi>j</mi>
<mi>a</mi>
</msubsup>
<msub>
<mi>x</mi>
<mi>j</mi>
</msub>
<mo>&rsqb;</mo>
<mo>}</mo>
</mrow>
</mtd>
</mtr>
</mtable>
<mo>-</mo>
<mo>-</mo>
<mo>-</mo>
<mrow>
<mo>(</mo>
<mn>29</mn>
<mo>)</mo>
</mrow>
</mrow>
Cons(sp), Cons (sr), Cons (zp), Cons (zr) (30)
<mrow>
<msub>
<mi>x</mi>
<mi>t</mi>
</msub>
<mo>=</mo>
<munderover>
<mo>&Sigma;</mo>
<mrow>
<mi>k</mi>
<mo>=</mo>
<mn>1</mn>
</mrow>
<mi>K</mi>
</munderover>
<msub>
<mi>x</mi>
<mrow>
<mi>t</mi>
<mi>k</mi>
</mrow>
</msub>
<mo>,</mo>
<mo>&ForAll;</mo>
<mi>t</mi>
<mo>&Element;</mo>
<msub>
<mi>T</mi>
<mn>1</mn>
</msub>
<mo>&cup;</mo>
<msub>
<mi>T</mi>
<mn>2</mn>
</msub>
<mo>-</mo>
<mo>-</mo>
<mo>-</mo>
<mrow>
<mo>(</mo>
<mn>31</mn>
<mo>)</mo>
</mrow>
</mrow>
<mrow>
<mn>0</mn>
<mo>&le;</mo>
<msub>
<mi>x</mi>
<mrow>
<mi>t</mi>
<mi>k</mi>
</mrow>
</msub>
<mo>&le;</mo>
<mfrac>
<mn>1</mn>
<mi>K</mi>
</mfrac>
<mo>,</mo>
<mo>&ForAll;</mo>
<mi>t</mi>
<mo>&Element;</mo>
<msub>
<mi>T</mi>
<mn>1</mn>
</msub>
<mo>&cup;</mo>
<msub>
<mi>T</mi>
<mn>2</mn>
</msub>
<mo>,</mo>
<mi>k</mi>
<mo>=</mo>
<mn>1</mn>
<mo>,</mo>
<mn>...</mn>
<mo>,</mo>
<mi>K</mi>
<mo>-</mo>
<mo>-</mo>
<mo>-</mo>
<mrow>
<mo>(</mo>
<mn>32</mn>
<mo>)</mo>
</mrow>
</mrow>
<mrow>
<msub>
<mi>y</mi>
<mrow>
<mi>t</mi>
<mi>k</mi>
</mrow>
</msub>
<mo>&Element;</mo>
<mo>{</mo>
<mn>0</mn>
<mo>,</mo>
<mn>1</mn>
<mo>}</mo>
<mo>,</mo>
<mo>&ForAll;</mo>
<mi>t</mi>
<mo>&Element;</mo>
<msub>
<mi>T</mi>
<mn>1</mn>
</msub>
<mo>&cup;</mo>
<msub>
<mi>T</mi>
<mn>2</mn>
</msub>
<mo>,</mo>
<mi>k</mi>
<mo>=</mo>
<mn>1</mn>
<mo>,</mo>
<mn>...</mn>
<mo>,</mo>
<mi>K</mi>
<mo>-</mo>
<mo>-</mo>
<mo>-</mo>
<mrow>
<mo>(</mo>
<mn>33</mn>
<mo>)</mo>
</mrow>
</mrow>
<mrow>
<msub>
<mi>y</mi>
<mrow>
<mi>t</mi>
<mi>k</mi>
</mrow>
</msub>
<mfrac>
<mn>1</mn>
<mi>K</mi>
</mfrac>
<mo>&le;</mo>
<msub>
<mi>x</mi>
<mrow>
<mi>t</mi>
<mi>k</mi>
</mrow>
</msub>
<mo>,</mo>
<mo>&ForAll;</mo>
<mi>t</mi>
<mo>&Element;</mo>
<msub>
<mi>T</mi>
<mn>1</mn>
</msub>
<mo>&cup;</mo>
<msub>
<mi>T</mi>
<mn>2</mn>
</msub>
<mo>,</mo>
<mi>k</mi>
<mo>=</mo>
<mn>1</mn>
<mo>,</mo>
<mn>...</mn>
<mo>,</mo>
<mi>K</mi>
<mo>-</mo>
<mo>-</mo>
<mo>-</mo>
<mrow>
<mo>(</mo>
<mn>34</mn>
<mo>)</mo>
</mrow>
</mrow>
<mrow>
<msub>
<mi>x</mi>
<mrow>
<mi>t</mi>
<mo>,</mo>
<mi>k</mi>
<mo>+</mo>
<mn>1</mn>
</mrow>
</msub>
<mo>&le;</mo>
<msub>
<mi>y</mi>
<mrow>
<mi>t</mi>
<mi>k</mi>
</mrow>
</msub>
<mo>,</mo>
<mo>&ForAll;</mo>
<mi>t</mi>
<mo>&Element;</mo>
<msub>
<mi>T</mi>
<mn>1</mn>
</msub>
<mo>&cup;</mo>
<msub>
<mi>T</mi>
<mn>2</mn>
</msub>
<mo>,</mo>
<mi>k</mi>
<mo>=</mo>
<mn>1</mn>
<mo>,</mo>
<mn>...</mn>
<mo>,</mo>
<mi>K</mi>
<mo>-</mo>
<mo>-</mo>
<mo>-</mo>
<mrow>
<mo>(</mo>
<mn>35</mn>
<mo>)</mo>
</mrow>
</mrow>
Subproblem is rewritten with the variable newly introduced in formula 26 to 29;Formula 30 represents the quadratic constraints of generation;Formula 31 and 32 is
To xtkConstraint;
MIQCQP is solved using the SCIP methods of Vigerske S.et al. propositions;
Theorem 2:NoteForWithThe distance between, then limK→+∞Dis=
0。
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201711324273.1A CN108053068A (en) | 2017-12-13 | 2017-12-13 | The method that mankind attacker cooperation behavior modeled and formulated corresponding Defending Policy |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201711324273.1A CN108053068A (en) | 2017-12-13 | 2017-12-13 | The method that mankind attacker cooperation behavior modeled and formulated corresponding Defending Policy |
Publications (1)
Publication Number | Publication Date |
---|---|
CN108053068A true CN108053068A (en) | 2018-05-18 |
Family
ID=62132545
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201711324273.1A Pending CN108053068A (en) | 2017-12-13 | 2017-12-13 | The method that mankind attacker cooperation behavior modeled and formulated corresponding Defending Policy |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN108053068A (en) |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103152345A (en) * | 2013-03-07 | 2013-06-12 | 南京理工大学常熟研究院有限公司 | Network safety optimum attacking and defending decision method for attacking and defending game |
CN106921666A (en) * | 2017-03-06 | 2017-07-04 | 中山大学 | A kind of ddos attack system of defense and method based on Synergy |
-
2017
- 2017-12-13 CN CN201711324273.1A patent/CN108053068A/en active Pending
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103152345A (en) * | 2013-03-07 | 2013-06-12 | 南京理工大学常熟研究院有限公司 | Network safety optimum attacking and defending decision method for attacking and defending game |
CN106921666A (en) * | 2017-03-06 | 2017-07-04 | 中山大学 | A kind of ddos attack system of defense and method based on Synergy |
Non-Patent Citations (3)
Title |
---|
BINRU WANG等: ""On Repeated Stackelberg Security Game with the Cooperative Human Behavior Model for Wildlife Protection"", 《PROCEEDINGS OF THE 16TH CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS》 * |
姜伟: "" 基于攻防博弈模型的主动防御关键技术研究"", 《中国博士学位论文全文数据库 信息科技辑》 * |
赵婧等: ""基于轨迹频率抑制的轨迹隐私保护方法"", 《计算机学报》 * |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Kolkowicz et al. | Soldiers, peasants, and bureaucrats: civil-military relations in Communist and modernizing societies | |
CN106682351A (en) | Fight simulation system generating military strength based on computer and simulation method | |
Montague | Stolen goods: Coltan and conflict in the Democratic Republic of Congo | |
CN109426567A (en) | A kind of node deployment and electoral machinery of block chain | |
CN110166428A (en) | Intelligence defence decision-making technique and device based on intensified learning and attacking and defending game | |
CN110035066B (en) | Attack and defense behavior quantitative evaluation method and system based on game theory | |
CN112883874A (en) | Active defense method aiming at deep face tampering | |
Kullberg et al. | Elites, institutions and democratisation in Russia and Eastern Europe | |
CN108053068A (en) | The method that mankind attacker cooperation behavior modeled and formulated corresponding Defending Policy | |
Russell | The Persian Gulf's collective-security mirage | |
Avdagic | Accounting for variations in trade union effectiveness: State-labor relations in East Central Europe | |
CN115983389A (en) | Attack and defense game decision method based on reinforcement learning | |
DeLong et al. | Teamskill evolved: Mixed classification schemes for team-based multi-player games | |
Ahmad et al. | Changing fortunes in anti‐poverty programmes in Bangladesh | |
CN113792331A (en) | Fair-resistance interference-based federated learning member reasoning attack defense method | |
CN114662655A (en) | Attention mechanism-based weapon and chess deduction AI hierarchical decision method and device | |
Mwangi | Mozambique's Foreign Policy: From Ideological Conflict to Pragmatic Cooperation | |
Robinson | Authoritarianism with a Palestinian face | |
Yamadera | Examining the myth of money with agent-based modelling | |
Atkinson et al. | On-the-fly generation of multi-robot team formation strategies based on game conditions | |
Martinez-Mejorado et al. | A multidimensional framework to uncover insights of group performance and outcomes in competitive environments with a case study of FIFA World Cups | |
Grieshop et al. | A Continuous-Time Stochastic Process for High-Resolution Network Data in Sports | |
Yang et al. | Modeling human bounded rationality to improve defender strategies in network security games | |
Grossmann et al. | Viability and sustainability of civilizations, corporations, institutions and ecological systems | |
Van Schendelen | The Netherlands: Parliamentary parties rival with pressure groups |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20180518 |