CN108053068A - The method that mankind attacker cooperation behavior modeled and formulated corresponding Defending Policy - Google Patents

The method that mankind attacker cooperation behavior modeled and formulated corresponding Defending Policy Download PDF

Info

Publication number
CN108053068A
CN108053068A CN201711324273.1A CN201711324273A CN108053068A CN 108053068 A CN108053068 A CN 108053068A CN 201711324273 A CN201711324273 A CN 201711324273A CN 108053068 A CN108053068 A CN 108053068A
Authority
CN
China
Prior art keywords
mrow
msub
msubsup
munder
attacker
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201711324273.1A
Other languages
Chinese (zh)
Inventor
王彬入
张渊
仲盛
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nanjing University
Original Assignee
Nanjing University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanjing University filed Critical Nanjing University
Priority to CN201711324273.1A priority Critical patent/CN108053068A/en
Publication of CN108053068A publication Critical patent/CN108053068A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/04Forecasting or optimisation specially adapted for administrative or management purposes, e.g. linear programming or "cutting stock problem"
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/06Resources, workflows, human or project management; Enterprise or organisation planning; Enterprise or organisation modelling
    • G06Q10/063Operations research, analysis or management
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q50/00Systems or methods specially adapted for specific business sectors, e.g. utilities or tourism
    • G06Q50/10Services
    • G06Q50/26Government or public services
    • G06Q50/265Personal security, identity or safety

Landscapes

  • Business, Economics & Management (AREA)
  • Engineering & Computer Science (AREA)
  • Human Resources & Organizations (AREA)
  • Strategic Management (AREA)
  • Economics (AREA)
  • Tourism & Hospitality (AREA)
  • Entrepreneurship & Innovation (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Marketing (AREA)
  • General Physics & Mathematics (AREA)
  • General Business, Economics & Management (AREA)
  • Development Economics (AREA)
  • Educational Administration (AREA)
  • Quality & Reliability (AREA)
  • Operations Research (AREA)
  • Game Theory and Decision Science (AREA)
  • Computer Security & Cryptography (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Primary Health Care (AREA)
  • Medicines Containing Antibodies Or Antigens For Use As Internal Diagnostic Agents (AREA)

Abstract

The present invention discloses the mankind attacker cooperation behavior modeling in a kind of security protection and the method for formulating corresponding Defending Policy, first, an adaptive human behavior model is proposed, for optimizing distribution of the tool there are two limited defence resource in the repetition SSG for the attacker that may cooperate;Secondly, devise an efficient algorithm and approximate solve the Nonlinear Nonconvex optimization problem obtained by model, final output Defending Policy;Finally, true man's simulated experiment has been carried out to test the validity of model of the present invention and method.

Description

The method that mankind attacker cooperation behavior modeled and formulated corresponding Defending Policy
Technical field
The present invention relates to the mankind attacker cooperation behavior modeling in a kind of security protection and formulate corresponding Defending Policy Method belongs to technical field of safety protection.
Background technology
The safe games of Stackelberg (SSG) are applied in a variety of safeguard protection scenes in real life, such as beautiful State's Coast Guard, federal air marshal, Los Angeles Airport etc..
Instantly it is to patrol to needing the most common defence method that target to be protected is taken in the case where defence resource is limited It patrols.Patrolman only has limited defence resource (such as human resources), so they cannot be a piece of very big in guard of same time Area every nook and cranny.In addition, attacker can be with the patrol rule of observation patrol person, and corresponding attack action is designed to keep away Exempt to be grabbed.Therefore, it is significant for defence organization's optimization Defending Policy.
In traditional single-wheel SSG, defender (leaders) arranges patrol strategy first, then attacker (follower) root Target attack is selected to the observation of Defending Policy according to it.
The main problem of traditional single-wheel SSG is the incomplete reasonability for not accounting for mankind attacker, and without abundant Utilize the passing data-optimized Defending Policy of attack.
The content of the invention
Goal of the invention:For problems of the prior art and deficiency, the present invention provides the people in a kind of security protection The method that class attacker cooperation behavior modeled and formulated corresponding Defending Policy, first, it is proposed that an adaptive human behavior Model, for optimizing distribution of the tool there are two limited defence resource in the repetition SSG for the attacker that may cooperate;Secondly, design One efficient algorithm approximate solve the Nonlinear Nonconvex optimization problem obtained by model, final output Defending Policy; Finally, true man's simulated experiment has been carried out to test the validity of model of the present invention and method.
Technical solution:A kind of side of the modeling of mankind attacker cooperation behavior and the corresponding Defending Policy of formulation in security protection Method, including following content:
(1) cooperative mechanism in SSG is repeated
Repeatedly defender and two attackers, the two attackers can choose whether to cooperate there are one cooperative games;It is anti- The person of keeping periodically arranges new Defending Policy as the leader in game, and attacker is as follower, according to it to defender Attack is made in the observation of the Defending Policy of arrangement;This game is formed more by taking turns, thus defender can periodically according to front-wheel The data that are collected into secondary change Defending Policy;Attacker independently attacks or cooperation attack;In the case where independently attacking, Whether the income that attacker attacks every time is only dependent upon the attacker this attack and succeeds.And in the case where cooperation is attacked, it attacks The income that the person of hitting attacks every time by be two people's income sums half;An extra returns ∈ is introduced simultaneously, i.e., when an attack When person has successfully carried out once cooperation attack, he will obtain extra returns ∈;
Attackers need to make two main selections in game:1) selection attack target and 2) choose whether with separately One attacker cooperates.Only when two attackers select cooperation in certain attack, a cooperative relationship just calculates foundation.
The defence resource summation M possessed by defender defends a target collection T={ 1 ..., | T | }, and T is by T1With T2Two are not overlapped subset and are combined into;T1Represent the target collection that first attacker may attack, T2Represent second attacker The target collection that may be attacked, and T1=T-T2;Vector x is defended, per one-dimensional element xiDefender is represented in each target The defended probability of the defence resource namely target i of input.
Income of one attacker in target of attack i depend on 1) this target it is whether defended and 2) two attack Whether person cooperates.When an attacker selects independent attack, the income that he attacks not defended target i isAttack quilt The income of the target i of defence isRelatively, defender is in the income of the previous caseIn the income of latter situation ForThe total revenue of defender for it in two incomes by target of attack and.AndIf two A attacker agrees to cooperate, they will obtain extra returns ∈ when attacking not defended target.
(2) incomplete reasonability and adaptivity
In some illegal activities, attacker is typically considered Imperfect Rationality.For the mankind's in SSG is repeated Incomplete reasonability proposes a model.
Tendentiousness of the attacker for some target is defined first.
Define 1:One attacker is in R wheels to the tendentiousness of target iIt is defined by following equation:
WhereinIt isWithDifference,Represent the single attack that attacker r wheels carry out at target i Average yield,Represent the single attack average yield for all attacks that attacker carries out in r wheels, QrFor a vector, Its i-th dimension isC be a constant, Var (Qr) represent { QrVariance.Therefore in fixation { QrDistribution in the case of, increase AddAlso can accordingly increase.First equation explanationForAverage (r=1 ..., R).
When attacker tends to target of attack iTo be just, on the contrary will be negative, and attacker be not revealed inclining for target Tropism is set to 0.
Adaptive subjectivity revenue function:
D=1/ (N are setr- r), wherein vector ω=(ω1, ω2, ω3) for parameter to be learned, NrAlways to take turns number, r is represented For learning the round of ω.
(3) optimization problem provides
With reference to human behavior model and from the parameter that attack data learning obtains, established by equation 1 to equation 19 Generate the optimization problem of optimal Defending Policy.
D=d1+d2 (2)
α1, α2, β ∈ { 0,1 } (3)
β≤α1 (17)
β≤α2 (18)
α12≤β+1 (19)
In all equatioies, Z represents a larger constant, and c represents cooperation, and nc represents uncooperative.
Formula 1 gives the expected revenus of the target of optimization problem, i.e. defender.From 2 visual target of formula by d1And d2Two parts It forms.We define three binary variables, wherein α in formula 31And α2Attacker is represented each for the selection of cooperation, β generations The final cooperation state of table.Only when two attackers agree to cooperation, β 1, β is 0 in all other cases.This It can be ensured by the constraints in formula 17 and 19.Formula 4 and 5 ensure that the every one-dimensional all between [0,1] of defence vector x, and The total resources of defender are not more than M.In formula 6 to 11, based on α1、α2Two parts of defender's expected revenus are defined with β --- d1And d2
With d1Exemplified by, if (i.e. β=0) is not reached in cooperation, for d1Constraints will be determined by formula 6 and 7.On the contrary, If (i.e. β=1), d are reached in cooperation1Constraints will be determined by formula 8.Similarly, formula 9 to 11 is constituted to d2Constraint.
Formula 12 to 14 respectively define attacker cooperate with it is uncooperative when expected revenus.Here two attackers are distinguished Modeling, so they have different expected revenusesWithBy formula 15 and 16, α ensure that1And α2It is to be received based on expectation What benefit determined.As previously mentioned, three constraintss in formula 17 to 19 are for ensuring that correctly running for cooperative mechanism, i.e., only There is the cooperation when two attackers select cooperation just to reach.
(4) optimal policy is calculated
Because above-mentioned optimization problem is complicated, this optimization problem is decomposed into 4 subproblems.By solving subproblem, and Globally optimal solution is selected in the optimal solution of subproblem, just can solve former problem, obtains defence vector x.
PROBLEM DECOMPOSITION
The angle of cooperation whether is selected from attacker, may be occurred there are four types of situation altogether.Table 3 summarizes defender at this Expected revenus (while being also the object function of subproblem) D in the case of four kinds1、D2、D3And D4
Table 3
Expected revenus in given table 3, just can correspondingly establish 4 subproblems.Without loss of generality, one of them is only provided The definition (formula 20 to 24) of subproblem.In this subproblem, first attacker selects cooperation and second attacker selects It is uncooperative.The mark used in subproblemWithCome from the definition in previous formula 12 to 14.
Based on approximate method
Four subproblems suffer from the simple form of comparison, but the globally optimal solution for finding subproblem is still difficult , therefore piecewise linear function is introduced, constraints is loosened, it is secondary that subproblem is converted into a MIXED INTEGER quadratic constraints Planning problem (MIQCQP) solves.Such issues that can be with any MIQCQP solutions come numerical solution.
Some variable replacements are done first.Order So D4A quadric form can be rewritten as, sees formula 26.Because introduce four A new variable to s and z according to definition, it is necessary to increase by four constraintss.Here as example is only provided for spPact Beam, others constraint can be provided similarly.
Formula 25 is a nonlinear restriction, therefore it is replaced with two groups and slightly loosened to obtain quadratic constraints condition.Now It illustrates how to find such two set of segmentation linear functions.
OrderIt needs to find two groups of suitably piecewise linear functionsWithSo thatHere f (xi) on [0,1] it is a monotone decreasing convex function, because For negative.
Use xikDefinitionWith
WhereinIt representsThe slope of middle kth section broken line,It representsThe slope of middle kth section broken line.So In order to defineWithOnly it needs to be determined thatWith
Theorem 1:According to piecewise linear functionWithMeet
After the piecewise linear function met the requirements is determined, the constraints in formula 25 can be loosened.Specifically, will Formula 25 is replaced with following two groups of inequality.
This two groups of inequality are denoted as Cons (sp).By sr、zpAnd zrThe similar other inequality introduced are denoted as respectively Cons(sr)、Cons(zp) and Cons (zr)。
Based on the modification above to subproblem, corresponding MIQCQP is established:
Cons(sp), Cons (sr), Cons (zp), Cons (zr) (30)
Subproblem is rewritten with the variable newly introduced in formula 26 to 29.Formula 30 represents the quadratic constraints of generation.31 He of formula 32 be to xtkConstraint.
MIQCQP there are many ∈-global optimization solution, come by the SCIP methods proposed using Vigerske S.et al. Solve MIQCQP.
Theorem 2:NoteForWithThe distance between, then limK→+∞ Dis=0.
Description of the drawings
Fig. 1 be as K=4,WithComparison diagram;
Fig. 2 be as K=10,WithComparison diagram;
Fig. 3 is the experiment interface of participant in experiment;
Fig. 4 is the earnings structure S1 and S2 used in two groups of experiments;
When Fig. 5 is K=10, using earnings structure S1 manoeuvre second wheel in " Maximin " method generate strategy and The strategy of model generation proposed by the present invention, wherein (a) is the strategy of " Maximin " method generation, (b) proposes for the present invention The strategy of model generation;
Fig. 6 is the loss comparison diagram that defender often takes turns;
Fig. 7 is the ratio chart that cooperation attack accounts for general offensive number.
Specific embodiment
With reference to specific embodiment, the present invention is furture elucidated, it should be understood that these embodiments are merely to illustrate the present invention Rather than limit the scope of the invention, after the present invention has been read, those skilled in the art are to the various equivalences of the present invention The modification of form falls within the application scope as defined in the appended claims.
A kind of method of the modeling of mankind attacker cooperation behavior and the corresponding Defending Policy of formulation in security protection, including such as Lower content:
(1) cooperative mechanism in SSG is repeated
Repeatedly defender and two attackers, the two attackers can choose whether to cooperate there are one cooperative games;One A defender or attacker can represent a tissue and not necessarily only represent personal;Defender as the leader in game, Periodically arrange new Defending Policy, and attacker does the observation of the Defending Policy of defender's arrangement according to it as follower Go out attack;This game is formed more by taking turns, so defender can periodically change according to the data that be collected into round before Defending Policy;Attacker independently attacks or cooperation attack;In the case where independently attacking, the income that attacker attacks every time is only Depending on the attacker, whether this attack succeeds.And in the case where cooperation is attacked, the income that attacker attacks every time will be The half of two people's income sums;An extra returns ∈ is introduced simultaneously, i.e., when an attacker is successfully once cooperated During attack, he will obtain extra returns ∈;
Generally speaking, attackers need to make two main selections in game:1) target of selection attack and 2) choosing It selects and whether cooperates with another attacker.Only when two attackers select cooperation in certain attack, once cooperation is closed System just calculates and establishes.
Table 1
In order to further describe this game, the mark being introduced into table 1.One target collection T=is defendd by M { 1 ..., | T | }, T is by T1And T2Two are not overlapped subset and are combined into.T1Represent the object set that first attacker may attack It closes, T2Represent the target collection that second attacker may attack, and T1=T-T2.The two subsets are non-overlapped, if becauseAttacker may attack identical target, this probability that them will be caused all to be arrested in once attacking Rise.Vector x is defended, per one-dimensional element xiRepresent defence resource namely the target i that defender is put into each target Defended probability.
Table 2
Income of one attacker in target of attack i depend on 1) this target it is whether defended and 2) two attack Whether person cooperates.When an attacker selects independent attack, the income that he attacks not defended target i isAttack quilt The income of the target i of defence isRelatively, defender is in the income of the previous caseIn the income of latter situation ForThe total revenue of defender for it in two incomes by target of attack and.Because income and mistake that attacker successfully obtains Loss and income of the loss being subject to respectively from defender are lost, so it is contemplated that zero-sum game.Then haveWith AndIf two attackers agree to cooperate, they will obtain extra returns when attacking not defended target ∈.The income of defender and each attacker are shown in Table 2 under cooperation.In table 2, i ∈ T1Represent first attacker's selection Target of attack, j ∈ T2Represent the target of attack of second attacker's selection." A1 " represents first attacker, and " A2 " is represented Second attacker." S " is represented successfully, and " F " represents failure.Such as " A1S, A2F " represent first attacker success and second A attacker's failure.
(2) incomplete reasonability and adaptivity
In some illegal activities, attacker is typically considered Imperfect Rationality.For the mankind's in SSG is repeated Incomplete reasonability proposes a model.
Tendentiousness of the attacker for some target is defined first.
Define 1:One attacker is in R wheels to the tendentiousness of target iIt is defined by following equation:
WhereinIt isWithDifference,Represent the single attack that attacker r wheels carry out at target i Average yield,Represent the single attack average yield for all attacks that attacker carries out in r wheels, QrFor a vector, Its i-th dimension isC be a constant, Var (Qr) represent { QrVariance.Therefore in fixation { QrDistribution in the case of, increase AddAlso can accordingly increase.First equation explanationForAverage (r=1 ..., R).
Generally speaking, when the average single attack income that attacker is obtained in some target i in passing round is more higher than institute There is the overall average single attack income of target,It will be higher.This is also consistent with intuition --- and attacker should be more likely to attack It hits him and has obtained the target of higher average single income.
It should be noted that it is likely to occur the feelings that some target is not all attacked in a certain even all rounds of wheel Condition, the information of these targets need to be disclosed.Because when attacker tends to target of attack iWill be just, otherwise will be it is negative, So attacker is set to 0 by us in the tendentiousness for not being revealed target.
There is above tendentious definition, we have proposed adaptive subjective revenue functions:
ASU(xi) in there are a kind of adaptation mechanism, it by introduce d andTo realize.Here d is one to be directed to The parameter that target information is insufficient and introduces.D=1/ (N are setr- r), wherein vector ω=(ω1, ω2, ω3) it is ginseng to be learned Number, NrAlways to take turns number, r represents the round for being used for learning ω.So will increase with the increase d of round, represent attacker and obtain Obtained more information.
In xiWeight before multiplyButWithMultiply beforeThis is because Attacker will be according to tendentiousness in modelIts subjective revenue function is adjusted, it is necessary to ensure after parameter substitution learn, Work as tendentiousnessDuring increase, ASU (xi) also corresponding increase.In an experiment, all ω learnt1And ω30 is all no more than, All ω20 is all greater than, so being made that such adjustment to weight.
Finally, the model of the present invention is summarized in the case where cooperation repeats SSG.In game, two attackers can select be No cooperation.At each occurrence, they have different attack probability at target i, are denoted asWithout loss of generality, only provide belowCalculating side Method, other three probability can also be calculated similarly.
Above formula subjectivity revenue functionIt is given by:
Wherein parameterIt is with non-conjunction of the maximum Likelihood (MLE) from first attacker Make what attack data learning arrived.For ease of use, we willIt is rewritten as xiLinear expression.
(3) optimization problem provides
With reference to human behavior model and from the parameter that attack data learning obtains, we arrive equation 19 by equation 1 Establish the optimization problem for generating optimal Defending Policy.
D=d1+d2 (2)
α1, α2, β ∈ { 0,1 } (3)
β≤α1 (17)
β≤α2 (18)
α12≤β+1 (19)
In all equatioies, it is (desirable that Z represents one big constantC generations Table cooperation, nc represent uncooperative.
Formula 1 gives the expected revenus of the target of optimization problem, i.e. defender.From 2 visual target of formula by d1And d2Two parts It forms.We define the integer variable that three values are only 0 or 1, wherein α in formula 31And α2It is each right to represent attacker In the selection of cooperation, β represents final cooperation state.Only when two attackers agree to cooperation, β 1, at any other In the case of β all be 0.This can be ensured by the constraints in formula 17 and 19.Formula 4 and 5 ensure that defence vector x per it is one-dimensional all Between [0,1], and the total resources of defender are not more than M.
In formula 6 to 11, based on α1、α2Two parts of defender's expected revenus are defined with β --- d1And d2.With d1For Example, if (i.e. β=0) is not reached in cooperation, for d1Constraints will be determined by formula 6 and 7.On the contrary, if cooperation is reached (i.e. β=1), d1Constraints will be determined by formula 8.Similarly, formula 9 to 11 is constituted to d2Constraint.
Formula 12 to 14 respectively define attacker cooperate with it is uncooperative when expected revenus.Here two attackers are distinguished Modeling, so they have different expected revenusesWithBy formula 15 and 16, α ensure that1And α2It is to be received based on expectation What benefit determined.As previously mentioned, three constraintss in formula 17 to 19 are for ensuring that correctly running for cooperative mechanism, i.e., only There is the cooperation when two attackers select cooperation just to reach.
(4) optimal policy is calculated
Because above-mentioned optimization problem is complicated, this optimization problem is decomposed into 4 subproblems.By solving subproblem, and Globally optimal solution is selected in the optimal solution of subproblem, just can solve former problem, obtains defence vector x.As described in table 1, x is represented (i.e. M is defended resource preventing every time to the specific defence scheme of each target any time defended probability, x and defender Which position should be appeared in keeping) it is of equal value, therefore be just equivalent to obtain Defending Policy we can say that having solved x.
PROBLEM DECOMPOSITION
The angle of cooperation whether is selected from attacker, may be occurred there are four types of situation altogether.Table 3 summarizes defender at this Expected revenus (while being also the object function of subproblem) D in the case of four kinds1、D2、D3And D4
Table 3
Expected revenus in given table 3, just can correspondingly establish 4 subproblems.Without loss of generality, one of them is only provided The definition (formula 20 to 24) of subproblem.In this subproblem, first attacker selects cooperation and second attacker selects It is uncooperative.The mark used in subproblemWithCome from the definition in previous formula 12 to 14.
Based on approximate method
Four subproblems suffer from the simple form of comparison, but the globally optimal solution for finding subproblem is still difficult , therefore piecewise linear function is introduced, constraints is loosened, it is secondary that subproblem is converted into a MIXED INTEGER quadratic constraints Planning problem (MIQCQP) solves.Such issues that can be with any MIQCQP solutions come numerical solution.
Some variable replacements are done first.Order So D4A quadric form can be rewritten as, sees formula 26.Because introduce four A new variable to s and z according to definition, it is necessary to increase by four constraintss.Here as example is only provided for spPact Beam, others constraint can be provided similarly.
Formula 25 is a nonlinear restriction, therefore it is replaced with two groups and slightly loosened to obtain quadratic constraints condition.Now It illustrates how to find such two set of segmentation linear functions.
OrderIt needs to find two groups of suitably piecewise linear functionsWithSo thatHere f (xi) on [0,1] it is a monotone decreasing convex function, because For negative.Based on descending and convexity, we are illustrated in Fig. 1 as K=10, f (xi) and piecewise linear function WithFunctional image on [0,1].
It will be seen that [0,1] is equally divided into K (K=4 here) part from Fig. 1, per portion by a new variable xik(k=1,2 ..., K) is represented.Obviously,And Use xikDefinitionWith
WhereinIt representsThe slope of middle kth section broken line,It representsThe slope of middle kth section broken line.So In order to defineWithOnly it needs to be determined thatWithIt will calculate WithStep Suddenly it is summarized as algorithm 1.
Above-mentioned algorithm 1 is specifically described as:
1) it is directly substituted into firstIt calculatesValue.
2) slope of kth section (k=1 ..., K) is then calculated.
3) Δ is madeI, kForOnMaximum, calculate ΔI, kThe value and basis of (k=1 ..., K) ΔI, kIt calculates
4) Δ is takenI, 0I, 1,
5) basisIt calculates
6) exportWith
Theorem 1:The piecewise linear function generated according to algorithm 1WithMeet
It proves:BecauseIt is [0, a 1] Convex Functions, it is clear that
Below based onIt provesGiven k ∈ 1, 2 ..., K }, orderHave
AndNotice g (xi) andAll it isOn linear function.Therefore,
On the other hand, becauseIt can be seen that
Thus obtainSo
After the piecewise linear function met the requirements is determined with algorithm 1, the constraints in formula 25 can be loosened.Tool Body, the following two groups of inequality of formula 25 are replaced.
This two groups of inequality are denoted as Cons (sp).By sr、zpAnd zrThe similar other inequality introduced are denoted as respectively Cons(sr)、Cons(zp) and Cons (zr)。
Based on the modification above to subproblem, corresponding MIQCQP is established:
Cons(sp), Cons (sr), Cons (zp), Cons (zr) (30)
Subproblem is rewritten with the variable newly introduced in formula 26 to 29.Formula 30 represent the method for the present invention generation it is secondary about Beam.Formula 31 and 32 is to xtkConstraint.
MIQCQP there are many ∈-global optimization solution, come by the SCIP methods proposed using Vigerske S.et al. Solve MIQCQP.
Theorem 2:NoteForWithThe distance between, then limK→+∞ Dis=0.
It proves:It is based onWithDefinition andIt is clear toOrder Thenφ (x) existsOn maximum be NoteThen
NoteIt obtains
ΔI, 1=wlnw-w+1
Because
It obtains
Theorem 2 shows that the method for the present invention is capable of providing and constraints arbitrary extent is loosened, this is because two set of segmentation The distance between linear function can be arbitrarily small.Fig. 2 is illustrated as K=10,、f(xi) andWhat can be reached connects Short range degree.Therefore, after a suitably K has been selected, the method for the present invention can obtain the approximate solution of four subproblems.
(5) test
In order to estimate the parameter in cooperation human behavior model and test its validity, tested, and analyze reality Test result.
Test general view
In an experiment, participant needs to play the part of the attacker for being carried out wildcat operation by Conservation areas at one piece, carries out manoeuvre. The panel of manoeuvre such as Fig. 3.
Protected area has been divided into 2 pieces of regions, and each region is divided into as 9 fritters, and each fritter represents a target. Manoeuvre needs 2 participants to simultaneously participate in, they are referred to as Alice and Bob.Alice and Bob can be with before manoeuvre starts Observe 18 all targets, income at each target when the defence probability of defender, attacker's independence success attack and Punishment during independent attack failure can all show them.As auxiliary, a temperature figure is provided to aid in participant's observation every The defence intensity of defender at a target.The color of target is deeper in figure, its defended probability is higher.In addition, also it is two A participant respectively provides a form, when form summarizes cooperation attack can obtainable income and its generation probability.Alice The target of half can only be all attacked with Bob, i.e. Alice can only attack 9, left side target, and Bob can only attack 9 targets in the right. Two participants need 1) from oneself one target of selection and can 2) to decide whether and another one participant closes in firing area Make.After two participants press attack (Attack) key, this attack terminates, and the result of attack can be illustrated in manoeuvre panel Portion, two respective remaining sums of participant can also be updated according to result.
Experiment setting
Participants, which need to play, amounts to 5 wheel manoeuvres, often takes turns manoeuvre and is made of several offices.In the first round, due to not yet Any attack data are collected into, we use Gholami, the Defending Policy for the method generation first round that S.et al. are proposed, with Under we term it " Maximin " methods.For the validity of evaluation method, the strategy for respectively generating " Maximin " method It is applied to the strategy of model proposed by the present invention generation in 5 wheel manoeuvres.Two groups of manoeuvres are provided with altogether, they have different Earnings structure S1 and S2, every group of manoeuvre tool correspond to the strategy that two methods generate respectively there are two 5 wheel manoeuvres.
The two kinds of earnings structure such as Fig. 4 (receipts of independence success attack at each target of digitized representation used in an experiment Benefit).Punishment at each target is set to -1.The extra returns ∈ of co-participant is set to 1.Total defence resource of defender M is set to 9, it means that the same time only has 9 targets can be defended.
Interpretation of result
In cooperation SSG, our main target is the cooperation destroyed between attacker.When Fig. 5 illustrates K=10, adopt It is generated with the tactful and proposed by the present invention model that " Maximin " method generates in the second wheel of the manoeuvre of earnings structure S1 Strategy (percentage that probability is defended at each target of digitized representation).As seen from Figure 5, the defence intensity that two attackers face has Significant difference.
For the incomplete reasonability of mankind attacker, we have collected data and are drilled according to often wheel (except last wheel) It practises data and has estimated model parameter.Here we listed in table 4 and table 5 from attack behavior learning to parameter.
Table 4
Table 5
In order to show validity of the model in terms of defender's average loss is reduced of the present invention, carry out in figure 6 pair Than.As seen from Figure 6, except the first round, the strategy that our model obtains in two kinds of earnings structures can provide in each round The significantly low defender loss than " Maximin " method.In addition, model of the invention is in the second wheel horse back initially applied With regard to defender's loss can be reduced.
Finally calculate cooperative rate, i.e., cooperation attack proportion in one wheel.Fig. 7 shows the experiment on two kinds of earnings structures As a result.When using the model of the present invention, the cooperative rate of the second wheel person of beginning participating in is decreased obviously and is kept low.With Upper result is consistent with the result in Fig. 6, and further illustrates the method for the present invention better than " Maximin " method.

Claims (7)

1. a kind of method of the modeling of mankind attacker cooperation behavior and the corresponding Defending Policy of formulation in security protection, feature exist In including following content:
(1) cooperative mechanism in SSG is repeated
Game is formed more by taking turns, and attackers need to make two selections in game:1) target of selection attack and 2) selection Whether cooperate with another attacker.Only when two attackers select cooperation in certain attack, a cooperative relationship It just calculates and establishes;In the case where cooperation is attacked, income that attacker attacks every time by be two people's income sums half;It introduces simultaneously One extra returns ∈, i.e., when an attacker has successfully carried out once cooperation attack, he will obtain extra returns ∈;
(2) incomplete reasonability and adaptivity
Incomplete reasonability for the mankind in SSG is repeated proposes a model;
Tendentiousness of the attacker for some target is defined first.
Define 1:One attacker is in R wheels to the tendentiousness of target iIt is defined by following equation:
<mrow> <msubsup> <mi>I</mi> <mi>i</mi> <mi>R</mi> </msubsup> <mo>=</mo> <mfrac> <mrow> <msubsup> <mo>&amp;Sigma;</mo> <mrow> <mi>r</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>R</mi> </msubsup> <msubsup> <mi>N</mi> <mi>i</mi> <mi>r</mi> </msubsup> </mrow> <mi>R</mi> </mfrac> </mrow>
<mrow> <msubsup> <mi>N</mi> <mi>i</mi> <mi>r</mi> </msubsup> <mo>=</mo> <mfrac> <msubsup> <mi>Q</mi> <mi>i</mi> <mi>r</mi> </msubsup> <mrow> <mi>C</mi> <mo>*</mo> <mi>V</mi> <mi>a</mi> <mi>r</mi> <mrow> <mo>(</mo> <msup> <mi>Q</mi> <mi>r</mi> </msup> <mo>)</mo> </mrow> </mrow> </mfrac> </mrow>
<mrow> <msubsup> <mi>Q</mi> <mi>i</mi> <mi>r</mi> </msubsup> <mo>=</mo> <msubsup> <mi>avg</mi> <mi>i</mi> <mi>r</mi> </msubsup> <mo>-</mo> <msubsup> <mi>avg</mi> <mrow> <mi>a</mi> <mi>l</mi> <mi>l</mi> </mrow> <mi>r</mi> </msubsup> </mrow>
WhereinIt isWithDifference,Represent being averaged for the single attack that attacker r wheels carry out at target i Income,Represent the single attack average yield for all attacks that attacker carries out in r wheels, QrFor a vector, I is tieed upC be a constant, Var (Qr) represent { QrVariance;
When attacker tends to target of attack iTo be just, on the contrary will be negative, and attacker be not revealed the tendentiousness of target It is set to 0.
Adaptive subjectivity revenue function:
<mrow> <msup> <mi>ASU</mi> <mi>R</mi> </msup> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mi>i</mi> </msub> <mo>)</mo> </mrow> <mo>=</mo> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <mi>d</mi> <mo>*</mo> <msubsup> <mi>I</mi> <mi>i</mi> <mi>R</mi> </msubsup> <mo>)</mo> </mrow> <msub> <mi>w</mi> <mn>1</mn> </msub> <msub> <mi>x</mi> <mi>i</mi> </msub> <mo>+</mo> <mrow> <mo>(</mo> <mn>1</mn> <mo>+</mo> <mi>d</mi> <mo>*</mo> <msubsup> <mi>I</mi> <mi>i</mi> <mi>R</mi> </msubsup> <mo>)</mo> </mrow> <msub> <mi>w</mi> <mn>2</mn> </msub> <msubsup> <mi>R</mi> <mi>i</mi> <mi>a</mi> </msubsup> <mo>+</mo> <mrow> <mo>(</mo> <mn>1</mn> <mo>+</mo> <mi>d</mi> <mo>*</mo> <msubsup> <mi>I</mi> <mi>i</mi> <mi>R</mi> </msubsup> <mo>)</mo> </mrow> <msub> <mi>w</mi> <mn>3</mn> </msub> <msubsup> <mi>P</mi> <mi>i</mi> <mi>a</mi> </msubsup> </mrow>
D=1/ (N are setr- r), wherein vector ω=(ω123) for parameter to be learned, NrAlways to take turns number, r representatives are used for Learn the round of ω;
(3) optimization problem provides
With reference to human behavior model and from the parameter that attack data learning obtains, the optimization for generating optimal Defending Policy is established Problem;
(4) optimal policy is calculated
Optimization problem is decomposed into 4 subproblems.By solving subproblem, and global optimum is selected in the optimal solution of subproblem Solution, just can solve former problem, obtain defence vector x.
2. the modeling of mankind attacker cooperation behavior and the corresponding Defending Policy of formulation in a kind of security protection as described in power requires 1 Method, which is characterized in that repeat cooperative game there are one defender and two attackers, the two attackers can select be No cooperation;Defender periodically arranges new Defending Policy as the leader in game, and attacker is as follower, according to It makes attack to the observation of the Defending Policy of defender's arrangement;This game is formed more by taking turns, so defender can be regular Defending Policy is changed according to the data being collected into round before;Attacker independently attacks or cooperation attack;Independently attacking In the case of hitting, whether income that attacker attacks every time is only dependent upon the attacker this attack and succeeds.
3. the modeling of mankind attacker cooperation behavior and the corresponding Defending Policy of formulation in a kind of security protection as described in power requires 1 Method, which is characterized in that the defence resource summation M possessed by defender come defend a target collection T=1 ..., | T |, T is by T1And T2Two are not overlapped subset and are combined into;T1Represent the target collection that first attacker may attack, T2Represent the The target collection that two attackers may attack, and T1=T-T2;Vector x is defended, per one-dimensional element xiDefender is represented to exist The defence resource of each target input namely the defended probability of target i.
4. the modeling of mankind attacker cooperation behavior and the corresponding Defending Policy of formulation in a kind of security protection as described in power requires 1 Method, which is characterized in that income of the attacker in target of attack i depend on 1) this target it is whether defended and 2) whether two attackers cooperate;When an attacker selects independent attack, the income that he attacks not defended target i isThe income for attacking defended target i isRelatively, defender is in the income of the previous caseIn latter feelings The income of condition isThe total revenue of defender for it in two incomes by target of attack and;AndIf two attackers agree to cooperate, they will obtain extra returns ∈ when attacking not defended target.
5. the modeling of mankind attacker cooperation behavior and the corresponding Defending Policy of formulation in a kind of security protection as described in power requires 1 Method, which is characterized in that with reference to human behavior model and from the obtained parameter of attack data learning, arrived by equation 1 Equation 19 establishes the optimization problem for generating optimal Defending Policy.
<mrow> <munder> <mi>max</mi> <mi>x</mi> </munder> <mi>D</mi> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>1</mn> <mo>)</mo> </mrow> </mrow>
D=d1+d2 (2)
α1, α2, β ∈ { 0,1 } (3)
<mrow> <munder> <mo>&amp;Sigma;</mo> <mrow> <mi>i</mi> <mo>&amp;Element;</mo> <msub> <mi>T</mi> <mn>1</mn> </msub> <mo>&amp;cup;</mo> <msub> <mi>T</mi> <mn>2</mn> </msub> </mrow> </munder> <msub> <mi>x</mi> <mi>i</mi> </msub> <mo>&amp;le;</mo> <mi>M</mi> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>4</mn> <mo>)</mo> </mrow> </mrow>
<mrow> <msub> <mi>x</mi> <mi>i</mi> </msub> <mo>&amp;Element;</mo> <mo>&amp;lsqb;</mo> <mn>0</mn> <mo>,</mo> <mn>1</mn> <mo>&amp;rsqb;</mo> <mo>,</mo> <mo>&amp;ForAll;</mo> <mi>i</mi> <mo>&amp;Element;</mo> <msub> <mi>T</mi> <mn>1</mn> </msub> <mo>&amp;cup;</mo> <msub> <mi>T</mi> <mn>2</mn> </msub> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>5</mn> <mo>)</mo> </mrow> </mrow>
<mrow> <msub> <mi>d</mi> <mn>1</mn> </msub> <mo>-</mo> <munder> <mo>&amp;Sigma;</mo> <mrow> <mi>i</mi> <mo>&amp;Element;</mo> <msub> <mi>T</mi> <mn>1</mn> </msub> </mrow> </munder> <msubsup> <mi>q</mi> <mi>i</mi> <mrow> <mi>n</mi> <mi>c</mi> </mrow> </msubsup> <mrow> <mo>(</mo> <mi>x</mi> <mo>)</mo> </mrow> <mo>&amp;lsqb;</mo> <msubsup> <mi>R</mi> <mi>i</mi> <mi>d</mi> </msubsup> <msub> <mi>x</mi> <mi>i</mi> </msub> <mo>+</mo> <msubsup> <mi>P</mi> <mi>i</mi> <mi>d</mi> </msubsup> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <msub> <mi>x</mi> <mi>i</mi> </msub> <mo>)</mo> </mrow> <mo>&amp;rsqb;</mo> <mo>&amp;le;</mo> <msub> <mi>&amp;alpha;</mi> <mn>1</mn> </msub> <mi>Z</mi> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>6</mn> <mo>)</mo> </mrow> </mrow>
<mrow> <msub> <mi>d</mi> <mn>1</mn> </msub> <mo>-</mo> <munder> <mo>&amp;Sigma;</mo> <mrow> <mi>i</mi> <mo>&amp;Element;</mo> <msub> <mi>T</mi> <mn>1</mn> </msub> </mrow> </munder> <msubsup> <mi>q</mi> <mi>i</mi> <mi>c</mi> </msubsup> <mrow> <mo>(</mo> <mi>x</mi> <mo>)</mo> </mrow> <mo>&amp;lsqb;</mo> <msubsup> <mi>R</mi> <mi>i</mi> <mi>d</mi> </msubsup> <msub> <mi>x</mi> <mi>i</mi> </msub> <mo>+</mo> <msubsup> <mi>P</mi> <mi>i</mi> <mi>d</mi> </msubsup> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <msub> <mi>x</mi> <mi>i</mi> </msub> <mo>)</mo> </mrow> <mo>&amp;rsqb;</mo> <mo>&amp;le;</mo> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <msub> <mi>&amp;alpha;</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> <mi>Z</mi> <mo>+</mo> <mi>&amp;beta;</mi> <mi>Z</mi> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>7</mn> <mo>)</mo> </mrow> </mrow>
<mrow> <msub> <mi>d</mi> <mn>1</mn> </msub> <mo>-</mo> <munder> <mo>&amp;Sigma;</mo> <mrow> <mi>i</mi> <mo>&amp;Element;</mo> <msub> <mi>T</mi> <mn>1</mn> </msub> </mrow> </munder> <msubsup> <mi>q</mi> <mi>i</mi> <mi>c</mi> </msubsup> <mrow> <mo>(</mo> <mi>x</mi> <mo>)</mo> </mrow> <mo>&amp;lsqb;</mo> <msubsup> <mi>R</mi> <mi>i</mi> <mi>d</mi> </msubsup> <msub> <mi>x</mi> <mi>i</mi> </msub> <mo>+</mo> <mrow> <mo>(</mo> <msubsup> <mi>P</mi> <mi>i</mi> <mi>d</mi> </msubsup> <mo>-</mo> <mo>&amp;Element;</mo> <mo>)</mo> </mrow> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <msub> <mi>x</mi> <mi>i</mi> </msub> <mo>)</mo> </mrow> <mo>&amp;rsqb;</mo> <mo>&amp;le;</mo> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <msub> <mi>&amp;alpha;</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> <mi>Z</mi> <mo>+</mo> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <mi>&amp;beta;</mi> <mo>)</mo> </mrow> <mi>Z</mi> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>8</mn> <mo>)</mo> </mrow> </mrow>
<mrow> <msub> <mi>d</mi> <mn>2</mn> </msub> <mo>-</mo> <munder> <mo>&amp;Sigma;</mo> <mrow> <mi>j</mi> <mo>&amp;Element;</mo> <msub> <mi>T</mi> <mn>2</mn> </msub> </mrow> </munder> <msubsup> <mi>q</mi> <mi>j</mi> <mrow> <mi>n</mi> <mi>c</mi> </mrow> </msubsup> <mrow> <mo>(</mo> <mi>x</mi> <mo>)</mo> </mrow> <mo>&amp;lsqb;</mo> <msubsup> <mi>R</mi> <mi>j</mi> <mi>d</mi> </msubsup> <msub> <mi>x</mi> <mi>j</mi> </msub> <mo>+</mo> <msubsup> <mi>P</mi> <mi>j</mi> <mi>d</mi> </msubsup> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <msub> <mi>x</mi> <mi>j</mi> </msub> <mo>)</mo> </mrow> <mo>&amp;rsqb;</mo> <mo>&amp;le;</mo> <msub> <mi>&amp;alpha;</mi> <mn>2</mn> </msub> <mi>Z</mi> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>9</mn> <mo>)</mo> </mrow> </mrow>
<mrow> <msub> <mi>d</mi> <mn>2</mn> </msub> <mo>-</mo> <munder> <mo>&amp;Sigma;</mo> <mrow> <mi>j</mi> <mo>&amp;Element;</mo> <msub> <mi>T</mi> <mn>2</mn> </msub> </mrow> </munder> <msubsup> <mi>q</mi> <mi>j</mi> <mi>c</mi> </msubsup> <mrow> <mo>(</mo> <mi>x</mi> <mo>)</mo> </mrow> <mo>&amp;lsqb;</mo> <msubsup> <mi>R</mi> <mi>j</mi> <mi>d</mi> </msubsup> <msub> <mi>x</mi> <mi>j</mi> </msub> <mo>+</mo> <msubsup> <mi>P</mi> <mi>j</mi> <mi>d</mi> </msubsup> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <msub> <mi>x</mi> <mi>j</mi> </msub> <mo>)</mo> </mrow> <mo>&amp;rsqb;</mo> <mo>&amp;le;</mo> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <msub> <mi>&amp;alpha;</mi> <mn>2</mn> </msub> <mo>)</mo> </mrow> <mi>Z</mi> <mo>+</mo> <mi>&amp;beta;</mi> <mi>Z</mi> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>10</mn> <mo>)</mo> </mrow> </mrow>
<mrow> <msub> <mi>d</mi> <mn>2</mn> </msub> <mo>-</mo> <munder> <mo>&amp;Sigma;</mo> <mrow> <mi>j</mi> <mo>&amp;Element;</mo> <msub> <mi>T</mi> <mn>2</mn> </msub> </mrow> </munder> <msubsup> <mi>q</mi> <mi>j</mi> <mi>c</mi> </msubsup> <mrow> <mo>(</mo> <mi>x</mi> <mo>)</mo> </mrow> <mo>&amp;lsqb;</mo> <msubsup> <mi>R</mi> <mi>j</mi> <mi>d</mi> </msubsup> <msub> <mi>x</mi> <mi>j</mi> </msub> <mo>+</mo> <mrow> <mo>(</mo> <msubsup> <mi>P</mi> <mi>j</mi> <mi>d</mi> </msubsup> <mo>-</mo> <mo>&amp;Element;</mo> <mo>)</mo> </mrow> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <msub> <mi>x</mi> <mi>j</mi> </msub> <mo>)</mo> </mrow> <mo>&amp;rsqb;</mo> <mo>&amp;le;</mo> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <msub> <mi>&amp;alpha;</mi> <mn>2</mn> </msub> <mo>)</mo> </mrow> <mi>Z</mi> <mo>+</mo> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <mi>&amp;beta;</mi> <mo>)</mo> </mrow> <mi>Z</mi> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>11</mn> <mo>)</mo> </mrow> </mrow>
<mrow> <mtable> <mtr> <mtd> <mrow> <msubsup> <mi>U</mi> <mi>a</mi> <mi>c</mi> </msubsup> <mo>=</mo> <mfrac> <mn>1</mn> <mn>2</mn> </mfrac> <mo>{</mo> <munder> <mo>&amp;Sigma;</mo> <mrow> <mi>i</mi> <mo>&amp;Element;</mo> <msub> <mi>T</mi> <mn>1</mn> </msub> </mrow> </munder> <msubsup> <mi>q</mi> <mi>i</mi> <mi>c</mi> </msubsup> <mrow> <mo>(</mo> <mi>x</mi> <mo>)</mo> </mrow> <mo>&amp;lsqb;</mo> <mrow> <mo>(</mo> <msubsup> <mi>R</mi> <mi>i</mi> <mi>a</mi> </msubsup> <mo>+</mo> <mo>&amp;Element;</mo> <mo>)</mo> </mrow> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <msub> <mi>x</mi> <mi>i</mi> </msub> <mo>)</mo> </mrow> <mo>+</mo> <msubsup> <mi>P</mi> <mi>i</mi> <mi>a</mi> </msubsup> <msub> <mi>x</mi> <mi>i</mi> </msub> <mo>&amp;rsqb;</mo> <mo>+</mo> </mrow> </mtd> </mtr> <mtr> <mtd> <mrow> <munder> <mo>&amp;Sigma;</mo> <mrow> <mi>j</mi> <mo>&amp;Element;</mo> <msub> <mi>T</mi> <mn>2</mn> </msub> </mrow> </munder> <msubsup> <mi>q</mi> <mi>j</mi> <mi>c</mi> </msubsup> <mrow> <mo>(</mo> <mi>x</mi> <mo>)</mo> </mrow> <mo>&amp;lsqb;</mo> <mrow> <mo>(</mo> <msubsup> <mi>R</mi> <mi>j</mi> <mi>a</mi> </msubsup> <mo>+</mo> <mo>&amp;Element;</mo> <mo>)</mo> </mrow> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <msub> <mi>x</mi> <mi>j</mi> </msub> <mo>)</mo> </mrow> <mo>+</mo> <msubsup> <mi>P</mi> <mi>j</mi> <mi>a</mi> </msubsup> <msub> <mi>x</mi> <mi>j</mi> </msub> <mo>&amp;rsqb;</mo> <mo>}</mo> </mrow> </mtd> </mtr> </mtable> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>12</mn> <mo>)</mo> </mrow> </mrow>
<mrow> <msubsup> <mi>U</mi> <mrow> <mi>a</mi> <mn>1</mn> </mrow> <mrow> <mi>n</mi> <mi>c</mi> </mrow> </msubsup> <mo>=</mo> <munder> <mo>&amp;Sigma;</mo> <mrow> <mi>i</mi> <mo>&amp;Element;</mo> <msub> <mi>T</mi> <mn>1</mn> </msub> </mrow> </munder> <msubsup> <mi>q</mi> <mi>i</mi> <mrow> <mi>n</mi> <mi>c</mi> </mrow> </msubsup> <mrow> <mo>(</mo> <mi>x</mi> <mo>)</mo> </mrow> <mo>&amp;lsqb;</mo> <msubsup> <mi>R</mi> <mi>i</mi> <mi>a</mi> </msubsup> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <msub> <mi>x</mi> <mi>i</mi> </msub> <mo>)</mo> </mrow> <mo>+</mo> <msubsup> <mi>P</mi> <mi>i</mi> <mi>a</mi> </msubsup> <msub> <mi>x</mi> <mi>i</mi> </msub> <mo>&amp;rsqb;</mo> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>13</mn> <mo>)</mo> </mrow> </mrow>
<mrow> <msubsup> <mi>U</mi> <mrow> <mi>a</mi> <mn>2</mn> </mrow> <mrow> <mi>n</mi> <mi>c</mi> </mrow> </msubsup> <mo>=</mo> <munder> <mo>&amp;Sigma;</mo> <mrow> <mi>j</mi> <mo>&amp;Element;</mo> <msub> <mi>T</mi> <mn>2</mn> </msub> </mrow> </munder> <msubsup> <mi>q</mi> <mi>j</mi> <mrow> <mi>n</mi> <mi>c</mi> </mrow> </msubsup> <mrow> <mo>(</mo> <mi>x</mi> <mo>)</mo> </mrow> <mo>&amp;lsqb;</mo> <msubsup> <mi>R</mi> <mi>j</mi> <mi>a</mi> </msubsup> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <msub> <mi>x</mi> <mi>j</mi> </msub> <mo>)</mo> </mrow> <mo>+</mo> <msubsup> <mi>P</mi> <mi>j</mi> <mi>a</mi> </msubsup> <msub> <mi>x</mi> <mi>j</mi> </msub> <mo>&amp;rsqb;</mo> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>14</mn> <mo>)</mo> </mrow> </mrow>
<mrow> <mo>-</mo> <msub> <mi>&amp;alpha;</mi> <mn>1</mn> </msub> <mi>Z</mi> <mo>&amp;le;</mo> <msubsup> <mi>U</mi> <mrow> <mi>a</mi> <mn>1</mn> </mrow> <mrow> <mi>n</mi> <mi>c</mi> </mrow> </msubsup> <mo>-</mo> <msubsup> <mi>U</mi> <mi>a</mi> <mi>c</mi> </msubsup> <mo>&amp;le;</mo> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <msub> <mi>&amp;alpha;</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> <mi>Z</mi> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>15</mn> <mo>)</mo> </mrow> </mrow>
<mrow> <mo>-</mo> <msub> <mi>&amp;alpha;</mi> <mn>2</mn> </msub> <mi>Z</mi> <mo>&amp;le;</mo> <msubsup> <mi>U</mi> <mrow> <mi>a</mi> <mn>2</mn> </mrow> <mrow> <mi>n</mi> <mi>c</mi> </mrow> </msubsup> <mo>-</mo> <msubsup> <mi>U</mi> <mi>a</mi> <mi>c</mi> </msubsup> <mo>&amp;le;</mo> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <msub> <mi>&amp;alpha;</mi> <mn>2</mn> </msub> <mo>)</mo> </mrow> <mi>Z</mi> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>16</mn> <mo>)</mo> </mrow> </mrow>
β≤α1 (17)
β≤α2 (18)
α12≤β+1 (19)
In all equatioies, Z represents a larger constant, and c represents cooperation, and nc represents uncooperative;
Formula 1 gives the expected revenus of the target of optimization problem, i.e. defender;From 2 visual target of formula by d1And d2Two parts structure Into;Three binary variables defined in formula 3, wherein α1And α2Attacker is represented each for the selection of cooperation, and β is represented most Whole cooperation state;Only when two attackers agree to cooperation, β 1, β is 0 in all other cases;This is by formula 17 Ensure with the constraints in 19;Formula 4 and 5 ensure that the every one-dimensional all between [0,1] of defence vector x, and defender's is total Resource is not more than M;
In formula 6 to 11, based on α1、α2Two parts of defender's expected revenus are defined with β --- d1And d2
Formula 12 to 14 respectively define attacker cooperate with it is uncooperative when expected revenus;Here two attackers are built respectively Mould, so they have different expected revenusesWithBy formula 15 and 16, α ensure that1And α2It is based on expected revenus It determines;As previously mentioned, three constraintss in formula 17 to 19 are for ensuring that correctly running for cooperative mechanism, that is, only have Cooperation is just reached when two attackers select cooperation.
6. the modeling of mankind attacker cooperation behavior and the corresponding Defending Policy of formulation in a kind of security protection as described in power requires 1 Method, which is characterized in that optimization problem is decomposed into 4 subproblems;By solving subproblem, and in the optimal solution of subproblem Middle selection globally optimal solution just can solve former problem, obtain defence vector x.
PROBLEM DECOMPOSITION
The angle of cooperation whether is selected from attacker, may be occurred there are four types of situation altogether;Following table summarize defender this four Expected revenus D in the case of kind1、D2、D3And D4
Table 3
Expected revenus in given table, just can correspondingly establish 4 subproblems.
7. the modeling of mankind attacker cooperation behavior and the corresponding Defending Policy of formulation in a kind of security protection as described in power requires 6 Method, which is characterized in that find the globally optimal solution of subproblem based on approximate method, introduce piecewise linear function, will about Beam condition is loosened, and subproblem is converted into a MIXED INTEGER quadratically constrained quadratic programming problem (MIQCQP) to solve;
Variable replacement is done first;Order That D4A quadric form can be rewritten as, sees formula 26;Because four new variables are introduced, it is necessary to according to definition to s and z Increase by four constraintss;Here only provide for spConstraint, others constraint can similarly provide;
<mrow> <msub> <mi>s</mi> <mi>p</mi> </msub> <munder> <mo>&amp;Sigma;</mo> <mrow> <mi>i</mi> <mo>&amp;Element;</mo> <msub> <mi>T</mi> <mn>1</mn> </msub> </mrow> </munder> <msup> <mi>e</mi> <mrow> <msubsup> <mi>ASU</mi> <mn>1</mn> <mi>c</mi> </msubsup> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mi>i</mi> </msub> <mo>)</mo> </mrow> </mrow> </msup> <mo>-</mo> <msup> <mi>e</mi> <mrow> <msubsup> <mi>ASU</mi> <mn>1</mn> <mi>c</mi> </msubsup> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mi>p</mi> </msub> <mo>)</mo> </mrow> </mrow> </msup> <mo>=</mo> <mn>0</mn> <mo>,</mo> <mo>&amp;ForAll;</mo> <mi>p</mi> <mo>&amp;Element;</mo> <msub> <mi>T</mi> <mn>1</mn> </msub> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>25</mn> <mo>)</mo> </mrow> </mrow>
Formula 25 is a nonlinear restriction, therefore it is replaced with two groups and slightly loosened to obtain quadratic constraints condition;Present explanation How such two set of segmentation linear functions are found.
OrderIt needs to find two groups of suitably piecewise linear functionsWithSo thatHere f (xi) on [0,1] it is a monotone decreasing convex function, because For negative;
AndUse xikDefinitionWith
<mrow> <msubsup> <mi>l</mi> <mi>i</mi> <mn>1</mn> </msubsup> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mi>i</mi> </msub> <mo>)</mo> </mrow> <mo>=</mo> <mn>1</mn> <mo>+</mo> <munderover> <mo>&amp;Sigma;</mo> <mrow> <mi>k</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>K</mi> </munderover> <msubsup> <mi>v</mi> <mrow> <mi>i</mi> <mi>k</mi> </mrow> <mrow> <mi>u</mi> <mi>p</mi> </mrow> </msubsup> <msub> <mi>x</mi> <mrow> <mi>i</mi> <mi>k</mi> </mrow> </msub> </mrow>
<mrow> <msubsup> <mi>l</mi> <mi>i</mi> <mn>2</mn> </msubsup> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mi>i</mi> </msub> <mo>)</mo> </mrow> <mo>=</mo> <msubsup> <mi>l</mi> <mi>i</mi> <mn>2</mn> </msubsup> <mrow> <mo>(</mo> <mn>0</mn> <mo>)</mo> </mrow> <mo>+</mo> <munderover> <mo>&amp;Sigma;</mo> <mrow> <mi>k</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>K</mi> </munderover> <msubsup> <mi>v</mi> <mrow> <mi>i</mi> <mi>k</mi> </mrow> <mrow> <mi>l</mi> <mi>o</mi> <mi>w</mi> </mrow> </msubsup> <msub> <mi>x</mi> <mrow> <mi>i</mi> <mi>k</mi> </mrow> </msub> </mrow>
WhereinIt representsThe slope of middle kth section broken line,It representsThe slope of middle kth section broken line;So in order to DefinitionWithOnly it needs to be determined thatWith
Theorem 1:According to piecewise linear functionWithMeetIt is full determining After the piecewise linear function required enough, the constraints in formula 25 is loosened;Specifically, formula 25 is replaced with following two groups of inequality It changes;
<mrow> <mi>C</mi> <mi>o</mi> <mi>n</mi> <mi>s</mi> <mrow> <mo>(</mo> <msub> <mi>s</mi> <mi>p</mi> </msub> <mo>)</mo> </mrow> <mo>:</mo> <mfenced open = "{" close = ""> <mtable> <mtr> <mtd> <mrow> <msub> <mi>s</mi> <mi>p</mi> </msub> <munder> <mo>&amp;Sigma;</mo> <mrow> <mi>i</mi> <mo>&amp;Element;</mo> <msub> <mi>T</mi> <mn>1</mn> </msub> </mrow> </munder> <msup> <mi>e</mi> <msubsup> <mi>b</mi> <mrow> <mn>1</mn> <mi>t</mi> </mrow> <mi>c</mi> </msubsup> </msup> <msubsup> <mi>l</mi> <mi>i</mi> <mn>1</mn> </msubsup> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mi>i</mi> </msub> <mo>)</mo> </mrow> <mo>-</mo> <msup> <mi>e</mi> <msubsup> <mi>b</mi> <mrow> <mn>1</mn> <mi>p</mi> </mrow> <mi>c</mi> </msubsup> </msup> <msubsup> <mi>l</mi> <mi>p</mi> <mn>2</mn> </msubsup> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mi>p</mi> </msub> <mo>)</mo> </mrow> <mo>&amp;GreaterEqual;</mo> <mn>0</mn> <mo>,</mo> </mrow> </mtd> <mtd> <mrow> <mo>&amp;ForAll;</mo> <mi>p</mi> <mo>&amp;Element;</mo> <msub> <mi>T</mi> <mn>1</mn> </msub> </mrow> </mtd> </mtr> <mtr> <mtd> <mrow> <msub> <mi>s</mi> <mi>p</mi> </msub> <munder> <mo>&amp;Sigma;</mo> <mrow> <mi>i</mi> <mo>&amp;Element;</mo> <msub> <mi>T</mi> <mn>1</mn> </msub> </mrow> </munder> <msup> <mi>e</mi> <msubsup> <mi>b</mi> <mrow> <mn>1</mn> <mi>t</mi> </mrow> <mi>c</mi> </msubsup> </msup> <msubsup> <mi>l</mi> <mi>i</mi> <mn>2</mn> </msubsup> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mi>i</mi> </msub> <mo>)</mo> </mrow> <mo>-</mo> <msup> <mi>e</mi> <msubsup> <mi>b</mi> <mrow> <mn>1</mn> <mi>p</mi> </mrow> <mi>c</mi> </msubsup> </msup> <msubsup> <mi>l</mi> <mi>p</mi> <mn>1</mn> </msubsup> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mi>p</mi> </msub> <mo>)</mo> </mrow> <mo>&amp;le;</mo> <mn>0</mn> <mo>,</mo> </mrow> </mtd> <mtd> <mrow> <mo>&amp;ForAll;</mo> <mi>p</mi> <mo>&amp;Element;</mo> <msub> <mi>T</mi> <mn>1</mn> </msub> </mrow> </mtd> </mtr> </mtable> </mfenced> </mrow>
This two groups of inequality are denoted as Cons (sp);By sr、zpAnd zrThe similar other inequality introduced are denoted as Cons respectively (sr)、Cons(zp) and Cons (zr);
Based on the modification above to subproblem, corresponding MIQCQP is established:
<mrow> <munder> <mi>max</mi> <mrow> <mi>x</mi> <mo>,</mo> <mi>s</mi> <mo>,</mo> <mi>z</mi> </mrow> </munder> <munder> <mo>&amp;Sigma;</mo> <mrow> <mi>i</mi> <mo>&amp;Element;</mo> <msub> <mi>T</mi> <mn>1</mn> </msub> </mrow> </munder> <msub> <mi>s</mi> <mi>i</mi> </msub> <mo>&amp;lsqb;</mo> <msubsup> <mi>R</mi> <mi>i</mi> <mi>d</mi> </msubsup> <msub> <mi>x</mi> <mi>i</mi> </msub> <mo>+</mo> <msubsup> <mi>P</mi> <mi>i</mi> <mi>d</mi> </msubsup> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <msub> <mi>x</mi> <mi>i</mi> </msub> <mo>)</mo> </mrow> <mo>&amp;rsqb;</mo> <mo>+</mo> <munder> <mo>&amp;Sigma;</mo> <mrow> <mi>j</mi> <mo>&amp;Element;</mo> <msub> <mi>T</mi> <mn>2</mn> </msub> </mrow> </munder> <msub> <mi>z</mi> <mi>j</mi> </msub> <mo>&amp;lsqb;</mo> <msubsup> <mi>R</mi> <mi>j</mi> <mi>d</mi> </msubsup> <msub> <mi>x</mi> <mi>j</mi> </msub> <mo>+</mo> <msubsup> <mi>P</mi> <mi>j</mi> <mi>d</mi> </msubsup> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <msub> <mi>x</mi> <mi>j</mi> </msub> <mo>)</mo> </mrow> <mo>&amp;rsqb;</mo> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>26</mn> <mo>)</mo> </mrow> </mrow>
<mrow> <munder> <mo>&amp;Sigma;</mo> <mrow> <mi>i</mi> <mo>&amp;Element;</mo> <msub> <mi>T</mi> <mn>1</mn> </msub> <mo>&amp;cup;</mo> <msub> <mi>T</mi> <mn>2</mn> </msub> </mrow> </munder> <msub> <mi>x</mi> <mi>i</mi> </msub> <mo>&amp;le;</mo> <mi>M</mi> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>27</mn> <mo>)</mo> </mrow> </mrow>
<mrow> <mtable> <mtr> <mtd> <mrow> <munder> <mo>&amp;Sigma;</mo> <mrow> <mi>i</mi> <mo>&amp;Element;</mo> <msub> <mi>T</mi> <mn>1</mn> </msub> </mrow> </munder> <msub> <mi>z</mi> <mi>i</mi> </msub> <mo>&amp;lsqb;</mo> <msubsup> <mi>R</mi> <mi>i</mi> <mi>a</mi> </msubsup> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <msub> <mi>x</mi> <mi>i</mi> </msub> <mo>)</mo> </mrow> <mo>+</mo> <msubsup> <mi>P</mi> <mi>i</mi> <mi>a</mi> </msubsup> <msub> <mi>x</mi> <mi>i</mi> </msub> <mo>&amp;rsqb;</mo> <mo>&amp;le;</mo> <mfrac> <mn>1</mn> <mn>2</mn> </mfrac> <mo>{</mo> <munder> <mo>&amp;Sigma;</mo> <mrow> <mi>i</mi> <mo>&amp;Element;</mo> <msub> <mi>T</mi> <mn>1</mn> </msub> </mrow> </munder> <msub> <mi>s</mi> <mi>i</mi> </msub> <mo>&amp;lsqb;</mo> <mrow> <mo>(</mo> <msubsup> <mi>R</mi> <mi>i</mi> <mi>a</mi> </msubsup> <mo>+</mo> <mo>&amp;Element;</mo> <mo>)</mo> </mrow> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <msub> <mi>x</mi> <mi>i</mi> </msub> <mo>)</mo> </mrow> <mo>+</mo> <msubsup> <mi>P</mi> <mi>i</mi> <mi>a</mi> </msubsup> <msub> <mi>x</mi> <mi>i</mi> </msub> <mo>&amp;rsqb;</mo> </mrow> </mtd> </mtr> <mtr> <mtd> <mrow> <mo>+</mo> <munder> <mo>&amp;Sigma;</mo> <mrow> <mi>j</mi> <mo>&amp;Element;</mo> <msub> <mi>T</mi> <mn>2</mn> </msub> </mrow> </munder> <msub> <mi>s</mi> <mi>j</mi> </msub> <mo>&amp;lsqb;</mo> <mrow> <mo>(</mo> <msubsup> <mi>R</mi> <mi>j</mi> <mi>a</mi> </msubsup> <mo>+</mo> <mo>&amp;Element;</mo> <mo>)</mo> </mrow> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <msub> <mi>x</mi> <mi>j</mi> </msub> <mo>)</mo> </mrow> <mo>+</mo> <msubsup> <mi>P</mi> <mi>j</mi> <mi>a</mi> </msubsup> <msub> <mi>x</mi> <mi>j</mi> </msub> <mo>&amp;rsqb;</mo> <mo>}</mo> </mrow> </mtd> </mtr> </mtable> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>28</mn> <mo>)</mo> </mrow> </mrow>
<mrow> <mtable> <mtr> <mtd> <mrow> <munder> <mo>&amp;Sigma;</mo> <mrow> <mi>j</mi> <mo>&amp;Element;</mo> <msub> <mi>T</mi> <mn>2</mn> </msub> </mrow> </munder> <msub> <mi>z</mi> <mi>j</mi> </msub> <mo>&amp;lsqb;</mo> <msubsup> <mi>R</mi> <mi>j</mi> <mi>a</mi> </msubsup> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <msub> <mi>x</mi> <mi>j</mi> </msub> <mo>)</mo> </mrow> <mo>+</mo> <msubsup> <mi>P</mi> <mi>j</mi> <mi>a</mi> </msubsup> <msub> <mi>x</mi> <mi>j</mi> </msub> <mo>&amp;rsqb;</mo> <mo>&amp;GreaterEqual;</mo> <mfrac> <mn>1</mn> <mn>2</mn> </mfrac> <mo>{</mo> <munder> <mo>&amp;Sigma;</mo> <mrow> <mi>i</mi> <mo>&amp;Element;</mo> <msub> <mi>T</mi> <mn>1</mn> </msub> </mrow> </munder> <msub> <mi>s</mi> <mi>i</mi> </msub> <mo>&amp;lsqb;</mo> <mrow> <mo>(</mo> <msubsup> <mi>R</mi> <mi>i</mi> <mi>a</mi> </msubsup> <mo>+</mo> <mo>&amp;Element;</mo> <mo>)</mo> </mrow> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <msub> <mi>x</mi> <mi>i</mi> </msub> <mo>)</mo> </mrow> <mo>+</mo> <msubsup> <mi>P</mi> <mi>i</mi> <mi>a</mi> </msubsup> <msub> <mi>x</mi> <mi>i</mi> </msub> <mo>&amp;rsqb;</mo> </mrow> </mtd> </mtr> <mtr> <mtd> <mrow> <mo>+</mo> <munder> <mo>&amp;Sigma;</mo> <mrow> <mi>j</mi> <mo>&amp;Element;</mo> <msub> <mi>T</mi> <mn>2</mn> </msub> </mrow> </munder> <msub> <mi>s</mi> <mi>j</mi> </msub> <mo>&amp;lsqb;</mo> <mrow> <mo>(</mo> <msubsup> <mi>R</mi> <mi>j</mi> <mi>a</mi> </msubsup> <mo>+</mo> <mo>&amp;Element;</mo> <mo>)</mo> </mrow> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <msub> <mi>x</mi> <mi>j</mi> </msub> <mo>)</mo> </mrow> <mo>+</mo> <msubsup> <mi>P</mi> <mi>j</mi> <mi>a</mi> </msubsup> <msub> <mi>x</mi> <mi>j</mi> </msub> <mo>&amp;rsqb;</mo> <mo>}</mo> </mrow> </mtd> </mtr> </mtable> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>29</mn> <mo>)</mo> </mrow> </mrow>
Cons(sp), Cons (sr), Cons (zp), Cons (zr) (30)
<mrow> <msub> <mi>x</mi> <mi>t</mi> </msub> <mo>=</mo> <munderover> <mo>&amp;Sigma;</mo> <mrow> <mi>k</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>K</mi> </munderover> <msub> <mi>x</mi> <mrow> <mi>t</mi> <mi>k</mi> </mrow> </msub> <mo>,</mo> <mo>&amp;ForAll;</mo> <mi>t</mi> <mo>&amp;Element;</mo> <msub> <mi>T</mi> <mn>1</mn> </msub> <mo>&amp;cup;</mo> <msub> <mi>T</mi> <mn>2</mn> </msub> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>31</mn> <mo>)</mo> </mrow> </mrow>
<mrow> <mn>0</mn> <mo>&amp;le;</mo> <msub> <mi>x</mi> <mrow> <mi>t</mi> <mi>k</mi> </mrow> </msub> <mo>&amp;le;</mo> <mfrac> <mn>1</mn> <mi>K</mi> </mfrac> <mo>,</mo> <mo>&amp;ForAll;</mo> <mi>t</mi> <mo>&amp;Element;</mo> <msub> <mi>T</mi> <mn>1</mn> </msub> <mo>&amp;cup;</mo> <msub> <mi>T</mi> <mn>2</mn> </msub> <mo>,</mo> <mi>k</mi> <mo>=</mo> <mn>1</mn> <mo>,</mo> <mn>...</mn> <mo>,</mo> <mi>K</mi> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>32</mn> <mo>)</mo> </mrow> </mrow>
<mrow> <msub> <mi>y</mi> <mrow> <mi>t</mi> <mi>k</mi> </mrow> </msub> <mo>&amp;Element;</mo> <mo>{</mo> <mn>0</mn> <mo>,</mo> <mn>1</mn> <mo>}</mo> <mo>,</mo> <mo>&amp;ForAll;</mo> <mi>t</mi> <mo>&amp;Element;</mo> <msub> <mi>T</mi> <mn>1</mn> </msub> <mo>&amp;cup;</mo> <msub> <mi>T</mi> <mn>2</mn> </msub> <mo>,</mo> <mi>k</mi> <mo>=</mo> <mn>1</mn> <mo>,</mo> <mn>...</mn> <mo>,</mo> <mi>K</mi> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>33</mn> <mo>)</mo> </mrow> </mrow>
<mrow> <msub> <mi>y</mi> <mrow> <mi>t</mi> <mi>k</mi> </mrow> </msub> <mfrac> <mn>1</mn> <mi>K</mi> </mfrac> <mo>&amp;le;</mo> <msub> <mi>x</mi> <mrow> <mi>t</mi> <mi>k</mi> </mrow> </msub> <mo>,</mo> <mo>&amp;ForAll;</mo> <mi>t</mi> <mo>&amp;Element;</mo> <msub> <mi>T</mi> <mn>1</mn> </msub> <mo>&amp;cup;</mo> <msub> <mi>T</mi> <mn>2</mn> </msub> <mo>,</mo> <mi>k</mi> <mo>=</mo> <mn>1</mn> <mo>,</mo> <mn>...</mn> <mo>,</mo> <mi>K</mi> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>34</mn> <mo>)</mo> </mrow> </mrow>
<mrow> <msub> <mi>x</mi> <mrow> <mi>t</mi> <mo>,</mo> <mi>k</mi> <mo>+</mo> <mn>1</mn> </mrow> </msub> <mo>&amp;le;</mo> <msub> <mi>y</mi> <mrow> <mi>t</mi> <mi>k</mi> </mrow> </msub> <mo>,</mo> <mo>&amp;ForAll;</mo> <mi>t</mi> <mo>&amp;Element;</mo> <msub> <mi>T</mi> <mn>1</mn> </msub> <mo>&amp;cup;</mo> <msub> <mi>T</mi> <mn>2</mn> </msub> <mo>,</mo> <mi>k</mi> <mo>=</mo> <mn>1</mn> <mo>,</mo> <mn>...</mn> <mo>,</mo> <mi>K</mi> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>35</mn> <mo>)</mo> </mrow> </mrow>
Subproblem is rewritten with the variable newly introduced in formula 26 to 29;Formula 30 represents the quadratic constraints of generation;Formula 31 and 32 is To xtkConstraint;
MIQCQP is solved using the SCIP methods of Vigerske S.et al. propositions;
Theorem 2:NoteForWithThe distance between, then limK→+∞Dis= 0。
CN201711324273.1A 2017-12-13 2017-12-13 The method that mankind attacker cooperation behavior modeled and formulated corresponding Defending Policy Pending CN108053068A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201711324273.1A CN108053068A (en) 2017-12-13 2017-12-13 The method that mankind attacker cooperation behavior modeled and formulated corresponding Defending Policy

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201711324273.1A CN108053068A (en) 2017-12-13 2017-12-13 The method that mankind attacker cooperation behavior modeled and formulated corresponding Defending Policy

Publications (1)

Publication Number Publication Date
CN108053068A true CN108053068A (en) 2018-05-18

Family

ID=62132545

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201711324273.1A Pending CN108053068A (en) 2017-12-13 2017-12-13 The method that mankind attacker cooperation behavior modeled and formulated corresponding Defending Policy

Country Status (1)

Country Link
CN (1) CN108053068A (en)

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103152345A (en) * 2013-03-07 2013-06-12 南京理工大学常熟研究院有限公司 Network safety optimum attacking and defending decision method for attacking and defending game
CN106921666A (en) * 2017-03-06 2017-07-04 中山大学 A kind of ddos attack system of defense and method based on Synergy

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103152345A (en) * 2013-03-07 2013-06-12 南京理工大学常熟研究院有限公司 Network safety optimum attacking and defending decision method for attacking and defending game
CN106921666A (en) * 2017-03-06 2017-07-04 中山大学 A kind of ddos attack system of defense and method based on Synergy

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
BINRU WANG等: ""On Repeated Stackelberg Security Game with the Cooperative Human Behavior Model for Wildlife Protection"", 《PROCEEDINGS OF THE 16TH CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS》 *
姜伟: "" 基于攻防博弈模型的主动防御关键技术研究"", 《中国博士学位论文全文数据库 信息科技辑》 *
赵婧等: ""基于轨迹频率抑制的轨迹隐私保护方法"", 《计算机学报》 *

Similar Documents

Publication Publication Date Title
Kolkowicz et al. Soldiers, peasants, and bureaucrats: civil-military relations in Communist and modernizing societies
CN106682351A (en) Fight simulation system generating military strength based on computer and simulation method
Montague Stolen goods: Coltan and conflict in the Democratic Republic of Congo
CN109426567A (en) A kind of node deployment and electoral machinery of block chain
CN110166428A (en) Intelligence defence decision-making technique and device based on intensified learning and attacking and defending game
CN110035066B (en) Attack and defense behavior quantitative evaluation method and system based on game theory
CN112883874A (en) Active defense method aiming at deep face tampering
Kullberg et al. Elites, institutions and democratisation in Russia and Eastern Europe
CN108053068A (en) The method that mankind attacker cooperation behavior modeled and formulated corresponding Defending Policy
Russell The Persian Gulf's collective-security mirage
Avdagic Accounting for variations in trade union effectiveness: State-labor relations in East Central Europe
CN115983389A (en) Attack and defense game decision method based on reinforcement learning
DeLong et al. Teamskill evolved: Mixed classification schemes for team-based multi-player games
Ahmad et al. Changing fortunes in anti‐poverty programmes in Bangladesh
CN113792331A (en) Fair-resistance interference-based federated learning member reasoning attack defense method
CN114662655A (en) Attention mechanism-based weapon and chess deduction AI hierarchical decision method and device
Mwangi Mozambique's Foreign Policy: From Ideological Conflict to Pragmatic Cooperation
Robinson Authoritarianism with a Palestinian face
Yamadera Examining the myth of money with agent-based modelling
Atkinson et al. On-the-fly generation of multi-robot team formation strategies based on game conditions
Martinez-Mejorado et al. A multidimensional framework to uncover insights of group performance and outcomes in competitive environments with a case study of FIFA World Cups
Grieshop et al. A Continuous-Time Stochastic Process for High-Resolution Network Data in Sports
Yang et al. Modeling human bounded rationality to improve defender strategies in network security games
Grossmann et al. Viability and sustainability of civilizations, corporations, institutions and ecological systems
Van Schendelen The Netherlands: Parliamentary parties rival with pressure groups

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20180518