CN106953879A - The cyber-defence strategy choosing method of best response dynamics Evolutionary Game Model - Google Patents
The cyber-defence strategy choosing method of best response dynamics Evolutionary Game Model Download PDFInfo
- Publication number
- CN106953879A CN106953879A CN201710335128.7A CN201710335128A CN106953879A CN 106953879 A CN106953879 A CN 106953879A CN 201710335128 A CN201710335128 A CN 201710335128A CN 106953879 A CN106953879 A CN 106953879A
- Authority
- CN
- China
- Prior art keywords
- defender
- strategy
- defence
- best response
- response dynamics
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L63/00—Network architectures or network communication protocols for network security
- H04L63/20—Network architectures or network communication protocols for network security for managing network security; network security policies in general
- H04L63/205—Network architectures or network communication protocols for network security for managing network security; network security policies in general involving negotiation or determination of the one or more network security mechanisms to be used, e.g. by negotiation between the client and the server or between peers or by selection according to the capabilities of the entities involved
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L63/00—Network architectures or network communication protocols for network security
- H04L63/14—Network architectures or network communication protocols for network security for detecting or protecting against malicious traffic
- H04L63/1441—Countermeasures against malicious traffic
Landscapes
- Engineering & Computer Science (AREA)
- Computer Security & Cryptography (AREA)
- Computer Hardware Design (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- Computer Networks & Wireless Communication (AREA)
- Signal Processing (AREA)
- Management, Administration, Business Operations System, And Electronic Commerce (AREA)
Abstract
The invention belongs to computer network security defense technique field, being specifically related to a kind of cyber-defence strategy choosing method of best response dynamics Evolutionary Game Model includes:Based on bounded rationality condition, using best response dynamics study mechanism, the attacking and defending Evolutionary Game Model based on best response dynamics is built;Dynamic Evolution and defence Evolutionary Equilibrium point are chosen using defender's strategy, defence policies On The Choice between different defenders is studied;On the basis of the best response dynamics Evolutionary Game Model of foundation, the model is analyzed and solved by specific example, promote Evolutionary Game Model.The present invention establishes the non-cooperative network attacking and defending Evolutionary Game Model under the conditions of bounded rationality, original state is chosen by arranging defender's strategy, by constantly evolution, best response dynamics will eventually make game playing system reach some stable state, so as to obtain optimal defence policies, method proposed by the present invention can be good at being applied to network security defence policies On The Choice, can provide network security research certain directive significance.
Description
Technical field
The invention belongs to computer network security defense technique field, a kind of best response dynamics evolution is specifically related to rich
Play chess the cyber-defence strategy choosing method of model.
Background technology
In recent years, the social life that the fast development of internet gives people brings huge change, particularly " internet
+ " strategy push the development of internet to a new climax.With the fast development of internet, cyberspace safety problem
Become increasingly conspicuous.Network security problem is very severe, and for local and overseas disparate networks attacks, how Strengthens network is pacified
The problem of full defence turns into current era urgent need to resolve, needing badly can be analyzed and be predicted to network-combination yarn behavior, Jin Ershi
Alms giver moves the new technology of Prevention-Security.Because network safe state is determined by the agonistic behavior and its result of attacking and defending both sides in itself
It is fixed, and target antagonism, tactful interdependence and the relation Non-synergic exactly game theory having in network-combination yarn confrontation
Essential characteristic, therefore game theory increasingly rises in the research and application of network safety filed, and with using classical traditional game
Based on model is analyzed network security behavior.
But, existing achievement in research sets up what is assumed in participant's rational mostly based on traditional game is theoretical
Under the premise of, and such hypothesis is not consistent with actual conditions.Its betting model and real deviation are larger, so as to reduce safety
The accuracy and directive significance of defence policies choosing method.For problem above, some scholars are used premised on bounded rationality
Evolutionary game theory is analyzed applied to network-combination yarn.By analysis, evolutionary Game more conforms to network-combination yarn confrontation dynamic evolution
Reality, turns to the gradual evolution process with certain adaptability learning ability, using typical by the behavior model of attacking and defending both sides
Replicator dynamics equation is solved and analyzed.But it is low that replica locating study mechanism has that pace of learning is slow, strategy chooses efficiency
Problem.
The content of the invention
The present invention mostly based on traditional game is theoretical, sets up false in participant's rational for existing achievement in research
If on the premise of, and such hypothesis is not consistent with actual conditions, is drilled if existing and being directly applied to network-combination yarn confrontation
Change game theory analysis, it will have that learning cycle is long, learning efficiency is not high, this is suitable by largely reduction model and method
With sex chromosome mosaicism, a kind of cyber-defence strategy choosing method of best response dynamics Evolutionary Game Model is proposed.
The technical scheme is that:A kind of cyber-defence strategy selection side of best response dynamics Evolutionary Game Model
Method, comprises the following steps:
Step 1:Based on bounded rationality condition, using best response dynamics study mechanism, build and be based on best response dynamics
Attacking and defending Evolutionary Game Model;
Step 2:Dynamic Evolution and defence Evolutionary Equilibrium point are chosen using defender's strategy, between different defenders
Defence policies On The Choice is studied;
Step 3:On the basis of the best response dynamics Evolutionary Game Model of foundation, the model is entered by specific example
Row analysis promotes Evolutionary Game Model with solving.
It is optimal in the cyber-defence strategy choosing method of described best response dynamics Evolutionary Game Model, the step 1
Reaction dynamic evolution betting model is represented by four-tuple, BRDEGM=(D, DS, P, U)
D={ d1,d2,…dnDefence participant space is represented, wherein, diRepresent defender i, different defender can be with
Choose different defence policies;
DS={ DS1,DS2,…DSmDefender's policy space is represented, different defenders enjoy the defence policies jointly
Collection;
P={ p1,p2,…pmDefender's conviction set is represented, wherein, piRepresent that defender chooses defence policies DSiIt is general
Rate;
U={ U1,U2,…UmRevenue function set is represented, wherein, UiRepresent that defender chooses defence policies DSiIt is acquired
Income.
The cyber-defence strategy choosing method of described best response dynamics Evolutionary Game Model, the best response dynamics
Equation isWherein NtRepresent Selection Strategy DS in n defender1Number, DS1It is optional
Any one defence policies in set of strategies.
Defendd in the cyber-defence strategy choosing method of described best response dynamics Evolutionary Game Model, the step 2
Strategy slightly chooses Dynamic Evolution:There is a kind of competition in network-combination yarn antagonistic process, between different defence policies to close
System, the defence policies of high yield will eliminate the relatively low strategy of income.
The cyber-defence strategy choosing method of described best response dynamics Evolutionary Game Model, the defence of the high yield
Strategy will be eliminated in the relatively low strategy of income, and gain matrix is:Wherein, u1、u2Point
Wei not strategy DS1、DS2Income, a is u1、u2Difference.
Promoted in the cyber-defence strategy choosing method of described best response dynamics Evolutionary Game Model, the step 3
Evolutionary Game Model is mainly, when defender has any n defender, based on best response dynamics Evolutionary Game Model,
To any two defence policies DSiAnd DSjCarry out evolutionary Game Analysis, it is assumed that DSiIt is relative to DSjDominating stragegy, and i ≠
J, over time, finally gives certain Evolution.
The cyber-defence strategy choosing method of described best response dynamics Evolutionary Game Model, the Evolution is:
For there is the defender of n defender, when all defenders choose defence policies DS in first gameiOr strategy DSj
When, using best response dynamics study mechanism, it is then institute that the strategy of whole network defender, which chooses the stable state being finally reached,
Some equal Selection Strategy DS of defenderiOr strategy DSj。
The cyber-defence strategy choosing method of described best response dynamics Evolutionary Game Model, the Evolution is:
For there is the defender of n defender, when n is odd number, in first game, as long as there is a defender to have chosen strategy
DSi, institute finally can all be converged on by the adjustment repeatedly in multiple periods to itself strategy by best response dynamics study mechanism
There is defender's Selection Strategy DSiStable state.
The cyber-defence strategy choosing method of described best response dynamics Evolutionary Game Model, the Evolution is:
For there is the defender of n defender, when n is even number, in first game, there is a defender to have chosen defence policies
DSi, the equal Selection Strategy DS of other defendersj, then, best response dynamics can not make all defenders converge on some stable shape
State, evolution over time, adjustment of each defender to strategy can only be absorbed in loop cycle and change.
The cyber-defence strategy choosing method of described best response dynamics Evolutionary Game Model, the Evolution is:
For there is the defender of n defender, in first game, strategy is have chosen simultaneously simply by the presence of two adjacent defenders
DSi, under best response dynamics study mechanism, evolution over time eventually converges on all defenders and all chooses plan
Slightly DSiStable state.
The beneficial effects of the invention are as follows:The present invention establishes the non-cooperative network attacking and defending evolutionary Game under the conditions of bounded rationality
Model, and computable general equilibrium has been carried out with solving to the model.On this basis, from defender's angle, for different defence
Policy learning adjustment process between person, using best response dynamics study mechanism, establishes the multistage weight between defender
Double action state Evolutionary Game Model, is studied defence policies On The Choice between different defenders.In the optimal anti-of foundation
Answer on the basis of dynamic evolution betting model, the model is analyzed and solved by specific example, and the model is made
Further genralrlization, improves the versatility of model.Chosen just for the difference of defender's number parity, and defender's strategy
The difference of beginning state, can all influence the final evolution result of whole game playing system.Initial shape is chosen by arranging defender's strategy
State, by constantly evolution, best response dynamics will eventually make game playing system reach some stable state, so as to obtain optimal anti-
Imperial strategy.Illustrate that method proposed by the present invention can be good at being applied to network security defence policies On The Choice, network is pacified
Full research can provide certain directive significance.
Brief description of the drawings
Fig. 1 method of the present invention step schematic block diagrams;
Fig. 2 cyber-defence person's game theory schematic diagrames;
1 DS of the first games of Fig. 31Best response dynamics schematic diagram;
Two DS of the first games of Fig. 41Best response dynamics Developing Tactics process schematic;
Three DS of the first games of Fig. 51Best response dynamics Developing Tactics process schematic;
Fig. 6 is odd number as n, and original state only one of which selects DS1Simulated effect schematic diagram;
Fig. 7 is odd number as n, and original state selects DS in the presence of two adjacent defenders1Simulated effect schematic diagram.
Embodiment
Embodiment 1, with reference to Fig. 1-Fig. 7, a kind of cyber-defence strategy selection side of best response dynamics Evolutionary Game Model
Method, comprises the following steps:
Step 1:Based on bounded rationality condition, using best response dynamics study mechanism, build and be based on best response dynamics
Attacking and defending Evolutionary Game Model;
Best response dynamics Evolutionary Game Model is represented by four-tuple, BRDEGM=(D, DS, P, U) in step 1
D={ d1,d2,…dnDefence participant space is represented, wherein, diRepresent defender i, different defender can be with
Choose different defence policies;
DS={ DS1,DS2,…DSmDefender's policy space is represented, different defenders enjoy the defence policies jointly
Collection;
P={ p1,p2,…pmDefender's conviction set is represented, wherein, piRepresent that defender chooses defence policies DSiIt is general
Rate;
U={ U1,U2,…UmRevenue function set is represented, wherein, UiRepresent that defender chooses defence policies DSiIt is acquired
Income.
Best response dynamics equation isWherein NtRepresent to choose plan in n defender
Slightly DS1Number, DS1It is any one defence policies in optional set of strategies.
Step 2:Dynamic Evolution and defence Evolutionary Equilibrium point are chosen using defender's strategy, between different defenders
Defence policies On The Choice is studied;
Defender's strategy selection Dynamic Evolution is in step 2:In network-combination yarn antagonistic process, different defence policies
Between there is a kind of competitive relation, the defence policies of high yield will eliminate the relatively low strategy of income;The defence plan of high yield
Summary will be eliminated in the relatively low strategy of income, and gain matrix is:Wherein, u1、u2Respectively
For tactful DS1、DS2Income, a is u1、u2Difference.
Step 3:On the basis of the best response dynamics Evolutionary Game Model of foundation, the model is entered by specific example
Row analysis promotes Evolutionary Game Model with solving.
Evolutionary Game Model is promoted in step 3 is mainly, when defender has any n defender, based on optimal anti-
Dynamic evolution betting model is answered, to any two defence policies DSiAnd DSjCarry out evolutionary Game Analysis, it is assumed that DSiBe relative to
DSjDominating stragegy, and i ≠ j over time, finally gives certain Evolution.
Further:Evolution is:For there is the defender of n defender, when all defenders are in first game
In all choose defence policies DSiOr strategy DSjWhen, using best response dynamics study mechanism, the strategy of whole network defender
It is then all equal Selection Strategy DS of defender to choose the stable state being finally reachediOr strategy DSj。
Further, Evolution is:For there is the defender of n defender, when n is odd number, in first game
In, as long as there is a defender to have chosen tactful DSi, when passing through multiple to itself strategy by best response dynamics study mechanism
The adjustment repeatedly of phase, finally can all converge on all defender's Selection Strategy DSiStable state.
Further, Evolution is:For there is the defender of n defender, when n is even number, in first game
In, there is a defender to have chosen defence policies DSi, the equal Selection Strategy DS of other defendersj, then, best response dynamics can not
All defenders are made to converge on some stable state, evolution over time, adjustment of each defender to strategy can only be absorbed in week
Phase cyclical variations.
Further, Evolution is:For there is the defender of n defender, in first game, simply by the presence of two
Individual adjacent defender have chosen tactful DS simultaneouslyi, under best response dynamics study mechanism, evolution over time, eventually
Converge on the whole Selection Strategy DS of all defendersiStable state.
Embodiment 2, with reference to Fig. 1-Fig. 7, a kind of cyber-defence strategy selection side of best response dynamics Evolutionary Game Model
Method, for analyzing network-combination yarn evolutionary Game process.Due to conventional replica locating study mechanism, to there is pace of learning slower, learns
The problems such as practising inefficient, the present invention still uses the thought of evolutionary Game, dynamic using peak optimization reaction based on bounded rationality condition
State study mechanism, builds the attacking and defending Evolutionary Game Model based on best response dynamics, and analysis defender's strategy chooses dynamic evolution
Process and defence Evolutionary Equilibrium point, are studied defence policies On The Choice between different defenders.In the optimal of foundation
React on the basis of dynamic evolution betting model, the model is analyzed and solved by specific example, and by the model
Make further genralrlization, improve the versatility of model.Method proposed by the present invention can be good at being applied to network security defence
Tactful On The Choice, can provide network security research certain directive significance.
In network-combination yarn game playing system, policymaker by continuous trial and error, imitation and Developing Tactics, from original state with
Time constantly develops, and can finally reach that some evolutionarily stable is balanced, and the direct shadow of policy learning method and process of policymaker
Ring and arrive final evolutionarily stable state.For defender, it is assumed that different defenders shares same defence policies collection.Due to not
Same defence policies can bring different incomes to defender, under the traction of yield variance and the driving of study mechanism, low to receive
Beneficial defender constantly learns the strategy of the high defender of income.Evolution over time, the strategy of low income will be by high yield
Strategy eliminate.For the attack of reality, from defence square degree, according to the collaboration between different defenders
Relation, the game that best response dynamics study mechanism is applied between different defence policies is set up multistage peak optimization reaction and moved
State repeats Evolutionary Game Model.Under the promotion of above-mentioned " study-improvement " mechanism, the selection probability of different defence policies is presented
Dynamic evolution trend, may finally analyze and obtain network security defence policies choosing method.
Best response dynamics Evolutionary Game Model
It is dynamic using peak optimization reaction under conditions of bounded rationality based on evolutionary game theory for network defense side
State Fast Learning mechanism, builds the best response dynamics Evolutionary Game Model between different defenders.
Define 1 best response dynamics Evolutionary Game Model BRDEGM (Best-response Dynamics
Evolutionary Game Model) it can be expressed as four-tuple, BRDEGM=(D, DS, P, U), wherein
1. D={ d1,d2,…dnRepresent defence participant space.Wherein, diDefender i is represented, different defenders can
To choose different defence policies.
2. DS={ DS1,DS2,…DSmRepresent defender's policy space.Different defenders enjoys the defence policies jointly
Collection.
3. P={ p1,p2,…pmRepresent defender's conviction set.Wherein, piRepresent that defender chooses defence policies DSi's
Probability.
4. U={ U1,U2,…UmRepresent revenue function set.Wherein, UiRepresent that defender chooses defence policies DSiObtained
The income taken.
In network-combination yarn antagonistic process, there is a kind of competitive relation, the defence plan of high yield between different defence policies
Summary will eliminate the relatively low strategy of income.For any two defender d1And d2, it is assumed that DS1、DS2In being optional set of strategies
Any two defence policies, wherein DS1It is to compare DS2Dominating stragegy, i.e. strategy DS1Compare DS2There is more preferable protection effect,
Higher defence income, but DS can be brought1Corresponding defence cost compares DS2It is high.Using best response dynamics Fast Learning machine
System, the Evolutionary Game Model set up under the conditions of bounded rationality.Game theory is as shown in Figure 2.
As game both sides difference Selection Strategy DS1And DS2When, choose DS1Higher defence income will be obtained, remember high cost
It is α, now, Selection Strategy DS to go out part1Defender obtain income u2- α, and Selection Strategy DS2Defender because hitchhike
Higher income is then obtained Deng behavior, u is designated as2+a.Wherein, u1-u2> > a.
From gain matrixAs can be seen that in the game in the presence of two pure strategies receive it is assorted
Weigh (DS1,DS1) and (DS2,DS2), wherein (DS1,DS1) it is that Pareto very wise move is balanced.But if it is also contemplated that between defender
Trusting relationship, or to factors such as Risk Sensitivities, then equilibrium (DS2,DS2) occur possibility can be bigger.
Based on above game condition, it is assumed that all defenders all in a circumference on, each defender with each
Left and right neighbours carry out repeated game, learn than itself high defence policies of tactful income.The income point that note game both sides are obtained
Wei not ∏1And ∏2If, pi(t) it is the Selection Strategy DS in t periods, game person i neighbours1Quantity, the quantity is possible to be taken
Value has 0,1,2 three kind of situation.It can thus be concluded that, game person's Selection Strategy DS1When obtain income for ∏1=u1×pi(t)+(u2-a)×
[2-pi(t)], Selection Strategy DS2When obtain income for ∏2=(u2-a)×pi(t)+u2×[2-pi(t)].According to peak optimization reaction
Dynamic mechanism is understood, works as ∏1>∏2, i.e.,When, game person will be in next game stage Selection Strategy DS2.Thus may be used
To obtain following best response dynamics equation.
Wherein, NtRepresent Selection Strategy DS in n defender1Number.By the game dynamical equation, network is prevented
The final stable state that imperial strategy is chosen will be certain trend for being chosen for defence policies.
Evolutionary Game Model is analyzed with solving
Based on best response dynamics Evolutionary Game Model established above, the policy learning process between defender is carried out
Detailed description and analysis.In network-combination yarn confrontation, because defender is bounded rationality, and with Fast Learning
Ability, can be analyzed and summarized to payoff on last stage, and make corresponding Developing Tactics at once, realize lower single order
The defence maximum revenue of section.Over time, the strategy of whole defender, which is chosen, will reach a stable state.
Below by using from thinking from the particular to the general, using best response dynamics study mechanism, circumference game is carried out
Concrete analysis.
Assuming that defender has 5 defenders, and 5 defenders are distributed in 5 different positions on circumference, such as Fig. 3 institutes
Show, the defender on each position both can be with Selection Strategy DS1, can also Selection Strategy DS2, therefore, the game has at the beginning of 32
Beginning state, including a whole Selection Strategy DS1, a whole Selection Strategy DS2, it is left 30 and includes DS1And DS2
Two kinds of strategies.
pi(t) it is the Selection Strategy DS in t periods, game person i neighbours1Quantity, the possible value of the quantity has 0,1,
2 three kinds of situations.Correspondingly, Selection Strategy DS2Neighbours' quantity be 1-pi(t), equally exist 0,1,2 three kind of value.According to optimal
React dynamic mechanism to understand, work as ∏1>∏2, i.e.,When, game person will be in next game stage Selection Strategy DS2.
Know u1-u2> > α, thenDue to pi(t) only exist 0,1,2 three kind of value, if at two of t period game persons i
As long as there is one to have chosen tactful DS in neighbours1, then game person i will choose DS in t+1 periods1Strategy;If two neighbours are
There is no Selection Strategy DS1, then game person i will choose DS in t+1 periods2Strategy.Therefore deduce that, when 5 defenders are first
All choose DS1Strategy (DS2Strategy) when, final stable state chooses DS for all defenders1Strategy (DS2Strategy).
If there is 1 defender to have chosen DS in first game1Strategy, and other defenders use DS2Strategy
When, then this 5 defenders have finally converged to all defenders and have used DS by the Developing Tactics repeatedly in 4 periods1Plan
Stable state slightly.As shown in figure 3, original state (the DS for giving defender1,DS2,DS2,DS2,DS2), by 4 times
Stage evolution, defender has been finally reached stable state (DS1,DS1,DS1,DS1,DS1)。
Two non-conterminous defenders are contained it can be seen from best response dynamics adjustment process in Fig. 3 to adopt
Use DS1, three non-adjacent game persons use DS1, four defenders use DS1The peak optimization reaction of this several first game situation is moved
State adjust process, they be respectively necessary for three, two and one stages adjustment can reach all use DS1The stabilization of strategy
State.Below to thering are two non-adjacent defenders and three adjacent defenders to use DS in first game1Situation analyzed.
As shown in Figure 4, two adjacent defenders use DS1Best response dynamics Developing Tactics process only need two stages
It can reach all defenders and choose DS1The stable state of strategy.As shown in Figure 5, three adjacent defenders use DS1It is optimal
Reaction dynamic strategy adjustment process only needs a stage to can reach all defenders and chooses DS1The stable state of strategy.
Analyzed more than, in 32 kinds of possible first game situations, only one kind is evolutionarily stable in all defence
Person's Selection Strategy DS2, remaining 31 kinds finally can all converge on all selection DS1State.Illustrate all defender's Selection Strategies
DS1Or DS2The stable state belonged in the gambling process, but converge on DS1Probability to be far longer than DS2。
Above-mentioned two evolutionarily stable state is further understood, if defender is reaching all defender's Selection Strategy DS1
Stable state under, there are a small number of defenders and deviate strategy DS1Situation, best response dynamics can make defender strategy quickly
Converge to and all choose DS1State.Therefore, all defenders choose DS1Stable state be with stability.On the contrary, working as
Reach that all defenders choose DS2Stable state be not but sane, once because some defender deviate DS2, peak optimization reaction
Dynamic can make the state of defender more and more remote from the stable state, therefore the equilibrium is not really stable.With it is long when
Between evolution, defender most at last can Selection Strategy DS1。
The popularization of Evolutionary Game Model
Because during actual cyber-defence, defender is made up of multiple defenders, it is therefore necessary to the game mould
Type makees further genralrlization, i.e., when defender has any n defender, based on above fast reaction dynamic evolution game mould
Type, to any two defence policies DSiAnd DSjEvolutionary Game Analysis is carried out (assuming that DSiIt is relative to DSjDominating stragegy, and i
≠ j), over time, finally give certain Evolution.Evolution for above particular number of networks defender is won
Analysis is played chess, by further analysis and summary, following proposition can be obtained.
Proposition 1:For there is the defender of n defender, when all defenders choose defence plan in first game
Slightly DSi(tactful DSj) when, using best response dynamics study mechanism, whole network defender strategy choose be finally reached it is steady
It is then all equal Selection Strategy DS of defender to determine statei(tactful DSj)。
Proposition 2:For there is the defender of n defender, when n is odd number, in first game, as long as there is one to prevent
Driver have chosen tactful DSi, pass through the adjustment repeatedly in multiple periods to itself strategy by best response dynamics study mechanism, most
It can all converge on all defender's Selection Strategy DS eventuallyiStable state.
Proposition 3:For there is the defender of n defender, when n is even number, in first game, there is a defender
It has chosen defence policies DSi, the equal Selection Strategy DS of other defendersj, then, best response dynamics can not receive all defenders
Hold back in some stable state, evolution over time, adjustment of each defender to strategy can only be absorbed in loop cycle and change.
Proposition 4:It is same simply by the presence of two adjacent defenders in first game for there is the defender of n defender
When have chosen tactful DSi, under best response dynamics study mechanism, evolution over time eventually converges on all defence
Person's whole Selection Strategy DSiStable state.
Proposition 5:For there is the defender of n defender, if by arranging strategy of the defender in first game,
By continuous dynamic evolution so that when the game playing system reaches a certain stage, occur in that some feelings in aforementioned four proposition
Same evolutionary process will occurs in shape, the later stage.
Numerical simulation
Based on network-combination yarn evolutionary Game established above, experiment simulation is carried out using system dynamics, checking network is attacked
The validity and reasonability of anti-Evolutionary Game Model and best response dynamics Evolutionary Game Model.Defender chooses different defence
Original state, whole game playing system will produce different evolution results.Below by for the different defence initial shapes of defender
State, carries out specific numerical simulation.The present invention will be used as simulation object using proposition 2 and proposition 4.
(1) when n be odd number, and defender's original state be only one of which defender's Selection Strategy DS1, other defenders are equal
Selection Strategy DS2When, n=21 is taken, then Selection Strategy DS1Defender's proportion beSelection Strategy DS2Defender institute
Accounting example beNow, there is the power that adjustment changes itself strategy in the game playing system between defender.It is imitative by system
Very, defender's Selection Strategy DS is found1Defender's ratio it is linear increase, and the DS of Selection Strategy2Defender's ratio is linear
Reduce, and just reached in the 10th simulation result final evolutionarily stable state.It is specific as shown in Figure 6.The evolution
As a result can be system stable stateIn one kind, now DS1For optimal defence policies.
(2) there is two adjacent defenders Selection Strategy DS simultaneously in n is odd number, and defender's original state1, other
The equal Selection Strategy DS of defender2When, n=21 is taken, then Selection Strategy DS1Defender's proportion beSelection Strategy DS2's
Defender's proportion isNow, there is the power that adjustment changes itself strategy in the game playing system between defender.Pass through
Constantly develop, defender's Selection Strategy DS1Defender's ratio it is linear increase, and the DS of Selection Strategy2Defender's ratio is into line
Property reduce, and reach in the 10th simulation result final evolutionarily stable state.It is specific as shown in Figure 7.The evolution result can be with
It is system stable stateIn one kind, now DS1For optimal defence policies.
It can be seen from above simulation result, chosen just for the difference of defender's number parity, and defender's strategy
The difference of beginning state, can all influence the final evolution result of whole game playing system.Initial shape is chosen by arranging defender's strategy
State, by constantly evolution, best response dynamics will eventually make game playing system reach some stable state.By experimental result and this
Literary model reasoning is contrasted, it can be seen that the evolution result in experimental system is consistent with the theory analysis in text, explanation
The Evolutionary Game Model meets reality system Evolution, so as to demonstrate the achieved availability of this model.It can be applied
In the network-combination yarn confrontation of reality, the Alliance Defense of defender is made a concrete analysis of and predicted, be the strategy choosing of defender
Take and strong support is provided.
Claims (10)
1. a kind of cyber-defence strategy choosing method of best response dynamics Evolutionary Game Model, it is characterised in that:Including following
Step:
Step 1:Based on bounded rationality condition, using best response dynamics study mechanism, attacking based on best response dynamics is built
Anti- Evolutionary Game Model;
Step 2:Dynamic Evolution and defence Evolutionary Equilibrium point are chosen using defender's strategy, to being defendd between different defenders
Tactful On The Choice is studied;
Step 3:On the basis of the best response dynamics Evolutionary Game Model of foundation, the model is divided by specific example
Analysis promotes Evolutionary Game Model with solving.
2. the cyber-defence strategy choosing method of best response dynamics Evolutionary Game Model according to claim 1, it is special
Levy and be:Best response dynamics Evolutionary Game Model is represented by four-tuple, BRDEGM=(D, DS, P, U) in the step 1
D={ d1,d2,…dnDefence participant space is represented, wherein, diDefender i is represented, different defenders can choose not
Same defence policies;
DS={ DS1,DS2,…DSmDefender's policy space is represented, different defenders enjoys the defence policies collection jointly;
P={ p1,p2,…pmDefender's conviction set is represented, wherein, piRepresent that defender chooses defence policies DSiProbability;
U={ U1,U2,…UmRevenue function set is represented, wherein, UiRepresent that defender chooses defence policies DSiAcquired receipts
Benefit.
3. the cyber-defence strategy choosing method of best response dynamics Evolutionary Game Model according to claim 1, it is special
Levy and be:The best response dynamics equation isWherein NtRepresent to select in n defender
Take tactful DS1Number, DS1It is any one defence policies in optional set of strategies.
4. the cyber-defence strategy choosing method of best response dynamics Evolutionary Game Model according to claim 1, it is special
Levy and be:Defender's strategy selection Dynamic Evolution is in the step 2:In network-combination yarn antagonistic process, difference defence
There is a kind of competitive relation between strategy, the defence policies of high yield will eliminate the relatively low strategy of income.
5. the cyber-defence strategy choosing method of best response dynamics Evolutionary Game Model according to claim 4, it is special
Levy and be:The defence policies of the high yield will be eliminated in the relatively low strategy of income, and gain matrix is:
Wherein, u1、u2Respectively strategy DS1、DS2Income, a is u1、u2Difference.
6. the cyber-defence strategy choosing method of best response dynamics Evolutionary Game Model according to claim 1, it is special
Levy and be:Evolutionary Game Model is promoted in the step 3 is mainly, when defender has any n defender, based on optimal
Dynamic evolution betting model is reacted, to any two defence policies DSiAnd DSjCarry out evolutionary Game Analysis, it is assumed that DSiIt is relative
In DSjDominating stragegy, and i ≠ j over time, finally gives certain Evolution.
7. the cyber-defence strategy choosing method of best response dynamics Evolutionary Game Model according to claim 6, it is special
Levy and be:The Evolution is:For there is the defender of n defender, when all defenders select in first game
Take defence policies DSiOr strategy DSjWhen, using best response dynamics study mechanism, the strategy of whole network defender is chosen most
The stable state reached eventually is then all equal Selection Strategy DS of defenderiOr strategy DSj。
8. the cyber-defence strategy choosing method of best response dynamics Evolutionary Game Model according to claim 6, it is special
Levy and be:The Evolution is:For there is the defender of n defender, when n is odd number, in first game, as long as
There is a defender to have chosen tactful DSi, the anti-of multiple periods is passed through to itself strategy by best response dynamics study mechanism
Polyphony is whole, finally can all converge on all defender's Selection Strategy DSiStable state.
9. the cyber-defence strategy choosing method of best response dynamics Evolutionary Game Model according to claim 6, it is special
Levy and be:The Evolution is:For there is the defender of n defender, when n is even number, in first game, there is one
Individual defender have chosen defence policies DSi, the equal Selection Strategy DS of other defendersj, then, best response dynamics can not make to own
Defender converges on some stable state, evolution over time, and adjustment of each defender to strategy can only be absorbed in loop cycle
Change.
10. the cyber-defence strategy choosing method of best response dynamics Evolutionary Game Model according to claim 6, it is special
Levy and be:The Evolution is:It is adjacent simply by the presence of two in first game for there is the defender of n defender
Defender have chosen tactful DS simultaneouslyi, under best response dynamics study mechanism, evolution over time is eventually converged on
The whole Selection Strategy DS of all defendersiStable state.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710335128.7A CN106953879A (en) | 2017-05-12 | 2017-05-12 | The cyber-defence strategy choosing method of best response dynamics Evolutionary Game Model |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710335128.7A CN106953879A (en) | 2017-05-12 | 2017-05-12 | The cyber-defence strategy choosing method of best response dynamics Evolutionary Game Model |
Publications (1)
Publication Number | Publication Date |
---|---|
CN106953879A true CN106953879A (en) | 2017-07-14 |
Family
ID=59479620
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201710335128.7A Pending CN106953879A (en) | 2017-05-12 | 2017-05-12 | The cyber-defence strategy choosing method of best response dynamics Evolutionary Game Model |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN106953879A (en) |
Cited By (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107483486A (en) * | 2017-09-14 | 2017-12-15 | 中国人民解放军信息工程大学 | Cyber-defence strategy choosing method based on random evolution betting model |
CN107566387A (en) * | 2017-09-14 | 2018-01-09 | 中国人民解放军信息工程大学 | Cyber-defence action decision method based on attacking and defending evolutionary Game Analysis |
CN108182536A (en) * | 2017-12-28 | 2018-06-19 | 东北大学 | A kind of power distribution network CPS safety defense methods based on bounded rationality |
CN108541071A (en) * | 2018-04-10 | 2018-09-14 | 清华大学 | Wireless communication system multi-user resource distribution system based on the double-deck game |
CN108833402A (en) * | 2018-06-11 | 2018-11-16 | 中国人民解放军战略支援部队信息工程大学 | A kind of optimal defence policies choosing method of network based on game of bounded rationality theory and device |
CN110087194A (en) * | 2019-04-25 | 2019-08-02 | 东华大学 | Position data poisoning attacks prototype system in car networking based on game |
CN110381020A (en) * | 2019-06-13 | 2019-10-25 | 长沙理工大学 | A kind of IDS resource allocation method, device and computer readable storage medium |
CN111224966A (en) * | 2019-12-31 | 2020-06-02 | 中国人民解放军战略支援部队信息工程大学 | Optimal defense strategy selection method based on evolutionary network game |
CN111245857A (en) * | 2020-01-17 | 2020-06-05 | 安徽师范大学 | Channel network steady state evolution game method in block link environment |
CN112989357A (en) * | 2021-03-09 | 2021-06-18 | 中国人民解放军空军工程大学 | Multi-stage platform dynamic defense method based on signal game model |
CN113868932A (en) * | 2021-06-09 | 2021-12-31 | 南京大学 | Task allocation method based on complete information bidding game |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101808020A (en) * | 2010-04-19 | 2010-08-18 | 吉林大学 | Intrusion response decision-making method based on incomplete information dynamic game |
CN103152345A (en) * | 2013-03-07 | 2013-06-12 | 南京理工大学常熟研究院有限公司 | Network safety optimum attacking and defending decision method for attacking and defending game |
CN103384384A (en) * | 2013-07-19 | 2013-11-06 | 哈尔滨工程大学 | Recognition relay network trust management device and method based on dynamic evolution |
CN105682174A (en) * | 2016-01-15 | 2016-06-15 | 哈尔滨工业大学深圳研究生院 | Opportunity network evolution algorithm and device for promoting node cooperation |
US9471777B1 (en) * | 2012-02-24 | 2016-10-18 | Emc Corporation | Scheduling of defensive security actions in information processing systems |
CN106550373A (en) * | 2016-09-30 | 2017-03-29 | 天津大学 | Wireless sensor network data fusion degree of accuracy model based on evolutionary Game |
-
2017
- 2017-05-12 CN CN201710335128.7A patent/CN106953879A/en active Pending
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101808020A (en) * | 2010-04-19 | 2010-08-18 | 吉林大学 | Intrusion response decision-making method based on incomplete information dynamic game |
US9471777B1 (en) * | 2012-02-24 | 2016-10-18 | Emc Corporation | Scheduling of defensive security actions in information processing systems |
CN103152345A (en) * | 2013-03-07 | 2013-06-12 | 南京理工大学常熟研究院有限公司 | Network safety optimum attacking and defending decision method for attacking and defending game |
CN103384384A (en) * | 2013-07-19 | 2013-11-06 | 哈尔滨工程大学 | Recognition relay network trust management device and method based on dynamic evolution |
CN105682174A (en) * | 2016-01-15 | 2016-06-15 | 哈尔滨工业大学深圳研究生院 | Opportunity network evolution algorithm and device for promoting node cooperation |
CN106550373A (en) * | 2016-09-30 | 2017-03-29 | 天津大学 | Wireless sensor network data fusion degree of accuracy model based on evolutionary Game |
Cited By (19)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107566387B (en) * | 2017-09-14 | 2020-01-10 | 中国人民解放军信息工程大学 | Network defense action decision method based on attack and defense evolution game analysis |
CN107566387A (en) * | 2017-09-14 | 2018-01-09 | 中国人民解放军信息工程大学 | Cyber-defence action decision method based on attacking and defending evolutionary Game Analysis |
CN107483486A (en) * | 2017-09-14 | 2017-12-15 | 中国人民解放军信息工程大学 | Cyber-defence strategy choosing method based on random evolution betting model |
CN107483486B (en) * | 2017-09-14 | 2020-04-03 | 中国人民解放军信息工程大学 | Network defense strategy selection method based on random evolution game model |
CN108182536A (en) * | 2017-12-28 | 2018-06-19 | 东北大学 | A kind of power distribution network CPS safety defense methods based on bounded rationality |
CN108182536B (en) * | 2017-12-28 | 2021-11-16 | 东北大学 | CPS security defense method for power distribution network based on finiteness |
CN108541071A (en) * | 2018-04-10 | 2018-09-14 | 清华大学 | Wireless communication system multi-user resource distribution system based on the double-deck game |
CN108541071B (en) * | 2018-04-10 | 2019-03-01 | 清华大学 | Wireless communication system multi-user resource distribution system based on the double-deck game |
CN108833402A (en) * | 2018-06-11 | 2018-11-16 | 中国人民解放军战略支援部队信息工程大学 | A kind of optimal defence policies choosing method of network based on game of bounded rationality theory and device |
CN108833402B (en) * | 2018-06-11 | 2020-11-24 | 中国人民解放军战略支援部队信息工程大学 | Network optimal defense strategy selection method and device based on limited theory game theory |
CN110087194B (en) * | 2019-04-25 | 2021-05-11 | 东华大学 | Game-based position data poisoning attack prototype system in Internet of vehicles |
CN110087194A (en) * | 2019-04-25 | 2019-08-02 | 东华大学 | Position data poisoning attacks prototype system in car networking based on game |
CN110381020A (en) * | 2019-06-13 | 2019-10-25 | 长沙理工大学 | A kind of IDS resource allocation method, device and computer readable storage medium |
CN111224966A (en) * | 2019-12-31 | 2020-06-02 | 中国人民解放军战略支援部队信息工程大学 | Optimal defense strategy selection method based on evolutionary network game |
CN111224966B (en) * | 2019-12-31 | 2021-11-02 | 中国人民解放军战略支援部队信息工程大学 | Optimal defense strategy selection method based on evolutionary network game |
CN111245857A (en) * | 2020-01-17 | 2020-06-05 | 安徽师范大学 | Channel network steady state evolution game method in block link environment |
CN111245857B (en) * | 2020-01-17 | 2021-11-26 | 安徽师范大学 | Channel network steady state evolution game method in block link environment |
CN112989357A (en) * | 2021-03-09 | 2021-06-18 | 中国人民解放军空军工程大学 | Multi-stage platform dynamic defense method based on signal game model |
CN113868932A (en) * | 2021-06-09 | 2021-12-31 | 南京大学 | Task allocation method based on complete information bidding game |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN106953879A (en) | The cyber-defence strategy choosing method of best response dynamics Evolutionary Game Model | |
Holcomb et al. | Overview on deepmind and its alphago zero ai | |
CN107483486B (en) | Network defense strategy selection method based on random evolution game model | |
CN107566387B (en) | Network defense action decision method based on attack and defense evolution game analysis | |
Brzezinski | In quest of national security | |
Deng et al. | Stratified and maximum information item selection procedures in computer adaptive testing | |
CN110083748A (en) | A kind of searching method based on adaptive Dynamic Programming and the search of Monte Carlo tree | |
CN115577874A (en) | Strategy model training method, device and equipment applied to war game deduction | |
CN114070655B (en) | Network flow detection rule generation method and device, electronic equipment and storage medium | |
Chartier et al. | Bracketology: How can math help? | |
CN112870722A (en) | Method, device, equipment and medium for generating fighting AI (AI) game model | |
CN115994484A (en) | Air combat countergame strategy optimizing system based on multi-population self-adaptive orthoevolutionary algorithm | |
Chowdhury et al. | The max-min group contest | |
Li et al. | Nonlinear Random Matrix Model and Research for Quantitative Representation of Volleyball Attacker’s Action Links | |
Li et al. | Amazon Chess Based on UCT-PVS Hybrid Algorithm | |
Li et al. | A game AI based on ID3 algorithm | |
CN115187409B (en) | Method and device for determining energy investment strategy, electronic equipment and storage medium | |
Ye et al. | COOPERATION AND COMPETITION IN HISTORY-DEPENDENT PARRONDO'S GAME ON NETWORKS | |
Zhang et al. | Discussion on the design of computer games system based on surakarta | |
CN118170003B (en) | PID parameter optimization method based on improved corner lizard optimization algorithm | |
Chen et al. | An Empirical Evaluation of Applying Deep Reinforcement Learning to Taiwanese Mahjong Programs | |
Zhao et al. | Particle Swarm Optimization Algorithm can Promote Cooperation in Public Goods Game with Costly Punishment | |
Nhengu | Feminisms, WPS Agenda and Women’s Peacebuilding and Peacemaking Networks in Africa: Solution or Quandary? | |
Tejani | Application of Neuroevolution in Blackjack | |
McLean | Challenges in the Asia-Pacific Strategic Environment & Imperatives for Cooperation. |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20170714 |
|
RJ01 | Rejection of invention patent application after publication |