CN110807230B - Method for autonomously learning and optimizing topological structure robustness of Internet of things - Google Patents
Method for autonomously learning and optimizing topological structure robustness of Internet of things Download PDFInfo
- Publication number
- CN110807230B CN110807230B CN201911036835.1A CN201911036835A CN110807230B CN 110807230 B CN110807230 B CN 110807230B CN 201911036835 A CN201911036835 A CN 201911036835A CN 110807230 B CN110807230 B CN 110807230B
- Authority
- CN
- China
- Prior art keywords
- network
- action
- learning
- topological structure
- model
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 20
- 238000002474 experimental method Methods 0.000 claims abstract description 20
- 238000012360 testing method Methods 0.000 claims abstract description 13
- 238000012549 training Methods 0.000 claims abstract description 10
- 230000002787 reinforcement Effects 0.000 claims abstract description 6
- 238000013135 deep learning Methods 0.000 claims abstract description 5
- 238000005457 optimization Methods 0.000 claims description 33
- 230000000694 effects Effects 0.000 claims description 17
- 238000004891 communication Methods 0.000 claims description 9
- 230000006870 function Effects 0.000 claims description 8
- 238000013507 mapping Methods 0.000 claims description 8
- 238000003062 neural network model Methods 0.000 claims description 8
- 230000006399 behavior Effects 0.000 claims description 7
- 239000011159 matrix material Substances 0.000 claims description 4
- NAWXUBYGYWOOIX-SFHVURJKSA-N (2s)-2-[[4-[2-(2,4-diaminoquinazolin-6-yl)ethyl]benzoyl]amino]-4-methylidenepentanedioic acid Chemical compound C1=CC2=NC(N)=NC(N)=C2C=C1CCC1=CC=C(C(=O)N[C@@H](CC(=C)C(O)=O)C(O)=O)C=C1 NAWXUBYGYWOOIX-SFHVURJKSA-N 0.000 claims description 3
- 238000009825 accumulation Methods 0.000 claims description 3
- 238000005070 sampling Methods 0.000 claims description 3
- 230000001186 cumulative effect Effects 0.000 claims description 2
- 230000005540 biological transmission Effects 0.000 abstract description 2
- 230000006835 compression Effects 0.000 description 3
- 238000007906 compression Methods 0.000 description 3
- 238000010586 diagram Methods 0.000 description 2
- 238000011160 research Methods 0.000 description 2
- 206010033799 Paralysis Diseases 0.000 description 1
- 230000002068 genetic effect Effects 0.000 description 1
- 238000004088 simulation Methods 0.000 description 1
Landscapes
- Data Exchanges In Wide-Area Networks (AREA)
Abstract
The invention discloses a method for optimizing the robustness of an internet of things topological structure through autonomous learning, which comprises the following steps of: initializing an Internet of things topological structure; step 2: compressing a topological structure; step 3: an autonomous learning model is initialized. According to the features of deep learning and reinforcement learning, constructing a deep deterministic learning strategy model to train the topological structure of the Internet of things; step 4: training and testing a model; step 5: periodically repeating step 4 in one independent repetition experiment, and periodically repeating steps 1, 2, 3 and 4 in multiple independent repetition experiments; up to a maximum number of iterations. In the process, the maximum iteration number is set, and the experiment is independently repeated each time, and the optimal result is selected. The experiment was repeated several times and the average value was chosen as the result of this experiment. The invention can obviously improve the capability of the initial topological structure for resisting attack; and the robust capability of the network topology structure is optimized through autonomous learning, so that high-reliability data transmission is ensured.
Description
Technical Field
The invention relates to the technical field of Internet of things networks, in particular to a method for optimizing topological structure robustness of the Internet of things.
Background
The internet of things is an important component in a smart city network, and large-scale equipment nodes are connected together through the internet of things to provide high-quality services for people. However, connected device nodes need to be tolerant of failure threats, such as random failures of devices, artificial vandalism and natural disasters, network part node failures caused by energy exhaustion, etc., so that the whole network is paralyzed. Under the wide application scene of the Internet of things, how to ensure that large-scale nodes ensure high-quality data service communication of the network on the premise of network part node failure has important research significance.
In conventional optimization of network topology of the internet of things, nodes are typically deployed at fixed locations with certain communication range limitations. The network topology initializes its network model according to a scaleless network model. In the network topology optimization strategy, most of researches adopt greedy edge-changing strategies or evolution algorithms to optimize the robustness of the network topology, so that the whole network has very high capability of resisting attacks. For example, journal Robustness optimization scheme with multi-delivery co-evolution for scale-free wireless sensor networks proposes a multi-population genetic algorithm to solve the problem of local optimization so as to obtain a globally optimal network topology, but the time cost for optimizing a network topology is large, and the algorithm cannot accumulate the optimization experience every time, so that the algorithm needs to be restarted every time. Secondly, some researchers use neural network models to represent learning behaviors before and after network topology optimization, so that the optimization time of the topology is reduced, but the method needs label target data, and the label target data defines the maximum value of optimization. Therefore, in the optimization of the topological structure of the Internet of things, the network topological structure strategy of autonomous learning optimization is utilized to improve the network topological robustness, the online of the optimization target value is eliminated, and the experience of each learning is accumulated to guide the subsequent optimization behavior.
Disclosure of Invention
In order to solve the problems in the prior art, the invention aims to provide a method for optimizing the robustness of the topological structure of the Internet of things by autonomous learning, wherein the topological structure of the Internet of things is used as an environment space according to the characteristics of reinforcement learning and deep learning, meanwhile, an action space is designed to search the environment, the effect of each action on the environment space is optimized, and the accumulated optimization effect is maximized, so that the robustness of the topological structure of the Internet of things is improved, and the autonomous learning behavior of the whole topological structure of the network is increased.
The invention discloses a method for optimizing the topological structure robustness of the Internet of things through autonomous learning, which comprises the following steps:
step 1: initializing an internet of things topological structure, namely randomly deploying nodes according to rules of a scaleless network model and an edge density parameter M, and fixing geographic positions, wherein the edge density parameter M is set to be M=2;
step 2: the method comprises the steps of compressing a topological structure, namely eliminating redundant node information which is not in a communication range, reserving only node connection relations in the communication range in a form of an adjacency matrix, compressing a storage space of the network topological structure, and taking the compressed network topology as an environment space S, wherein the environment space S is a row vector, and the row vector is changed along with the change of a network topology state;
step 3: initializing an autonomous learning model, namely constructing a deep deterministic learning strategy model according to the features of deep learning and reinforcement learning to train the topological structure of the Internet of things: adopting a depth deterministic Q-learning network model, simulating a selection strategy pi of actions and an optimization strategy Q of a network, mapping a continuous action space into a discrete action space, and designing an updating rule of a target optimization function O and the whole training model; wherein:
the action selection policy pi is defined by equation (1):
a t =π(s t |θ) (1)
wherein a is t Representing the selected deterministic action s t Representing the current network topology state, and theta represents the parameters of the action network;
the network optimization strategy Q is defined by equation (2):
Q(s t ,a t )=E(r(s t ,a t )+γQ(s t+1 ,π(s t+1 ))) (2)
wherein r represents the current action a t For the current network state s t And (c) the immediate return value of gamma represents the discount factor, the accumulated learning experience, Q(s) t+1 ,π(s t+1 ) A future return value representing an action taken in the next network state, and thus the effect Q(s) of the current action on the current network state t ,a t ) The method comprises the steps of forming an immediate return value and a future return value, wherein E () represents an expected value, and selecting an effect before strategy accumulation for a series of actions;
the objective function O of the autonomous learning model is defined by equation (3) according to the above description:
O(θ)=Ε(r 1 +γr 2 +γ 2 r 3 +...|π(,θ)) (3)
where r represents the sum of the actions for environment O (θ) =e (r 1 +γr 2 +γ 2 r 3 +. |pi (, θ)) produced effects, i.e., a return value, γ represents a discount factor, cumulative learning experience, pi (, θ) represents a policy of action selection, θ represents a parameter of an action policy network, and E represents an average expected value;
the update rules of the network are defined by equation (4):
wherein T is i Represents a target expected value, defined by the formula (5):
T i =r i +γQ'(s i ,π'(s i+1 |θ π’ )|θ Q’ ) (5)
in the formula, Q ', pi' represent target networks of action selection strategies and optimization strategies, and errors of the whole autonomous learning model are calculated;
step 4: training and testing a model, namely in a training stage, randomly obtaining a discrete action a through an action selection neural network model, evaluating the effect of the action on the current environment through a network optimization neural network model strategy, accumulating the previous learning experience and updating the whole network model, and finally obtaining an optimal result; in the test stage, testing sample data to obtain a test result; wherein:
the output of the discrete action is defined by equation (6):
d=MAP(a) (6)
wherein MAP represents a mapping relationship between a continuous motion space and a discrete motion space, and a is defined by formula (7):
a=π(s)=π(s|θ)+N (7)
wherein N represents a random sampling rule, and more effective action behaviors in the action space are explored;
wherein, the action selecting strategy network updating principle is updated towards the direction of making the strategy selecting network value maximum, so the selected action makes the strategy selecting network maximum;
the updating rule of the target network is defined as formula (10);
θ Q’ ←τθ Q +(1-τ)θ Q’
θ π’ ←τθ π +(1-τ)θ π’ (10)
where τ represents the update rate of the target network;
step 5: periodically repeating step 4 in one independent repetition experiment, and periodically repeating steps 1, 2, 3 and 4 in multiple independent repetition experiments; up to the maximum iteration times;
in the process, the maximum iteration times are set, the experiment is independently repeated each time, the optimal result is selected, the experiment is repeated for a plurality of times, and the average value is selected as the result of the experiment.
The positive technical effects obtained by the invention include:
(1) The invention designs an autonomous learning optimization strategy for the topological structure robustness of the Internet of things by utilizing the deep reinforcement learning neural network model, and can obviously improve the capability of the initial topological structure for resisting attack;
(2) The invention utilizes the state representation of the topological structure of the Internet of things, discrete action space mapping relation, the non-scale characteristic and compression characteristic of the network to come from the robust capability of main learning optimization network topological structure, thereby ensuring high-reliability data transmission.
Drawings
FIG. 1 is a flowchart of an overall method for optimizing the topological robustness of an Internet of things by autonomous learning;
FIG. 2 is a schematic diagram of a mapping relationship between continuous and discrete actions of an autonomous learning optimization model;
fig. 3 is a schematic diagram of a topological structure compression model of the internet of things.
Detailed Description
The following describes in detail the specific manner, structure, features and actions of a node deployment strategy designed according to the present invention with reference to the accompanying drawings.
As shown in FIG. 1, the overall flow chart of the method for optimizing the topological structure robustness of the Internet of things by autonomous learning is disclosed, and the method comprehensively considers the mapping relation between a large-scale continuous action space and a discrete action space, the compression mode of a network topological structure and the connection relation of nodes, effectively improves the network robustness, enhances the autonomous learning behavior of the overall network, balances the node connection distribution and ensures the high-quality communication capability of the network. The flow of the method specifically comprises the following steps:
step 1: initializing an Internet of things topological structure. And randomly deploying nodes according to the rule of the scale-free network model and the edge density parameter M, and fixing the geographic position. Most nodes have few degrees, few nodes have large degrees, and the real world Internet of things topological structure is described to the greatest extent. Each node has the same attributes.
Wherein the edge density parameter is set to m=2, indicating that the number of edges in the network is 2 times the number of nodes.
Step 2: the topology is compressed. Different from the adjacent matrix representation mode of the network topology structure, the invention eliminates redundant node information which is not in the communication range, only reserves the node connection relation in the communication range on the form of the adjacent matrix, compresses the storage space of the network topology structure, and takes the compressed network topology as the environment space S.
The environment space S is a row vector, which changes with the change of the network topology state.
Step 3: an autonomous learning model is initialized. And constructing a deep deterministic learning strategy model to train the topological structure of the Internet of things according to the deep learning and reinforcement learning characteristics. A depth deterministic Q-learning network model is adopted, a selection strategy pi of a simulation action and an optimization strategy Q of a network are adopted, a continuous action space is mapped into a discrete action space, and an updating rule of a target optimization function O and the whole training model is designed.
Wherein the action selection policy pi is defined by formula (1).
a t =π(s t |θ) (1)
Wherein a is t Representing the selected deterministic action s t Representing the current network topology state, θ represents a parameter of the action network. Current network topology state s t And obtaining a deterministic action through the action policy function pi, wherein the deterministic action can directly operate on the current network topology structure.
The network optimization strategy Q is defined by a formula (2), and the effect of the selected action on the environment space is measured.
Q(s t ,a t )=E(r(s t ,a t )+γQ(s t+1 ,π(s t+1 ))) (2)
Where r represents the current action a t For the current network state s t Gamma is the discount factor, and the accumulated learning experience. Q(s) t+1 ,π(s t+1 ) A future return value representing an action taken in the next network state, and thus the effect Q(s) of the current action on the current network state t ,a t ) Consists of an immediate return value and a future return value, E () represents the expected value, and the effect before policy accumulation is selected for a series of actions.
Wherein the objective function O of the autonomous learning model can be defined specifically as formula (3) according to the above description
(3)
Where r represents the sum of the actions for environment O (θ) =e (r 1 +γr 2 +γ 2 r 3 +. |pi (, θ)), the effect produced, i.e., the return value. Gamma denotes the discount factor, and the learning experience is accumulated. Pi (, θ) represents the policy of action selection, θ is a parameter of the action policy network. E represents the average expected value as an objective function of the entire autonomous learning model.
Wherein the update rule of the network is defined by equation (4).
Wherein T is i The target expected value is represented by the expression (5).
T i =r i +γQ'(s i ,π'(s i+1 |θ π’ )|θ Q’ ) (5)
In the formula, Q ', pi' represents a target network of an action selection strategy and an optimization strategy, and the error of the whole autonomous learning model is calculated.
Step 4: training and testing the model. In the training stage, a discrete action a is randomly obtained through action selection neural network model, the effect of the action on the current environment is evaluated through a network optimization neural network model strategy, the previous learning experience is accumulated, the whole network model is updated, and finally the optimal result is obtained. And in the test stage, testing the sample data to obtain a test result.
Wherein the output of the discrete action is defined by equation (6).
d=MAP(a) (6)
Wherein MAP represents a mapping relationship between a continuous motion space and a discrete motion space, and a is defined by formula (7):
a=π(s)=π(s|θ)+N
(7)
where N represents a random sampling rule, exploring more effective action behaviors in the action space.
After obtaining the effect of the action on the current environment, namely the return value, the return value is stored in a memory, the convergence rate of the autonomous learning model is accelerated by using the previous learning experience in the candidate optimization learning, and the rule is defined as formula (8).
(s t ,a t ,r t ,s t+1 )→D (8)
Wherein D represents a memory in the network model, and stores the current network state, action, instant return value, and next network state.
Wherein, in the stage of updating the autonomous learning optimization model, the action selection policy network updating rule is defined by the formula (9).
▽π=E π' [▽ a Q(s,a)▽ θ π(s)] (9)
In the formula, the action selection policy network updating principle is updated towards the direction of maximizing the value of the policy selection network, so that the selected action maximizes the policy selection network.
Wherein the update rule of the target network is defined as formula (10).
θ Q’ ←τθ Q +(1-τ)θ Q’
θ π’ ←τθ π +(1-τ)θ π’ (10)
Where τ represents the update rate of the target network.
Step 5: periodically repeating step 4 in one independent repetition experiment, and periodically repeating steps 1, 2, 3 and 4 in multiple independent repetition experiments; up to a maximum number of iterations. In the process, the maximum iteration number is set, and the experiment is independently repeated each time, and the optimal result is selected. The experiment was repeated several times and the average value was chosen as the result of this experiment.
Claims (1)
1. The method for autonomously learning and optimizing the topological structure robustness of the Internet of things is characterized by comprising the following steps of:
step 1: initializing an internet of things topological structure, namely randomly deploying nodes according to rules of a scaleless network model and an edge density parameter M, and fixing geographic positions, wherein the edge density parameter M is set to be M=2;
step 2: the method comprises the steps of compressing a topological structure, namely eliminating redundant node information which is not in a communication range, reserving only node connection relations in the communication range in a form of an adjacency matrix, compressing a storage space of the network topological structure, and taking the compressed network topology as an environment space S, wherein the environment space S is a row vector, and the row vector is changed along with the change of a network topology state;
step 3: initializing an autonomous learning model, namely constructing a deep deterministic learning strategy model according to the features of deep learning and reinforcement learning to train the topological structure of the Internet of things: adopting a depth deterministic Q-learning network model, simulating a selection strategy pi of actions and an optimization strategy Q of a network, mapping a continuous action space into a discrete action space, and designing an updating rule of a target optimization function O and the whole training model; wherein:
the action selection policy pi is defined by equation (1):
a t =π(s t |θ) (1)
wherein a is t Representing the selected deterministic action s t Representing the current network topology state, and theta represents the parameters of the action network;
the network optimization strategy Q is defined by equation (2):
Q(s t ,a t )=E(r(s t ,a t )+γQ(s t+1 ,π(s t+1 ))) (2)
wherein r represents the current action a t For the current network state s t And (c) the immediate return value of gamma represents the discount factor, the accumulated learning experience, Q(s) t+1 ,π(s t+1 ) A future return value representing an action taken in the next network state, and thus the effect Q(s) of the current action on the current network state t ,a t ) The method comprises the steps of forming an immediate return value and a future return value, wherein E () represents an expected value, and selecting an effect before strategy accumulation for a series of actions;
the objective function O of the autonomous learning model is defined by equation (3) according to the above description:
O(θ)=Ε(r 1 +γr 2 +γ 2 r 3 +...|π(,θ)) (3)
where r represents the sum of the actions for environment O (θ) =e (r 1 +γr 2 +γ 2 r 3 +. |pi (, θ)) produced effects, i.e., a return value, γ represents a discount factor, cumulative learning experience, pi (, θ) represents a policy of action selection, θ represents a parameter of an action policy network, and E represents an average expected value; θ
The update rules of the network are defined by equation (4):
wherein T is i Represents a target expected value, which is determined by the formula (5)The meaning is as follows:
T i =r i +γQ'(s i ,π'(s i+1 |θ π′ )|θ Q′ ) (5)
in the formula, Q ', pi' represent target networks of action selection strategies and optimization strategies, and errors of the whole autonomous learning model are calculated;
step 4: training and testing a model, namely in a training stage, randomly obtaining a discrete action a through an action selection neural network model, evaluating the effect of the action on the current environment by a network optimization neural network model strategy, accumulating the previous learning experience and updating the whole network model, and finally obtaining an optimal result; in the test stage, testing sample data to obtain a test result; wherein:
the output of the discrete action is defined by equation (6):
d=MAP(a) (6)
wherein MAP represents a mapping relationship between a continuous motion space and a discrete motion space, and a is defined by formula (7):
a=π(s)=π(s|θ)+N (7)
wherein N represents a random sampling rule, more effective action behaviors in an action space are explored, and s represents the current network state;
wherein, the action selecting strategy network updating principle is updated towards the direction of making the strategy selecting network value maximum, so the selected action makes the strategy selecting network maximum;
the updating rule of the target network is defined as formula (10);
θ Q′ ←τθ Q +(1-τ)θ Q′
θ π′ ←τθ π +(1-τ)θ π′ (10)
where τ represents the update rate of the target network;
step 5: periodically repeating step 4 in one independent repetition experiment, and periodically repeating steps 1, 2, 3 and 4 in multiple independent repetition experiments; up to the maximum iteration times;
in the process, the maximum iteration times are set, the experiment is independently repeated each time, the optimal result is selected, the experiment is repeated for a plurality of times, and the average value is selected as the result of the experiment.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201911036835.1A CN110807230B (en) | 2019-10-29 | 2019-10-29 | Method for autonomously learning and optimizing topological structure robustness of Internet of things |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201911036835.1A CN110807230B (en) | 2019-10-29 | 2019-10-29 | Method for autonomously learning and optimizing topological structure robustness of Internet of things |
Publications (2)
Publication Number | Publication Date |
---|---|
CN110807230A CN110807230A (en) | 2020-02-18 |
CN110807230B true CN110807230B (en) | 2024-03-12 |
Family
ID=69489419
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201911036835.1A Active CN110807230B (en) | 2019-10-29 | 2019-10-29 | Method for autonomously learning and optimizing topological structure robustness of Internet of things |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN110807230B (en) |
Families Citing this family (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111929641B (en) * | 2020-06-19 | 2022-08-09 | 天津大学 | Rapid indoor fingerprint positioning method based on width learning |
CN111935724B (en) * | 2020-07-06 | 2022-05-03 | 天津大学 | Wireless sensor network topology optimization method based on asynchronous deep reinforcement learning |
CN113422695B (en) * | 2021-06-17 | 2022-05-24 | 天津大学 | Optimization method for improving robustness of topological structure of Internet of things |
CN113435567B (en) * | 2021-06-25 | 2023-07-07 | 广东技术师范大学 | Intelligent topology reconstruction method based on flow prediction, electronic equipment and storage medium |
CN113923123B (en) * | 2021-09-24 | 2023-06-09 | 天津大学 | Underwater wireless sensor network topology control method based on deep reinforcement learning |
CN114567563B (en) * | 2022-03-31 | 2024-04-12 | 北京邮电大学 | Training method of network topology model, and reconstruction method and device of network topology |
CN115225509B (en) * | 2022-07-07 | 2023-09-22 | 天津大学 | Internet of things topological structure generation method based on neural evolution |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102789593A (en) * | 2012-06-18 | 2012-11-21 | 北京大学 | Intrusion detection method based on incremental GHSOM (Growing Hierarchical Self-organizing Maps) neural network |
CN102868972A (en) * | 2012-09-05 | 2013-01-09 | 河海大学常州校区 | Internet of things (IoT) error sensor node location method based on improved Q learning algorithm |
CN103490413A (en) * | 2013-09-27 | 2014-01-01 | 华南理工大学 | Intelligent electricity generation control method based on intelligent body equalization algorithm |
Family Cites Families (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7130807B1 (en) * | 1999-11-22 | 2006-10-31 | Accenture Llp | Technology sharing during demand and supply planning in a network-based supply chain environment |
US10715395B2 (en) * | 2017-11-27 | 2020-07-14 | Massachusetts Institute Of Technology | Methods and apparatus for communication network |
-
2019
- 2019-10-29 CN CN201911036835.1A patent/CN110807230B/en active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102789593A (en) * | 2012-06-18 | 2012-11-21 | 北京大学 | Intrusion detection method based on incremental GHSOM (Growing Hierarchical Self-organizing Maps) neural network |
CN102868972A (en) * | 2012-09-05 | 2013-01-09 | 河海大学常州校区 | Internet of things (IoT) error sensor node location method based on improved Q learning algorithm |
CN103490413A (en) * | 2013-09-27 | 2014-01-01 | 华南理工大学 | Intelligent electricity generation control method based on intelligent body equalization algorithm |
Non-Patent Citations (3)
Title |
---|
伯努利节点网络模型的拓扑鲁棒性分析方法;冯涛;李洪涛;袁占亭;马建峰;电子学报;第39卷(第7期);全文 * |
基于强化学习的无线网络自组织性研究;王超;王芷阳;沈聪;;中国科学技术大学学报(第12期);全文 * |
面向SDN网络虚拟化平台的控制器放置算法;董晓东;郭志强;陈胜;周晓波;齐恒;李克秋;电信科学(第004期);全文 * |
Also Published As
Publication number | Publication date |
---|---|
CN110807230A (en) | 2020-02-18 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110807230B (en) | Method for autonomously learning and optimizing topological structure robustness of Internet of things | |
CN112541302B (en) | Air quality prediction model training method, air quality prediction method and device | |
Straatman et al. | Towards an automatic calibration procedure for constrained cellular automata | |
CN111935724B (en) | Wireless sensor network topology optimization method based on asynchronous deep reinforcement learning | |
CN110267292A (en) | Cellular network method for predicting based on Three dimensional convolution neural network | |
CN103905246B (en) | Link prediction method based on grouping genetic algorithm | |
CN104636801A (en) | Transmission line audible noise prediction method based on BP neural network optimization | |
CN110826244B (en) | Conjugated gradient cellular automaton method for simulating influence of rail transit on urban growth | |
CN116207739B (en) | Optimal scheduling method and device for power distribution network, computer equipment and storage medium | |
CN110619082B (en) | Project recommendation method based on repeated search mechanism | |
CN117498400A (en) | Distributed photovoltaic and energy storage data processing method and system | |
Gallego et al. | Maintaining flexibility in smart grid consumption through deep learning and deep reinforcement learning | |
CN113824641A (en) | Internet of vehicles routing method, system, equipment and storage medium | |
CN114861917A (en) | Knowledge graph inference model, system and inference method for Bayesian small sample learning | |
CN117810987A (en) | Prediction method, prediction device, prediction terminal and prediction storage medium for residential payload section | |
CN112116305A (en) | Power grid probability visualization model construction method and system for machine learning | |
CN116167254A (en) | Multidimensional city simulation deduction method and system based on city big data | |
CN115395502A (en) | Photovoltaic power station power prediction method and system | |
CN114372418A (en) | Wind power space-time situation description model establishing method | |
CN114566048A (en) | Traffic control method based on multi-view self-adaptive space-time diagram network | |
CN112884129A (en) | Multi-step rule extraction method and device based on teaching data and storage medium | |
Cong et al. | Research on Satellite Network Traffic Prediction Algorithm Based on Gray Wolf Algorithm Optimizing GRU and Spatiotemporal Analysis | |
Fang et al. | Towards data-centric control of sensor networks through bayesian dynamic linear modelling | |
CN116540678B (en) | Industrial equipment state deduction evaluation method based on digital twin | |
CN118171585B (en) | Super-long pile foundation structure design and quality control method and system |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |