CN110807230B - Method for autonomously learning and optimizing topological structure robustness of Internet of things - Google Patents

Method for autonomously learning and optimizing topological structure robustness of Internet of things Download PDF

Info

Publication number
CN110807230B
CN110807230B CN201911036835.1A CN201911036835A CN110807230B CN 110807230 B CN110807230 B CN 110807230B CN 201911036835 A CN201911036835 A CN 201911036835A CN 110807230 B CN110807230 B CN 110807230B
Authority
CN
China
Prior art keywords
network
action
learning
topological structure
model
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201911036835.1A
Other languages
Chinese (zh)
Other versions
CN110807230A (en
Inventor
邱铁
陈宁
李克秋
周晓波
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tianjin University
Original Assignee
Tianjin University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tianjin University filed Critical Tianjin University
Priority to CN201911036835.1A priority Critical patent/CN110807230B/en
Publication of CN110807230A publication Critical patent/CN110807230A/en
Application granted granted Critical
Publication of CN110807230B publication Critical patent/CN110807230B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Landscapes

  • Data Exchanges In Wide-Area Networks (AREA)

Abstract

The invention discloses a method for optimizing the robustness of an internet of things topological structure through autonomous learning, which comprises the following steps of: initializing an Internet of things topological structure; step 2: compressing a topological structure; step 3: an autonomous learning model is initialized. According to the features of deep learning and reinforcement learning, constructing a deep deterministic learning strategy model to train the topological structure of the Internet of things; step 4: training and testing a model; step 5: periodically repeating step 4 in one independent repetition experiment, and periodically repeating steps 1, 2, 3 and 4 in multiple independent repetition experiments; up to a maximum number of iterations. In the process, the maximum iteration number is set, and the experiment is independently repeated each time, and the optimal result is selected. The experiment was repeated several times and the average value was chosen as the result of this experiment. The invention can obviously improve the capability of the initial topological structure for resisting attack; and the robust capability of the network topology structure is optimized through autonomous learning, so that high-reliability data transmission is ensured.

Description

Method for autonomously learning and optimizing topological structure robustness of Internet of things
Technical Field
The invention relates to the technical field of Internet of things networks, in particular to a method for optimizing topological structure robustness of the Internet of things.
Background
The internet of things is an important component in a smart city network, and large-scale equipment nodes are connected together through the internet of things to provide high-quality services for people. However, connected device nodes need to be tolerant of failure threats, such as random failures of devices, artificial vandalism and natural disasters, network part node failures caused by energy exhaustion, etc., so that the whole network is paralyzed. Under the wide application scene of the Internet of things, how to ensure that large-scale nodes ensure high-quality data service communication of the network on the premise of network part node failure has important research significance.
In conventional optimization of network topology of the internet of things, nodes are typically deployed at fixed locations with certain communication range limitations. The network topology initializes its network model according to a scaleless network model. In the network topology optimization strategy, most of researches adopt greedy edge-changing strategies or evolution algorithms to optimize the robustness of the network topology, so that the whole network has very high capability of resisting attacks. For example, journal Robustness optimization scheme with multi-delivery co-evolution for scale-free wireless sensor networks proposes a multi-population genetic algorithm to solve the problem of local optimization so as to obtain a globally optimal network topology, but the time cost for optimizing a network topology is large, and the algorithm cannot accumulate the optimization experience every time, so that the algorithm needs to be restarted every time. Secondly, some researchers use neural network models to represent learning behaviors before and after network topology optimization, so that the optimization time of the topology is reduced, but the method needs label target data, and the label target data defines the maximum value of optimization. Therefore, in the optimization of the topological structure of the Internet of things, the network topological structure strategy of autonomous learning optimization is utilized to improve the network topological robustness, the online of the optimization target value is eliminated, and the experience of each learning is accumulated to guide the subsequent optimization behavior.
Disclosure of Invention
In order to solve the problems in the prior art, the invention aims to provide a method for optimizing the robustness of the topological structure of the Internet of things by autonomous learning, wherein the topological structure of the Internet of things is used as an environment space according to the characteristics of reinforcement learning and deep learning, meanwhile, an action space is designed to search the environment, the effect of each action on the environment space is optimized, and the accumulated optimization effect is maximized, so that the robustness of the topological structure of the Internet of things is improved, and the autonomous learning behavior of the whole topological structure of the network is increased.
The invention discloses a method for optimizing the topological structure robustness of the Internet of things through autonomous learning, which comprises the following steps:
step 1: initializing an internet of things topological structure, namely randomly deploying nodes according to rules of a scaleless network model and an edge density parameter M, and fixing geographic positions, wherein the edge density parameter M is set to be M=2;
step 2: the method comprises the steps of compressing a topological structure, namely eliminating redundant node information which is not in a communication range, reserving only node connection relations in the communication range in a form of an adjacency matrix, compressing a storage space of the network topological structure, and taking the compressed network topology as an environment space S, wherein the environment space S is a row vector, and the row vector is changed along with the change of a network topology state;
step 3: initializing an autonomous learning model, namely constructing a deep deterministic learning strategy model according to the features of deep learning and reinforcement learning to train the topological structure of the Internet of things: adopting a depth deterministic Q-learning network model, simulating a selection strategy pi of actions and an optimization strategy Q of a network, mapping a continuous action space into a discrete action space, and designing an updating rule of a target optimization function O and the whole training model; wherein:
the action selection policy pi is defined by equation (1):
a t =π(s t |θ) (1)
wherein a is t Representing the selected deterministic action s t Representing the current network topology state, and theta represents the parameters of the action network;
the network optimization strategy Q is defined by equation (2):
Q(s t ,a t )=E(r(s t ,a t )+γQ(s t+1 ,π(s t+1 ))) (2)
wherein r represents the current action a t For the current network state s t And (c) the immediate return value of gamma represents the discount factor, the accumulated learning experience, Q(s) t+1 ,π(s t+1 ) A future return value representing an action taken in the next network state, and thus the effect Q(s) of the current action on the current network state t ,a t ) The method comprises the steps of forming an immediate return value and a future return value, wherein E () represents an expected value, and selecting an effect before strategy accumulation for a series of actions;
the objective function O of the autonomous learning model is defined by equation (3) according to the above description:
O(θ)=Ε(r 1 +γr 22 r 3 +...|π(,θ)) (3)
where r represents the sum of the actions for environment O (θ) =e (r 1 +γr 22 r 3 +. |pi (, θ)) produced effects, i.e., a return value, γ represents a discount factor, cumulative learning experience, pi (, θ) represents a policy of action selection, θ represents a parameter of an action policy network, and E represents an average expected value;
the update rules of the network are defined by equation (4):
wherein T is i Represents a target expected value, defined by the formula (5):
T i =r i +γQ'(s i ,π'(s i+1π’ )|θ Q’ ) (5)
in the formula, Q ', pi' represent target networks of action selection strategies and optimization strategies, and errors of the whole autonomous learning model are calculated;
step 4: training and testing a model, namely in a training stage, randomly obtaining a discrete action a through an action selection neural network model, evaluating the effect of the action on the current environment through a network optimization neural network model strategy, accumulating the previous learning experience and updating the whole network model, and finally obtaining an optimal result; in the test stage, testing sample data to obtain a test result; wherein:
the output of the discrete action is defined by equation (6):
d=MAP(a) (6)
wherein MAP represents a mapping relationship between a continuous motion space and a discrete motion space, and a is defined by formula (7):
a=π(s)=π(s|θ)+N (7)
wherein N represents a random sampling rule, and more effective action behaviors in the action space are explored;
wherein, the action selecting strategy network updating principle is updated towards the direction of making the strategy selecting network value maximum, so the selected action makes the strategy selecting network maximum;
the updating rule of the target network is defined as formula (10);
θ Q’ ←τθ Q +(1-τ)θ Q’
θ π’ ←τθ π +(1-τ)θ π’ (10)
where τ represents the update rate of the target network;
step 5: periodically repeating step 4 in one independent repetition experiment, and periodically repeating steps 1, 2, 3 and 4 in multiple independent repetition experiments; up to the maximum iteration times;
in the process, the maximum iteration times are set, the experiment is independently repeated each time, the optimal result is selected, the experiment is repeated for a plurality of times, and the average value is selected as the result of the experiment.
The positive technical effects obtained by the invention include:
(1) The invention designs an autonomous learning optimization strategy for the topological structure robustness of the Internet of things by utilizing the deep reinforcement learning neural network model, and can obviously improve the capability of the initial topological structure for resisting attack;
(2) The invention utilizes the state representation of the topological structure of the Internet of things, discrete action space mapping relation, the non-scale characteristic and compression characteristic of the network to come from the robust capability of main learning optimization network topological structure, thereby ensuring high-reliability data transmission.
Drawings
FIG. 1 is a flowchart of an overall method for optimizing the topological robustness of an Internet of things by autonomous learning;
FIG. 2 is a schematic diagram of a mapping relationship between continuous and discrete actions of an autonomous learning optimization model;
fig. 3 is a schematic diagram of a topological structure compression model of the internet of things.
Detailed Description
The following describes in detail the specific manner, structure, features and actions of a node deployment strategy designed according to the present invention with reference to the accompanying drawings.
As shown in FIG. 1, the overall flow chart of the method for optimizing the topological structure robustness of the Internet of things by autonomous learning is disclosed, and the method comprehensively considers the mapping relation between a large-scale continuous action space and a discrete action space, the compression mode of a network topological structure and the connection relation of nodes, effectively improves the network robustness, enhances the autonomous learning behavior of the overall network, balances the node connection distribution and ensures the high-quality communication capability of the network. The flow of the method specifically comprises the following steps:
step 1: initializing an Internet of things topological structure. And randomly deploying nodes according to the rule of the scale-free network model and the edge density parameter M, and fixing the geographic position. Most nodes have few degrees, few nodes have large degrees, and the real world Internet of things topological structure is described to the greatest extent. Each node has the same attributes.
Wherein the edge density parameter is set to m=2, indicating that the number of edges in the network is 2 times the number of nodes.
Step 2: the topology is compressed. Different from the adjacent matrix representation mode of the network topology structure, the invention eliminates redundant node information which is not in the communication range, only reserves the node connection relation in the communication range on the form of the adjacent matrix, compresses the storage space of the network topology structure, and takes the compressed network topology as the environment space S.
The environment space S is a row vector, which changes with the change of the network topology state.
Step 3: an autonomous learning model is initialized. And constructing a deep deterministic learning strategy model to train the topological structure of the Internet of things according to the deep learning and reinforcement learning characteristics. A depth deterministic Q-learning network model is adopted, a selection strategy pi of a simulation action and an optimization strategy Q of a network are adopted, a continuous action space is mapped into a discrete action space, and an updating rule of a target optimization function O and the whole training model is designed.
Wherein the action selection policy pi is defined by formula (1).
a t =π(s t |θ) (1)
Wherein a is t Representing the selected deterministic action s t Representing the current network topology state, θ represents a parameter of the action network. Current network topology state s t And obtaining a deterministic action through the action policy function pi, wherein the deterministic action can directly operate on the current network topology structure.
The network optimization strategy Q is defined by a formula (2), and the effect of the selected action on the environment space is measured.
Q(s t ,a t )=E(r(s t ,a t )+γQ(s t+1 ,π(s t+1 ))) (2)
Where r represents the current action a t For the current network state s t Gamma is the discount factor, and the accumulated learning experience. Q(s) t+1 ,π(s t+1 ) A future return value representing an action taken in the next network state, and thus the effect Q(s) of the current action on the current network state t ,a t ) Consists of an immediate return value and a future return value, E () represents the expected value, and the effect before policy accumulation is selected for a series of actions.
Wherein the objective function O of the autonomous learning model can be defined specifically as formula (3) according to the above description
(3)
Where r represents the sum of the actions for environment O (θ) =e (r 1 +γr 22 r 3 +. |pi (, θ)), the effect produced, i.e., the return value. Gamma denotes the discount factor, and the learning experience is accumulated. Pi (, θ) represents the policy of action selection, θ is a parameter of the action policy network. E represents the average expected value as an objective function of the entire autonomous learning model.
Wherein the update rule of the network is defined by equation (4).
Wherein T is i The target expected value is represented by the expression (5).
T i =r i +γQ'(s i ,π'(s i+1π’ )|θ Q’ ) (5)
In the formula, Q ', pi' represents a target network of an action selection strategy and an optimization strategy, and the error of the whole autonomous learning model is calculated.
Step 4: training and testing the model. In the training stage, a discrete action a is randomly obtained through action selection neural network model, the effect of the action on the current environment is evaluated through a network optimization neural network model strategy, the previous learning experience is accumulated, the whole network model is updated, and finally the optimal result is obtained. And in the test stage, testing the sample data to obtain a test result.
Wherein the output of the discrete action is defined by equation (6).
d=MAP(a) (6)
Wherein MAP represents a mapping relationship between a continuous motion space and a discrete motion space, and a is defined by formula (7):
a=π(s)=π(s|θ)+N
(7)
where N represents a random sampling rule, exploring more effective action behaviors in the action space.
After obtaining the effect of the action on the current environment, namely the return value, the return value is stored in a memory, the convergence rate of the autonomous learning model is accelerated by using the previous learning experience in the candidate optimization learning, and the rule is defined as formula (8).
(s t ,a t ,r t ,s t+1 )→D (8)
Wherein D represents a memory in the network model, and stores the current network state, action, instant return value, and next network state.
Wherein, in the stage of updating the autonomous learning optimization model, the action selection policy network updating rule is defined by the formula (9).
▽π=E π' [▽ a Q(s,a)▽ θ π(s)] (9)
In the formula, the action selection policy network updating principle is updated towards the direction of maximizing the value of the policy selection network, so that the selected action maximizes the policy selection network.
Wherein the update rule of the target network is defined as formula (10).
θ Q’ ←τθ Q +(1-τ)θ Q’
θ π’ ←τθ π +(1-τ)θ π’ (10)
Where τ represents the update rate of the target network.
Step 5: periodically repeating step 4 in one independent repetition experiment, and periodically repeating steps 1, 2, 3 and 4 in multiple independent repetition experiments; up to a maximum number of iterations. In the process, the maximum iteration number is set, and the experiment is independently repeated each time, and the optimal result is selected. The experiment was repeated several times and the average value was chosen as the result of this experiment.

Claims (1)

1. The method for autonomously learning and optimizing the topological structure robustness of the Internet of things is characterized by comprising the following steps of:
step 1: initializing an internet of things topological structure, namely randomly deploying nodes according to rules of a scaleless network model and an edge density parameter M, and fixing geographic positions, wherein the edge density parameter M is set to be M=2;
step 2: the method comprises the steps of compressing a topological structure, namely eliminating redundant node information which is not in a communication range, reserving only node connection relations in the communication range in a form of an adjacency matrix, compressing a storage space of the network topological structure, and taking the compressed network topology as an environment space S, wherein the environment space S is a row vector, and the row vector is changed along with the change of a network topology state;
step 3: initializing an autonomous learning model, namely constructing a deep deterministic learning strategy model according to the features of deep learning and reinforcement learning to train the topological structure of the Internet of things: adopting a depth deterministic Q-learning network model, simulating a selection strategy pi of actions and an optimization strategy Q of a network, mapping a continuous action space into a discrete action space, and designing an updating rule of a target optimization function O and the whole training model; wherein:
the action selection policy pi is defined by equation (1):
a t =π(s t |θ) (1)
wherein a is t Representing the selected deterministic action s t Representing the current network topology state, and theta represents the parameters of the action network;
the network optimization strategy Q is defined by equation (2):
Q(s t ,a t )=E(r(s t ,a t )+γQ(s t+1 ,π(s t+1 ))) (2)
wherein r represents the current action a t For the current network state s t And (c) the immediate return value of gamma represents the discount factor, the accumulated learning experience, Q(s) t+1 ,π(s t+1 ) A future return value representing an action taken in the next network state, and thus the effect Q(s) of the current action on the current network state t ,a t ) The method comprises the steps of forming an immediate return value and a future return value, wherein E () represents an expected value, and selecting an effect before strategy accumulation for a series of actions;
the objective function O of the autonomous learning model is defined by equation (3) according to the above description:
O(θ)=Ε(r 1 +γr 22 r 3 +...|π(,θ)) (3)
where r represents the sum of the actions for environment O (θ) =e (r 1 +γr 22 r 3 +. |pi (, θ)) produced effects, i.e., a return value, γ represents a discount factor, cumulative learning experience, pi (, θ) represents a policy of action selection, θ represents a parameter of an action policy network, and E represents an average expected value; θ
The update rules of the network are defined by equation (4):
wherein T is i Represents a target expected value, which is determined by the formula (5)The meaning is as follows:
T i =r i +γQ'(s i ,π'(s i+1π′ )|θ Q′ ) (5)
in the formula, Q ', pi' represent target networks of action selection strategies and optimization strategies, and errors of the whole autonomous learning model are calculated;
step 4: training and testing a model, namely in a training stage, randomly obtaining a discrete action a through an action selection neural network model, evaluating the effect of the action on the current environment by a network optimization neural network model strategy, accumulating the previous learning experience and updating the whole network model, and finally obtaining an optimal result; in the test stage, testing sample data to obtain a test result; wherein:
the output of the discrete action is defined by equation (6):
d=MAP(a) (6)
wherein MAP represents a mapping relationship between a continuous motion space and a discrete motion space, and a is defined by formula (7):
a=π(s)=π(s|θ)+N (7)
wherein N represents a random sampling rule, more effective action behaviors in an action space are explored, and s represents the current network state;
wherein, the action selecting strategy network updating principle is updated towards the direction of making the strategy selecting network value maximum, so the selected action makes the strategy selecting network maximum;
the updating rule of the target network is defined as formula (10);
θ Q′ ←τθ Q +(1-τ)θ Q′
θ π′ ←τθ π +(1-τ)θ π′ (10)
where τ represents the update rate of the target network;
step 5: periodically repeating step 4 in one independent repetition experiment, and periodically repeating steps 1, 2, 3 and 4 in multiple independent repetition experiments; up to the maximum iteration times;
in the process, the maximum iteration times are set, the experiment is independently repeated each time, the optimal result is selected, the experiment is repeated for a plurality of times, and the average value is selected as the result of the experiment.
CN201911036835.1A 2019-10-29 2019-10-29 Method for autonomously learning and optimizing topological structure robustness of Internet of things Active CN110807230B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911036835.1A CN110807230B (en) 2019-10-29 2019-10-29 Method for autonomously learning and optimizing topological structure robustness of Internet of things

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911036835.1A CN110807230B (en) 2019-10-29 2019-10-29 Method for autonomously learning and optimizing topological structure robustness of Internet of things

Publications (2)

Publication Number Publication Date
CN110807230A CN110807230A (en) 2020-02-18
CN110807230B true CN110807230B (en) 2024-03-12

Family

ID=69489419

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911036835.1A Active CN110807230B (en) 2019-10-29 2019-10-29 Method for autonomously learning and optimizing topological structure robustness of Internet of things

Country Status (1)

Country Link
CN (1) CN110807230B (en)

Families Citing this family (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111929641B (en) * 2020-06-19 2022-08-09 天津大学 Rapid indoor fingerprint positioning method based on width learning
CN111935724B (en) * 2020-07-06 2022-05-03 天津大学 Wireless sensor network topology optimization method based on asynchronous deep reinforcement learning
CN113422695B (en) * 2021-06-17 2022-05-24 天津大学 Optimization method for improving robustness of topological structure of Internet of things
CN113435567B (en) * 2021-06-25 2023-07-07 广东技术师范大学 Intelligent topology reconstruction method based on flow prediction, electronic equipment and storage medium
CN113923123B (en) * 2021-09-24 2023-06-09 天津大学 Underwater wireless sensor network topology control method based on deep reinforcement learning
CN114567563B (en) * 2022-03-31 2024-04-12 北京邮电大学 Training method of network topology model, and reconstruction method and device of network topology
CN115225509B (en) * 2022-07-07 2023-09-22 天津大学 Internet of things topological structure generation method based on neural evolution

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102789593A (en) * 2012-06-18 2012-11-21 北京大学 Intrusion detection method based on incremental GHSOM (Growing Hierarchical Self-organizing Maps) neural network
CN102868972A (en) * 2012-09-05 2013-01-09 河海大学常州校区 Internet of things (IoT) error sensor node location method based on improved Q learning algorithm
CN103490413A (en) * 2013-09-27 2014-01-01 华南理工大学 Intelligent electricity generation control method based on intelligent body equalization algorithm

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7130807B1 (en) * 1999-11-22 2006-10-31 Accenture Llp Technology sharing during demand and supply planning in a network-based supply chain environment
US10715395B2 (en) * 2017-11-27 2020-07-14 Massachusetts Institute Of Technology Methods and apparatus for communication network

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102789593A (en) * 2012-06-18 2012-11-21 北京大学 Intrusion detection method based on incremental GHSOM (Growing Hierarchical Self-organizing Maps) neural network
CN102868972A (en) * 2012-09-05 2013-01-09 河海大学常州校区 Internet of things (IoT) error sensor node location method based on improved Q learning algorithm
CN103490413A (en) * 2013-09-27 2014-01-01 华南理工大学 Intelligent electricity generation control method based on intelligent body equalization algorithm

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
伯努利节点网络模型的拓扑鲁棒性分析方法;冯涛;李洪涛;袁占亭;马建峰;电子学报;第39卷(第7期);全文 *
基于强化学习的无线网络自组织性研究;王超;王芷阳;沈聪;;中国科学技术大学学报(第12期);全文 *
面向SDN网络虚拟化平台的控制器放置算法;董晓东;郭志强;陈胜;周晓波;齐恒;李克秋;电信科学(第004期);全文 *

Also Published As

Publication number Publication date
CN110807230A (en) 2020-02-18

Similar Documents

Publication Publication Date Title
CN110807230B (en) Method for autonomously learning and optimizing topological structure robustness of Internet of things
CN112541302B (en) Air quality prediction model training method, air quality prediction method and device
Straatman et al. Towards an automatic calibration procedure for constrained cellular automata
CN111935724B (en) Wireless sensor network topology optimization method based on asynchronous deep reinforcement learning
CN110267292A (en) Cellular network method for predicting based on Three dimensional convolution neural network
CN103905246B (en) Link prediction method based on grouping genetic algorithm
CN104636801A (en) Transmission line audible noise prediction method based on BP neural network optimization
CN110826244B (en) Conjugated gradient cellular automaton method for simulating influence of rail transit on urban growth
CN116207739B (en) Optimal scheduling method and device for power distribution network, computer equipment and storage medium
CN110619082B (en) Project recommendation method based on repeated search mechanism
CN117498400A (en) Distributed photovoltaic and energy storage data processing method and system
Gallego et al. Maintaining flexibility in smart grid consumption through deep learning and deep reinforcement learning
CN113824641A (en) Internet of vehicles routing method, system, equipment and storage medium
CN114861917A (en) Knowledge graph inference model, system and inference method for Bayesian small sample learning
CN117810987A (en) Prediction method, prediction device, prediction terminal and prediction storage medium for residential payload section
CN112116305A (en) Power grid probability visualization model construction method and system for machine learning
CN116167254A (en) Multidimensional city simulation deduction method and system based on city big data
CN115395502A (en) Photovoltaic power station power prediction method and system
CN114372418A (en) Wind power space-time situation description model establishing method
CN114566048A (en) Traffic control method based on multi-view self-adaptive space-time diagram network
CN112884129A (en) Multi-step rule extraction method and device based on teaching data and storage medium
Cong et al. Research on Satellite Network Traffic Prediction Algorithm Based on Gray Wolf Algorithm Optimizing GRU and Spatiotemporal Analysis
Fang et al. Towards data-centric control of sensor networks through bayesian dynamic linear modelling
CN116540678B (en) Industrial equipment state deduction evaluation method based on digital twin
CN118171585B (en) Super-long pile foundation structure design and quality control method and system

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant