AU2019101412A4 - A distributed adaptive clustering strategy algorithm - Google Patents

A distributed adaptive clustering strategy algorithm Download PDF

Info

Publication number
AU2019101412A4
AU2019101412A4 AU2019101412A AU2019101412A AU2019101412A4 AU 2019101412 A4 AU2019101412 A4 AU 2019101412A4 AU 2019101412 A AU2019101412 A AU 2019101412A AU 2019101412 A AU2019101412 A AU 2019101412A AU 2019101412 A4 AU2019101412 A4 AU 2019101412A4
Authority
AU
Australia
Prior art keywords
node
task
algorithm
mcc
clustering
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Ceased
Application number
AU2019101412A
Inventor
Feng Chen
Rui Hu
Zhifeng Liu
Qing Shi
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Southwest University
Original Assignee
Southwest University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Southwest University filed Critical Southwest University
Priority to AU2019101412A priority Critical patent/AU2019101412A4/en
Application granted granted Critical
Publication of AU2019101412A4 publication Critical patent/AU2019101412A4/en
Ceased legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • G06N20/10Machine learning using kernel methods, e.g. support vector machines [SVM]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F15/00Digital computers in general; Data processing equipment in general
    • G06F15/16Combinations of two or more digital computers each having at least an arithmetic unit, a program unit and a register, e.g. for a simultaneous processing of several programs
    • G06F15/161Computing infrastructure, e.g. computer clusters, blade chassis or hardware partitioning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/48Program initiating; Program switching, e.g. by interrupt
    • G06F9/4806Task transfer initiation or dispatching
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W4/00Services specially adapted for wireless communication networks; Facilities therefor
    • H04W4/70Services for machine-to-machine communication [M2M] or machine type communication [MTC]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W84/00Network topologies
    • H04W84/18Self-organising networks, e.g. ad-hoc networks or sensor networks

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Software Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Hardware Design (AREA)
  • Mathematical Physics (AREA)
  • Evolutionary Computation (AREA)
  • Medical Informatics (AREA)
  • Data Mining & Analysis (AREA)
  • Computing Systems (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Artificial Intelligence (AREA)
  • Data Exchanges In Wide-Area Networks (AREA)

Abstract

Abstract: A distributed adaptive clustering algorithm over dynamic multi-task networks, the algorithm mainly comprising two processes including normal task adaptation and the same task clustering. In the process of abnormal task adaptation, the algorithm use the task anomaly detection based on the non cooperative least mean square (NC-LMS) algorithm; in the same task clustering process, the algorithm use the task based on the diffusion maximum entropy criterion (D-MCC) algorithm. Switch detection. In multi-task networks, a series of scenarios, such as dynamic network, time varying tasks and non-stationary (Gaussian and pulse interference). The algorithm also discuss optimization schemes to design the NC-LMS and D-MCC weights and examine the estimate performance and clustering effects of the proposed algorithm by simulation results. Compared to the prior art the present invention provides an effective distributed adaptive clustering strategy to solve distributed adaptive estimation over dynamic multi-task networks. Besides, since nodes in the network are constrained by communication power consumption and external interference in a non stationary environment, the objective pursued by the node is prone to change or abnormality. Fig.1 40- 1 a 0 20 40 60 80 100 120 x-coord nate Fig.2 x7o70 0 20 40 60 80 100 120 x-coord inate

Description

1. Background and Purpose
Distributed estimation for adaptation, learning, modeling, and optimization through cooperation between nodes plays a key role in reinforcement learning, signal processing, and online supervised learning and many other application areas, which aims to estimate a single parameter vector collaboratively. However, in reality, there are many parameters of interest happening to be multitask-oriented. In other words, there are multiple optimum parameter vectors that are simultaneously inferred in a collaborative manner. Multi-task problems have been studied in many important applications, such as multi-task clustering, multi-target tracking, and multi-model classification. In our work, we consider the situation where there are connected clusters of nodes, and each cluster has a parameter vector to estimate.
Recent years, several useful distributed strategies have been proposed, including incremental strategies, consensus strategies and diffusion strategies. In particular, diffusion strategies are attracted by many researchers because of their scalability and reliability. It is worth noting that has proved that the diffusion strategy in data processing on adaptive networks has better stability and performance range than consensus-based strategies. Accordingly, the diffusion adaptive learning algorithm is mainly considered in our work. Adaptive net-works are well-suited for decentralized inference, filtering and clustering tasks. However, previous work on topology design and tuning techniques included research that was not dynamic, and in the sense that they cannot track changes in the network. Motivated by the problem, we develop an adaptive clustering algorithm over dynamic multi-task network in this paper, which can reduce the impact of weak links on network estimation by selecting data subsets from neighbor nodes with normal tasks.
The purpose of this work is to motivate and develop a distributed clustering strategy algorithm based on diffusion MCC, for robust distributed multi-task network estimation in a nonstationary environment. Motivated by, the algorithm consider a general situation where there are connected clusters of nodes, and each cluster has a parameter vector to estimate. In addition, all nodes in the network do not know in advance which cluster they belong to, and they also do not know which neighbor are interested in their task. In summary, the main contributions are as follows:
1) Algorithm implementation: task anomaly detection and task switching detection steps in the clustering strategy to improve clustering accuracy are added to improve the accuracy of clustering;
2) Simulation: we simulate a variety of scenarios and investigate the estimated performance and clustering effect of the proposed algorithm.
2. System Model
Each node k in the connected network observes random measurements {dki,uki}, where dk i
2019101412 22 Nov 2019 is a scalar data and uk i is 1XL regression vector data, which are assumed to be related to some unknown Lx 1 parameter vector wki by a linear regression model of the form:
dk,i=ukA,i+vk,i (1)
Where uki at time instant i, is temporally white and independent over space with covariance matrix Ruk= Euk > 0 and zero means. vki is an additive temporally and spatially independent zero-mean noise process with a time-independent variance fc, and it is independent of every other signal over space. It is considered that nodes of different clusters track different objectives (which also call tasks), and there is one task per cluster, namely, w°k=wcpfor^kE C q (2)
The task time-varying model is:
= skA-i + (1 - SXWCP + Zk,i-1 · (3)
Where zki_k is the process noise for node k at time instant i — 1, which is independent of measurement noise vki and regression vector uk i. Let sk i denote a random indicator variable, i.e., ί l,if cki_l >cr otherwise ’ W where ck i_k is a positive potential communication cost for each node k at time instant i — 1, and cr represents the network tolerable communication cost threshold. It is given by Ck ;_j — Conk ;_j, where c0 is the communication cost between connected nodes.
3. Distributed Clustering Algorithm Description
3.1 The task anomaly detection based on non-cooperative least-mean-squares (NC-LMS) algorithm
Step 1 : Each node k in the network can updates the estimate learning strategy, namely,
K = Ka -/Pk^k,-nk,Kk,i-M,i (5)
Due to process noise interference, especially impulse noise, task of nodes may be abnormal, sending or receiving abnormal tasks not only increases power consumption but also reduces the accuracy of parameter estimation. A hypothesis test based on the updated estimate Wk; is developed to ascertain whether the task of node k is abnormal.
Step 2: Detection for abnormal tasks (6) where the Ho hypothesis denotes the task of node k is normal, and node k sends data {dk i, uk i } to
2019101412 22 Nov 2019 neighbors I. Conversely, the hypothesis Hk denotes the task of node k is abnormal, and node k does not send data {dki, uki} to neighbors I . The threshold θ0 is predefined. Besides, no exchange of data for abnormal task is needed during the adaptation, which makes the communication cost relatively low.
3.2 The task switching detection based on diffusion maximum correntropy criterion (D-MCC) algorithm
Step 1 : To reduce the impact of noise on the estimate, each node k in the network can update the intermediate estimate through the diffusion learning strategy over MCC.
V^i = ^k,i-l + Pk Σ j
I. = Σ cikWi,i
I (7)
Step 2: The correntropy between two random variables X and y is associated with a generalized correlation function, which scales the similarity of X and y via ^)=E[^exp(_w<)], (8) where β is the Gaussian kernel size.
Step 3: With Gaussian kernel and local error ek t — dki~ Uk iWk ._j , the instantaneous correntropy cost function is e2
6>CC(%) = JkCC(^ = E[^=exp(-^)]. (9)
By using instantaneous approximation, an approximation of the gradient vector is (10)
The diffusion algorithm based on MCC (D-MCC) is ¥k,i = wk,i-t + Σ aikGffCC (ei2eu uli’
P kNt where G^cc(q1 f is a Gaussian kernel, as kernel size /? —>oo, then G^cc(ez;) —> 1. Note that the kernel size for correntropy function is quite important, and the Gaussian kernel G^cc(e(z)) versus local error e(z) for different values of size.
Step 4: The neighborhood set Nk will be time-dependent and expressed as , and
2019101412 22 Nov 2019 combination coefficient alk t is ||2<6»0 \(Q, otherwise
The algorithm D-MCC is
Ψ k,i = Wk,i-l+rlk aik,i&p (Ql,i)Ql,iUl,i’ t=Nkj where ηk = —is the step size.
(12) (13)
Step 5: The weights can be used to minimize the instantaneous mean-square deviation (MSD) of the network:
minMW)^£E||wy-wJ| ]y Λ=1 (14)
The combination coefficients Clk t can be obtained, which can be approximated by
otherwise where Nk t = Nk \ {k} . The combination rule gives larger weights to neighbors with common cluster and smaller weights to neighbors that come from different clusters. Then, the combination step is rewritten by wkj = Σ feN.,· (16)
Step 6: Using these dynamically-evolving estimates, we introduce another hypothesis test based on the updated estimate wk t, which is developed to ascertain whether the tasks of node k and node I are the same at time z.
II II2 §01 (17)
Hj
Where the threshold θ3 is predefined. The Ho hypothesis denotes the tasks of node k and node
I are the same, and the link between node k with neighbor I are active. Conversely, the hypothesis
2019101412 22 Nov 2019
Hj denotes the tasks of node k and node I are different, and the links between node k with neighbor I are dropped. Then, the cluster connection coefficient lkl; given by i \,if HL success lc .= lc . = 1 J 0 a,, ik,i \Q,otherwise
4. Brief Description of the Drawings
Fig. 1 is network initial topology.
Fig.2 is the resulting topology of the subnetwork over no task switching or exception.
Fig.3 is the subnetwork average MSDs for NC-LMS and D-MCC over no task switching or exception.
Fig.4 is the resulting topology of the subnetwork over task switching and exception.
Fig.5 is the subnetwork average MSDs for NC-LMS and D-MCC overtask switching and exception.
5. Detailed Description
5.1 Model Validation
The topology of the network consisting of N = 20 nodes divided into Q = 3 clusters, i.e., 6!= {1, 2, 3, 4, 5}, C2= {7, 8, 9, 10, 11, 12, 13, 14}, and C3= {15, 16, 17, 18, 19, 20}, with connection is generated as a random geometric graph model as shown in Fig. 1. The location coordinates (xk], yk j) of each node k in the square region [0, 110] x [0, 110]. In the time-varying scenario, they vary according to the first-order Markov vector process:
xk(i) = bxk(i - Ϊ) + h(i), yk (X = byk O’-1) + h(i).
Where b = 0.98 and /z(z) is an independent zero-mean Gaussian vector process with variance:
2 Γ 0.01 —> Small interference h [1 —> Big interference
5.2 Illustrative Example
Computer simulations are carried out to evaluate the performance of the proposed algorithm, under the assumption that all nodes have no prior knowledge about the clusters, clustering effects of the proposed method in the following two scenarios are illustrated. Scenario 1: The location coordinates of nodes in network fluctuate is gentle (the network structure changes slightly, i.e.,o^ = 0.01), then for zero-mean Gaussian interference with k, = 0 and σ|. = 0.5, the task of nodes does not switch and cannot evolve to exceptions. Scenario 2: The location coordinates of nodes in network fluctuate is dramatic (the network structure changes wildly, i.e., = 1), then for Impulse
2019101412 22 Nov 2019 interference with and — 103, the task of nodes is switched and abnormal.
Scenario l(No task switching or exception)
The network node position fluctuates slightly, and it suffers from constrained Gaussian interference. After approximate 400 iterations of MSD curves, the clustering decision of the proposed algorithm does not change with time. The neighboring links within the same cluster are active whereas the neighboring links, which come from different clusters, are dropped. Fig. 2 illustrates the resulting topology when the network is in steady-state. From the simulation results, we find that there is no task switching or exception when the nodes are under Gaussian interference with small constraint. The MSD learning curves for the proposed clustering algorithm consisting of the recursions NC-LMS and D-MCC are plotted in Fig. 3. It is obvious that three clusters take MCC cooperation clustering policy to improve their MSD performance on average.
Scenario 2(Task switching and exception)
The network node position fluctuates wildly, and it suffers from Impulse interference. After approximate 400 iterations of MSD curves, similarly, the clustering decision of the proposed algorithm does not change with time. In addition, the neighboring links within the same cluster are active, whereas the neighboring links that come from different clusters, are dropped. Fig. 4 illustrates that the three subnetworks are themselves connected when net-work is at steady-state. There is a similar implied result with scenario 1 that the proposed clustering strategy can suppress the interference between clusters. From Fig. 4, we can see that there are task switching and exceptions when the nodes are Impulse interference. The MSD learning curves for the proposed clustering algorithm consisting of the recursions NC-LMS and D-MCC are plotted in Fig. 5. Obviously, the proposed algorithm for recursion D-MCC has a superior performance in comparison with recursion NC-LMS.

Claims (1)

  1. The claims defining the invention are as follows:
    A distributed adaptive clustering strategy algorithm
    1. Distributed Clustering Algorithm Description
    1.1 The task anomaly detection based on non-cooperative least-mean-squares (NC-LMS) algorithm
    Step 1 : Each node k in the network can updates the estimate learning strategy, namely, =+pk (<k - ^kykymT k. (5)
    Due to process noise interference, especially impulse noise, task of nodes may be abnormal, sending or receiving abnormal tasks not only increases power consumption but also reduces the accuracy of parameter estimation. A hypothesis test based on the updated estimate Wk t is developed to ascertain whether the task of node k is abnormal.
    Step 2: Detection for abnormal tasks
    11 II2 (6) where the Ho hypothesis denotes the task of node k is normal, and node k sends data {dki, uki} to neighbors I. Conversely, the hypothesis denotes the task of node k is abnormal, and node k does not send data {dk i, uk i} to neighbors I . The threshold θ0 is predefined. Besides, no exchange of data for abnormal task is needed during the adaptation, which makes the communication cost relatively low.
    1.2 The task switching detection based on diffusion maximum correntropy criterion (D-MCC) algorithm
    Step 1 : To reduce the impact of noise on the estimate, each node k in the network can update the intermediate estimate through the diffusion learning strategy over MCC.
    = +A- Σ j
    I ™k,i = Σ wu l fciVi (7)
    Step 2: The correntropy between two random variables X and y is associated with a generalized correlation function, which scales the similarity of X and y via (8) where β is the Gaussian kernel size.
    Step 3: With Gaussian kernel and local error ek, — dki~ Uk iWk , the instantaneous correntropy cost function is
    2019101412 22 Nov 2019
    1e = JkCC^fl = Ε[-^=οχρ(-^_)].(9)
    By using instantaneous approximation, an approximation of the gradient vector is )» G“cc (e t,) e,., <(10)
    The diffusion algorithm based on MCC (D-MCC) is = wk,i-i + ui,i’ (11)
    P ‘N where is a Gaussian kernel, as kernel size /?—>oo, then G^cc(Qt;.) —> 1. Note that the kernel size for correntropy function is quite important, and the Gaussian kernel G^CC(e(z’)) versus local error e(z) for different values of size.
    Step 4: The neighborhood set Nk will be time-dependent and expressed as Nki, and combination coefficient Cllk t is
    Il2<#o
    110, otherwise (12)
    The algorithm D-MCC is
    Ψk,i Wk,i-l+rlk Σ aik,iGf (13) where ηk = —y is the step size.
    Step 5: The weights can be used to minimize the instantaneous mean-square deviation (MSD) of the network:
    minMSD(i) & ±£e || <- wk, ||. (14)
    The combination coefficients C/k t can be obtained, which can be approximated by
    ΙΙ^,,·-^,,·-! ΙΓ
    Σ II Ψί,ι- ^,/-ι ΙΓ2 ί^, ’if Ιε Nk.
    (15) ι- Σ me. NT i
    otherwise
    Where Nk t = Nk \ {A:}. The combination rule gives larger weights to neighbors with common cluster and smaller weights to neighbors that come from different clusters. Then, the combination step is rewritten by = Σ ^,/^,,(16)
    Step 6: Using these dynamically-evolving estimates, we introduce another hypothesis test based on the updated estimate wk;, which is developed to ascertain whether the tasks of node k and node I are the same at time z.
    II wk,i~wi,i II2 (17)
    Where the threshold θ3 is predefined. The Ho hypothesis denotes the tasks of node k and node I are the same, and the link between node k with neighbor I are active. Conversely, the hypothesis Hj denotes the tasks of node k and node I are different, and the links between node k with neighbor I are dropped. Then, the cluster connection coefficient lkI t given by ( 1, if HL success lc , = lc . = 1 J 0 kl,lk,‘ |θ, otherwise (18)
AU2019101412A 2019-11-17 2019-11-17 A distributed adaptive clustering strategy algorithm Ceased AU2019101412A4 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
AU2019101412A AU2019101412A4 (en) 2019-11-17 2019-11-17 A distributed adaptive clustering strategy algorithm

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
AU2019101412A AU2019101412A4 (en) 2019-11-17 2019-11-17 A distributed adaptive clustering strategy algorithm

Publications (1)

Publication Number Publication Date
AU2019101412A4 true AU2019101412A4 (en) 2020-01-02

Family

ID=68982453

Family Applications (1)

Application Number Title Priority Date Filing Date
AU2019101412A Ceased AU2019101412A4 (en) 2019-11-17 2019-11-17 A distributed adaptive clustering strategy algorithm

Country Status (1)

Country Link
AU (1) AU2019101412A4 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113987860A (en) * 2021-10-08 2022-01-28 中山大学 Structure optimization method and device based on dynamic Gaussian kernel convolution filtering

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113987860A (en) * 2021-10-08 2022-01-28 中山大学 Structure optimization method and device based on dynamic Gaussian kernel convolution filtering

Similar Documents

Publication Publication Date Title
Casbeer et al. Distributed information filtering using consensus filters
Pin et al. Robust deadbeat continuous-time observer design based on modulation integrals
Farina et al. Identification of polynomial input/output recursive models with simulation error minimisation methods
Gholami et al. Denoising and detection of bad data in distribution phasor measurements using filtering, clustering, and Koopman mode analysis
AU2019101412A4 (en) A distributed adaptive clustering strategy algorithm
Panigrahi et al. Error saturation nonlinearities for robust incremental LMS over wireless sensor networks
Bazzi et al. A robust diffusion adaptive network based on the maximum correntropy criterion
Yu et al. Temporal and spatial correlation based distributed fault detection in wireless sensor networks
Liu et al. Risk of Cascading Failures in Multi-agent Rendezvous with Communication Time Delay
Mahmoud Distributed estimation based on information‐based covariance intersection algorithms
Simonetto et al. Distributed time-varying stochastic optimization and utility-based communication
Liu et al. Gaussian process learning for distributed sensor networks under false data injection attacks
Mustafa et al. Analysis and detection of cyber-physical attacks in distributed sensor networks
Jianhong et al. A neural network based approach for modeling of severity of defects in function based software systems
Zhang et al. Quantifying the bias dynamics in a mode-based Kalman filter for stochastic hybrid systems
Wang et al. Distributed two‐stage state estimation with event‐triggered strategy for multirate sensor networks
Kawano et al. Revisit input observability: A new approach to attack detection and privacy preservation
Chen et al. Performance analysis of diffusion LMS in multitask networks
Pillonetto et al. Kernel-based model order selection for identification and prediction of linear dynamic systems
Hajshirmohamadi et al. Event-triggered simultaneous fault detection and consensus control for linear multi-agent systems
Chen et al. Universal structural estimator and dynamics approximator for complex networks
Stanković et al. Distributed change detection based on a randomized consensus algorithm
Boem et al. Distributed model-based fault diagnosis with stochastic uncertainties
Zhao et al. Dynamic joint outage identification and state estimation in power systems
Yazidi et al. A stochastic search on the line-based solution to discretized estimation

Legal Events

Date Code Title Description
FGI Letters patent sealed or granted (innovation patent)
MK22 Patent ceased section 143a(d), or expired - non payment of renewal fee or expiry