CN113947210A - Cloud side end federal learning method in mobile edge computing - Google Patents

Cloud side end federal learning method in mobile edge computing Download PDF

Info

Publication number
CN113947210A
CN113947210A CN202111169752.7A CN202111169752A CN113947210A CN 113947210 A CN113947210 A CN 113947210A CN 202111169752 A CN202111169752 A CN 202111169752A CN 113947210 A CN113947210 A CN 113947210A
Authority
CN
China
Prior art keywords
round
optimal
model
training
representing
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202111169752.7A
Other languages
Chinese (zh)
Other versions
CN113947210B (en
Inventor
毕远国
项天敖
闫飞宏
王博洋
郭朋超
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Northeastern University China
Original Assignee
Northeastern University China
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Northeastern University China filed Critical Northeastern University China
Priority to CN202111169752.7A priority Critical patent/CN113947210B/en
Publication of CN113947210A publication Critical patent/CN113947210A/en
Application granted granted Critical
Publication of CN113947210B publication Critical patent/CN113947210B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Software Systems (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Medical Informatics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Artificial Intelligence (AREA)
  • Mobile Radio Communication Systems (AREA)

Abstract

The invention belongs to the technical field of edge computing, and provides a cloud edge side federal learning method in mobile edge computing, which is used for reducing communication turns and improving the reliability of a trained machine learning model in a three-layer edge computing framework. By having the edge server participate in the federal learning training process, the communication round between the edge server and the terminal device is minimized by dynamically adjusting the training round in the terminal device. The problem of model parameter loss caused by channel bandwidth and other factors in the process of transmitting the model is reduced by minimizing the communication turn between the edge server and the terminal equipment.

Description

Cloud side end federal learning method in mobile edge computing
Technical Field
The invention belongs to the technical field of edge computing, and relates to a cloud edge federal learning method in mobile edge computing.
Background
With the development of smart devices and communication technologies, over 1250 billions of devices were deployed in various fields such as industrial internet, smart car, health care, etc. by 2030 according to the prediction of IHS Markit. Applications in these areas often have strict requirements on response delay compared to traditional multimedia services, making deployment of such services at the edge of the network an on-going need.
Mobile Edge Computing (MEC) has become a promising technology to support the Quality of Service (QoS) requirements described above, which can provide real-time services with local Computing power, low latency, and large bandwidth communications. Meanwhile, as a distributed machine Learning paradigm, Federal Learning (FL) has a great potential in MEC to provide edge intelligence and privacy protection. And the terminal devices participating in the joint learning process cooperatively establish a global high-performance model, and simultaneously retain training data locally. One FL procedure consists of several rounds of training procedure, each round only selecting a part of the terminal devices to participate in the training procedure. The participating terminal equipment trains the model by using local data, and then the cloud server summarizes the received model to form a global model.
The primary advantage of FL is to decouple model training from direct access to the original training data. Without data exchange, the FL procedure preserves data privacy, reducing the risk of data leakage. Based on this feature, FL is widely used in various fields from business companies to academic organizations, such as google, Webank, IEEE, and the like. Meanwhile, FL integrated with MEC has been studied in many research areas, such as blockchain, deep reinforcement learning, and virtual network function auto scaling.
However, integrating the FL with the MEC presents many challenges due to limited bandwidth, unreliable wireless channels, and frequent handovers. On the one hand, many devices contend for access to the wireless channel for data transmission, resulting in severe interference or packet collisions. On the other hand, doppler effect and frequent handovers caused by mobility of the terminal equipment also greatly increase the probability of packet loss in the MEC. Since the terminal device participates in the FL process and exchanges parameters with the edge server, the unstable wireless connection reduces the reliability of the model and even makes the model unavailable, thereby degrading the quality of the global aggregation model.
Disclosure of Invention
In order to improve the reliability of the model, a positive correlation relationship is established between the possibility of model parameter loss and the communication turn between the terminal equipment and the edge server by utilizing the three-layer structure of the MEC. The existence of an optimal communication turn and the corresponding number of hours of duration that minimizes the possibility of parameter loss are demonstrated by theoretical analysis. The minimum communication turn and the corresponding estimation of the optimal local training turn are obtained by adjusting some key parameters, and then the edge server is used for dynamically controlling the training turn of the selected terminal equipment to reduce the communication turn.
The optimal communication round and corresponding duration are derived theoretically based on a positive correlation of communication round and packet loss, and further utilized to reduce the likelihood of parameter loss due to a fragile wireless connection. By expanding the optimization parameters, the optimal communication round and the corresponding local iteration round which is convenient to calculate are estimated, and an effective local iteration round adjustment algorithm is designed to improve the reliability of the model. Through extensive simulations, the results show that over 80% of the estimated hours are close to the optimal iteration round in practice, and the enhanced FL scheme effectively reduces the communication round to achieve the required accuracy.
The technical scheme of the invention is as follows:
a cloud edge federal learning method in mobile edge computing comprises the following steps:
firstly, a mobile terminal device estimates and transmits parameters required for calculating an optimal local iteration turn to a server;
the mobile terminal device estimates and transmits parameters required for calculating the optimal local iteration turn to the server, and the parameters are specifically as follows: and selecting the mobile terminal equipment participating in the FL process, and using the local data to train the initial machine learning model of the current round transmitted by the edge server, and controlling the local training round of the selected terminal equipment by the edge server. And the terminal equipment records the initialized machine learning model in the training process and collects data in the training process, estimates the Ripocez coefficient and the smooth coefficient of the model trained by the terminal equipment by utilizing the collected data and the initialized machine learning model after the training is finished, and returns the Ripocez coefficient and the smooth coefficient of the model to the edge server together with the model.
Secondly, the server dynamically adjusts the local training round based on the minimum communication round;
2-1, selecting control parameters; the method comprises the following specific steps:
first, a theoretically optimal local training round e is given*
Figure BDA0003292422360000031
Eta is a step length selected during training, L is a second-order RipShtz constant of a loss function, beta is a smooth coefficient of the loss function, ν is an upper bound of a reciprocal difference between a machine learning model weight and an optimal machine learning model weight obtained by a centralized random gradient descent method, μ is a lower bound of a difference between a machine learning loss function and an optimal loss function value trained by a federal mean algorithm, and δ is a gradient divergence generated by the model after the federation of the federal mean algorithm is aggregated, which is defined as follows;
Figure BDA0003292422360000032
Figure BDA0003292422360000033
wherein
Figure BDA0003292422360000034
Representing the first order gradient of the aggregated federal learning model,
Figure BDA0003292422360000035
representing the first order gradient, δ, of the Federal learning model through terminal iiRepresenting the difference in the gradients of the aggregate model and the terminal i model, |, representing the potential of the set, niRepresenting a data set, n, owned by a terminal iSumRepresenting the sum of the terminals in the edge server range,
Figure BDA0003292422360000036
wherein minkMinimum value expressed in k-th round of training is equivalent to
Figure BDA0003292422360000037
And μ ═ min | | | F (w)(c)(t))-F(w*)||2
Figure BDA0003292422360000038
Because v and mu in the formula (1) are difficult to obtain under the actual application scene, v mu in the formula (1) is converted into mu by utilizing the first-order Riposetz continuous condition2The scale conversion is as follows:
Figure BDA0003292422360000039
zeta represents the unknown parameter in equation (3) i.e.:
Figure BDA0003292422360000041
w(c)(t) representing the machine learning model parameters of the centralized stochastic gradient descent method at the corresponding t moment; w is a*Representing the optimal model parameters, i.e. the optimal solution of the problem, that minimize the model loss function; maxkRepresents the maximum value in the k-th round of training; zeta is smaller than 1, and the value is controlled to be close to the theoretical value when the optimal local iteration round is calculated.
2-2, estimating an optimal local training turn and a corresponding minimum communication turn;
the method comprises the following specific steps: from equations (1) and (4), the estimate of the optimal local iteration round corresponding to the minimum communication round is:
Figure BDA0003292422360000042
the minimum communication round corresponding to the theoretical optimal iteration number is:
Figure BDA0003292422360000043
wherein Δ is an upper bound of the difference between the optimal weight and the suboptimal weight of the desired machine learning model; e.g. of the type*Representing the optimal local training round within the communication round.
And (3) combining the formula (5) and the formula (6) to obtain the estimation of the minimum communication turn:
Figure BDA0003292422360000044
when the algorithm runs beyond the estimate of the predicted minimum communication turn, the algorithm is forced to terminate to prevent the algorithm from stopping in the event of an inability to converge.
The invention has the beneficial effects that: the training round of the selected terminal equipment is dynamically adjusted through the minimum communication round estimated by the edge server and the corresponding optimal training round of the terminal equipment, so that the reliability of a transmission model in the FL process is improved, and the problem of model parameter loss caused by data packet loss caused by an unreliable wireless communication environment is solved.
Drawings
FIG. 1 is a FL overall system architecture in an MEC of the present invention;
fig. 2(a) is a flow chart of a mobile terminal device;
fig. 2(b) is a flow chart of the edge server.
FIG. 3(a) is a logistic regression training synthetic dataset.
FIG. 3(b) is a two-layer neural network regression training synthetic dataset.
Fig. 3(c) is a logistic regression training MNIST dataset.
Fig. 3(d) is a two-layer neural network training MNIST dataset.
In the figure:
Figure BDA0003292422360000051
Detailed Description
The technical solution of the present invention is described in detail below with reference to specific examples.
The invention provides a cloud side end federal learning method in mobile edge computing, which is used for reducing the problem of reliability reduction of a machine learning training model in a communication environment. In the implementation process, the purpose of reducing the communication round is achieved by collecting parameters to estimate the local optimal training round of the mobile terminal, so that the reliability of the machine learning model is guaranteed. The method combines the specific flow of a federal learning training machine learning model in a mobile edge computing environment and the current communication situation in a wireless network, and is divided into the following two parts.
(1) The mobile terminal device estimates and passes to the server the parameters needed to compute the optimal local iteration round, since the edge server needs to estimate some key parameters like β, L and δ, which the terminal device needs to estimate first in the local epoch, unlike the normal FL procedure. At the beginning of each round, the terminal device receives the initial model and the iteration rounds from the edge server. Then, the terminal device updates the model thereof and trains the model. After completing the local training round, the terminal device estimates betai、LiAnd send the parameters and updated model to the edge server. The training process in the terminal device then stops and, unless selected again, it no longer continues to participate in the training process.
(2) Transmitting the model and the estimated value L at the terminal deviceiAnd betaiAfter the edge server, it collects the received models to obtain a region model. The aggregated model satisfies the stopping condition delta and the edge server will send to the cloud server. And sending the result to the cloud server as a final result, and finishing the training process. The edge server obtains by using the received parameters
Figure BDA0003292422360000052
And
Figure BDA0003292422360000053
and calculate
Figure BDA0003292422360000054
As the optimum calendar hours e*(ii) an estimate of (d); will be provided with
Figure BDA0003292422360000055
And sending the model aggregated by the areas to the selected terminal equipment as the training turn and the initial model of the terminal equipment in the next training turn. According to equation (5), the edge server computes
Figure BDA0003292422360000061
First, estimate RminBy using
Figure BDA0003292422360000062
Maximum value of as communication round RlimitThe upper limit of (3). Due to the fact that in the formula (5)
Figure BDA0003292422360000063
Estimation of optimal communication turns
Figure BDA0003292422360000064
Less than Rmin. To avoid the problem of underestimating RminAnd the number of communication rounds R is determined by using the parameter gamma > 1limitTo γ Rlimit. If number of communication rounds RcountGreater than γ RlimitThe edge server considers the algorithm to be difficult to converge and terminates the training process.
The following is introduced from three aspects of a training process of the mobile terminal device, an estimation of the optimal training round of the local device in the edge server, and a control mechanism of non-convergence.
1. Mobile terminal device training process
During federal learning, end devices generate and maintain data, while they are also responsible for training tasks. By utilizing data maintained by a single device, the trained model is not suitable for all devices. Therefore, the selected terminal deviceThe models are ready to be transmitted to the edge server for aggregation to alleviate this problem in the FL process. The mobile terminal device estimates and passes to the server the parameters needed to compute the optimal local iteration round, the edge server needs to estimate the key parameters, such as β, L and δ, which the terminal device estimates first in the local epoch. At the beginning of each round, the terminal device receives the initial model and the iteration round from the edge server. The terminal device then updates its model and trains the model. After completing the local training round, the terminal device estimates betai、LiAnd send the parameters and updated model to the edge server. The training process in the terminal device then stops and, unless selected again, it no longer continues to participate in the training process.
(1) The mobile terminal device receives the initial machine learning model and the local training times sent by the edge server, records the initial model and initializes the local model with the initial model.
(2) And the mobile terminal device trains the model by using local data according to a random gradient descent method, and the training turn is determined by the local training turn estimated by the edge server received in the last step.
(3) When the local training round reaches the upper limit, estimating the RipSetz coefficient L of the loss function by using the loss function of the local model and the gradient thereofiAnd the smoothing coefficient beta of the functioni
Figure BDA0003292422360000065
(4) Ripocitz coefficient L estimated by mobile terminal deviceiAnd a coefficient of smoothness betaiAnd sending the trained model and the gradient of the loss function of the trained model to an edge server and stopping training.
2. Dynamic local device training turn adjustment mechanism in edge server
The aim of the present mechanism is to reduce the likelihood of model degradation due to packet loss in the communication between the terminal device and the edge server. In the FL procedure, the terminal deviceThe device maintains own data and uses the data to train the model without data exchange. The quality of the aggregated model depends on the updated model from the terminal device. Therefore, the loss of data packets during communication results in the loss of model parameters, thereby degrading the quality of the aggregation model in the edge server. Furthermore, model degradation in the edge server can slow down the entire FL process and even make the model unable to converge. The expected value of model loss depends on the communication round R and the number n of devices in the regionSumAnd the packet loss probability p of the device iiWhere i ∈ {1,2.. nSum}. The expected time of packet loss can severely impact the quality of the aggregated model in the edge server, since loss of important parameters can reduce the reliability of the model. In order to reduce the possibility of parameter loss, improve model quality,
Figure BDA0003292422360000071
should be minimized.
Figure BDA0003292422360000072
Probability of packet loss p due to device iiBeing independent and assumed to be fixed, increasing the reliability of the model means minimizing the number of communication rounds R of the objective function in equation (8).
The minimum communication round can be achieved by adjusting the local training round theoretically through analysis, and the corresponding local optimal training round is as follows:
Figure BDA0003292422360000073
the minimum communication round corresponding to the theoretical optimal iteration number is:
Figure BDA0003292422360000081
wherein Δ is an upper bound of the difference between the optimal weight and the suboptimal weight of the desired machine learning model;
because v and mu in the formula (1) are difficult to obtain in the practical application scene, v mu in the formula (1) can be converted into v mu by utilizing the first-order Riposetz continuous condition2Scaling yields an estimate for the local optimal training round:
Figure BDA0003292422360000082
combining equation (5) with equation (6) yields an estimate of the minimum communication turn:
Figure BDA0003292422360000083
the mechanism comprises the following specific flows:
(1) the edge server receives the machine learning model, the gradient of the loss function, the Ripocez coefficient and the smoothing coefficient of the loss function transmitted from the selected terminal device.
(2) The models received from the terminal devices are weighted and aggregated according to a federal averaging algorithm and new machine learning models within the region are generated.
(3) If the generated model does not meet the requirement on the precision of the loss function, the edge server calculates and estimates the Ripocitz coefficient and the smooth coefficient of the loss function corresponding to the model in the edge server at the moment according to the gradient of the loss function received from the terminal equipment, the Ripocitz coefficient and the smooth coefficient of the loss function, and estimates the gradient divergence in the round of communication.
Figure BDA0003292422360000084
(4) The edge server calculates an optimal local training round corresponding to the minimum communication round according to formula (5), and estimates the minimum communication round according to the training round.
(5) And updating the communication round upper limit of the minimum communication round representing the maximum estimation in the training process, so as to prevent the condition that the training process cannot be terminated when the model cannot be converged.
(6) And sending the aggregated region model as a terminal equipment initialization model of the next round, and sending the estimated optimal local training round to the selected terminal equipment.
The dynamic adjustment mechanism was verified by two data sets, MNIST and composite data set, and two machine learning models, two-layer fully connected neural network and logistic model, with the results shown in fig. 3. The estimates of the optimal local training rounds in the figures are mainly distributed in the intervals [8,13 ] in fig. 3(a) and 3(b)]. In FIG. 3.(a), are distributed over [8,13 ]]The estimated rate of the interval was 86%, while in fig. 3(b), the estimated rate was 85.5%. The above simulation results show that the estimation in equation (5) is valid. The estimation results are dynamically adjusted to fit the best training round. Fig. 3.(c) and 3.(d) show that the estimates are mainly distributed in two intervals. In FIG. 3(c), 84% of the estimates are distributed over [12,14 ]]And [23,26 ]]In FIG. 3(d), are distributed in [8,9 ]]And [18,19 ]]The estimated value of the interval exceeds 89%. The reason for this estimated distribution is the delta fluctuation per round in the proposed algorithm. Since the partitioning of MNIST is more unbalanced compared to the synthetic dataset, δ is from 0.8 · 10-7Fluctuating to 0.3. Fluctuations in the parameter δ lead to estimated transitions, which demonstrate that the gradient divergence δ can reflect the effect of the difference data in gradient divergence on the training process. Furthermore, such a sharp fluctuation of δ expands the estimated interval by only 10, which indicates that the estimation in equation (5) is stable.

Claims (8)

1. A cloud side federal learning method in mobile edge computing is characterized by comprising the following steps:
firstly, a mobile terminal device estimates and transmits parameters required for calculating an optimal local iteration turn to a server;
and secondly, dynamically adjusting the server based on the local training turn for realizing the minimum communication turn.
2. The cloud edge federal learning method in mobile edge computing according to claim 1, wherein in the second step, the server dynamically adjusts based on a local training turn that realizes a minimum communication turn; the method specifically comprises the following steps:
2-1, selecting control parameters;
2-2 estimating an optimal local training turn and a corresponding minimum communication turn.
3. The method according to claim 1 or 2, wherein in the first step, the mobile terminal device estimates and transmits parameters required for calculating the optimal local iteration turn to the server, specifically: selecting the mobile terminal equipment participating in the FL process, and using local data to train the initial machine learning model of the current round sent by the edge server, and controlling the local training round of the selected terminal equipment by the edge server; and the terminal equipment records the initialized machine learning model in the training process and collects data in the training process, and estimates the Rippletz coefficient and the smooth coefficient of the self-trained model by utilizing the collected data and the initialized machine learning model after the training is finished and returns the Rippletz coefficient and the smooth coefficient together with the model to the edge server.
4. The method for learning the cloud side federation in the mobile edge computing according to claim 1 or 2, wherein the 2-1 selects the control parameters, and the specific steps are as follows:
first, a theoretically optimal local training round e is given*
Figure FDA0003292422350000011
Eta is a step length selected during training, L is a second-order RipShtz constant of a loss function, beta is a smooth coefficient of the loss function, ν is an upper bound of a reciprocal difference between a machine learning model weight and an optimal machine learning model weight obtained by a centralized random gradient descent method, μ is a lower bound of a difference between a machine learning loss function and an optimal loss function value trained by a federal mean algorithm, and δ is a gradient divergence generated by the model after the federation of the federal mean algorithm is aggregated, which is defined as follows;
Figure FDA0003292422350000021
Figure FDA0003292422350000022
wherein
Figure FDA0003292422350000023
Representing the first order gradient of the aggregated federal learning model,
Figure FDA0003292422350000024
representing the first order gradient, δ, of the Federal learning model through terminal iiRepresenting the difference in the gradients of the aggregate model and the terminal i model, |, representing the potential of the set, niRepresenting a data set, n, owned by a terminal iSumRepresenting the sum of the terminals in the edge server range,
Figure FDA0003292422350000025
wherein minkMinimum value expressed in k-th round of training is equivalent to
Figure FDA0003292422350000026
And μ ═ min | | | F (w)(c)(t))-F(w*)||2
Figure FDA0003292422350000027
Because v and mu in the formula (1) are difficult to obtain under the actual application scene, v mu in the formula (1) is converted into mu by utilizing the first-order Riposetz continuous condition2The scale conversion is as follows:
Figure FDA0003292422350000028
zeta represents the unknown parameter in equation (3) i.e.:
Figure FDA0003292422350000029
w(c)(t) representing the machine learning model parameters of the centralized stochastic gradient descent method at the corresponding t moment; w is a*Representing the optimal model parameters, i.e. the optimal solution of the problem, that minimize the model loss function; maxkRepresents the maximum value in the k-th round of training; zeta is smaller than 1, and the value is controlled to be close to the theoretical value when the optimal local iteration round is calculated.
5. The method for cloud edge federal learning in mobile edge computing according to claim 3, wherein the 2-1 selection of control parameters comprises the following specific steps:
first, a theoretically optimal local training round e is given*
Figure FDA0003292422350000031
Eta is a step length selected during training, L is a second-order RipShtz constant of a loss function, beta is a smooth coefficient of the loss function, ν is an upper bound of a reciprocal difference between a machine learning model weight and an optimal machine learning model weight obtained by a centralized random gradient descent method, μ is a lower bound of a difference between a machine learning loss function and an optimal loss function value trained by a federal mean algorithm, and δ is a gradient divergence generated by the model after the federation of the federal mean algorithm is aggregated, which is defined as follows;
Figure FDA0003292422350000032
Figure FDA0003292422350000033
wherein
Figure FDA0003292422350000034
Representing the first order gradient of the aggregated federal learning model,
Figure FDA0003292422350000035
representing the first order gradient, δ, of the Federal learning model through terminal iiRepresenting the difference in the gradients of the aggregate model and the terminal i model, |, representing the potential of the set, niRepresenting a data set, n, owned by a terminal iSumRepresenting the sum of the terminals in the edge server range,
Figure FDA0003292422350000036
wherein minkMinimum value expressed in k-th round of training is equivalent to
Figure FDA0003292422350000037
And μ ═ min | | | F (w)(c)(t))-F(w*)||2
Figure FDA0003292422350000038
Because v and mu in the formula (1) are difficult to obtain under the actual application scene, v mu in the formula (1) is converted into mu by utilizing the first-order Riposetz continuous condition2The scale conversion is as follows:
Figure FDA0003292422350000039
zeta represents the unknown parameter in equation (3) i.e.:
Figure FDA0003292422350000041
w(c)(t) representing the machine learning model parameters of the centralized stochastic gradient descent method at the corresponding t moment; w is a*Representing the optimal model parameters, i.e. the optimal solution of the problem, that minimize the model loss function; maxkRepresents the maximum value in the k-th round of training; zeta is smaller than 1, and the value is controlled to be close to the theoretical value when the optimal local iteration round is calculated.
6. The method for learning the cloud side federal in the mobile edge computing according to claim 1,2 or 5, wherein the 2-2 estimation optimal local training turns and the corresponding minimum communication turn comprise the following specific steps: from equations (1) and (4), the estimate of the optimal local iteration round corresponding to the minimum communication round is:
Figure FDA0003292422350000042
the minimum communication round corresponding to the theoretical optimal iteration number is:
Figure FDA0003292422350000043
wherein Δ is an upper bound of the difference between the optimal weight and the suboptimal weight of the desired machine learning model; e.g. of the type*The communication round is an optimal local training round;
and (3) combining the formula (5) and the formula (6) to obtain the estimation of the minimum communication turn:
Figure FDA0003292422350000044
when the algorithm runs beyond the estimate of the predicted minimum communication turn, the algorithm is forced to terminate to prevent the algorithm from stopping in the event of an inability to converge.
7. The method for cloud edge federal learning in mobile edge computing according to claim 3, wherein the 2-2 estimated optimal local training rounds and the corresponding minimum communication rounds specifically comprise the following steps: from equations (1) and (4), the estimate of the optimal local iteration round corresponding to the minimum communication round is:
Figure FDA0003292422350000045
the minimum communication round corresponding to the theoretical optimal iteration number is:
Figure FDA0003292422350000051
wherein Δ is an upper bound of the difference between the optimal weight and the suboptimal weight of the desired machine learning model; e.g. of the type*The communication round is an optimal local training round;
and (3) combining the formula (5) and the formula (6) to obtain the estimation of the minimum communication turn:
Figure FDA0003292422350000052
when the algorithm runs beyond the estimate of the predicted minimum communication turn, the algorithm is forced to terminate to prevent the algorithm from stopping in the event of an inability to converge.
8. The method for cloud edge federal learning in mobile edge computing according to claim 4, wherein the 2-2 estimated optimal local training rounds and the corresponding minimum communication rounds specifically comprise the following steps: from equations (1) and (4), the estimate of the optimal local iteration round corresponding to the minimum communication round is:
Figure FDA0003292422350000053
the minimum communication round corresponding to the theoretical optimal iteration number is:
Figure FDA0003292422350000054
wherein Δ is an upper bound of the difference between the optimal weight and the suboptimal weight of the desired machine learning model; e.g. of the type*The communication round is an optimal local training round;
and (3) combining the formula (5) and the formula (6) to obtain the estimation of the minimum communication turn:
Figure FDA0003292422350000055
when the algorithm runs beyond the estimate of the predicted minimum communication turn, the algorithm is forced to terminate to prevent the algorithm from stopping in the event of an inability to converge.
CN202111169752.7A 2021-10-08 2021-10-08 Cloud edge end federation learning method in mobile edge calculation Active CN113947210B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111169752.7A CN113947210B (en) 2021-10-08 2021-10-08 Cloud edge end federation learning method in mobile edge calculation

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111169752.7A CN113947210B (en) 2021-10-08 2021-10-08 Cloud edge end federation learning method in mobile edge calculation

Publications (2)

Publication Number Publication Date
CN113947210A true CN113947210A (en) 2022-01-18
CN113947210B CN113947210B (en) 2024-05-10

Family

ID=79329973

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111169752.7A Active CN113947210B (en) 2021-10-08 2021-10-08 Cloud edge end federation learning method in mobile edge calculation

Country Status (1)

Country Link
CN (1) CN113947210B (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115099420A (en) * 2022-08-26 2022-09-23 香港中文大学(深圳) Model aggregation weight dynamic distribution method for wireless federal learning
CN116582840A (en) * 2023-07-13 2023-08-11 江南大学 Level distribution method and device for Internet of vehicles communication, storage medium and electronic equipment
CN116644802A (en) * 2023-07-19 2023-08-25 支付宝(杭州)信息技术有限公司 Model training method and device

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20200293887A1 (en) * 2019-03-11 2020-09-17 doc.ai, Inc. System and Method with Federated Learning Model for Medical Research Applications
CN112100659A (en) * 2020-09-14 2020-12-18 电子科技大学 Block chain federal learning system and Byzantine attack detection method
CN112181666A (en) * 2020-10-26 2021-01-05 华侨大学 Method, system, equipment and readable storage medium for equipment evaluation and federal learning importance aggregation based on edge intelligence
US20210073639A1 (en) * 2018-12-04 2021-03-11 Google Llc Federated Learning with Adaptive Optimization
CN113406974A (en) * 2021-08-19 2021-09-17 南京航空航天大学 Learning and resource joint optimization method for unmanned aerial vehicle cluster federal learning

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20210073639A1 (en) * 2018-12-04 2021-03-11 Google Llc Federated Learning with Adaptive Optimization
US20200293887A1 (en) * 2019-03-11 2020-09-17 doc.ai, Inc. System and Method with Federated Learning Model for Medical Research Applications
CN112100659A (en) * 2020-09-14 2020-12-18 电子科技大学 Block chain federal learning system and Byzantine attack detection method
CN112181666A (en) * 2020-10-26 2021-01-05 华侨大学 Method, system, equipment and readable storage medium for equipment evaluation and federal learning importance aggregation based on edge intelligence
CN113406974A (en) * 2021-08-19 2021-09-17 南京航空航天大学 Learning and resource joint optimization method for unmanned aerial vehicle cluster federal learning

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
刘耕;赵立君;陈庆勇;唐小勇;游正鹏;: "联邦学习在5G云边协同场景中的原理和应用综述", 通讯世界, no. 07, 25 July 2020 (2020-07-25) *
周俊;沈华杰;林中允;曹珍富;董晓蕾;: "边缘计算隐私保护研究进展", 计算机研究与发展, no. 10, 9 October 2020 (2020-10-09) *
朱剑;赵海;孙佩刚;毕远国;: "基于RSSI均值的等边三角形定位算法", 东北大学学报(自然科学版), no. 08, 15 August 2007 (2007-08-15) *

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115099420A (en) * 2022-08-26 2022-09-23 香港中文大学(深圳) Model aggregation weight dynamic distribution method for wireless federal learning
CN116582840A (en) * 2023-07-13 2023-08-11 江南大学 Level distribution method and device for Internet of vehicles communication, storage medium and electronic equipment
CN116644802A (en) * 2023-07-19 2023-08-25 支付宝(杭州)信息技术有限公司 Model training method and device

Also Published As

Publication number Publication date
CN113947210B (en) 2024-05-10

Similar Documents

Publication Publication Date Title
CN113947210A (en) Cloud side end federal learning method in mobile edge computing
Peng et al. Deep reinforcement learning based resource management for multi-access edge computing in vehicular networks
Qiao et al. Proactive caching for mobile video streaming in millimeter wave 5G networks
CN111800828B (en) Mobile edge computing resource allocation method for ultra-dense network
WO2023040022A1 (en) Computing and network collaboration-based distributed computation offloading method in random network
CN111526592B (en) Non-cooperative multi-agent power control method used in wireless interference channel
Bozkaya et al. QoE-based flow management in software defined vehicular networks
CN114595632A (en) Mobile edge cache optimization method based on federal learning
CN101248619A (en) Distributed wireless network with dynamic bandwidth allocation
CN112511197A (en) Unmanned aerial vehicle auxiliary elastic video multicast method based on deep reinforcement learning
Zhao et al. Online distributed optimization for energy-efficient computation offloading in air-ground integrated networks
Wang et al. Distributed incentives and digital twin for resource allocation in air-assisted internet of vehicles
CN114626306A (en) Method and system for guaranteeing freshness of regulation and control information of park distributed energy
CN104581918B (en) Satellite layer-span combined optimization power distribution method based on non-cooperative game
CN115209426A (en) Dynamic deployment method of digital twin servers in edge internet of vehicles
CN114885340A (en) Ultra-dense wireless network power distribution method based on deep transfer learning
CN113132497A (en) Load balancing and scheduling method for mobile edge operation
CN112929412A (en) Method, device and storage medium for joint bandwidth allocation and data stream unloading based on MEC single-vehicle single-cell
CN117221951A (en) Task unloading method based on deep reinforcement learning in vehicle-mounted edge environment
CN109600793B (en) Social relationship-based D2D communication dynamic relay selection method
CN107659940A (en) A kind of multipriority multi-channel medium access control protocol based on channel-aware
CN101917753B (en) Method for determining joint call control strategy of heterogeneous network
CN115767703A (en) Long-term power control method for SWIPT-assisted de-cellular large-scale MIMO network
CN112601256B (en) MEC-SBS clustering-based load scheduling method in ultra-dense network
CN114980039A (en) Random task scheduling and resource allocation method in MEC system of D2D cooperative computing

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant