WO2024027164A1 - Adaptive personalized federated learning method supporting heterogeneous model - Google Patents

Adaptive personalized federated learning method supporting heterogeneous model Download PDF

Info

Publication number
WO2024027164A1
WO2024027164A1 PCT/CN2023/082145 CN2023082145W WO2024027164A1 WO 2024027164 A1 WO2024027164 A1 WO 2024027164A1 CN 2023082145 W CN2023082145 W CN 2023082145W WO 2024027164 A1 WO2024027164 A1 WO 2024027164A1
Authority
WO
WIPO (PCT)
Prior art keywords
model
pri
sha
federated learning
global shared
Prior art date
Application number
PCT/CN2023/082145
Other languages
French (fr)
Chinese (zh)
Inventor
邓水光
秦臻
Original Assignee
浙江大学
浙江大学中原研究院
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 浙江大学, 浙江大学中原研究院 filed Critical 浙江大学
Publication of WO2024027164A1 publication Critical patent/WO2024027164A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F21/00Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F21/60Protecting data
    • G06F21/62Protecting access to data via a platform, e.g. using keys or access control rules
    • G06F21/6218Protecting access to data via a platform, e.g. using keys or access control rules to a system of files or objects, e.g. local or distributed file system or database
    • G06F21/6245Protecting personal data, e.g. for financial or medical purposes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/764Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects

Definitions

  • the invention belongs to the field of artificial intelligence technology, and specifically relates to an adaptive personalized federated learning method that supports heterogeneous models.
  • Deep Mutual Learning technology provides the technical basis for training two different models at the same time based on the same data.
  • some researchers have proposed the Federated Mutual Learning (Federated Mutual Learning) method, and the participation of federated learning Participants train private models and global shared models at the same time.
  • the private model remains local and its model structure and parameters are not shared.
  • the structure and parameters of the global shared model are consistent among each participant, and the central server is responsible for periodicity. It can be aggregated and distributed locally as a medium for knowledge sharing among various participants.
  • each participant holds two different models: a private model and a global shared model.
  • a simple approach is to directly average the output predictions of the two models and use the average prediction result as the final result.
  • the performance of the two models on different data has certain differences: In the case of highly heterogeneous data, the private model learns the distribution of the corresponding participant's private data set well, thereby having better accuracy on the corresponding participant's private data set, while the global shared model suffers Due to the impact of data heterogeneity, accuracy is usually poor.
  • the global shared model benefits from the knowledge sharing of multiple participants and has better accuracy, while the private model mainly relies on the knowledge of the corresponding participants. In this case, the accuracy Poorly, directly integrating two models will cause the accuracy of the integration to be severely affected by the low-accuracy model.
  • the present invention provides an adaptive personalized federated learning method that supports heterogeneous models to carry out adaptive personalized federated learning on different data when the private model structure and parameters of the participants are unknown. Participants can benefit from federated learning in scenarios with varying degrees of heterogeneity.
  • An adaptive personalized federated learning method that supports heterogeneous models, including the following steps:
  • the central server initializes the parameters of the global shared model
  • the central server distributes the global shared model parameters to each participant of federated learning. After receiving the global shared model parameters, the participants use the parameters to update the global shared model they hold;
  • the central server After the central server collects enough global shared model parameters, it aggregates these model parameters to obtain new global shared model parameters, and then returns to step (2) to distribute the new global shared model parameters to each participant. , and loop in this manner until the loss functions of all models converge or reach the maximum number of iterations.
  • the global shared model is trained by the participants of federated learning, and the central server is responsible for aggregation. Each participant holds a copy of the global shared model. On the one hand, the model is available to each participant after the federated learning training is completed. It is used for reasoning, and on the other hand, it serves as a medium for sharing knowledge among participants.
  • the private model is a model held by each participant of the federated learning and the structure and parameters are not made public.
  • the structure of the private model held by each participant is different.
  • the participants are terminal devices in the federated learning system.
  • they In order to profit from the federated learning system, that is, to obtain higher accuracy model parameters, they upload the model parameters to the central server and download the aggregated model parameters from the central server. model parameters.
  • step (3) the participant first divides a small part (for example, 5% of the training data) from the obtained private training data as the verification set, and combines the private model and the global shared model in Inference is performed on the verification set to obtain the predicted output result p pri of the private model and the predicted output result p sha of the global shared model; then the participants update the weight of the private model through the stochastic gradient descent method, and the update expression is as follows:
  • ⁇ i is the weight of the private model before updating
  • ⁇ ′ i is the weight of the private model after updating
  • eta represents the learning rate
  • L CE (p aen ,y) obtains the gradient of ⁇ i
  • L CE (p aen ,y) indicates the cross entropy of p aen and y
  • p aen indicates the result of the weighted average of p pri and p sha
  • y is the true value Label.
  • L pri L CE (p pri ,y)+D KL (p pri
  • L pri is the loss function of the private model
  • L CE (p pri ,y) represents the cross entropy of p pri and y
  • L CE (p aen ,y) represents the cross entropy of p aen and y
  • p sha ) represents the KL divergence of p pri relative to p sha
  • p aen represents the weighted average result of p pri and p sha
  • y is the true value label
  • p pri is the prediction output result of the private model
  • psha is the prediction output result of the global shared model.
  • L sha is the loss function of the global shared model
  • L CE (p sha ,y) represents the cross entropy of p sha and y
  • L CE (p aen ,y) represents the cross entropy of p aen and y
  • p pri ) represents the KL divergence of p sha relative to p pri
  • p aen represents the weighted average result of p pri and p sha
  • y is the true value label
  • p pri is the prediction output result of the private model
  • p sha The prediction output results for the global shared model.
  • step (6) after collecting enough global shared model parameters, the central server executes the federated averaging algorithm to aggregate these model parameters, and then issues the aggregated new global shared model parameters to each participant.
  • the method of the present invention achieves high accuracy by learning dynamic weights for model integration and introducing optimization goals for model integration in the process of training model parameters.
  • Personalized federated learning that is adaptive to data heterogeneity can enable participants to benefit from federated learning in scenarios with varying degrees of data heterogeneity.
  • the adaptive personalized federated learning method of the present invention does not require the introduction of new hyperparameters and can be easily deployed in existing federated learning systems.
  • the present invention has the following beneficial technical effects:
  • the present invention enables federated learning that supports model heterogeneity. On the basis of protecting participants' private training data from being leaked, it further protects the privacy of participants' model structures and achieves broader privacy protection.
  • the present invention enables an adaptive personalized federated learning method, which enables federated learning participants to benefit from federated learning in scenarios with different degrees of data heterogeneity (compared to using only local In the case of private data, a higher accuracy model is obtained).
  • the present invention solves the problem that the existing personalized federated learning method is only effective in scenarios with a specific degree of data heterogeneity; compared with the traditional personalized federated learning method, the present invention has stronger adaptability. .
  • Figure 1 is a schematic diagram of the architecture of the adaptive personalized federated learning system of the present invention.
  • Figure 2 is a schematic flow chart of the adaptive personalized federated learning method of the present invention.
  • the system architecture of the adaptive personalized federated learning method that supports heterogeneous models is shown in Figure 1.
  • the system mainly includes two parts: a central server and participants.
  • the central server is responsible for coordinating each participant to run the federated learning method, including the overall situation. It is responsible for the initialization of the shared model, the reception, aggregation and delivery of the global shared model, and is also responsible for checking whether the global shared model has converged or whether the adaptive personalized federated learning method has cycled for a sufficient number of rounds to decide whether to terminate the method.
  • each participant uses the method of the present invention to collaboratively train an image classification model, and uses the private model and global shared model obtained by training to perform subsequent reasoning.
  • the central server initializes the parameters of the selected global shared model.
  • the initialization algorithm can be coordinated by each participant in advance, such as through the Xavier initialization method or the Kaiming initialization method. This embodiment does not impose restrictions.
  • each participant in federated learning holds a private training set composed of several training private data, in which each training data sample is a labeled picture.
  • Each participant in federated learning randomly samples 5% of the training data from the private training set held by them as the verification set. For each data sample in the verification set, it is used as input and sent to the private model and the global shared model for inference. , obtain the classification result p pri output by the private model and the classification result p sha output by the global shared model, and obtain the weighted average classification result p aen according to the following formula:
  • the participant's private model weight coefficient ⁇ i is then updated through the stochastic gradient descent algorithm, as shown in the following formula:
  • y represents the label of the image.
  • mini-batch gradient descent is used to update ⁇ i , that is, several pictures are packaged into a batch of data and input into two models at once to obtain a Classification results of a batch of data, and update the weight ⁇ i according to the above formula based on the classification results of a batch of data. After several rounds of iterations, ⁇ i will converge to a suitable value, and the adaptive force learning step ends.
  • ⁇ i is iteratively updated on the verification set for several epochs. It should be noted that the solution of modifying the number of iterative updates of ⁇ i is still within the scope of the present invention.
  • Each participant runs this step independently; for one of the participants, it uses its own private training data to simultaneously train the private model (Private Model) and the global shared model based on the stochastic gradient descent algorithm, and the process of training the private model
  • the goal is to minimize the loss function L pri defined as follows:
  • L pri L CE (p pri ,y)+D KL ( ppri
  • L CE (p, y) represents the cross-entropy loss function calculated based on the image classification result p output by the model and the real label y of the image
  • p sha ) represents the classification result p pri output by the private model KL divergence calculated relative to the classification result p sha output by the global shared model
  • L sha L CE (p sha ,y)+D KL (p sha
  • this embodiment uses the small-batch gradient descent method for training. Specifically: assuming that the k-th batch of data is used in the t-th training, first based on the private data after the t-1th training model and the global shared model, using the k-th batch of data as input to obtain the classification results p pri and p sha , then update the private model according to the definition of L pri , and then update the global shared model according to the definition of L sha ; repeat the above steps for several cycles. , the learning integration step ends.
  • Global shared model aggregation and delivery The central server receives enough global shared models Finally, federated averaging is performed to aggregate these global shared models. Considering that the participants of federated learning are usually not in a local area network, and the device performance of each participant is different, the central server will set a certain waiting time, and the global shared model received within the waiting time window will be used Aggregation, after the time window ends, the global shared model of the current round will no longer be received. After the central server ends the time window of the current round, it aggregates a new global shared model through the federated averaging algorithm.
  • the aggregation process is as follows:
  • w sha represents the new global shared model after aggregation, Represents the global shared model uploaded by the i-th participant.
  • the central server issues the aggregated new global shared model to each participant; each time step (6) is executed, the central server will check whether the number of method loops has reached the preset number of rounds of overall iteration, or whether The accuracy of the model has not been further improved after several consecutive rounds of aggregation; if one of the above two determination conditions is met, the method is terminated, otherwise it will be re-executed from step (3).

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Software Systems (AREA)
  • General Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • Medical Informatics (AREA)
  • General Health & Medical Sciences (AREA)
  • Physics & Mathematics (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Computation (AREA)
  • Computing Systems (AREA)
  • Bioethics (AREA)
  • Databases & Information Systems (AREA)
  • Mathematical Physics (AREA)
  • Data Mining & Analysis (AREA)
  • Computer Hardware Design (AREA)
  • Computer Security & Cryptography (AREA)
  • Multimedia (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

Disclosed in the present invention is an adaptive personalized federated learning method supporting a heterogeneous model. In the method, on the basis of supporting participants of federated learning in using models having different structures, high-accuracy data heterogeneity adaptive personalized federated learning is realized by means of learning a dynamic weight for model integration, and introducing an optimization objective for model integration during the process of training model parameters, such that the participants can benefit from federated learning in scenarios having different degrees of data heterogeneity. The adaptive personalized federated learning method of the present invention does not need to have a new hyper-parameter introduced and can be conveniently deployed in existing federated learning systems; and compared with traditional personalized federated learning methods, the present invention is more adaptable.

Description

一种支持异构模型的自适应个性化联邦学习方法An adaptive personalized federated learning method that supports heterogeneous models 技术领域Technical field
本发明属于人工智能技术领域,具体涉及一种支持异构模型的自适应个性化联邦学习方法。The invention belongs to the field of artificial intelligence technology, and specifically relates to an adaptive personalized federated learning method that supports heterogeneous models.
背景技术Background technique
人工智能已成为带动社会经济发展的重要技术之一,深度融入到人们生活的各个角落。随着以深度学习为代表的人工智能核心技术不断取得全新突破,人工智能技术逐步依赖于大量的数据进行模型训练,但这带来了个人隐私数据被过度收集和使用的问题,致使人们对数据隐私的认知和担忧也在不断增长。数据监管政策的出台以及相关监管技术的涌现推动了隐私保护的人工智能技术的发展,促进了联邦学习(federated learning)这一在保护数据隐私的前提下协同多个参与者训练机器学习模型的计算范式的进步。Artificial intelligence has become one of the important technologies driving social and economic development and has been deeply integrated into every corner of people's lives. As the core artificial intelligence technology represented by deep learning continues to make new breakthroughs, artificial intelligence technology gradually relies on a large amount of data for model training. However, this has brought about the problem of excessive collection and use of personal privacy data, causing people to misunderstand the data. Privacy awareness and concerns are also growing. The introduction of data regulatory policies and the emergence of related regulatory technologies have promoted the development of privacy-protecting artificial intelligence technology and promoted federated learning, a computing method that collaborates with multiple participants to train machine learning models while protecting data privacy. Paradigm progress.
然而,现有的联邦学习方法面临数据异构和模型异构两个问题;一方面,分布在各个参与设备上的训练数据的非独立同分布特性(non-independent and identical distributed,non-IID)将严重制约联邦学习的有效性,许多研究表明,传统的联邦平均方法在各个参与者的持有的数据的分布不同的情况下收敛较慢,甚至不收敛。尽管许多研究者针对联邦学习的面临数据异构问题提出了多种个性化联邦学习方法,例如基于正则化(regularization)、本地调优(tuning)、模型插值(model interpolation)和多任务学习(multi-task learning)等,然而这些方法仅适用于某些数据异构程度的场景。实际运行中,由于训练数据广泛分布在各个参与设备上,数据异构程度通常不可知,故难以针对性地选用合适的个性化联邦学习方法,这催生了对自适应的个性化联邦学习技术的需求。另一方面,现有的个性化联邦学习方法更多面向的是模型同构的场景,即各个参与者需要使用结构相同的模型,而在联邦学习的各个参与者来自不同的商业组织的情况下,各个参与者可能倾向于使用更适合自己业务数据的模型,而模型结构可能是各商 业组织的机密。因此,使能差异化的模型结构的联邦学习方法能够更进一步地保护参与者的隐私,提供更高程度的个性化。However, existing federated learning methods face two problems: data heterogeneity and model heterogeneity; on the one hand, the non-independent and identically distributed (non-IID) characteristics of the training data distributed on various participating devices This will seriously restrict the effectiveness of federated learning. Many studies have shown that the traditional federated averaging method converges slowly or even does not converge when the distribution of data held by each participant is different. Although many researchers have proposed a variety of personalized federated learning methods for the problem of data heterogeneity faced by federated learning, such as based on regularization (regularization), local tuning (tuning), model interpolation (model interpolation) and multi-task learning (multi -task learning), etc. However, these methods are only suitable for scenarios with certain degrees of data heterogeneity. In actual operation, since the training data is widely distributed on various participating devices, and the degree of data heterogeneity is usually unknown, it is difficult to select an appropriate personalized federated learning method. This has given rise to the need for adaptive personalized federated learning technology. need. On the other hand, the existing personalized federated learning methods are more oriented to the scenario of model isomorphism, that is, each participant needs to use a model with the same structure, and in the case of federated learning, each participant comes from different commercial organizations. , each participant may tend to use a model that is more suitable for their own business data, and the model structure may be different for each business owner. Confidentiality of professional organizations. Therefore, federated learning methods that enable differentiated model structures can further protect the privacy of participants and provide a higher degree of personalization.
深度互学习技术(Deep Mutual Learning)提供了基于相同的数据,同时训练两个不同的模型的技术基础,在此基础上有研究者提出了联邦互学习(Federated Mutual Learning)方法,联邦学习的参与者同时训练私有模型和全局共享模型,其中私有模型保留在本地,其模型结构和参数均不共享,而全局共享模型的结构和参数在各个参与者上是保持一致的,由中心服务器负责周期性地聚合和下发,作为各个参与者知识共享的媒介。Deep Mutual Learning technology (Deep Mutual Learning) provides the technical basis for training two different models at the same time based on the same data. On this basis, some researchers have proposed the Federated Mutual Learning (Federated Mutual Learning) method, and the participation of federated learning Participants train private models and global shared models at the same time. The private model remains local and its model structure and parameters are not shared. However, the structure and parameters of the global shared model are consistent among each participant, and the central server is responsible for periodicity. It can be aggregated and distributed locally as a medium for knowledge sharing among various participants.
在联邦学习系统中,每个参与者都持有私有模型和全局共享模型两个不同的模型。为了提高模型准确性,一种简单的做法是直接将两个模型的输出预测取平均,将平均后的预测结果作为最终的结果,然而两个模型在不同数据上的表现具有一定的差异性:在数据高度异构的情况下,私有模型很好地学习了对应的参与者的私有数据集的分布,从而在对应的参与者的私有数据集上具有较好的准确度,而全局共享模型受到数据异构的影响,准确性通常较差。在数据趋于同构的情况下,全局共享模型得益于多个参与者的知识共享,具有较好的准确性,而私有模型主要依赖于对应参与者的知识,在这种情况下准确性较差,直接集成两个模型将使得集成的准确性受到准确度低的模型的严重影响。In a federated learning system, each participant holds two different models: a private model and a global shared model. In order to improve the accuracy of the model, a simple approach is to directly average the output predictions of the two models and use the average prediction result as the final result. However, the performance of the two models on different data has certain differences: In the case of highly heterogeneous data, the private model learns the distribution of the corresponding participant's private data set well, thereby having better accuracy on the corresponding participant's private data set, while the global shared model suffers Due to the impact of data heterogeneity, accuracy is usually poor. In the case where the data tends to be isomorphic, the global shared model benefits from the knowledge sharing of multiple participants and has better accuracy, while the private model mainly relies on the knowledge of the corresponding participants. In this case, the accuracy Poorly, directly integrating two models will cause the accuracy of the integration to be severely affected by the low-accuracy model.
发明内容Contents of the invention
鉴于上述,本发明提供了一种支持异构模型的自适应个性化联邦学习方法,以在参与者的私有模型结构和参数不可知的情况下,开展自适应的个性化联邦学习,在不同数据异构程度的场景下均能使参与者从联邦学习中获益。In view of the above, the present invention provides an adaptive personalized federated learning method that supports heterogeneous models to carry out adaptive personalized federated learning on different data when the private model structure and parameters of the participants are unknown. Participants can benefit from federated learning in scenarios with varying degrees of heterogeneity.
一种支持异构模型的自适应个性化联邦学习方法,包括如下步骤:An adaptive personalized federated learning method that supports heterogeneous models, including the following steps:
(1)由中心服务器初始化全局共享模型的参数;(1) The central server initializes the parameters of the global shared model;
(2)中心服务器将全局共享模型参数下发给联邦学习的各个参与者,参与者收到全局共享模型参数后,利用参数更新自己持有的全局共享模型;(2) The central server distributes the global shared model parameters to each participant of federated learning. After receiving the global shared model parameters, the participants use the parameters to update the global shared model they hold;
(3)参与者进行自适应力学习以更新私有模型的权重;(3) Participants perform adaptive force learning to update the weights of private models;
(4)参与者利用新获得的私有训练数据基于随机梯度下降算法同时训练私有模型和全局共享模型; (4) Participants use the newly obtained private training data to simultaneously train private models and global shared models based on the stochastic gradient descent algorithm;
(5)参与者将经过一轮迭代训练后的全局共享模型参数上传给中心服务器;(5) Participants upload the global shared model parameters after a round of iterative training to the central server;
(6)中心服务器收集到足够多的全局共享模型参数后对这些模型参数进行聚合,得到新的全局共享模型参数,进而返回执行步骤(2)将新的全局共享模型参数下发给各个参与者,依此循环直至所有模型的损失函数收敛或达到最大迭代次数。(6) After the central server collects enough global shared model parameters, it aggregates these model parameters to obtain new global shared model parameters, and then returns to step (2) to distribute the new global shared model parameters to each participant. , and loop in this manner until the loss functions of all models converge or reach the maximum number of iterations.
进一步地,所述全局共享模型由联邦学习的参与者负责训练,中心服务器负责聚合,各参与者均持有一份全局共享模型的拷贝,该模型一方面在联邦学习训练完成后供各参与者推理时使用,另一方面则作为各参与者共享知识的媒介。Further, the global shared model is trained by the participants of federated learning, and the central server is responsible for aggregation. Each participant holds a copy of the global shared model. On the one hand, the model is available to each participant after the federated learning training is completed. It is used for reasoning, and on the other hand, it serves as a medium for sharing knowledge among participants.
进一步地,所述私有模型为联邦学习的参与者各自持有的且结构和参数均不公开的模型,各参与者持有的私有模型结构不尽相同。Furthermore, the private model is a model held by each participant of the federated learning and the structure and parameters are not made public. The structure of the private model held by each participant is different.
进一步地,所述参与者为联邦学习系统中的终端设备,其为了从联邦学习系统中获利即获得更高准确度的模型参数,向中心服务器上传模型参数并从中心服务器上下载聚合后的模型参数。Further, the participants are terminal devices in the federated learning system. In order to profit from the federated learning system, that is, to obtain higher accuracy model parameters, they upload the model parameters to the central server and download the aggregated model parameters from the central server. model parameters.
进一步地,所述步骤(3)的具体实现方式为:参与者首先从获得的私有训练数据中划分出一小部分(例如5%的训练数据)作为验证集,将私有模型和全局共享模型在验证集上进行推理,得到私有模型的预测输出结果ppri和全局共享模型的预测输出结果psha;随后参与者通过随机梯度下降法更新私有模型的权重,更新表达式如下:
Further, the specific implementation method of step (3) is: the participant first divides a small part (for example, 5% of the training data) from the obtained private training data as the verification set, and combines the private model and the global shared model in Inference is performed on the verification set to obtain the predicted output result p pri of the private model and the predicted output result p sha of the global shared model; then the participants update the weight of the private model through the stochastic gradient descent method, and the update expression is as follows:
其中:λi为更新前私有模型的权重,λ′i为更新后私有模型的权重,η表示学习率,表示LCE(paen,y)对λi求梯度,LCE(paen,y)表示paen与y的交叉熵,paen表示ppri与psha加权平均后的结果,y为真值标签。Among them: λ i is the weight of the private model before updating, λ′ i is the weight of the private model after updating, eta represents the learning rate, Indicates that L CE (p aen ,y) obtains the gradient of λ i , L CE (p aen ,y) indicates the cross entropy of p aen and y, p aen indicates the result of the weighted average of p pri and p sha , and y is the true value Label.
进一步地,所述步骤(4)中对私有模型训练采用的损失函数表达式如下:
Lpri=LCE(ppri,y)+DKL(ppri||psha)+LCE(paen,y)
Further, the loss function expression used for private model training in step (4) is as follows:
L pri =L CE (p pri ,y)+D KL (p pri ||p sha )+L CE (p aen ,y)
其中:Lpri为私有模型的损失函数,LCE(ppri,y)表示ppri与y的交叉熵,LCE(paen,y)表示paen与y的交叉熵,DKL(ppri||psha)表示ppri相对于psha的KL散度,paen表示ppri与psha加权平均后的结果,y为真值标签,ppri为私有模型的预测输出结果, psha为全局共享模型的预测输出结果。Among them: L pri is the loss function of the private model, L CE (p pri ,y) represents the cross entropy of p pri and y, L CE (p aen ,y) represents the cross entropy of p aen and y, D KL (p pri ||p sha ) represents the KL divergence of p pri relative to p sha , p aen represents the weighted average result of p pri and p sha , y is the true value label, p pri is the prediction output result of the private model, psha is the prediction output result of the global shared model.
进一步地,所述步骤(4)中对全局共享模型训练采用的损失函数表达式如下:
Lsha=LCE(psha,y)+DKL(psha||ppri)+LCE(paen,y)
Further, the loss function expression used for global shared model training in step (4) is as follows:
L sha =L CE (p sha ,y)+D KL (p sha ||p pri )+L CE (p aen ,y)
其中:Lsha为全局共享模型的损失函数,LCE(psha,y)表示psha与y的交叉熵,LCE(paen,y)表示paen与y的交叉熵,DKL(psha||ppri)表示psha相对于ppri的KL散度,paen表示ppri与psha加权平均后的结果,y为真值标签,ppri为私有模型的预测输出结果,psha为全局共享模型的预测输出结果。Among them: L sha is the loss function of the global shared model, L CE (p sha ,y) represents the cross entropy of p sha and y, L CE (p aen ,y) represents the cross entropy of p aen and y, D KL (p sha ||p pri ) represents the KL divergence of p sha relative to p pri , p aen represents the weighted average result of p pri and p sha , y is the true value label, p pri is the prediction output result of the private model, p sha The prediction output results for the global shared model.
进一步地,所述步骤(6)中在收集到足够多的全局共享模型参数后,中心服务器执行联邦平均算法对这些模型参数进行聚合,随后将聚合后新的全局共享模型参数下发给各参与者。Further, in step (6), after collecting enough global shared model parameters, the central server executes the federated averaging algorithm to aggregate these model parameters, and then issues the aggregated new global shared model parameters to each participant. By.
本发明方法在支持联邦学习的各个参与者使用结构不同的模型的基础上,通过学习用于模型集成的动态权重和在训练模型参数的过程中引入针对模型集成的优化目标,实现高准确性的数据异构自适应的个性化联邦学习,能够在不同程度的数据异构的场景下使参与方从联邦学习中获益。此外,本发明自适应个性化联邦学习方法无需引入新的超参数,可以便捷地部署在现有联邦学习系统中。具体地,本发明具有以下有益技术效果:On the basis of supporting each participant in federated learning to use models with different structures, the method of the present invention achieves high accuracy by learning dynamic weights for model integration and introducing optimization goals for model integration in the process of training model parameters. Personalized federated learning that is adaptive to data heterogeneity can enable participants to benefit from federated learning in scenarios with varying degrees of data heterogeneity. In addition, the adaptive personalized federated learning method of the present invention does not require the introduction of new hyperparameters and can be easily deployed in existing federated learning systems. Specifically, the present invention has the following beneficial technical effects:
1.本发明使能了一种支持模型异构的联邦学习,在保护参与者私有训练数据不泄露的基础上,进一步保护参与者的模型结构的隐私,实现更广义的隐私保护。1. The present invention enables federated learning that supports model heterogeneity. On the basis of protecting participants' private training data from being leaked, it further protects the privacy of participants' model structures and achieves broader privacy protection.
2.本发明使能了一种自适应的个性化联邦学习方法,能够在数据异构程度不同的场景下,使得联邦学习的参与者能够从联邦学习中获益(在相比仅利用本地的私有数据的情况下,获得更高准确度的模型)。2. The present invention enables an adaptive personalized federated learning method, which enables federated learning participants to benefit from federated learning in scenarios with different degrees of data heterogeneity (compared to using only local In the case of private data, a higher accuracy model is obtained).
3.本发明解决了现有的个性化联邦学习方法只在面向特定程度的数据异构性的场景下有效的问题;相较于传统的个性化联邦学习方法,本发明具有更强的适应性。 3. The present invention solves the problem that the existing personalized federated learning method is only effective in scenarios with a specific degree of data heterogeneity; compared with the traditional personalized federated learning method, the present invention has stronger adaptability. .
附图说明Description of drawings
图1为本发明自适应个性化联邦学习系统架构示意图。Figure 1 is a schematic diagram of the architecture of the adaptive personalized federated learning system of the present invention.
图2为本发明自适应个性化联邦学习方法的流程示意图。Figure 2 is a schematic flow chart of the adaptive personalized federated learning method of the present invention.
具体实施方式Detailed ways
为了更为具体地描述本发明,下面结合附图及具体实施方式对本发明的技术方案进行详细说明。In order to describe the present invention more specifically, the technical solution of the present invention will be described in detail below with reference to the accompanying drawings and specific embodiments.
本发明支持异构模型的自适应个性化联邦学习方法所运行的系统架构如图1所示,系统主要包括中心服务器和参与者两部分,中心服务器负责协调各个参与者运行联邦学习方法,包括全局共享模型的初始化、全局共享模型的接收、聚合和下发,同时负责检查全局共享模型是否已经收敛或自适应个性化联邦学习方法是否已经循环到达了足够的轮数,以决定是否终止方法。The system architecture of the adaptive personalized federated learning method that supports heterogeneous models is shown in Figure 1. The system mainly includes two parts: a central server and participants. The central server is responsible for coordinating each participant to run the federated learning method, including the overall situation. It is responsible for the initialization of the shared model, the reception, aggregation and delivery of the global shared model, and is also responsible for checking whether the global shared model has converged or whether the adaptive personalized federated learning method has cycled for a sufficient number of rounds to decide whether to terminate the method.
在本实施例中,各个参与者利用本发明方法协同训练图像分类模型,并利用训练所得到的私有模型和全局共享模型进行后续的推理。In this embodiment, each participant uses the method of the present invention to collaboratively train an image classification model, and uses the private model and global shared model obtained by training to perform subsequent reasoning.
首先,参与者协调选定一个用于图像分类的模型作为全局共享模型,并共同商定方法整体迭代的轮数等参数,随后在中心服务器的协调下如图2所示运行以下流程步骤:First, participants coordinate to select a model for image classification as a global shared model, and jointly agree on parameters such as the number of rounds of the overall iteration of the method. Then, under the coordination of the central server, the following process steps are run as shown in Figure 2:
(1)初始化全局共享模型:中心服务器初始化选定的全局共享模型的参数,初始化算法可以由各个参与者提前协调好,例如通过Xavier初始化方法或Kaiming初始化方法,本实施例不做约束。(1) Initialize the global shared model: The central server initializes the parameters of the selected global shared model. The initialization algorithm can be coordinated by each participant in advance, such as through the Xavier initialization method or the Kaiming initialization method. This embodiment does not impose restrictions.
(2)全局共享模型下发:在完成全局共享模型的参数初始化后,中心服务器将全局共享模型的参数下发给联邦学习的各个参与者,每个参与者在收到全局共享模型的参数后,更新自己持有的全局共享模型。(2) Global shared model delivery: After completing the parameter initialization of the global shared model, the central server will deliver the parameters of the global shared model to each participant in the federated learning. After receiving the parameters of the global shared model, each participant , update the global shared model held by itself.
(3)学习自适应力:在本实施例中联邦学习的各个参与者持有由若干训练私有数据构成的私有训练集,其中每一个训练数据样本都是一张标注标签的图片。联邦学习的各个参与者随机从自己持有的私有训练集中采样5%的训练数据作为验证集,对验证集中的每一个数据样本,将其作为输入,送入私有模型和全局共享模型以进行推理,得到私有模型输出的分类结果ppri和全局共享模型输出的分类结果psha,并根据下式得到加权平均的分类结果paen(3) Learning adaptability: In this embodiment, each participant in federated learning holds a private training set composed of several training private data, in which each training data sample is a labeled picture. Each participant in federated learning randomly samples 5% of the training data from the private training set held by them as the verification set. For each data sample in the verification set, it is used as input and sent to the private model and the global shared model for inference. , obtain the classification result p pri output by the private model and the classification result p sha output by the global shared model, and obtain the weighted average classification result p aen according to the following formula:
paen=λi·ppri+(1-λi)·psha p aeni ·p pri +(1-λ i )·p sha
随后通过随机梯度下降算法更新该参与者的私有模型权重系数λi,如下式所示:
The participant's private model weight coefficient λ i is then updated through the stochastic gradient descent algorithm, as shown in the following formula:
其中:y表示图片的标签。Among them: y represents the label of the image.
在本实施例中,为了提升λi学习过程的稳定性,采用小批量(mini-batch)梯度下降的方式更新λi,即将若干图片打包成一批数据一次性输入到两个模型中以获得一批数据的分类结果,并根据一批数据的分类结果按照上式更新权重λi。经过若干轮迭代,λi将收敛到一个合适的值,学习自适应力步骤结束。本实施例中λi在验证集上迭代更新若干个周期(epoch),需要注意的是,对λi迭代更新的次数做出修改的方案仍在本发明的保护范围之内。In this embodiment, in order to improve the stability of the λ i learning process, mini-batch gradient descent is used to update λ i , that is, several pictures are packaged into a batch of data and input into two models at once to obtain a Classification results of a batch of data, and update the weight λ i according to the above formula based on the classification results of a batch of data. After several rounds of iterations, λ i will converge to a suitable value, and the adaptive force learning step ends. In this embodiment, λ i is iteratively updated on the verification set for several epochs. It should be noted that the solution of modifying the number of iterative updates of λ i is still within the scope of the present invention.
(4)学习集成:各个参与方独立运行此步骤;对其中一个参与方,其使用自己的私有训练数据,基于随机梯度下降算法同时训练私有模型(Private Model)和全局共享模型,训练私有模型过程的目标为最小化定义如下的损失函数Lpri(4) Learning integration: Each participant runs this step independently; for one of the participants, it uses its own private training data to simultaneously train the private model (Private Model) and the global shared model based on the stochastic gradient descent algorithm, and the process of training the private model The goal is to minimize the loss function L pri defined as follows:
Lpri=LCE(ppri,y)+DKL(ppri||psha)+LCE(paen,y)L pri =L CE (p pri ,y)+D KL ( ppri ||p sha )+L CE (p aen ,y)
其中:LCE(p,y)表示基于模型输出的图像分类结果p和图像的真实标签y计算的交叉熵损失函数,DKL(ppri||psha)表示私有模型输出的分类结果ppri相对全局共享模型输出的分类结果psha计算出的KL散度;Among them: L CE (p, y) represents the cross-entropy loss function calculated based on the image classification result p output by the model and the real label y of the image, D KL (p pri ||p sha ) represents the classification result p pri output by the private model KL divergence calculated relative to the classification result p sha output by the global shared model;
训练全局共享模型的目标为最小化定义如下的损失函数LshaThe goal of training the global shared model is to minimize the loss function Lsha defined as follows:
Lsha=LCE(psha,y)+DKL(psha||ppri)+LCE(paen,y)L sha =L CE (p sha ,y)+D KL (p sha ||p pri )+L CE (p aen ,y)
为了完成上述训练任务,本实施例采用小批量梯度下降方式进行训练,具体地:假设在第t次训练的时候用到第k批(batch)数据,先基于第t-1次训练后的私有模型和全局共享模型,以第k批数据作为输入得到分类结果ppri和psha,然后根据Lpri的定义更新私有模型,随后根据Lsha的定义更新全局共享模型;重复上述步骤若干个周期后,学习集成步骤结束。In order to complete the above training tasks, this embodiment uses the small-batch gradient descent method for training. Specifically: assuming that the k-th batch of data is used in the t-th training, first based on the private data after the t-1th training model and the global shared model, using the k-th batch of data as input to obtain the classification results p pri and p sha , then update the private model according to the definition of L pri , and then update the global shared model according to the definition of L sha ; repeat the above steps for several cycles. , the learning integration step ends.
(5)全局共享模型上传:联邦学习的参与者在完成了第(3)、(4)步训练之后,向中心服务器上传经过自己训练的全局共享模型,同时将私有模型保留在本地。(5) Global shared model upload: After completing steps (3) and (4) of training, participants in federated learning upload their own trained global shared model to the central server, while keeping the private model locally.
(6)全局共享模型聚合和下发:中心服务器在收到了足够的全局共享模型 后,执行联邦平均以聚合这些全局共享模型。考虑到联邦学习的参与者通常不处于一个局域网中,并且各个参与者设备性能具有差异性,中心服务器将设置一定的等待时间,在等待时间的时间窗口内收到的全局共享模型将被用于聚合,时间窗口结束后将不再接收当前轮的全局共享模型。中心服务器在结束当前轮的时间窗口后,通过联邦平均算法聚合得到一个新的全局共享模型,聚合过程如下式所示:
(6) Global shared model aggregation and delivery: The central server receives enough global shared models Finally, federated averaging is performed to aggregate these global shared models. Considering that the participants of federated learning are usually not in a local area network, and the device performance of each participant is different, the central server will set a certain waiting time, and the global shared model received within the waiting time window will be used Aggregation, after the time window ends, the global shared model of the current round will no longer be received. After the central server ends the time window of the current round, it aggregates a new global shared model through the federated averaging algorithm. The aggregation process is as follows:
其中:wsha表示聚合后的新的全局共享模型,表示第i个参与者上传的全局共享模型。Among them: w sha represents the new global shared model after aggregation, Represents the global shared model uploaded by the i-th participant.
随后,中心服务器向各个参与者下发聚合后的新的全局共享模型;每当执行完步骤(6)后,中心服务器将检查方法循环次数是否已达到预设的整体迭代的轮数,或者是否已经连续若干轮聚合后模型的准确度都没有得到进一步提升;若上述两个判定条件有一个满足,则方法终止,否则将从步骤(3)开始重新执行。Subsequently, the central server issues the aggregated new global shared model to each participant; each time step (6) is executed, the central server will check whether the number of method loops has reached the preset number of rounds of overall iteration, or whether The accuracy of the model has not been further improved after several consecutive rounds of aggregation; if one of the above two determination conditions is met, the method is terminated, otherwise it will be re-executed from step (3).
上述对实施例的描述是为便于本技术领域的普通技术人员能理解和应用本发明,熟悉本领域技术的人员显然可以容易地对上述实施例做出各种修改,并把在此说明的一般原理应用到其他实施例中而不必经过创造性的劳动。因此,本发明不限于上述实施例,本领域技术人员根据本发明的揭示,对于本发明做出的改进和修改都应该在本发明的保护范围之内。 The above description of the embodiments is to facilitate those skilled in the art to understand and apply the present invention. It is obvious that those skilled in the art can easily make various modifications to the above embodiments, and the general descriptions here can be made. The principles may be applied to other embodiments without undue inventive effort. Therefore, the present invention is not limited to the above embodiments. Improvements and modifications made by those skilled in the art based on the disclosure of the present invention should be within the protection scope of the present invention.

Claims (8)

  1. 一种支持异构模型的自适应个性化联邦学习方法,包括如下步骤:An adaptive personalized federated learning method that supports heterogeneous models, including the following steps:
    (1)由中心服务器初始化全局共享模型的参数;(1) The central server initializes the parameters of the global shared model;
    (2)中心服务器将全局共享模型参数下发给联邦学习的各个参与者,参与者收到全局共享模型参数后,利用参数更新自己持有的全局共享模型;(2) The central server distributes the global shared model parameters to each participant of federated learning. After receiving the global shared model parameters, the participants use the parameters to update the global shared model they hold;
    (3)参与者进行自适应力学习以更新私有模型的权重;(3) Participants perform adaptive force learning to update the weights of private models;
    (4)参与者利用新获得的私有训练数据基于随机梯度下降算法同时训练私有模型和全局共享模型;(4) Participants use the newly obtained private training data to simultaneously train private models and global shared models based on the stochastic gradient descent algorithm;
    (5)参与者将经过一轮迭代训练后的全局共享模型参数上传给中心服务器;(5) Participants upload the global shared model parameters after a round of iterative training to the central server;
    (6)中心服务器收集到足够多的全局共享模型参数后对这些模型参数进行聚合,得到新的全局共享模型参数,进而返回执行步骤(2)将新的全局共享模型参数下发给各个参与者,依此循环直至所有模型的损失函数收敛或达到最大迭代次数。(6) After the central server collects enough global shared model parameters, it aggregates these model parameters to obtain new global shared model parameters, and then returns to step (2) to distribute the new global shared model parameters to each participant. , and loop in this manner until the loss functions of all models converge or reach the maximum number of iterations.
  2. 根据权利要求1所述的自适应个性化联邦学习方法,其特征在于:所述全局共享模型由联邦学习的参与者负责训练,中心服务器负责聚合,各参与者均持有一份全局共享模型的拷贝,该模型一方面在联邦学习训练完成后供各参与者推理时使用,另一方面则作为各参与者共享知识的媒介。The adaptive personalized federated learning method according to claim 1, characterized in that: the global shared model is trained by federated learning participants, the central server is responsible for aggregation, and each participant holds a copy of the global shared model Copy, on the one hand, this model can be used by each participant for reasoning after the federated learning training is completed, and on the other hand, it serves as a medium for each participant to share knowledge.
  3. 根据权利要求1所述的自适应个性化联邦学习方法,其特征在于:所述私有模型为联邦学习的参与者各自持有的且结构和参数均不公开的模型,各参与者持有的私有模型结构不尽相同。The adaptive personalized federated learning method according to claim 1, characterized in that: the private model is a model held by each participant of federated learning and the structure and parameters are not public. Model structures vary.
  4. 根据权利要求1所述的自适应个性化联邦学习方法,其特征在于:所述参与者为联邦学习系统中的终端设备,其为了从联邦学习系统中获利即获得更高准确度的模型参数,向中心服务器上传模型参数并从中心服务器上下载聚合后的模型参数。The adaptive personalized federated learning method according to claim 1, characterized in that: the participants are terminal devices in the federated learning system, and they obtain higher accuracy model parameters in order to profit from the federated learning system. , upload model parameters to the central server and download aggregated model parameters from the central server.
  5. 根据权利要求1所述的自适应个性化联邦学习方法,其特征在于:所述步骤(3)的具体实现方式为:参与者首先从获得的私有训练数据中划分出一小部分作为验证集,将私有模型和全局共享模型在验证集上进行推理,得到私有 模型的预测输出结果ppri和全局共享模型的预测输出结果psha;随后参与者通过随机梯度下降法更新私有模型的权重,更新表达式如下:
    The adaptive personalized federated learning method according to claim 1, characterized in that: the specific implementation method of step (3) is: participants first divide a small part from the obtained private training data as a verification set, The private model and the global shared model are inferred on the verification set to obtain the private model. The predicted output result p pri of the model and the predicted output result p sha of the global shared model; then the participants update the weight of the private model through the stochastic gradient descent method, and the update expression is as follows:
    其中:λi为更新前私有模型的权重,λ′i为更新后私有模型的权重,η表示学习率,表示LCE(paen,y)对λi求梯度,LCE(paen,y)表示paen与y的交叉熵,paen表示ppri与psha加权平均后的结果,y为真值标签。Among them: λ i is the weight of the private model before updating, λ′ i is the weight of the private model after updating, eta represents the learning rate, Indicates that L CE (p aen ,y) obtains the gradient of λ i , L CE (p aen ,y) indicates the cross entropy of p aen and y, p aen indicates the result of the weighted average of p pri and p sha , and y is the true value Label.
  6. 根据权利要求1所述的自适应个性化联邦学习方法,其特征在于:所述步骤(4)中对私有模型训练采用的损失函数表达式如下:
    Lpri=LCE(ppri,y)+DKL(ppri||psha)+LCE(paen,y)
    The adaptive personalized federated learning method according to claim 1, characterized in that: the loss function expression used for private model training in step (4) is as follows:
    L pri =L CE (p pri ,y)+D KL (p pri ||p sha )+L CE (p aen ,y)
    其中:Lpri为私有模型的损失函数,LCE(ppri,y)表示ppri与y的交叉熵,LCE(paen,y)表示paen与y的交叉熵,DKL(ppri||psha)表示ppri相对于psha的KL散度,paen表示ppri与psha加权平均后的结果,y为真值标签,ppri为私有模型的预测输出结果,psha为全局共享模型的预测输出结果。Among them: L pri is the loss function of the private model, L CE (p pri ,y) represents the cross entropy of p pri and y, L CE (p aen ,y) represents the cross entropy of p aen and y, D KL (p pri ||p sha ) represents the KL divergence of p pri relative to p sha , p aen represents the weighted average result of p pri and p sha , y is the true value label, p pri is the prediction output result of the private model, and p sha is Globally share the model’s prediction output.
  7. 根据权利要求1所述的自适应个性化联邦学习方法,其特征在于:所述步骤(4)中对全局共享模型训练采用的损失函数表达式如下:
    Lsha=LCE(psha,y)+DKL(psha||ppri)+LCE(paen,y)
    The adaptive personalized federated learning method according to claim 1, characterized in that: the loss function expression used for global shared model training in step (4) is as follows:
    L sha =L CE (p sha ,y)+D KL (p sha ||p pri )+L CE (p aen ,y)
    其中:Lsha为全局共享模型的损失函数,LCE(psha,y)表示psha与y的交叉熵,LCE(paen,y)表示paen与y的交叉熵,DKL(psha||ppri)表示psha相对于ppri的KL散度,paen表示ppri与psha加权平均后的结果,y为真值标签,ppri为私有模型的预测输出结果,psha为全局共享模型的预测输出结果。Among them: L sha is the loss function of the global shared model, L CE (p sha ,y) represents the cross entropy of p sha and y, L CE (p aen ,y) represents the cross entropy of p aen and y, D KL (p sha ||p pri ) represents the KL divergence of p sha relative to p pri , p aen represents the weighted average result of p pri and p sha , y is the true value label, p pri is the prediction output result of the private model, p sha The prediction output results for the global shared model.
  8. 根据权利要求1所述的自适应个性化联邦学习方法,其特征在于:所述步骤(6)中在收集到足够多的全局共享模型参数后,中心服务器执行联邦平均算法对这些模型参数进行聚合,随后将聚合后新的全局共享模型参数下发给各参与者。 The adaptive personalized federated learning method according to claim 1, characterized in that: in step (6), after collecting enough global shared model parameters, the central server executes a federated averaging algorithm to aggregate these model parameters. , and then distribute the aggregated new global shared model parameters to each participant.
PCT/CN2023/082145 2022-08-01 2023-03-17 Adaptive personalized federated learning method supporting heterogeneous model WO2024027164A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202210916817.8 2022-08-01
CN202210916817.8A CN115271099A (en) 2022-08-01 2022-08-01 Self-adaptive personalized federal learning method supporting heterogeneous model

Publications (1)

Publication Number Publication Date
WO2024027164A1 true WO2024027164A1 (en) 2024-02-08

Family

ID=83746862

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2023/082145 WO2024027164A1 (en) 2022-08-01 2023-03-17 Adaptive personalized federated learning method supporting heterogeneous model

Country Status (2)

Country Link
CN (1) CN115271099A (en)
WO (1) WO2024027164A1 (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117808129A (en) * 2024-02-29 2024-04-02 浪潮电子信息产业股份有限公司 Heterogeneous distributed learning method, device, equipment, system and medium
CN117808128A (en) * 2024-02-29 2024-04-02 浪潮电子信息产业股份有限公司 Image processing method, federal learning method and device under heterogeneous data condition
CN117829274A (en) * 2024-02-29 2024-04-05 浪潮电子信息产业股份有限公司 Model fusion method, device, equipment, federal learning system and storage medium
CN117910600A (en) * 2024-03-15 2024-04-19 山东省计算中心(国家超级计算济南中心) Meta-continuous federal learning system and method based on fast learning and knowledge accumulation

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115271099A (en) * 2022-08-01 2022-11-01 浙江大学中原研究院 Self-adaptive personalized federal learning method supporting heterogeneous model
CN116361398B (en) * 2023-02-21 2023-12-26 北京大数据先进技术研究院 User credit assessment method, federal learning system, device and equipment

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112329940A (en) * 2020-11-02 2021-02-05 北京邮电大学 Personalized model training method and system combining federal learning and user portrait
CN114357067A (en) * 2021-12-15 2022-04-15 华南理工大学 Personalized federal meta-learning method for data isomerism
CN114429219A (en) * 2021-12-09 2022-05-03 之江实验室 Long-tail heterogeneous data-oriented federal learning method
CN115271099A (en) * 2022-08-01 2022-11-01 浙江大学中原研究院 Self-adaptive personalized federal learning method supporting heterogeneous model

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112329940A (en) * 2020-11-02 2021-02-05 北京邮电大学 Personalized model training method and system combining federal learning and user portrait
CN114429219A (en) * 2021-12-09 2022-05-03 之江实验室 Long-tail heterogeneous data-oriented federal learning method
CN114357067A (en) * 2021-12-15 2022-04-15 华南理工大学 Personalized federal meta-learning method for data isomerism
CN115271099A (en) * 2022-08-01 2022-11-01 浙江大学中原研究院 Self-adaptive personalized federal learning method supporting heterogeneous model

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117808129A (en) * 2024-02-29 2024-04-02 浪潮电子信息产业股份有限公司 Heterogeneous distributed learning method, device, equipment, system and medium
CN117808128A (en) * 2024-02-29 2024-04-02 浪潮电子信息产业股份有限公司 Image processing method, federal learning method and device under heterogeneous data condition
CN117829274A (en) * 2024-02-29 2024-04-05 浪潮电子信息产业股份有限公司 Model fusion method, device, equipment, federal learning system and storage medium
CN117829274B (en) * 2024-02-29 2024-05-24 浪潮电子信息产业股份有限公司 Model fusion method, device, equipment, federal learning system and storage medium
CN117808129B (en) * 2024-02-29 2024-05-24 浪潮电子信息产业股份有限公司 Heterogeneous distributed learning method, device, equipment, system and medium
CN117808128B (en) * 2024-02-29 2024-05-28 浪潮电子信息产业股份有限公司 Image processing method and device under heterogeneous data condition
CN117910600A (en) * 2024-03-15 2024-04-19 山东省计算中心(国家超级计算济南中心) Meta-continuous federal learning system and method based on fast learning and knowledge accumulation
CN117910600B (en) * 2024-03-15 2024-05-28 山东省计算中心(国家超级计算济南中心) Meta-continuous federal learning system and method based on fast learning and knowledge accumulation

Also Published As

Publication number Publication date
CN115271099A (en) 2022-11-01

Similar Documents

Publication Publication Date Title
WO2024027164A1 (en) Adaptive personalized federated learning method supporting heterogeneous model
US11836615B2 (en) Bayesian nonparametric learning of neural networks
Khodak et al. Federated hyperparameter tuning: Challenges, baselines, and connections to weight-sharing
Yao et al. Safeguarded dynamic label regression for noisy supervision
Abdelmoniem et al. Refl: Resource-efficient federated learning
Yu et al. Training deep energy-based models with f-divergence minimization
US20190385019A1 (en) Systems and Methods for Conditional Generative Models
CN113850272A (en) Local differential privacy-based federal learning image classification method
US20220318412A1 (en) Privacy-aware pruning in machine learning
Wu et al. Federated unlearning: Guarantee the right of clients to forget
CN110689136B (en) Deep learning model obtaining method, device, equipment and storage medium
CN113626866B (en) Federal learning-oriented localization differential privacy protection method, system, computer equipment and storage medium
Mesquita et al. Embarrassingly parallel MCMC using deep invertible transformations
CN116627970A (en) Data sharing method and device based on blockchain and federal learning
CN112235062A (en) Federal learning method and system for resisting communication noise
Shen et al. Leveraging cross-network information for graph sparsification in influence maximization
Wang et al. Federated semi-supervised learning with class distribution mismatch
CN115879542A (en) Federal learning method oriented to non-independent same-distribution heterogeneous data
CN115359298A (en) Sparse neural network-based federal meta-learning image classification method
Hu et al. Federated one-class collaborative filtering via privacy-aware non-sampling matrix factorization
Yi et al. pFedLHNs: Personalized Federated Learning via Local Hypernetworks
Usmanova et al. Federated continual learning through distillation in pervasive computing
CN113705724B (en) Batch learning method of deep neural network based on self-adaptive L-BFGS algorithm
Lee et al. NanoBatch DPSGD: Exploring Differentially Private learning on ImageNet with low batch sizes on the IPU
Jankowiak et al. Neural likelihoods for multi-output Gaussian processes

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 23848879

Country of ref document: EP

Kind code of ref document: A1