CN115860135A - Method, apparatus, and medium for solving heterogeneous federated learning using a super network - Google Patents

Method, apparatus, and medium for solving heterogeneous federated learning using a super network Download PDF

Info

Publication number
CN115860135A
CN115860135A CN202211432123.3A CN202211432123A CN115860135A CN 115860135 A CN115860135 A CN 115860135A CN 202211432123 A CN202211432123 A CN 202211432123A CN 115860135 A CN115860135 A CN 115860135A
Authority
CN
China
Prior art keywords
model
network
client
heterogeneous
training
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202211432123.3A
Other languages
Chinese (zh)
Other versions
CN115860135B (en
Inventor
任皓
李达
刘敏超
刘通泽
段振飞
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Chinese PLA General Hospital
Original Assignee
Chinese PLA General Hospital
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Chinese PLA General Hospital filed Critical Chinese PLA General Hospital
Priority to CN202211432123.3A priority Critical patent/CN115860135B/en
Publication of CN115860135A publication Critical patent/CN115860135A/en
Application granted granted Critical
Publication of CN115860135B publication Critical patent/CN115860135B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Landscapes

  • Image Analysis (AREA)

Abstract

The invention discloses a method, equipment and a medium for solving heterogeneous federated learning by using an ultra-network, which are used for generating model weights for heterogeneous models on edge equipment through the ultra-network, helping different convolutional neural network models on equipment such as a camera, a smart phone and the like to accurately and quickly complete a federated learning task of the models under the condition of different model structures, and reducing memory space required to be occupied in a federated learning training process. Reasonably dividing a cifar10 data set; generating a heterogeneous model; performing heterogeneous federated learning; and repeating the steps S2 and S3, when the heterogeneous model of the client is converged, the super network generates model weight for the client, so that the client can accurately complete the target identification task, and the super network H is directly deployed in new edge equipment to generate weight for the new heterogeneous model and complete a new image processing task. Effectively finishing the image processing task, reducing the memory space and greatly reducing the communication cost.

Description

Method, apparatus, and medium for solving heterogeneous federated learning using a super network
Technical Field
The invention belongs to the technical field of federal learning, and particularly relates to a method, electronic equipment and a computer storage medium for solving heterogeneous federal learning by using an ultra-network.
Background
Mobile devices and internet of things devices are becoming the primary computing resource for billions of users worldwide. These devices generate large amounts of data that can be used to retrofit many existing applications. From a privacy and economic perspective, storing data and training models locally is becoming increasingly attractive as the computing power of these devices continues to grow. Federal Learning (FL) is a distributed machine learning framework that enables many customers to generate global inference models without sharing local data by aggregating locally trained model parameters. A widely accepted assumption is that the local model must be of the same architecture as the global model to enable the use of federal learning methods to produce a single global inference model. With this basic assumption, we must limit the complexity of the global model of the least computationally powerful customer to train their data. In practice, each client has different computing resources, computing power and communication bandwidth, and a heterogeneous deep learning model needs to be deployed at different clients so as to meet performance requirements of the different clients, so that solving the problem of the heterogeneous client is very important.
Heterogeneous federated learning (which may also be referred to as personalized federated learning, PFL) aims to provide a heterogeneous local model for each client in the federation, which may effectively solve heterogeneous client problems. Unlike global model heterogeneous strategies that train a single global model, this type of approach trains a single heterogeneous model on each client. The goal is to achieve efficient aggregation of heterogeneous models by modifying the aggregation process of the FL model. The most advanced PFL at present is a heterogeneous Federal super network, also called pFedHN, and the method realizes parameter sharing across clients by training a central super network model, generates a unique heterogeneous model for each client and keeps the capability of generating unique and diversified heterogeneous models. However, pFedHN has an obvious defect, in the client training stage, parameters before and after network model training need to be stored, the gradient of the network model is calculated for updating the extranet, and once the client network model is a complex model, the storage network model occupies a large memory space and even cannot be trained due to insufficient memory space. Therefore, there is a need to reduce the memory resources occupied by each edge device during the training process while using the super-network to provide a heterogeneous model satisfying the current computing resources for each edge device.
Disclosure of Invention
In view of the defects of the existing solutions, an object of the embodiments of the present invention is to provide a method for solving heterogeneous federated learning by using an ultra-network, so as to generate model weights for heterogeneous models on edge devices through the ultra-network, help models on devices such as cameras and smart phones to complete image processing tasks more accurately and quickly, and reduce memory space required to be occupied in the federated learning training process.
Another object of the embodiments of the present invention is to provide an electronic device.
It is a further object of embodiments of the present invention to provide a computer storage medium.
In order to solve the technical problems, the technical scheme adopted by the invention is that a method for solving heterogeneous federated learning by using an ultra-network is carried out according to the following steps:
s1, generating a heterogeneous model and initializing model parameters;
s2, generating heterogeneous model parameter weights;
s3, heterogeneous federation learning is carried out;
and S4, repeating the steps S2 and S3, when the heterogeneous model of the client is converged, generating model weight for the client by the super network, so that the client model can accurately complete the target identification task, and at the moment, directly deploying the super network H in new edge equipment to generate weight for the new heterogeneous model and complete a new image processing task.
Another technical solution adopted by the present invention is an electronic device, including a memory and a processor, where the memory stores a computer program, and the processor implements the steps of the method when executing the computer program.
Yet another technical solution adopted by the present invention is a computer storage medium, in which at least one program instruction is stored, and the at least one program instruction is loaded and executed by a processor to implement the steps of the method.
The invention has the beneficial effects that:
(1) The method realizes end-to-end gradient descent training of the ultra-network and the client heterogeneous model, improves the accuracy of the client heterogeneous model, helps the heterogeneous model to effectively complete an image processing task, and reduces the memory space required to be occupied in the training process compared with the pFedHN.
(2) The method only aggregates and updates the super-network parameters at the server end, so that the client end only needs to send the super-network parameters to the server end, and the communication cost is greatly reduced because the super-network is only a simple linear network.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below, it is obvious that the drawings in the following description are only some embodiments of the present invention, and for those skilled in the art, other drawings can be obtained according to the drawings without creative efforts.
FIG. 1 is a line graph of accuracy of the present invention compared to the pFadHN method.
Fig. 2 is a flow chart of heterogeneous federated learning.
FIG. 3 is a schematic diagram of a training process framework according to an embodiment of the invention.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
The invention aims at intelligent edge devices with limited computing resources and aims to provide a heterogeneous model meeting the current computing resources for each edge device. The idea of the invention is as follows:
before training begins, the server sets the heterogeneous model of each client in advance at the cloud end and sends a uniform super-network model to each client, and an independent model weight is generated for the heterogeneous model of each client by using the super-network. Each local client has a unique set of embedded vectors that are passed as input to the super-net to generate heterogeneous model weights. The client can train own heterogeneous model according to the currently held data, and the hyper-network and the heterogeneous model can perform gradient descent training end to end in the training process, so that the method is more effective. After the local training is finished, the client side sends the updated extranet parameters to the server side, and updates the extranet parameters of the server side.
The method specifically comprises the following steps:
s1, generating a heterogeneous model and initializing model parameters:
in this embodiment, the server side initially selects and generates a heterogeneous model according to the computing capability of the edge device and initializes heterogeneous model parameters, and then distributes the corresponding heterogeneous model to each edge device, where each edge device has its own image data set, and these image data sets are only used for training of the heterogeneous model and the super-network on the current device and are not shared with other edge devices and servers.
S2, generating heterogeneous model parameters:
the method specifically comprises the following steps:
s2.1, initializing a federal learning training environment: setting an integral training round E, local data T and the number M of integral clients participating in federal learning, wherein the number of the clients participating in the federal learning selected in each round is M, and initializing a super network model H;
s2.2, the server is used for each client M at the cloud end i Setting a corresponding client heterogeneous model C i The main process is as shown in fig. 2, the server side sends a unified hyper-network model H to the client side i After receiving the extranet, the client starts the federal learningTraining, namely, after one round of training is finished, the supernet model H is put in i Returning to the server side;
s2.3, generating heterogeneous model parameters;
the method specifically comprises the following substeps:
s2.3.1, at the current client, the super network model H i Is a linear network with L layers, and L is set according to the complexity and the structure of a client model, such as 2 layers. Client heterogeneous model C i The convolution network accords with the ith client computing resource, and the convolution kernel of the convolution network contains most of model parameters, so the ultra-network model H i Only the parameter weights of the convolution kernels of the convolution network need to be generated. Each convolution kernel in the client network contains K in ×K out Filters, each filter having a dimension f size ×f size Size. It is assumed that these convolution kernel parameter weights are all stored in a matrix
Figure BDA0003945552120000041
In the network, each layer of the client network is l =1 i The number of layers of (a). Based on each layer of the client network>
Figure BDA0003945552120000042
Extra-net model H i Will receive a layer insert->
Figure BDA0003945552120000043
As input and generates +>
Figure BDA0003945552120000044
Written as follows:
Figure BDA0003945552120000045
wherein R is a real number, and R is a real number,
Figure BDA0003945552120000046
is to save the parameter weight of the current layer convolution kernel of the client heterogeneous model>
Figure BDA0003945552120000047
F of (a) size K in ×f size K out A dimension real number parameter, and>
Figure BDA0003945552120000048
is described in description C i Embedded vector of each layer of network information, supernet H i By learning>
Figure BDA0003945552120000049
And generating the weight of the current model l layer of the client.
S2.3.2, the previous step has described the supernet model H in detail i Weight of how to generate l-layer convolution kernel of client heterogeneous model
Figure BDA00039455521200000410
The process of generating the entire heterogeneous model weight will be described in detail next.
Super net H i Will embed the vector
Figure BDA00039455521200000411
Output after linear operation->
Figure BDA00039455521200000412
Figure BDA00039455521200000413
Will be as H i And inputting the next layer. H i The last layer of (A) is a simple linear operation, which takes the form of H i Output on level L-1->
Figure BDA00039455521200000414
As input to the last layer, through H i The last layer will->
Figure BDA00039455521200000415
Linear projection onto->
Figure BDA00039455521200000416
Middle, or>
Figure BDA00039455521200000417
Is the l-th layer weight of the i-th client heterogeneous model. Therefore, the temperature of the molten metal is controlled,
Figure BDA00039455521200000418
the following formula can be written:
Figure BDA00039455521200000419
...
Figure BDA00039455521200000420
Figure BDA00039455521200000421
Figure BDA00039455521200000422
wherein
Figure BDA00039455521200000423
Is H i Based on the weight of->
Figure BDA00039455521200000424
Is the weight of the ith entire client heterogeneous model.
In the formula, the learnable parameter is
Figure BDA00039455521200000425
And all>
Figure BDA00039455521200000426
Therefore/based on>
Figure BDA00039455521200000427
Is H i Therefore, in the training process, the super-net performs gradient descent training end to end along with the client heterogeneous model.
S3, performing heterogeneous federated learning
The method specifically comprises the following steps:
s3.1, training the client by taking the current heterogeneous model as a target network, and generating weights for the heterogeneous model on one side of a super network, wherein the weights are represented as shown in a figure 3; while carrying out gradient descent training end to end together with the heterogeneous model and updating the parameter H of the ultra-net model i The objective function is formulated as follows:
Figure BDA0003945552120000051
the extranet parameter updating formula is as follows:
Figure BDA0003945552120000052
wherein, L (-) is a loss function, X and Y respectively correspond to the image and the label of the data set, N is the number of the data sets held by the client, and ζ is the learning rate of the extranet model.
S3.2, after the client training is finished, the client returns the parameter H of the hyper-network model i And wherein i = 1.,. M, performing extranet aggregation and updating extranet model parameters at the server side, and issuing the updated extranet model parameters to the corresponding client side:
Figure BDA0003945552120000053
h is a server-side ultra-network model;
and S4, repeating the steps S2 and S3, wherein the weight is generated for the client heterogeneous model when the client heterogeneous model is trained, so that when the client heterogeneous model is converged, the fact that the client heterogeneous model generates the model weight for the client by the super network is proved to enable the client model to accurately complete the target recognition task, and the super network H can be directly deployed in new edge equipment to generate the weight for the new heterogeneous model so as to complete a new image processing task.
The embodiment provides a program stored on an edge device, and when executed, the program is used for implementing the method for solving heterogeneous federated learning by using the ultra-network. The hyper-network can directly generate model weight for the heterogeneous model on the current edge equipment, and help the model on equipment such as a camera and a smart phone to complete an image processing task.
The method of the present embodiment, if implemented in the form of a software functional module and sold or used as a standalone product, may be stored in a computer-readable storage medium. Based on such understanding, the technical solution of the present invention or a part of the technical solution that substantially contributes to the prior art may be embodied in the form of a software product, where the computer software product is stored in a storage medium and includes several instructions to enable a computer device (which may be a personal computer, a server, or a network device, etc.) to execute all or part of the steps of the method for identifying an underground anomaly interface according to the embodiment of the present invention. And the aforementioned storage medium includes: various media capable of storing program codes, such as a U disk, a removable hard disk, a ROM, a RAM, a magnetic disk, or an optical disk.
In this embodiment, a CIFAR dataset is used for training and testing, so as to further verify that the embodiment of the present invention not only reduces the memory space required to be occupied in the training process, but also improves the accuracy of the client heterogeneous model, compared with the pFedHN.
This example tests the ResNet network against pFedHN on the CIFAR10 dataset in terms of accuracy. As shown in fig. 1, it can be seen that the accuracy of the present invention is consistently higher than that of pFedHN at the same setting.
All the embodiments in the present specification are described in a related manner, and the same and similar parts among the embodiments may be referred to each other, and each embodiment focuses on differences from other embodiments. In particular, for the system embodiment, since it is substantially similar to the method embodiment, the description is simple, and for the relevant points, reference may be made to the partial description of the method embodiment.
The above description is only for the preferred embodiment of the present invention, and is not intended to limit the scope of the present invention. Any modification, equivalent replacement, or improvement made within the spirit and principle of the present invention shall fall within the protection scope of the present invention.

Claims (7)

1. The method for solving heterogeneous federated learning by using the ultra-network is characterized by comprising the following steps of:
s1, generating a heterogeneous model and initializing model parameters;
s2, generating heterogeneous model parameters;
s3, heterogeneous federated learning is carried out;
and S4, repeating the steps S2 and S3, when the heterogeneous model of the client is converged, generating model weight for the client by the super network, so that the client model can accurately complete the target identification task, and at the moment, directly deploying the super network H in new edge equipment to generate weight for the new heterogeneous model and complete a new image processing task.
2. The method for solving heterogeneous federated learning using the ultra-network as claimed in claim 1, wherein in S1, the server end initially generates a heterogeneous model and initializes heterogeneous model parameters according to the computing power of the edge devices, and then distributes the corresponding heterogeneous model to each edge device, each edge device has its own image data set, and these image data sets are only used for training the heterogeneous model and the ultra-network on the current device and are not shared with other edge devices and servers.
3. The method for solving heterogeneous federated learning using the ultra-network of claim 1, wherein the S2 comprises:
s2.1, initializing a federal learning training environment: setting an integral training round E, local data T and the number M of integral clients participating in federal learning, wherein the number of the clients participating in the federal learning selected in each round is M, and initializing a super network model H;
s2.2, the server is used for each client M at the cloud end i Setting a corresponding client heterogeneous model C i The server side sends a unified hyper-network model H to the client side i After receiving the extranet, the client starts the federal learning training, and after one round of training is finished, the client starts the extranet model H i Returning to the server side;
and S2.3, generating parameters of the heterogeneous model.
4. The method for solving heterogeneous federated learning using the ultra-network of claim 3, wherein the S2.3 comprises:
s2.3.1, at the current client, the number of layers of the linear network is set according to the complexity and the structure of a client model, and a super network model H i Generating parameter weights for convolution kernels of a convolutional network, each convolution kernel in a client network comprising K in ×K out Filters, each filter size having f size ×f size Assuming that the convolution kernel parameter weights are all stored in a matrix
Figure FDA0003945552110000011
In the client network, each layer l =1 i For each layer l of the client network, the extranet model H i Will receive a layer embedding>
Figure FDA0003945552110000012
As input and generates->
Figure FDA0003945552110000013
Written as follows:
Figure FDA0003945552110000021
wherein R is a real number, and R is a real number,
Figure FDA0003945552110000022
the method is to save the parameter weight of the convolution kernel of the current layer of the client heterogeneous model
Figure FDA0003945552110000023
F of (a) size K in ×f size K out A dimension real number parameter, and>
Figure FDA0003945552110000024
is description of C i Embedded vector of each layer of network information, supernet H i By learning>
Figure FDA0003945552110000025
Generating the weight of the current model layer I of the client;
s2.3.2, super net H i Will embed the vector
Figure FDA0003945552110000026
Output after linear operation->
Figure FDA0003945552110000027
Figure FDA0003945552110000028
Will be as H i Input of the next layer, H i The last layer of (2) is a linear operation, taken as H i Output on level L-1->
Figure FDA0003945552110000029
As input to the last layer, by H i Last layer will>
Figure FDA00039455521100000210
Linear projection onto->
Figure FDA00039455521100000211
Middle, or>
Figure FDA00039455521100000212
Is the l-th tier weight of the i-th client heterogeneous model,
therefore, the temperature of the molten metal is controlled,
Figure FDA00039455521100000213
the following formula can be written:
Figure FDA00039455521100000214
...
Figure FDA00039455521100000215
Figure FDA00039455521100000216
Figure FDA00039455521100000217
wherein
Figure FDA00039455521100000218
Is H i In based on the weight of (c), in>
Figure FDA00039455521100000219
Is the weight of the ith entire client heterogeneous model.
5. The method for solving heterogeneous federal learning using a hyper network as claimed in claim 1, wherein said S3 comprises:
s3.1, training by the client by taking the current heterogeneous model as a target network, generating weight for the heterogeneous model by the super network, performing gradient descent training end to end together with the heterogeneous model and updating parameters Hi of the super network model, wherein an objective function formula is as follows:
Figure FDA00039455521100000220
the extranet parameter updating formula is as follows:
Figure FDA00039455521100000221
wherein L (-) is a loss function, X and Y respectively correspond to an image and a label of a data set, N is the number of the data sets held by a client, and ζ is the learning rate of the extranet model;
s3.2, after the client training is finished, the client returns the parameter H of the hyper-network model i Wherein
Figure FDA00039455521100000222
Figure FDA00039455521100000223
Carrying out hyper-network aggregation and updating hyper-network model parameters at a server side, and issuing the updated hyper-network model parameters to corresponding client sides:
Figure FDA0003945552110000031
h is a server-side ultra-network model.
6. An electronic device comprising a memory and a processor, the memory storing a computer program, characterized in that the processor realizes the steps of the method according to any of claims 1 to 5 when executing the computer program.
7. A computer storage medium, characterized in that at least one program instruction is stored in the storage medium, which at least one program instruction is loaded and executed by a processor to implement the steps of the method according to any of claims 1 to 5.
CN202211432123.3A 2022-11-16 2022-11-16 Heterogeneous federation learning method, equipment and medium based on super network Active CN115860135B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202211432123.3A CN115860135B (en) 2022-11-16 2022-11-16 Heterogeneous federation learning method, equipment and medium based on super network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202211432123.3A CN115860135B (en) 2022-11-16 2022-11-16 Heterogeneous federation learning method, equipment and medium based on super network

Publications (2)

Publication Number Publication Date
CN115860135A true CN115860135A (en) 2023-03-28
CN115860135B CN115860135B (en) 2023-08-01

Family

ID=85663617

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202211432123.3A Active CN115860135B (en) 2022-11-16 2022-11-16 Heterogeneous federation learning method, equipment and medium based on super network

Country Status (1)

Country Link
CN (1) CN115860135B (en)

Citations (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20210110140A1 (en) * 2020-12-22 2021-04-15 Juan Munoz Environment specific model delivery
CN113554201A (en) * 2020-04-23 2021-10-26 山东大学 Grading prediction system and method based on hyper-network and federal learning
CN113570027A (en) * 2021-06-24 2021-10-29 华为技术有限公司 Method, apparatus, system, device, medium and program product for generating a neural network model
US20210374502A1 (en) * 2020-06-01 2021-12-02 Nvidia Corporation Technique to perform neural network architecture search with federated learning
US20220036194A1 (en) * 2021-10-18 2022-02-03 Intel Corporation Deep neural network optimization system for machine learning model scaling
CN114358316A (en) * 2022-01-14 2022-04-15 中国人民解放军总医院 Federal learning system and large-scale image training method and device thereof
CN114404977A (en) * 2022-01-25 2022-04-29 腾讯科技(深圳)有限公司 Training method of behavior model and training method of structure expansion model
CN114548353A (en) * 2020-11-25 2022-05-27 共达地创新技术(深圳)有限公司 Model training method, electronic device and storage medium
CN114819152A (en) * 2022-01-19 2022-07-29 浙江大学 Graph embedding expert entity alignment method based on reinforcement learning enhancement
CN114882335A (en) * 2022-05-05 2022-08-09 河北工业大学 Intelligent image sensing device for sensing and computing cloud integration based on federal learning framework
CN115065728A (en) * 2022-06-13 2022-09-16 福州大学 Multi-strategy reinforcement learning-based multi-target content storage method
CN115086399A (en) * 2022-07-28 2022-09-20 深圳前海环融联易信息科技服务有限公司 Federal learning method and device based on hyper network and computer equipment
CN115238879A (en) * 2022-08-16 2022-10-25 南京大学 Architecture search method of deep neural network and hardware accelerator
CN115271101A (en) * 2022-08-26 2022-11-01 天津大学 Personalized federal learning method based on graph convolution hyper-network

Patent Citations (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113554201A (en) * 2020-04-23 2021-10-26 山东大学 Grading prediction system and method based on hyper-network and federal learning
US20210374502A1 (en) * 2020-06-01 2021-12-02 Nvidia Corporation Technique to perform neural network architecture search with federated learning
CN114548353A (en) * 2020-11-25 2022-05-27 共达地创新技术(深圳)有限公司 Model training method, electronic device and storage medium
US20210110140A1 (en) * 2020-12-22 2021-04-15 Juan Munoz Environment specific model delivery
CN113570027A (en) * 2021-06-24 2021-10-29 华为技术有限公司 Method, apparatus, system, device, medium and program product for generating a neural network model
US20220036194A1 (en) * 2021-10-18 2022-02-03 Intel Corporation Deep neural network optimization system for machine learning model scaling
CN114358316A (en) * 2022-01-14 2022-04-15 中国人民解放军总医院 Federal learning system and large-scale image training method and device thereof
CN114819152A (en) * 2022-01-19 2022-07-29 浙江大学 Graph embedding expert entity alignment method based on reinforcement learning enhancement
CN114404977A (en) * 2022-01-25 2022-04-29 腾讯科技(深圳)有限公司 Training method of behavior model and training method of structure expansion model
CN114882335A (en) * 2022-05-05 2022-08-09 河北工业大学 Intelligent image sensing device for sensing and computing cloud integration based on federal learning framework
CN115065728A (en) * 2022-06-13 2022-09-16 福州大学 Multi-strategy reinforcement learning-based multi-target content storage method
CN115086399A (en) * 2022-07-28 2022-09-20 深圳前海环融联易信息科技服务有限公司 Federal learning method and device based on hyper network and computer equipment
CN115238879A (en) * 2022-08-16 2022-10-25 南京大学 Architecture search method of deep neural network and hardware accelerator
CN115271101A (en) * 2022-08-26 2022-11-01 天津大学 Personalized federal learning method based on graph convolution hyper-network

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
AVIV SHAMSIAN等: "Personalized Federated Learning using Hypernetworks", 《ARXIV》, pages 1 - 14 *
DIXI YAO等: "Federated Model Search via Reinforcement Learning", 《2021 IEEE 41ST INTERNATIONAL CONFERENCE ON DISTRIBUTED COMPUTING SYSTEMS (ICDCS)》, pages 830 - 840 *
HANGYU ZHU等: "From federated learning to federated neural architecture search: a survey", 《COMPLEX & INTELLIGENT SYSTEMS》, vol. 7, pages 639 - 657, XP093015035, DOI: 10.1007/s40747-020-00247-z *
MIKHAIL KHODAK等: "Weight-Sharing for Hyperparameter Optimization in Federated Learning", 《INTERNATIONAL WORKSHOP ON FEDERATED LEARNING FOR USER PRIVACY AND DATA CONFIDENTIALITY IN CONJUNCTION WITH ICML 2020》, pages 1 - 9 *

Also Published As

Publication number Publication date
CN115860135B (en) 2023-08-01

Similar Documents

Publication Publication Date Title
Chen et al. DNNOff: offloading DNN-based intelligent IoT applications in mobile edge computing
US12008445B2 (en) Black-box optimization using neural networks
CN107609652B (en) Execute the distributed system and its method of machine learning
CN111406264B (en) Neural architecture search
CN110366734A (en) Optimization neural network framework
CN107292352B (en) Image classification method and device based on convolutional neural network
WO2019018375A1 (en) Neural architecture search for convolutional neural networks
CN107958285A (en) The mapping method and device of the neutral net of embedded system
CN111178517B (en) Model deployment method, system, chip, electronic equipment and medium
CN109360097A (en) Prediction of Stock Index method, apparatus, equipment and storage medium based on deep learning
CN109983480A (en) Use cluster loss training neural network
US11423307B2 (en) Taxonomy construction via graph-based cross-domain knowledge transfer
CN110428046A (en) Acquisition methods and device, the storage medium of neural network structure
CN110689136B (en) Deep learning model obtaining method, device, equipment and storage medium
CN111191789B (en) Model optimization deployment system, chip, electronic equipment and medium
CN116450312A (en) Scheduling strategy determination method and system for pipeline parallel training
CN109903100A (en) A kind of customer churn prediction technique, device and readable storage medium storing program for executing
CN112132279A (en) Convolutional neural network model compression method, device, equipment and storage medium
CN111292262A (en) Image processing method, image processing apparatus, electronic device, and storage medium
Chen et al. Computing offloading decision based on DDPG algorithm in mobile edge computing
CN106169961A (en) The network parameter processing method and processing device of neutral net based on artificial intelligence
CN114595815A (en) Transmission-friendly cloud-end cooperation training neural network model method
CN109670579A (en) Model generating method and device
Tong et al. Study on mindspore deep learning framework
CN115860135A (en) Method, apparatus, and medium for solving heterogeneous federated learning using a super network

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant