CN115278709B - Communication optimization method based on federal learning - Google Patents
Communication optimization method based on federal learning Download PDFInfo
- Publication number
- CN115278709B CN115278709B CN202210906790.4A CN202210906790A CN115278709B CN 115278709 B CN115278709 B CN 115278709B CN 202210906790 A CN202210906790 A CN 202210906790A CN 115278709 B CN115278709 B CN 115278709B
- Authority
- CN
- China
- Prior art keywords
- gradient
- matrix
- local
- model
- client
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 35
- 238000004891 communication Methods 0.000 title claims abstract description 24
- 238000005457 optimization Methods 0.000 title claims abstract description 16
- 238000012549 training Methods 0.000 claims abstract description 21
- 230000006835 compression Effects 0.000 claims abstract description 18
- 238000007906 compression Methods 0.000 claims abstract description 18
- 230000002776 aggregation Effects 0.000 claims abstract description 16
- 238000004220 aggregation Methods 0.000 claims abstract description 16
- 238000000354 decomposition reaction Methods 0.000 claims abstract description 8
- 238000006116 polymerization reaction Methods 0.000 claims abstract description 8
- 238000007781 pre-processing Methods 0.000 claims abstract description 5
- 239000011159 matrix material Substances 0.000 claims description 80
- 239000013598 vector Substances 0.000 claims description 22
- 230000008569 process Effects 0.000 claims description 11
- 238000005070 sampling Methods 0.000 claims description 9
- 238000004364 calculation method Methods 0.000 claims description 8
- 238000013441 quality evaluation Methods 0.000 claims description 6
- 230000006870 function Effects 0.000 claims description 2
- 238000003062 neural network model Methods 0.000 claims description 2
- 238000000638 solvent extraction Methods 0.000 claims description 2
- 230000009466 transformation Effects 0.000 claims description 2
- 241001396798 Aldama Species 0.000 claims 1
- 238000010801 machine learning Methods 0.000 abstract description 2
- 230000000694 effects Effects 0.000 description 4
- 230000004048 modification Effects 0.000 description 3
- 238000012986 modification Methods 0.000 description 3
- 239000000284 extract Substances 0.000 description 2
- 238000001303 quality assessment method Methods 0.000 description 2
- 230000035945 sensitivity Effects 0.000 description 2
- 238000009825 accumulation Methods 0.000 description 1
- 230000006978 adaptation Effects 0.000 description 1
- 230000003044 adaptive effect Effects 0.000 description 1
- 230000001186 cumulative effect Effects 0.000 description 1
- 238000004880 explosion Methods 0.000 description 1
- 239000000126 substance Substances 0.000 description 1
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04W—WIRELESS COMMUNICATION NETWORKS
- H04W16/00—Network planning, e.g. coverage or traffic planning tools; Network deployment, e.g. resource partitioning or cells structures
- H04W16/22—Traffic simulation tools or models
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N20/00—Machine learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04W—WIRELESS COMMUNICATION NETWORKS
- H04W24/00—Supervisory, monitoring or testing arrangements
- H04W24/02—Arrangements for optimising operational condition
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Software Systems (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Mathematical Physics (AREA)
- Evolutionary Computation (AREA)
- Computer Networks & Wireless Communication (AREA)
- Data Mining & Analysis (AREA)
- Artificial Intelligence (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- Signal Processing (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Medical Informatics (AREA)
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- General Health & Medical Sciences (AREA)
- Molecular Biology (AREA)
- Management, Administration, Business Operations System, And Electronic Commerce (AREA)
- Compression, Expansion, Code Conversion, And Decoders (AREA)
Abstract
The invention provides a communication optimization method based on federal learning, which belongs to the technical field of federal machine learning and comprises the following steps: acquiring local gradient core characteristics uploaded by each client; the local gradient core features refer to data extracted by preprocessing target data by each client, training a local model and carrying out singular value decomposition on the local model; carrying out sketch gradient compression and sketch gradient aggregation on the local gradient core characteristics, extracting first k sparse parameter matrixes, carrying out a federal average aggregation algorithm on the first k sparse parameter matrixes to obtain a global model, and sending the global model to each client; and calculating an accumulated gradient error value and a gradient error compensation value of each client, adaptively adjusting the gradient to be compensated of the local gradient core characteristics and the weight of the local gradient core characteristics when sketch gradient polymerization is carried out, and then iterating. The method can effectively compress the model gradient, reduce communication overhead generated in model training and improve the accuracy of model prediction.
Description
Technical Field
The invention relates to the technical field of federal machine learning, in particular to a federal learning-based communication optimization method.
Background
The existing federal learning generally uploads intermediate parameters of model training to a parameter server to realize model aggregation, and two problems exist in the process: firstly, large-scale distributed training requires a large amount of communication bandwidth for gradient exchange, which limits the expandability of multi-node training; and secondly, requires expensive high bandwidth network infrastructure.
When distributed training is performed on mobile devices, there is higher network delay, lower system throughput, and intermittent connectivity failures for such training.
In order to solve the above problems, if the conventional SVD singular value decomposition method is directly adopted, the accuracy of model prediction is improved, but the gradient compression effect is not ideal. If a sketch compression method is adopted, although the gradient compression effect is ideal, the accuracy of model prediction is reduced.
Disclosure of Invention
The technical problem of the invention is to provide a federal learning-based communication optimization method, which can effectively compress model gradients, reduce communication overhead generated in model training and improve model prediction accuracy.
In order to achieve the above purpose, the technical scheme adopted by the invention is as follows:
A federal learning-based communication optimization method, comprising the steps of: s1, acquiring local gradient core characteristics uploaded by each client; the local gradient core features refer to model gradients extracted by preprocessing target data, training a local model and carrying out singular value decomposition on the local model by each client; the local model is a federal model; s2, carrying out sketch gradient compression on the local gradient core features, carrying out sketch gradient aggregation, extracting first k sparse parameter matrixes, carrying out a federal average aggregation algorithm on the first k sparse parameter matrixes to obtain a global model, and distributing the global model to each client; s3, calculating an accumulated gradient error value and a gradient error compensation value of each client, adaptively adjusting the gradient to be compensated of the local gradient core feature and the weight of the local gradient core feature when sketch gradient polymerization is carried out, and then iteratively executing S1 to S3.
Further, preprocessing the target data includes: dividing target data into a plurality of data sets, and executing uniform random sampling operation to select the batch size of the data sets in each iteration to obtain the target data sets; performing normal form boundary calculation on the target data set; acquiring initialized global model parameters, training a local model through a target data set calculated by a normal form boundary, and calculating a local gradient of the local model; and cutting the local gradient to obtain a gradient matrix of the local model.
Further, singular value decomposition is performed on the local model into: and each client executes a rapid gradient descent algorithm to decompose a gradient matrix of the local model into a left singular vector, a right singular vector and a diagonal matrix containing gradients, so as to obtain local gradient core characteristics.
Further, each client executes a rapid gradient descent algorithm to decompose a gradient matrix of the local model into a left singular vector, a right singular vector and a diagonal matrix containing gradients, and obtain a local gradient core feature, including: sequentially storing the data of the gradient matrix of the local model in a zero matrix, and compressing until zero-value rows do not exist in the zero matrix to obtain a first matrix; constructing a sub-sampling random Hadamard transform matrix, calculating the product of the sub-sampling random Hadamard transform matrix and a first matrix, and compressing the size of the first matrix from m to mObtaining a second matrix; and executing an SVD algorithm on the second matrix, and decomposing the gradient matrix of the local model into a left singular vector, a right singular vector and a diagonal matrix containing gradients to obtain the core characteristics of the local gradient.
Further, S3 includes: calculating a local model loss value of each client in the current iteration process; calculating cross entropy of the loss value of the local model, and calculating the quality evaluation weight of the local model according to the cross entropy; calculating local gradient errors generated by sketch gradient compression in the sketch gradient aggregation process, and obtaining accumulated gradient error values and gradient error compensation values of all clients; according to the quality evaluation weight and the accumulated gradient error value of the local model of each client, the weight of the local gradient core characteristics of each client when sketch gradient polymerization is carried out is adjusted; and adaptively adjusting the gradient to be compensated of the local gradient core characteristics of each client according to the gradient error compensation value, and then entering the next iteration.
Drawings
The invention and its features, aspects and advantages will become more apparent from the detailed description of non-limiting embodiments with reference to the following drawings. Like numbers refer to like parts throughout. The drawings are not intended to be drawn to scale, emphasis instead being placed upon illustrating the principles of the invention.
FIG. 1 is a general flow chart of a federal learning-based communications optimization method provided by the present invention;
FIG. 2 is a specific flowchart of a preprocessing algorithm in a federal learning-based communication optimization method provided by the invention;
FIG. 3 is a specific flowchart of a fast gradient descent algorithm in a federal learning-based communication optimization method provided by the invention;
Fig. 4 is a specific flowchart for gradient error compensation in a federal learning-based communication optimization method provided by the invention.
Detailed Description
The invention will now be further described with reference to the accompanying drawings and specific examples, which are not intended to limit the invention.
In order to solve the problem that in the prior art, when intermediate parameters trained by a model are uploaded to a parameter server to realize model polymerization in a federal learning algorithm, a traditional SVD singular value decomposition method is directly adopted, and the accuracy of model prediction is improved, but the gradient compression effect is not ideal enough; the method is characterized in that a sketch compression method is directly adopted, the gradient compression effect is ideal, but the problem of model prediction accuracy is reduced, so as shown in fig. 1, the invention provides a communication optimization method based on federal learning, in particular to a mobile edge network communication optimization method, when distributed training is performed under a mobile edge network, the resources of a mobile end are limited, the training can bring the problems of higher network delay, lower system throughput and intermittent network connection failure, the communication optimization is performed on the mobile end, and huge communication expenditure is reduced in the model training process.
Firstly, each client needs to preprocess local target data, and then trains out a local model which is a federal model; this step is shown in fig. 2, and specifically includes:
partitioning target data into minibatch datasets And performing uniform random sampling operation to select the batch size of the data set in each iteration, so as to obtain a target data set d' i=di⊙Sc; wherein i is the ith client, d i is the data of the ith client, c is the number of matrix columns, and S c inputs a random uniform sampling matrix with the same dimension as the sensitive original data;
since each client runs a federal learning model on the target dataset d ' i to extract features, i.e., d ' l=F(d′i), to limit certain global sensitivity, a paradigm boundary calculation is performed on the target dataset d ' i, i.e. Wherein/>Is the minimum data batch size; d l represents an infinite norm of the first layer output of the neural network model; from this formula, the upper limit of d' l is S f=max||f(d)-f(d′)||2; wherein f is a real-valued function, provided that/>The sensitivity of d l can be protected, and sensitive data can be protected from leakage to a certain extent;
initializing global model parameters w global and distributing the parameters to each client, training a local model by each client through a target data set after the calculation of a normal form boundary, and calculating a local gradient g (i) of the local model; clipping local gradients, i.e. So as to prevent gradient explosion and finally obtain the gradient matrix A of the local model.
Then each client executes a rapid gradient descent algorithm, as shown in fig. 3, the gradient matrix of the local model is decomposed into a left singular vector, a right singular vector and a diagonal matrix containing gradients to approximately describe matrix information of a certain percentage of the gradient matrix, and then the gradients are compressed in a matrix form to obtain local gradient core characteristics, so that high communication cost generated by federal training is reduced; the method specifically comprises the following steps:
N data points in the m-th row in a gradient matrix A epsilon R m×n of the local model are collected in sequence and stored in a zero matrix F epsilon R m×n, and all zero-value rows in the zero matrix are compressed to obtain a first matrix F'; wherein R represents a matrix and R m×n represents a dimension of the matrix is m×n;
When zero value exists in the F, constructing a sub-sampling random Hadamard transformation matrix phi=N.H.D; where N ε R q×m represents a scaled random matrix that is uniformly sampled per row without replacing it from m rows of an m identity matrix, rescaling to Wherein q represents the number of rows of the random matrix N; d represents an m×m diagonal matrix with elements of I.I.D, which is a random variable (equal probability-1 or 1); h.epsilon. {. + -.1 } m×m is expressed as a Hadamard matrix (Hadamard), meaning that H ij=(-1)<i-1,j-1>, < i-1, j-1> is the dot product of b-bit binary vectors of integers i-1 and j-1,/> Where i, j represents the ith row and jth column where it is located, the Hadamard matrix may also be expressed as/>Where m is the size of the matrix;
Computing the product of the subsampled random hadamard transform matrix Φ and the first matrix F 'to compress the size of F' from m to Connecting the newly compressed gradient and the contracted gradient to obtain a second matrix B epsilon R k×n;
Performing an SVD algorithm on the second matrix, decomposing the gradient matrix of the local model into a left singular vector, a right singular vector and a diagonal matrix containing gradients according to [ U, Λ, V ] = SVD (B), wherein U represents the left singular vector, V represents the right singular vector, Λ represents the diagonal matrix, and passing through Compression of raw information to front of B/>Line, under the condition that the obtained local gradient core is slightly influenced, the communication efficiency of local model parameter uploading is improved, wherein I k refers to AND/>Unit vector of the same dimension,/>The diagonal matrix of the singular values of the k/2 rows before is obtained by subtracting the meaning of the representation and then deleting the previous calculation, and the diagonal matrix of the singular values of the k/2 rows after the diagonal matrix is reserved.
Carrying out sketch gradient compression on the local gradient core features of each client, carrying out sketch gradient aggregation, extracting first k sparse parameter matrixes, carrying out a federal average aggregation algorithm on the first k sparse parameter matrixes to obtain a global model, and distributing the global model to each client;
then as shown in fig. 4, calculating an accumulated gradient error value and a gradient error compensation value of each client, adaptively adjusting the gradient to be compensated by the local gradient core feature and the weight of the local gradient core feature when carrying out sketch gradient polymerization, and substituting the weights into the next iteration calculation global model; the method specifically comprises the following steps:
Calculating local model loss values of all clients in current iteration process Wherein DO i is the ith client,/>Is a parameter of the local model of the ith client in the t-th iteration;
Calculating cross entropy H (f (x i),yi)=-∑xiyilogf(xi) of local model loss values, by which to evaluate local model loss values If the cross entropy H (f (x i),yi) value is smaller, the predicted result of the representative local model is closer to the real result, and the model training quality is higher;
Then, calculating the quality assessment weight of the local model according to the cross entropy Where (x i,yi) is the local data tag of the i-th client, x i is the input data, y i is the expected output, and f (x i) is the local model prediction result of the i-th client;
Each time the sketch gradient aggregates the quality evaluation weight of each client, the client D i generates a local gradient error by the sketch gradient compression in the sketch gradient aggregation process, wherein D i represents the data set of the ith client and can use the accumulated gradient error value Representation of/> The gradient error is expressed as the gradient error accumulated by the ith client in the t-th iteration, beta is an error factor, and the accumulated gradient error value can be updated continuously along with the progress of the iteration; according to/>Wherein α is a compensation coefficient; /(I)Is the gradient after SVD and Sketch;
According to the quality evaluation weight and the accumulated gradient error value of the local model of each client, the weight of the local gradient core characteristics of each client when sketch gradient polymerization is carried out is adjusted; according to the gradient error compensation value, self-adaptively adjusting the gradient to be compensated of the local gradient core characteristics of each client, selecting optimal parameters alpha and beta, substituting the optimal parameters alpha and beta into the next iterative computation, thereby improving the accuracy of the global model, and particularly representing the optimal parameters in the global model updating computation process as follows Where n is the number of clients participating in model training, ψ is the number of high quality clients uniformly and randomly selected, v is the number of data samples owned by the ith client, η is the local model learning rate, P i(Di) is the quality assessment weight of the local model of the client D i in the ith model aggregation, and the model aggregation is to perform federal average on the uploaded gradient updates of all clients for updating the global model); /(I)Is the local gradient of client i not compensated in iteration t, alpha is the compensation coefficient,/>Is the cumulative gradient error value of the ith client in the t-th iteration.
In summary, according to the federal learning-based communication optimization method provided by the invention, each client side firstly trains a local model independently, then performs singular value decomposition (Singular Value Decomposition, SVD) on the local model, extracts local gradient core features, sends the local gradient core features to a parameter server, performs Sketch (Sketch) gradient compression by the parameter server, performs Sketch (Sketch) gradient aggregation, extracts k approximate gradient values before the aggregated parameters, performs a federal average aggregation algorithm on the k approximate gradient values to obtain a global model, and sends the global model to each client side to perform the next round of iteration process. And for the local gradient error generated after the depletion compression of the Sketch gradient, the adaptive gradient error accumulation and gradient error compensation are supported, the gradient to be compensated is adaptively adjusted and added to the next iteration, the model gradient can be effectively compressed by combining SVD with the depletion compression, the communication overhead generated in the model training is reduced, and the gradient error between the SVD operation and the execution of only SVD operation caused by combining SVD with the depletion operation is solved by gradient error compensation, so that the accuracy of global model prediction is improved. The invention can greatly reduce the communication cost between the client and the server in the whole training process, quicken the model convergence, and the accuracy of model prediction is almost the same as the original prediction accuracy.
The foregoing describes preferred embodiments of the present invention; it is to be understood that the invention is not limited to the specific embodiments described above, wherein devices and structures not described in detail are to be understood as being implemented in a manner common in the art; any person skilled in the art will make many possible variations and modifications, or adaptations to equivalent embodiments without departing from the technical solution of the present invention, which do not affect the essential content of the present invention; therefore, any simple modification, equivalent variation and modification of the above embodiments according to the technical substance of the present invention still fall within the scope of the technical solution of the present invention.
Claims (2)
1. A federal learning-based communication optimization method, comprising the steps of:
S1, acquiring local gradient core characteristics uploaded by each client; the local gradient core features refer to model gradients extracted by performing preprocessing on target data, training a local model and performing singular value decomposition on the local model by each client; the local model is a federal model;
s2, carrying out sketch gradient compression on the local gradient core features, carrying out sketch gradient aggregation, extracting first k sparse parameter matrixes, carrying out a federal average aggregation algorithm on the first k sparse parameter matrixes to obtain a global model, and distributing the global model to each client;
s3, calculating an accumulated gradient error value and a gradient error compensation value of each client, adaptively adjusting the gradient to be compensated of the local gradient core feature and the weight of the local gradient core feature when sketch gradient polymerization is carried out, and then iteratively executing S1 to S3;
each client preprocesses local target data, trains a local model, and the local model is a federal model; the implementation steps specifically comprise:
partitioning target data into minibatch datasets And performing uniform random sampling operation to select the batch size of the data set in each iteration, so as to obtain a target data set d' i=di⊙Sc; wherein i is the ith client, d i is the data of the ith client, c is the number of matrix columns, and S c inputs a random uniform sampling matrix with the same dimension as the sensitive original data;
Each client runs a federal learning model on the target dataset d ' i to extract features, d ' l=F(d′i), performs a paradigm boundary calculation on the target dataset d ' i, i.e Wherein/>Is the minimum data batch size; d l represents an infinite norm of the first layer output of the neural network model; the upper limit of d' l is S f=max||f(d)-f(d′)||2; wherein f is a real-valued function;
Initializing global model parameters w global and distributing the parameters to each client, training a local model by each client through a target data set after the calculation of a normal form boundary, and calculating a local gradient g (i) of the local model; the local gradient is cut out, Obtaining a gradient matrix A of a local model;
Each client executes a rapid gradient descent algorithm to decompose a gradient matrix of a local model into a matrix information which approximately describes a certain percentage of the gradient matrix and comprises a left singular vector, a right singular vector and a diagonal matrix of the gradient, compresses the gradient in a matrix form, and obtains a local gradient core feature, and the method specifically comprises the following steps:
N data points in the m-th row in a gradient matrix A epsilon R m×n of the local model are collected in sequence and stored in a zero matrix F epsilon R m×n, and all zero-value rows in the zero matrix are compressed to obtain a first matrix F'; wherein R represents a matrix and R m×n represents a dimension of the matrix is m×n;
When zero value exists in the F, constructing a sub-sampling random Hadamard transformation matrix phi=N.H.D; where N ε R q×m represents a scaled random matrix that is uniformly sampled per row without replacing it from m rows of an m identity matrix, rescaling to Wherein q represents the number of rows of the random matrix N; d represents m multiplied by m diagonal matrix with the element of I.I.D, and is random variable, equal probability-1 or 1; h.epsilon.+ -. 1} m×m is represented as a Aldama matrix, meaning H ij=(-1)<i-1,j-1>, < i-1, j-1> is the dot product of b-bit binary vectors of integers i-1 and j-1,/>Where i, j represents the ith row and jth column where it is located, the Hadamard matrix may also be expressed as/> Where m is the size of the matrix;
Calculating the product of the subsampled random Hadamard transform matrix phi and the first matrix F', and connecting the newly compressed gradient and the contracted gradient to obtain a second matrix B epsilon R k×n;
Performing an SVD algorithm on the second matrix, decomposing the gradient matrix of the local model into a left singular vector, a right singular vector and a diagonal matrix containing gradients according to [ U, Λ, V ] = SVD (B), wherein U represents the left singular vector, V represents the right singular vector, Λ represents the diagonal matrix, and passing through Compression of raw information to front of B/>Line, where I k refers to AND/>Unit vector of the same dimension,/>The diagonal matrix of the singular values of the k/2 rows is obtained by deleting the singular values of the k/2 rows before calculation, and the diagonal matrix of the singular values of the k/2 rows after the subtraction is reserved.
2. A federally learning-based communications optimization method according to claim 1, wherein S3 comprises:
Calculating a local model loss value of each client in the current iteration process;
Calculating cross entropy of the local model loss value, and calculating quality evaluation weight of the local model according to the cross entropy;
Calculating local gradient errors generated by sketch gradient compression in the sketch gradient aggregation process, and obtaining accumulated gradient error values and gradient error compensation values of all clients;
According to the quality evaluation weight and the accumulated gradient error value of the local model of each client, the weight of the local gradient core characteristics of each client when sketch gradient polymerization is carried out is adjusted; and adaptively adjusting the gradient to be compensated of the local gradient core characteristics of each client according to the gradient error compensation value, and then entering the next iteration.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202210906790.4A CN115278709B (en) | 2022-07-29 | 2022-07-29 | Communication optimization method based on federal learning |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202210906790.4A CN115278709B (en) | 2022-07-29 | 2022-07-29 | Communication optimization method based on federal learning |
Publications (2)
Publication Number | Publication Date |
---|---|
CN115278709A CN115278709A (en) | 2022-11-01 |
CN115278709B true CN115278709B (en) | 2024-04-26 |
Family
ID=83772330
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202210906790.4A Active CN115278709B (en) | 2022-07-29 | 2022-07-29 | Communication optimization method based on federal learning |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN115278709B (en) |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117216596B (en) * | 2023-08-16 | 2024-04-30 | 中国人民解放军总医院 | Federal learning optimization communication method, system and storage medium based on gradient clustering |
CN117692939B (en) * | 2024-02-02 | 2024-04-12 | 南京邮电大学 | Client scheduling method in dynamic communication environment |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111507481A (en) * | 2020-04-17 | 2020-08-07 | 腾讯科技(深圳)有限公司 | Federated learning system |
CN112199702A (en) * | 2020-10-16 | 2021-01-08 | 鹏城实验室 | Privacy protection method, storage medium and system based on federal learning |
CN112288097A (en) * | 2020-10-29 | 2021-01-29 | 平安科技(深圳)有限公司 | Federal learning data processing method and device, computer equipment and storage medium |
CN112862011A (en) * | 2021-03-31 | 2021-05-28 | 中国工商银行股份有限公司 | Model training method and device based on federal learning and federal learning system |
CN113222179A (en) * | 2021-03-18 | 2021-08-06 | 北京邮电大学 | Federal learning model compression method based on model sparsification and weight quantization |
CN113645197A (en) * | 2021-07-20 | 2021-11-12 | 华中科技大学 | Decentralized federal learning method, device and system |
CN114339252A (en) * | 2021-12-31 | 2022-04-12 | 深圳大学 | Data compression method and device |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10754764B2 (en) * | 2018-04-22 | 2020-08-25 | Sas Institute Inc. | Validation sets for machine learning algorithms |
US20190087713A1 (en) * | 2017-09-21 | 2019-03-21 | Qualcomm Incorporated | Compression of sparse deep convolutional network weights |
US11074317B2 (en) * | 2018-11-07 | 2021-07-27 | Samsung Electronics Co., Ltd. | System and method for cached convolution calculation |
-
2022
- 2022-07-29 CN CN202210906790.4A patent/CN115278709B/en active Active
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111507481A (en) * | 2020-04-17 | 2020-08-07 | 腾讯科技(深圳)有限公司 | Federated learning system |
CN112199702A (en) * | 2020-10-16 | 2021-01-08 | 鹏城实验室 | Privacy protection method, storage medium and system based on federal learning |
CN112288097A (en) * | 2020-10-29 | 2021-01-29 | 平安科技(深圳)有限公司 | Federal learning data processing method and device, computer equipment and storage medium |
CN113222179A (en) * | 2021-03-18 | 2021-08-06 | 北京邮电大学 | Federal learning model compression method based on model sparsification and weight quantization |
CN112862011A (en) * | 2021-03-31 | 2021-05-28 | 中国工商银行股份有限公司 | Model training method and device based on federal learning and federal learning system |
CN113645197A (en) * | 2021-07-20 | 2021-11-12 | 华中科技大学 | Decentralized federal learning method, device and system |
CN114339252A (en) * | 2021-12-31 | 2022-04-12 | 深圳大学 | Data compression method and device |
Non-Patent Citations (3)
Title |
---|
Laizhong Cui ; Xiaoxin Su ; Yipeng Zhou ; Lei Zhang."ClusterGrad: Adaptive Gradient Compression by Clustering in Federated Learning".《GLOBECOM 2020 - 2020 IEEE Global Communications Conference》.2021,全文. * |
P. Ravichandran ; C. Saravanakumar ; J. Dafni Rose ; M. Vijayakumar ; V. Muthu Lakshmi."Efficient Multilevel Federated Compressed Reinforcement Learning of Smart Homes Using Deep Learning Methods".《2021 International Conference on Innovative Computing, Intelligent Communication and Smart Electrical Systems (ICSES)》.2021,全文. * |
物联网中数据压缩与资源分配的研究;杨阳;《中国博士学位论文全文数据库(电子期刊)》;20200515;全文 * |
Also Published As
Publication number | Publication date |
---|---|
CN115278709A (en) | 2022-11-01 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN115278709B (en) | Communication optimization method based on federal learning | |
CN113222179B (en) | Federal learning model compression method based on model sparsification and weight quantification | |
CN111555781B (en) | Large-scale MIMO channel state information compression and reconstruction method based on deep learning attention mechanism | |
Khashman et al. | Image compression using neural networks and Haar wavelet | |
CN110677651A (en) | Video compression method | |
Aaron et al. | Dynamic incremental k-means clustering | |
CN104199627B (en) | Gradable video encoding system based on multiple dimensioned online dictionary learning | |
Saravanan et al. | Intelligent Satin Bowerbird Optimizer Based Compression Technique for Remote Sensing Images. | |
CN114581544A (en) | Image compression method, computer device and computer storage medium | |
CN110753225A (en) | Video compression method and device and terminal equipment | |
CN110929798A (en) | Image classification method and medium based on structure optimization sparse convolution neural network | |
CN112183742A (en) | Neural network hybrid quantization method based on progressive quantization and Hessian information | |
CN115829027A (en) | Comparative learning-based federated learning sparse training method and system | |
CN116192209A (en) | Gradient uploading method for air computing federal learning under MIMO channel | |
CN116168197A (en) | Image segmentation method based on Transformer segmentation network and regularization training | |
CN111010222A (en) | Deep learning-based large-scale MIMO downlink user scheduling method | |
Ndubuaku et al. | Edge-enhanced analytics via latent space dimensionality reduction | |
CN112492312B (en) | Image compression recovery method, device, equipment and medium based on wavelet transform | |
CN114677545B (en) | Lightweight image classification method based on similarity pruning and efficient module | |
Chen et al. | Rate distortion optimization for adaptive gradient quantization in federated learning | |
CN114419341A (en) | Convolutional neural network image identification method based on transfer learning improvement | |
Limmer et al. | Optimal deep neural networks for sparse recovery via Laplace techniques | |
Qiao et al. | Dcs-risr: Dynamic channel splitting for efficient real-world image super-resolution | |
Qian et al. | Deep Image Semantic Communication Model for Artificial Intelligent Internet of Things | |
CN117236201B (en) | Diffusion and ViT-based downscaling method |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |