WO2023029944A1 - 联邦学习的方法和装置 - Google Patents

联邦学习的方法和装置 Download PDF

Info

Publication number
WO2023029944A1
WO2023029944A1 PCT/CN2022/112055 CN2022112055W WO2023029944A1 WO 2023029944 A1 WO2023029944 A1 WO 2023029944A1 CN 2022112055 W CN2022112055 W CN 2022112055W WO 2023029944 A1 WO2023029944 A1 WO 2023029944A1
Authority
WO
WIPO (PCT)
Prior art keywords
operator
operators
federated learning
composite
data
Prior art date
Application number
PCT/CN2022/112055
Other languages
English (en)
French (fr)
Inventor
郑子木
李冉
孙铭
单雨昕
普杰
马坚伟
Original Assignee
华为云计算技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 华为云计算技术有限公司 filed Critical 华为云计算技术有限公司
Publication of WO2023029944A1 publication Critical patent/WO2023029944A1/zh

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • G06N20/20Ensemble learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent

Definitions

  • This application relates to the field of artificial intelligence, and more specifically, to a method and device for federated learning.
  • This application provides a method and device for federated learning, which can improve the efficiency and quality of federated learning.
  • An embodiment of the present application provides a federated learning method, including: obtaining N operators from an operator library according to the target federated learning task, the operator library includes K operators associated with federated learning, the N and the K are integers greater than 1, and the N is less than or equal to the K; multiple operators in the N operators are combined to obtain a first composite operator; through the first composite operator Process the first data set to obtain a second data set, the first data set is a training set for the target federated learning task; use the second data set to perform federated learning to obtain the target federated learning task The corresponding first federated learning model.
  • the federated learning method provided by the embodiment of this application obtains a composite operator for solving multi-objective problems by compounding multiple operators, and processes the training set of this federated learning through the composite operator, which can Realize the simultaneous optimization of multiple target problems, and solve the challenge of multi-problem superposition in the actual scene of federated learning.
  • the K operators associated with federated learning can be understood as that the K operators can be used to process the local training set in federated learning, or that the K operators are related to various federated learning An operator corresponding to a single technology.
  • the target federated learning task may include client requirement information.
  • the client requirement information here may be a comprehensive requirement negotiated and determined by the client before model training.
  • obtaining N operators from the operator library according to the target federated learning task may include: obtaining the target federated learning task, and then according to the target federated learning task from N operators are selected from the operator library, and the N operators are associated with the target federated learning task, and the target federated learning task includes customer demand information.
  • the operators used for this federated learning are operators used to solve customer needs, that is, operators used to perform this target federated learning task, so through the The composite operator obtained by compounding some or all of the N operators can meet the challenge of multi-problem superposition in federated learning.
  • the combining multiple operators among the N operators to obtain a first composite operator includes: according to the N operators Composite operator relationship between any two operators in the subgroup, multiple operators among the N operators are composited to obtain the first composite operator.
  • the compound operator is obtained based on the guidance of the operator compound relationship, that is, the adjustment direction of the structure of the compound operator is guided according to the operator compound relationship, so that the efficiency can be improved, and the practicability and performance of the compound operator can be improved at the same time. availability.
  • the operator composite relationship here can be represented by an operator composite relationship matrix.
  • the operator composite relationship matrix is an N-order matrix, and the rows of the matrix represent the output, and the columns represent the input (or reversed). For each element in the matrix, 1 means it can be combined, and 0 means it cannot be combined.
  • the first composite operator here refers to a collective name for multiple operators in the composite structure after multiple operators in the N operators are combined in series to form a certain operator composite structure.
  • the composite operator can also be called by other names
  • the method further includes: determining the matching relationship of input and output spaces between any two operators among the N operators; according to the The matching relationship of input and output spaces between any two operators in the N operators determines the operator composition relationship between any two operators in the N operators.
  • the composite relationship between operators can be obtained according to the matching relationship between the input and output spaces between operators, so that the composability between arbitrary operators can be obtained more accurately, and the efficiency of operator composite can be improved.
  • the method further includes: determining whether the first condition is met; If the first condition is not satisfied, the structure of the first compound operator is adjusted to obtain the second compound operator.
  • the structure of the first compound operator can be adjusted to obtain a compound operator that better meets user needs or has better performance.
  • the determining whether the first condition is met includes: acquiring measurement parameters of the first federated learning model; calculating an evaluation value according to the measurement parameters, The measurement parameter is associated with the target federated learning task; if the evaluation value is smaller than a first threshold, it is determined that the first condition is satisfied.
  • the calculating the evaluation value according to the measurement parameter includes: calculating the evaluation value according to the measurement parameter and a cost calculation formula, and the cost
  • the calculation formula is associated with the target federated learning task, or in other words, the form of the cost calculation formula is related to customer requirements.
  • the cost calculation formula may be in the form of a linear sum of measured parameters, the coefficients of which are associated with the target federated learning task.
  • the first threshold may be set according to the target federated learning task, for example, the first threshold is set according to the target parameter of the target federated learning model.
  • the calculation result of the cost calculation formula is the performance index, and the performance index of the federated learning model meets the set threshold by continuously adjusting the structure of the compound operator, so that the quality of this federated learning can be guaranteed and the federated learning can be improved.
  • the usefulness of the model is the performance index, and the performance index of the federated learning model meets the set threshold by continuously adjusting the structure of the compound operator, so that the quality of this federated learning can be guaranteed and the federated learning can be improved.
  • the determining whether the first condition is met includes: determining the number of times to adjust the structure of the compound operator; If it is greater than or equal to the second threshold, it is determined that the first condition is met.
  • the number of times of adjusting the structure of the compound operator is used as the termination condition, which can improve the efficiency of operator compounding and save resources.
  • the adjusting the structure of the first compound operator to obtain the second compound operator includes: according to the measurement parameters and the target federated model The difference between the parameters is used to calculate a loss value, the loss value is backpropagated, and the structure of the first composite operator is adjusted according to the loss value to obtain a second composite operator.
  • the loss value (or called loss function value, the application does not limit the specific form of the loss function) may also be calculated according to the difference between the evaluation value and the target evaluation value.
  • the adjustment process of the compound operator structure can be carried out in a directional manner guided by the loss value, thereby improving the adjustment efficiency of the operator compound structure and saving costs.
  • adjusting the structure of the first composite operator to obtain the second composite operator includes: when the first federated learning model When the first parameter of is less than the third threshold, the first operator among the N operators is combined with the first compound operator to obtain the second compound operator, and the first operator and the associated with the first parameter.
  • adjusting the structure of the first composite operator to obtain the second composite operator includes: The structure is automatically tuned to obtain the second composite operator.
  • the operator library includes at least one of a dataset distillation operator, a client grouping operator, a virtual sample balance operator, and a meta-learning operator. kind.
  • the embodiment of the present application provides a federated learning device, which includes a unit for implementing any possible implementation manner in the method design of the first aspect above.
  • the embodiment of the present application provides a federated learning device, including a transceiver and a processor.
  • the device further includes a memory.
  • the processor is used to control the transceiver to send and receive signals
  • the memory is used to store a computer program
  • the processor is used to call and run the computer program from the memory, so that the computer device executes any possible method in the method design of the first aspect above. method in the implementation.
  • an embodiment of the present application provides an electronic device, which may be a device for implementing federated learning in the method design of the first aspect above, or a chip disposed in the device.
  • the electronic device includes: a processor, coupled to a memory, and configured to execute instructions and/or program codes in the memory, so as to implement the method in any possible implementation manner in the method design of the first aspect above.
  • the electronic device further includes a memory.
  • the electronic device further includes a communication interface, and the processor is coupled to the communication interface.
  • the communication interface may be a transceiver, or an input/output interface.
  • the communication interface may be an input/output interface.
  • the transceiver may be a transceiver circuit.
  • the input/output interface may be an input/output circuit.
  • the embodiment of the present application provides a computer program product, the computer program product including: computer program code, when the computer program code is run on the computer, the computer is made to execute any of the above-mentioned method designs of the first aspect.
  • the embodiments of the present application provide a computer-readable medium, the computer-readable medium stores program codes, and when the computer program codes run on the computer, the computer executes the above-mentioned method in the first aspect.
  • Figure 1 is a schematic diagram of an artificial intelligence main frame.
  • Fig. 2 is a schematic diagram of a system architecture of federated learning.
  • Figure 3 is a flowchart of the data set distillation federated learning technology.
  • Fig. 4 is a flow chart of client group federated learning technology.
  • Figure 5 is a flowchart of federated meta-learning technology.
  • Figure 6 is a flowchart of data generation federated learning technology.
  • Fig. 7 is a schematic block diagram of a federated learning method according to an embodiment of the present application.
  • Fig. 8 is a system framework diagram of federated learning provided by the embodiment of the present application.
  • FIG. 9 is a flow chart of a method for executing an operator composite relational model provided by an embodiment of the present application.
  • Fig. 10 is a flow chart of the execution method of the operator compound decision module provided by the embodiment of the present application.
  • Fig. 11 is a schematic structural block diagram of an apparatus for federated learning according to an embodiment of the present application.
  • Fig. 12 is a schematic structural block diagram of an apparatus for federated learning according to an embodiment of the present application.
  • Figure 1 is a schematic diagram of an artificial intelligence main framework, which describes the overall workflow of an artificial intelligence system, and is applicable to general artificial intelligence field requirements.
  • Intelligent information chain reflects a series of processes from data acquisition to processing. For example, it can be the general process of intelligent information perception, intelligent information representation and formation, intelligent reasoning, intelligent decision-making, intelligent execution and output. In this process, the data has undergone a condensed process of "data-information-knowledge-wisdom".
  • IT value chain reflects the value brought by artificial intelligence to the information technology industry from the underlying infrastructure of artificial intelligence, information (provided and processed by technology) to the systematic industrial ecological process.
  • the infrastructure provides computing power support for the artificial intelligence system, realizes communication with the outside world, and realizes support through the basic platform.
  • the infrastructure can communicate with the outside through sensors, and the computing power of the infrastructure can be provided by smart chips.
  • the smart chip here can be a central processing unit (central processing unit, CPU), a neural network processor (neural-network processing unit, NPU), a graphics processing unit (graphics processing unit, GPU), an application specific integrated circuit (application specific) Integrated circuit, ASIC) and field programmable gate array (field programmable gate array, FPGA) and other hardware acceleration chips.
  • CPU central processing unit
  • NPU neural network processor
  • NPU graphics processing unit
  • ASIC application specific integrated circuit
  • FPGA field programmable gate array
  • the basic platform of infrastructure can include related platform guarantees and supports such as distributed computing framework and network, and can include cloud storage and computing, interconnection and interworking network, etc.
  • data can be obtained through sensors and external communication, and then these data can be provided to smart chips in the distributed computing system provided by the basic platform for calculation.
  • Data from the upper layer of the infrastructure is used to represent data sources in the field of artificial intelligence.
  • the data involves graphics, images, voice, text, sequences, and IoT data of traditional equipment, including business data of existing systems and sensory data such as force, displacement, liquid level, temperature, and humidity.
  • a sequence can be understood as a data sequence, the most common time-series sequence data, that is, time-series data, such as weather forecast data (temperature, wind direction, etc.) within a period of time, and physiological data such as stock market data, human blood sugar change data, etc. sequence etc.
  • the above data processing usually includes data training, machine learning, deep learning, search, reasoning, decision-making and other processing methods.
  • machine learning and deep learning can symbolize and formalize intelligent information modeling, extraction, preprocessing, training, etc. of data.
  • Reasoning refers to the process of simulating human intelligent reasoning in a computer or intelligent system, and using formalized information to carry out machine thinking and solve problems according to reasoning control strategies.
  • the typical functions are search and matching.
  • Decision-making refers to the process of decision-making after intelligent information is reasoned, and usually provides functions such as classification, sorting, and prediction.
  • some general-purpose capabilities can be formed based on the results of data processing, such as algorithms or a general-purpose system, such as translation, text analysis, computer vision processing, speech recognition, image processing identification, etc.
  • Intelligent products and industry applications refer to the products and applications of artificial intelligence systems in various fields. It is the packaging of the overall solution of artificial intelligence, which commercializes intelligent information decision-making and realizes landing applications. Its application fields mainly include: intelligent manufacturing, intelligent transportation, Smart home, smart medical care, smart security, automatic driving, safe city, smart terminals, etc.
  • the embodiments of the present application can be applied in many fields of artificial intelligence, for example, intelligent manufacturing, intelligent transportation, intelligent home, intelligent medical care, intelligent security, automatic driving, safe city and other fields.
  • the embodiments of the present application can be specifically applied in various data processing fields, such as image classification, image retrieval, image semantic segmentation, image super-resolution and other image data-related processing tasks, such as speech synthesis, speech recognition, Processing tasks related to voice data such as voice search, processing tasks related to text data such as intelligent translation and semantic extraction, and processing tasks related to graphic data such as sketch drawing and gesture control.
  • the processing tasks related to image data can also be video image data, for example, it can be applied to tasks such as video search, video classification, and key person tracking.
  • the embodiments of the present application may also be applied to data sequence processing tasks, such as time series data.
  • tasks are only exemplified according to the data types of the processed data involved, but some tasks may involve multiple data types, for example, it may be a task of searching images by voice, that is, input a piece of voice, according to this paragraph Speech finds the corresponding image from the image library, and this task involves both image data and speech data.
  • FIG. 2 is a schematic diagram of a federated learning system architecture applicable to the embodiment of the present application.
  • the federated learning system is only one of various federated learning architectures. This system is only an example, and the embodiment of the present application is also It can be applied to federated learning systems in other architectural forms, which is not limited in this application.
  • this scenario includes user equipment #1, user equipment #2, and cloud device #1.
  • user equipment #1 In order to ensure user data privacy and security, data interaction between user equipment cannot be performed.
  • cloud device #1 In order to ensure user data privacy and security, data interaction between user equipment cannot be performed.
  • there may also be three or more user equipments in FIG. 2 and the principle is the same, which is omitted here.
  • Local data #1 and local data #2 represent the data that user equipment #1 and user equipment #2 can provide respectively, or it can be understood that local data #1 and local data #2 are user equipment #1 and user equipment #2 respectively Respective local real data, in federated learning, these data do not want to be obtained by other user devices. These data can be image data, speech data, text data, sequence data, etc., for example.
  • User equipment #1 uses local data #1 to perform model training to obtain local model #1; user equipment #2 uses local data #2 to perform model training to obtain local model #2. Afterwards, the user equipment #1 sends the local model #1 or the parameters of the local model #1 to the cloud device #1, and the user equipment #2 sends the local model #2 or the parameters of the local model #2 to the cloud device #1.
  • cloud device #1 After cloud device #1 obtains all local models (including local model #1 and local model #2), it aggregates all obtained local models into a federated model.
  • the federated model can also be understood as a unified model, that is to say, all user devices Able to generalize the model.
  • cloud device #1 sends the obtained federated model to all user devices.
  • the user device After receiving the federated model, the user device replaces the local model with the federated model, and then repeats the next round of training: training on the basis of the federated learning model to obtain A new local model is uploaded, a new local model is uploaded, and the model converges to a new federated model, and so on, until certain preset requirements are met.
  • the entire training process can be regarded as a continuous adjustment of the model parameters of the federated model, so that the parameters of the federated model can be applied to all user devices at the same time.
  • Local data #1 is the image sample data (image with the label of animal category) of user equipment #1
  • local data #2 is the image sample data (image with the label of animal category) of user equipment #2, that is Say, the sample dimensions of local data #1 and local data #2 can be aligned, but the feature dimensions of local data #1 and local data #2 are different. Then, during the training process, user equipment #1 uses local data #1 to train to obtain local model #1, and user equipment #2 uses local data #2 to train to obtain local model #2.
  • user equipment #1 uploads the model parameters of local model #1 to cloud equipment #1
  • user equipment #2 uploads the model parameters of local model #2 to cloud equipment #1
  • cloud equipment #1 uploads local model #1 and local Model #2 is aggregated into a federated model
  • the cloud device distributes the federated model to user equipment #1 and user equipment #2
  • user equipment #1 and user equipment #2 use local data to continue training on the basis of the federated model to obtain local model, repeat the above process until the parameters of the federated model can be applied to both user equipment #1 and user equipment #2
  • cloud equipment #1 distributes the federation model to user equipment #1 and user equipment #2
  • user equipment #1 and UE #2 receive the federated model and replace the original local model with it.
  • This application provides a method of federated learning, hoping to solve the challenge of multi-problem superposition in federated learning, and improve the quality and efficiency of federated learning.
  • Federated learning is an encrypted distributed machine learning technology, which can also be called federated learning or alliance learning. Federated learning can make full use of the data and computing power of the participants, so that multiple parties can collaborate to build a general and robust machine learning model without sharing data. Therefore, federated learning can effectively help multiple institutions meet user privacy protection Data usage and learning modeling under the requirements of , data security and government regulations. Federated learning includes horizontal federated learning and vertical federated learning. It can be said that federated learning is to enable the sharing of knowledge and parameters without exchanging any of its own data. In the context of increasingly stringent data regulation, federated learning can solve key issues such as data ownership, data privacy, data access rights, and access to heterogeneous data.
  • the horizontal row of the data matrix represents a training sample
  • the vertical column represents a data feature.
  • Horizontal federated learning is to combine multiple rows of samples with the same characteristics of multiple participants for federated learning, that is, the training data of each participant is divided horizontally.
  • Horizontal federated learning is also called feature-aligned federated learning, that is, the data features of participants are aligned, and horizontal federated learning can increase the total number of training samples.
  • Vertical federated learning is to combine different data features of common samples of multiple participants for federated learning, that is, the training data of each participant is divided vertically.
  • Vertical federated learning is also called sample-aligned federated learning (sample-aligned federated learning), that is, the training samples of participants are aligned, and vertical federated learning can increase the feature dimension of training data.
  • Federated learning has been applied in the industry.
  • Google applied it to its own GBoard project
  • the FATE federated learning framework proposed by WeBank, etc. are typical examples of the practical application of the federated learning method of the above-mentioned parameter server architecture. , they all require very high consistency between different devices participating in the learning, the model structure is consistent, and the data can be aligned. This makes the existing federated learning methods not applicable to some scenarios that do not meet the above consistency requirements.
  • dataset distillation does not transfer knowledge from complex models to small models, but distillation at the data level.
  • Dataset distillation is the process of compressing knowledge from a large training dataset into a small amount of synthetic data.
  • the model can be quickly trained to achieve high accuracy by several steps of gradient descent on synthetic data.
  • the data obtained after dataset distillation can achieve a training effect similar to that of the original data.
  • Dataset distillation is equivalent to reducing data, so although the distilled data contains as much data knowledge as possible, it will be affected by the way the model is initialized. Influence.
  • Group federated learning refers to a federated learning method that groups clients to reduce the problem of large differences in the application effects of each side. Since the federated learning of a single model is difficult to meet the conditions that the cloud model is applicable to many clients, and the application effect on each edge is quite different, it is considered to group clients, for example, group clients according to similarity, and then perform federation within the group Learning forms multiple models to suit the needs of multiple client groups.
  • Meta learning (meta learning)
  • Meta-learning is a method of learning how to do machine learning, and its core idea is the process of learning to train.
  • Meta-learning trains a model from a large number of tasks and learns faster in new tasks with a small amount of data.
  • Meta-learning aims to find internal representations that can be easily adapted to new tasks.
  • internal representations can be viewed as a set of parameters for a neural network. Training the network requires only a small amount of new data to fine-tune the network. An effective model for the new task can be obtained.
  • a hyperparameter is a parameter whose value is set before starting the learning process, rather than the parameter data obtained through training.
  • This technology effectively reduces the communication cost in the process of federated learning by replacing the process of multiple rounds of side-cloud transmission model parameters in traditional federated learning with the transmission of one-time distillation data.
  • the schematic diagram of the technology implementation process is shown in Figure 1.
  • This technology divides the groups according to the similarity of the parameter updates generated by the client during the federated learning process, and performing federated optimization in each group can greatly improve the test accuracy of the model in each group and improve efficiency.
  • the schematic diagram of the technology implementation process is shown in Figure 2.
  • This technology achieves model personalization by combining meta-learning ideas.
  • Each client meta-learns a good initialization weight and obtains a common initial model as a starting point. The purpose is to enable the model to quickly migrate and adapt to new tasks. Then each client updates the model based on the initial model and a small amount of local data, and finally obtains the personalized model.
  • the schematic diagram of the technology implementation process is shown in Figure 3.
  • This technology aims at the tags corresponding to a small amount of data in the local data of the client.
  • the client downloads the GAN generator to supplement the local target tag data until each client data set is independent and simultaneous. distribution (i.e. IID).
  • IID independent and simultaneous. distribution
  • FIG. 7 is an exemplary flow chart of a federated learning method 100 provided by an embodiment of the present application. The steps described in FIG. 7 will be introduced below. It should be understood that the example in FIG. 7 is only intended to help those skilled in the art understand the embodiment of the present application, and is not intended to limit the embodiment of the present application to the specific values or specific scenarios in FIG. 7 . Those skilled in the art can obviously make various equivalent modifications or changes based on the given examples, and such modifications and changes also fall within the scope of the embodiments of the present application.
  • the operator library includes K operators associated with federated learning, where K is an integer greater than 1.
  • the operator library here may be an operator library regularly maintained by humans.
  • the K operators associated with federated learning here can be understood as that the K operators can be used to process the local training set in federated learning, or that the K operators are a single technology related to various federated learning corresponding operators, so the K operators can be called K federated operators.
  • the operator stored in the operator library may include a function corresponding to the operator and a normalized input and output.
  • the K operators include: a data set distillation operator, whose input is original data, and whose output is distillation data; and a client grouping operator, whose input is an original client set and data, and whose output is a new client set.
  • the operator library may only include the K operators, that is, the operator library may be a federated operator library; or, the operator library may include other different types of operators in addition to the K operators. Operators, that is, in addition to multiple federated operators, the operator library may also include other types of operators. Different types of operators can be classified and stored with different labels. This application does not limit the structure of the original operator library.
  • the target federated learning task may include client requirement information.
  • the client requirement information here may be a comprehensive requirement negotiated and determined by the client before model training.
  • the requirement includes that the amount of data communication is small without performance improvement, or the requirement includes that the requirement for performance is higher than the requirement for the amount of communication.
  • Obtaining N operators from the operator library according to the target federated learning task may include: obtaining the target federated learning task, and then selecting N operators from the operator library according to the target federated learning task, and the N operators are related to the target Federated learning task associations. For example, according to the target federated learning task, it is determined that the client’s requirements are small communication and data balance, then select N operators for the target federated learning task in the operator library, or select N operators in the operator library Operators used to implement customer requirements, for example, the N operators are dataset distillation operators and client grouping operators. Where N is an integer greater than 1 and N is less than or equal to K, that is, the N operators should be a subset of the K operators.
  • the set of N operators can be called an operator library for this federated learning.
  • the first composite operator here refers to a collective term for multiple operators in the composite structure after multiple operators in the N operators are combined in series to form a certain operator composite structure.
  • the composite operator may be called by other names, which is not limited in this application.
  • multiple operators among the N operators can be composited according to the operator composite relationship of the N operators, where the operator composite relationship can be used to represent any two operators among the N operators Composite relationship between operators, that is, it can be determined whether any two operators among the N operators can be composited according to the operator composite relationship.
  • the operator composite relationship may be represented by an operator composite relationship matrix.
  • an N ⁇ N zero matrix is generated, and then the compound relationship between operator i and operator j is judged.
  • the operator i and operator j are any two operators among the N operators, and the operator The i operators j can be the same or different. For example, by determining the matching relationship between the input and output spaces of any two operators, it is determined whether the two operators can be combined.
  • the present application does not limit the specific method for judging whether the input and output spaces match between the two operators, and any scheme in the prior art for determining whether the two operators match should fall within the protection scope of the present application.
  • an operator matching function (equal function) may be used to determine whether the input and output spaces between two operators match.
  • the equal function can also have a variety of design forms, for example, by traversing all element types in the input and output dict of two operators, if all element types are the same, it is considered that the input and output spaces between the two operators can match .
  • Updating the operator composite relationship matrix here refers to updating a ij in the matrix M to 1, where a ij represents the element in the i-th row and the j-th column of the matrix M.
  • the row of the matrix M is used as an example to represent the output and the column to represent the input. This application does not limit other representation forms of the matrix M.
  • the obtained matrix M is, for example:
  • a11 is 1, which means that the output of operator A can be used as the input of operator A (that is, the output space of operator A matches the input space of operator A), that is, operator A can iterate and compound itself multiple times; a12 If it is 0, it means that the output of operator A cannot be used as the input of operator B, that is, operator A cannot be connected to operator B in series.
  • the matching relationship between two operators is required in sequence, that is, the input space of operator i matches the output space of operator j, which does not mean that the output space of operator i matches the output space of operator j.
  • the input space of j must match.
  • the matching between the output space of operator i and the input space of operator j means that the combination of operator i+j can be realized, and the matching of the input space of operator i and the output space of operator j means that the combination of j+i can be realized relation.
  • the matrix M above if a23 is 0 and a32 is 1, it means that operator B cannot be composited to operator C, but operator C can be composited to operator B.
  • this embodiment uses an operator composite relationship matrix as an example to describe the operator composite relationship, but this application is not limited thereto, that is, in other possible implementation manners, other feasible ways can be used to express Composite relationship between operators.
  • an operator composite list may also be used, which includes all feasible operator composite methods between two operators, or includes all infeasible operator composite methods between two operators, etc., which is not limited in this application .
  • the guidance information that can be obtained includes but is not limited to: operator A can iterate itself multiple times, such as A+A+A+...; operator A+C or A+D or C+D can be used as a stage Combine blocks for multiple iterative compounding, such as A+C+A+C..., or A+D+A+D..., or C+D+C+D...; A, B, C, D can realize A+C +B or A+D+C+B and other serial composite relationships.
  • the operator composite decision-making module will involve the sequence composite between multiple operators, and will also involve two After the operator is compounded, it is regarded as a new operator, and the new operator is iteratively compounded, and so on.
  • the operator compound structures that can be generated include but are not limited to: A+C, A+C+D, A+A+D, A+D+D, C+B... .
  • the operator composite structure here refers to the structural form obtained after combining different operators, or it can be understood as a series structure of different operators.
  • Generating the operator composite structure according to the operator composite relationship matrix refers to determining the composability between two operators according to the operator composite relationship matrix, and combining the composable operators in series to obtain different operator composites. structure.
  • the operator composite structure corresponding to the first composite operator is one of the multiple operator composite structures.
  • the operator composite structure of the first composite operator may be generated and determined according to certain rules, or an operator composite structure as the first composite operator may be randomly determined among various feasible operator composite structures. There is no limit to this.
  • the first data set is a training set for a target federated learning task, or in other words, the first data set is a training set (or data set) used for this federated learning.
  • the local training set used for this federated learning is processed by the first composite operator to generate a new training set, which is a collection of new data obtained after being processed by the first composite operator, or A data set with a new participant composition structure after being processed by the first composite operator (for example, client grouping).
  • the structure of the first composite operator can be Adjust to obtain the second composite operator.
  • the first condition here can be regarded as the termination condition for adjusting the compound structure of the operator.
  • the first condition includes, for example: the evaluation value of the first federated learning model is less than or equal to the first threshold, or adjusting the structure of the compound operator The number of times is greater than or equal to the second threshold.
  • the evaluation value of the first federated learning model is calculated according to the measurement parameters of the first federated learning model, and the measurement parameters are associated with the target federated learning task, or in other words, the measurement parameters are used to evaluate this federated learning Related parameters, these parameters can be user-defined or system default.
  • the evaluation parameter may be, for example, one or more of communication volume (Mb), model accuracy (%), model training duration (s), and communication duration (s).
  • the evaluation value here may be calculated by substituting the evaluation parameters into the evaluation formula.
  • the present application does not limit the specific form of the evaluation formula, and the evaluation formula is described here as an example in which the evaluation formula is a linear sum of parameters.
  • the evaluation formula is: Among them, acc represents the average accuracy rate of the federation test, size(i) represents the communication traffic transmitted in the i-th communication, coefficient a and coefficient b are associated with user needs, and are used to represent the weight of each parameter.
  • a and b may be user-defined, or determined by the system according to user requirements. For example, a and b are 0.7 and 0.3 respectively, and this weight can indicate that the user is more inclined to maintain the accuracy of the model.
  • the termination condition of the adjustment is that the evaluation value is less than 10 or the number of adjustments reaches 30 times.
  • the structure of the first composite operator is adjusted.
  • the structure of the first composite operator may be adjusted through the structure of the first composite operator (or the first federated learning model) and the target federated learning task, where the target federated learning task may include user requirements.
  • the measurement parameters of the first federated learning are obtained, and when the first parameter of the measurement parameters is smaller than the third threshold, the first operator among the N operators is combined with the first compound operator to obtain the second A composite operator, wherein the first operator is associated with the first parameter, and the third threshold is associated with a target federated learning task.
  • the first operator is associated with the first parameter, which may indicate that the first operator can be used to adjust the first parameter.
  • the first parameter is the model accuracy of the first federated learning model.
  • the first parameter is less than the third threshold, it means that the model accuracy of the first federated learning model cannot meet the user's needs, and the first operator and the first The composite operator is composited (or the first operator is iterated multiple times in the first composite operator), and at this time, the first operator may be an operator used to improve the accuracy of the model.
  • the first parameter is the cost of training the first federated learning model. When the first parameter is less than the third threshold, it means that the cost of training the first federated learning model cannot meet the needs of users. Then the first operator and the second operator can be selected.
  • a compound operator is compounded (or the first operator is iterated multiple times in the first compound operator), and at this time, the first operator may be an operator for cost saving.
  • the composite structure of operators can be adjusted to meet user needs.
  • the structure of the first compound operator can be adjusted by means of automatic parameter tuning.
  • the automatic parameter tuning can be, for example, a grid search.
  • the grid search can be performed on the operator compound structure in a certain order.
  • Adjusted exhaustive search method For example, determine all possible operator composite structures according to the operator composite relationship matrix, and sort these operator composite structures in a certain order.
  • the grid search here may also be an exhaustive search method for adjusting the composite structure of operators in a random order, which is not limited in this application. It should be noted that when the number of adjustments reaches 30, but the evaluation values within these 30 times are all greater than or equal to 10, you can choose the operator composite structure used in the federated learning with the evaluation value closest to 10 among the 30 times .
  • the structure of the first composite operator may be adjusted through multiple iterations.
  • the loss value can be calculated according to the difference between the measured parameters and the parameters of the target federated model, and then the loss value can be backpropagated, and the structure of the first compound operator can be adjusted according to the loss value to obtain the second compound operator, here
  • the loss value of can indicate the size of the error, and the target federated model parameters here can be determined according to user requirements.
  • This application does not limit the calculation method of the loss value. It should be understood that other iterative optimization methods may also be used, such as an unsupervised algorithm for calculating functions without loss, or even a feed-forward algorithm, which is not limited in this application.
  • the federated learning method provided by the embodiment of this application cascades multiple operators, each operator is used to solve different problems, and connects different operators in series to interact, which will solve a certain federated learning technical problem
  • the error of the operator is weakened by other operators.
  • this method can combine methods for different single problems, and through a cascading meta-learning structure, the output method combination structure can be adapted to the problem of optimizing the actual scene of federated learning.
  • the existing technologies for solving a single problem of federated learning still have some defects, and cannot cope with various problems in actual federated learning scenarios.
  • this technology does not involve the heterogeneous data imbalance problem in federated learning, and there is a lack of discussion and solutions for heterogeneous data;
  • the multi-model federated learning technology based on client grouping because the difference of model parameter updates is used to guide the grouping, multiple federated learning processes are required, which will bring more communication costs;
  • the federated learning technology based on meta-learning simple Optimizing the model based on the accuracy of the global model will lose the subsequent personalization ability of the model; for the generation of federated learning technology for data with heterogeneous labels, this method is aimed at the case of heterogeneous labels, but cannot be used for the case of heterogeneous client data features.
  • federated learning itself is the superposition of multiple problems.
  • the method provided by the embodiment of this application is very important because it can adapt to the problem.
  • FIG. 8 shows a schematic block diagram of a federated learning system architecture 200 provided by an embodiment of the present application, and the system can be used to execute the method 100 .
  • the system mainly includes four modules: the first module, the second module, the third module and the fourth module. Each module is described as an example below.
  • the first module is used to generate the composite relationship of federated operators, so as to generate the composite operators of this federated learning.
  • the first module may also be called a federated operator composite relationship module.
  • the process 300 executed by the first module will be exemplarily described below with reference to FIG. 9 .
  • the first module obtains the operator library of this federated learning, and checks whether the input and output spaces between every two operators in the operator library match.
  • the first module receives the operator library of this federated learning from the central server of the federated learning system (for example, it may correspond to the N operators in the method 100).
  • the central server receives a target federated learning task, and the target federated learning task may include client requirement information, for example.
  • the central server obtains N operators in the operator library according to the client's demand information, and the N operators form the operator library of this federated learning.
  • the user directly selects the operator library of this federated learning according to the requirement, and sends or indicates the operator library to the central server. Further, the central server transmits the operator library of this federated learning to the first module.
  • the first module generates an operator composite relationship zero matrix, and whenever the input and output spaces of two operators in the operator library match, the first module updates the operator composite relationship matrix, otherwise it does not update. By traversing any two operators among the N operators (including each operator and itself), the final operator composite relationship matrix can be obtained, and the operator composite relationship matrix can represent the composability between different operators .
  • the specific scheme is similar to the scheme in method 100 in which the operator composite relationship is represented by the operator composite relationship matrix, and will not be repeated here.
  • the first module outputs the finally generated operator composite relationship matrix to the second module.
  • the second module adjusts the operator composite structure according to the operator composite relationship matrix to obtain the composite operator.
  • the second module may be called an operator composite decision module, or a federated operator composite decision module.
  • first module and the second module may be mutually independent modules, or may belong to the same module, for example, the first module and the second module belong to the decision maker module, or in other words, the first module and the second module may be Different submodules in the decider module. This application does not limit this.
  • the process 400 executed by the second module will be exemplarily described below with reference to FIG. 10 .
  • the second module obtains the operator library of this federated learning, and receives the operator composite relationship matrix from the first module, determines the structure of the first composite operator according to the operator composite relationship matrix, and converts the first composite operator sub output to the third module, which is a module for federated learning.
  • the second module may be an operator library obtained from the first module, or an operator library directly obtained from the central server, which is not limited in this application.
  • the third module processes the first data set through the first composite operator to obtain the second data set, and the first data set is a training set for this federated learning. Then the fourth module obtains the measurement parameters of this federated learning, and calculates the evaluation value according to the measurement parameters, and the fourth module sends the evaluation value to the second module through feedback information.
  • the second module can adjust the operator composite structure according to the feedback information. For specific schemes, reference may be made to related descriptions in method 100, and details are not repeated here for brevity.
  • the fourth module and the third module may be the same module or different modules, which is not limited in the present application.
  • the federated learning system uses the decision maker module (including the first module and the second module in the above-mentioned embodiment) to find the target structure of maximizing value for a single method compound, that is, to find out how to target the real business Scenarios to apply a better composite method to meet the specific needs of customers.
  • the decision maker module is responsible for dynamically adjusting and evaluating the system structure throughout the system, and learns a cascaded federated learning structure that meets customer needs through the decision maker module.
  • the decision maker module uses the evaluation value (the calculation result of the user cost formula) as a performance indicator (benchmark), and uses the operator compound relationship matrix as a guide in the entire operator library to dynamically adjust the operator compound structure in a direction, so that it can be obtained A cascaded federated learning system that meets user needs.
  • the federated learning method provided by the embodiment of the present application will be described in detail below by taking the training process of the federated model for controlling air-conditioning comfort as an example.
  • the user initializes some basic parameters and their data types.
  • Table 1 lists a parameter list and an example that the user needs to input in the initialization phase in a case.
  • Name user-defined federated learning project instance name
  • Meta-federated learning operator type list users can select algorithms from the operator library, such as data set distillation algorithm, client grouping algorithm, virtual sample balancing algorithm, meta-learning algorithm, etc.;
  • Base model User-defined federated learning base model, such as neural network model, Boost model, Bagging model, etc.;
  • Data example users need to submit at least one piece of example data to reflect the format of the data, describe its characteristics and label structure;
  • Training set users need to submit data sets for federated learning
  • Evaluation parameter list Users can customize the evaluation indicators about federated learning, and the evaluation indicators here are used to evaluate the models trained through federated learning.
  • optional intermediate variables in the federated learning process are communication volume (Mb), model accuracy (%), model training duration (s), communication duration (s), etc.;
  • Custom evaluation formula users can customize the evaluation formula.
  • the default evaluation formula of the system can be set in the form of linear sum, and the user can directly set the weight of each parameter. It should be noted that the parameters in the evaluation formula should be one or more of the parameters in the parameter evaluation list above.
  • Termination condition the user can set the termination condition for operator compound structure adjustment, for example, the user sets the compound times of operator compound structure adjustment to 30.
  • a default maximum number of adjustments may be used, for example, the default is 20 times.
  • the public data set ATCII (referred to as a federated simulation data set) is used as a training set for federated learning.
  • the federal simulation data set includes 69 countries and the collected data subdivided into different cities in each country. There are a total of about 76,000 pieces of collected data, including a total of 70 features, based on the consideration of the actual scene and the difficulty of the actual feature collection. , and finally select 8 common features for testing. These 8 common features are the numerical form information of temperature and velocity, and the latter three features are calculated from the average value of the original features.
  • the collection values of the feature Thermal preference in the ATCII data set are 'cooler', 'no change', and 'warmer', so the feature Thermal preference is selected as the label of the experiment to train the three-category prediction model.
  • the five cities Berkeley, Darwin, Kalgoorlie, Oxford, and Townsvill in the ACTII dataset are selected as five clients to participate in the entire federated learning, and the data saved in each city is used as the training set for the corresponding client.
  • the evaluation parameters set by the user are model test accuracy acc and traffic size
  • the evaluation standard is set to (100-acc) and size linear combination cost
  • the units of the two parameters in the cost calculation formula are artificially set is % and Mb, and the weights are 0.7 and 0.3 respectively, which means that users are more inclined to maintain the accuracy of the model.
  • the formula is only a test, and it does not mean that the units and costs in the cost calculation formula have a reasonable calculation relationship.
  • the calculation result unit of the abstract calculation formula is cost (yuan), and the abbreviated form of the calculation formula is:
  • Acc represents the average accuracy rate of the federation test
  • size(i) represents the communication volume transmitted in the i-th communication.
  • the system determines the operator library information of this federated learning according to the information input by the user.
  • Table 2 lists the list of parameters input by the user and their examples in one case.
  • Meta federated learning operator list the user selects an operator from the original operator library, for example, the user selects a dataset distillation operator and a client grouping operator as operators for this federated learning according to user needs;
  • Table 3 lists the parameter list output by the system in one case. The following is an exemplary description of the output parameters in Table 3:
  • Operator function list The system adjusts the standard function library obtained by adjusting the meta-federated learning operator selected by the user according to the structure of the data example provided by the user.
  • the corresponding function can be called through the function name, for example, "funA" can be called Operator A.
  • the input and output parameters listed in Table 2 and Table 3 are just examples, and the embodiment of the present application may also use other methods to output the operator library of this federated learning.
  • users can input user requirements to the system, for example, users input data encryption, data balance and other requirements (for example, select data encryption, data balance and other requirements on the system interface), and the system selects this federation from the original operator library according to user requirements Operators for learning.
  • the operator selected by the system for this federated learning should correspond to user requirements, or in other words, the operator selected by the system for this federated learning should be an operator for realizing user requirements.
  • the operator library of this federated learning is ⁇ dataset distillation operator (E operator), client grouping operator (F operator) ⁇ as an example for illustration.
  • E operator dataset distillation operator
  • F operator client grouping operator
  • Data set distillation operator select the artificial neural network ANN containing two hidden layers; select the ReLU function as the activation function; use a fixed initialization initial model for distillation; set the number of training rounds to 20 rounds, and each round has 30 steps of gradient descent ( The 10-step gradient descent was repeated 3 times); the number of distillation data was set to 5 for each class.
  • Client-side grouping operator set the maximum number of divided groups to the number of input cities, and the minimum number of divided groups to 1; set the division method to use a custom Non-iid metric algorithm to describe the similarity of tasks; set the iteration termination condition to inter-group measurement The sum is the largest, the sum of the measures within the group is the smallest; set the maximum number of iterations to 10.
  • the operator library information output by the system includes the input and output data (and the format of the data) of the operator and a table of parameter settings.
  • the input is the original dataset of each client, and the output is the client distillation dataset;
  • the input is the original client set and data, and the output is the new client gather.
  • the federated operator composite relationship module matches the input and output spaces between any two operators according to the operator library information of this federated learning to generate a federated operator composite relationship matrix.
  • the federated operator compound relationship module inputs the operator library information of this federated learning (for example, the operator library information is ⁇ E operator, F operator ⁇ , the input and output data format of the E operator, and the input and output of the F operator Combine the data format, the parameter setting form of the E operator, the parameter setting form of the F operator, etc.) and hyperparameters (the hyperparameters cannot be self-composited), and then judge whether the input and output spaces of the E operator and the F operator match.
  • the operator library information is ⁇ E operator, F operator ⁇ , the input and output data format of the E operator, and the input and output of the F operator Combine the data format, the parameter setting form of the E operator, the parameter setting form of the F operator, etc.
  • hyperparameters the hyperparameters cannot be self-composited
  • One way is to judge whether the input and output spaces of the E operator and the F operator match according to the data form of the input and output spaces of the E operator and the F operator.
  • BaseAlgorithm which has two base class methods get_input_type() and get_output_type(), and operators input to the system must inherit from this BaseAlgorithm base class, and the operator contains a run() method to implement operations such as distillation and customer grouping. Then the system obtains the input and output data format of the operator by calling the get_input_type() and get_output_type() methods, and stores it in the operator library.
  • the data format can be list/set or user-defined object.
  • the equal() calculation can be performed on the data forms of the E operator and the F operator to judge whether the input and output spaces of the E operator and the F operator can be connected.
  • the process of obtaining the data form of the E operator and the F operator and the process of judging whether the input and output spaces of the E operator and the F operator can be connected can be executed in the federated compound relationship module, or can be It is executed in other modules, which is not limited in this application.
  • the operator composite decision-making module obtains the operator composite relationship matrix and the federated operator library information, and generates the first composite operator. Then the first compound operator is compounded into the federated learning system, and the local training set of each client is federated through the federated learning system to obtain the first federated model.
  • Table 4 lists a list of parameters that users need to input in the federated learning process and their examples in one case. It should be understood that these parameters may be input into the system by the user during the initial setting stage.
  • Base model users can choose the type of base model used in this federation training, such as neural network model, Boost model, Bagging model, etc.;
  • Training set The user needs to submit the data set for federated learning.
  • Federation model the user can select the type of federation model, such as neural network model, Boost model, Bagging model, etc.
  • the measurement parameters can be obtained, and the evaluation value can be calculated according to the measurement parameters, and the measurement parameters and evaluation values can be fed back to the operator composite decision-making module, so that the operator composite decision-making module can Adjust the substructure.
  • Table 5 lists the list of parameters input by the user to the operator composite decision-making module and its examples in one case. It should be understood that these parameters may be input into the system by the user during the initial setting stage.
  • the structure of the first composite operator is adjusted.
  • Table 6 lists the search process of the operator composite structure in one case.
  • G the process of connecting the operator compound structure to the traditional federated system is recorded as G. It should be noted that G here is not an operator.
  • the federated learning process meta-learner adopts three-category XGBoost for federated learning, and XGBoost parameter control remains unchanged.
  • the specific federated learning process is combined training of federated distillation data and local data.
  • the termination condition set by the user is 30 times to adjust the structure/the result of the cost calculation formula is less than 10.
  • the operator composite structure can be adjusted in the order shown in Table 6.
  • the final operator composite decision module outputs the operator composite federated learning structure of E+F+E+F+E+G, which is The structure that satisfies the evaluation formula with the smallest value during the search.
  • E+F+E+F+E+G is collectively referred to as a compound operator. Further, the training set is processed through the composite operator:
  • 5Customer data set input data set distillation operator outputs distillation data and compresses the data; the distillation data is transmitted to the cloud, and the cloud performs Non-iid measurement on the distillation data, and divides into new groups (Berkeley, Kalgoorlie, Oxford) (Darwin, Townsvill); Continue to perform multi-cloud controlled federated learning in the group, and the customers in the group distill the data set and transmit it to the cloud; the cloud in the group performs Non-iid measurement on the distilled data, and re-divides the group structure (Berkeley) (Kalgoorlie, Oxford) ( Darwin, Townsvill); Distill the data set in the new group and transmit it to the cloud; the cloud summarizes the distillation data in the group, and the users in the group combine the shared distillation data and local data to train the Xgboost model and test it.
  • this embodiment adopts the method of adaptively migrating the instance of the distillation data set to increase the proportion of the distillation data combined with the local data, emphasizing the importance of the distillation data to The guiding significance of client grouping.
  • Fig. 11 is a schematic structural block diagram of a computer device provided according to an embodiment of the present application.
  • the computer device 500 shown in FIG. 11 includes an acquisition unit 501 and a processing unit 502 .
  • the obtaining unit 501 is configured to obtain N operators from the operator library according to the target federated learning task, the operator library includes K operators associated with federated learning, and the N and the K are greater than 1 Integer, said N is less than or equal to said K.
  • the processing unit 502 is configured to combine multiple operators in the N operators to obtain a first composite operator; process the first data set through the first composite operator to obtain a second data set, so The first data set is a training set for the target federated learning task; the federated learning is performed using the second data set to obtain a first federated learning model.
  • the processing unit 502 is specifically configured to: according to the operator composite relationship between any two operators in the N operators, multiple operators in the N operators Operators are compounded to obtain the first compound operator.
  • the processing unit 502 is further configured to: determine a matching relationship between input and output spaces between any two operators in the N operators; The matching relationship of the input and output spaces between any two operators determines the operator composition relationship between any two operators among the N operators.
  • the processing unit 502 is further configured to: determine whether the first condition is met; if the first condition is not met, adjust the structure of the first compound operator to obtain the first Two composite operators.
  • the acquiring unit 501 is specifically configured to: acquire the measurement parameters of the first federated learning model; the processing unit 502 is specifically configured to: calculate an evaluation value according to the measurement parameters, the The measurement parameter is associated with the target federated learning task; if the evaluation value is less than a first threshold, it is determined that the first condition is met.
  • the processing unit 502 is specifically configured to: determine the number of times to adjust the structure of the compound operator; when the number of times to adjust the structure of the compound operator is greater than or equal to a second threshold, It is determined that the first condition is met.
  • the processing unit 502 is specifically configured to: calculate a loss value according to the difference between the measurement parameter and the target federated model parameter, perform backpropagation on the loss value, and according to the The loss value adjusts the structure of the first composite operator to obtain the second composite operator.
  • the operator library includes at least one of a dataset distillation operator, a client grouping operator, a virtual sample balancing operator, and a meta-learning operator.
  • the specific functions and beneficial effects of the acquiring unit 501 and the processing unit 502 can refer to the embodiments shown in FIG. 7-FIG.
  • Fig. 12 is a schematic structural block diagram of an apparatus for federated learning according to an embodiment of the present application.
  • the device 600 includes a bus 601 , a processor 602 , a communication interface 603 and a memory 604 .
  • the processor 602 , the memory 604 and the communication interface 603 communicate through the bus 601 .
  • the processor 702 may be a programmable gate array (field programmable gate array, FPGA), may be an application specific integrated circuit (ASIC), may also be a system chip (system on chip, SoC), or may be a central Processor (central processor unit, CPU), can also be a network processor (network processor, NP), can also be a digital signal processing circuit (digital signal processor, DSP), can also be a microcontroller (micro controller unit, MCU ), can also be a programmable controller (programmable logic device, PLD), other programmable logic devices, discrete gate or transistor logic devices, discrete hardware components, or other integrated chips.
  • FPGA field programmable gate array
  • ASIC application specific integrated circuit
  • SoC system on chip
  • CPU central Processor
  • NP network processor
  • DSP digital signal processing circuit
  • microcontroller micro controller unit, MCU
  • PLD programmable logic device
  • PLD programmable logic device
  • the executable code included in the pathological image processing system is stored in the memory 604 , and the processor 602 reads the executable code in the memory 604 to execute the method shown in FIG. 3 .
  • the memory 604 may also include an operating system and other software modules required for running processes.
  • the operating system can be LINUX TM , UNIX TM , WINDOWS TM and so on.
  • the embodiment of the present application also provides a chip system, which is characterized in that it includes: a logic circuit, the logic circuit is used to couple with the input/output interface, and transmit data through the input/output interface, so as to execute the Methods.
  • each step of the above method can be completed by an integrated logic circuit of hardware in a processor or an instruction in the form of software.
  • the steps of the methods disclosed in connection with the embodiments of the present application may be directly implemented by a hardware processor, or implemented by a combination of hardware and software modules in the processor.
  • the software module can be located in a mature storage medium in the field such as random access memory, flash memory, read-only memory, programmable read-only memory or electrically erasable programmable memory, register.
  • the storage medium is located in the memory, and the processor reads the information in the memory, and completes the steps of the above method in combination with its hardware. To avoid repetition, no detailed description is given here.
  • the processor in the embodiment of the present application may be an integrated circuit chip, which has a signal processing capability.
  • each step of the above-mentioned method embodiments may be completed by an integrated logic circuit of hardware in a processor or instructions in the form of software.
  • a general-purpose processor may be a microprocessor, or the processor may be any conventional processor, or the like.
  • the steps of the method disclosed in connection with the embodiments of the present application may be directly implemented by a hardware decoding processor, or implemented by a combination of hardware and software modules in the decoding processor.
  • the software module can be located in a mature storage medium in the field such as random access memory, flash memory, read-only memory, programmable read-only memory or electrically erasable programmable memory, register. The storage medium is located in the memory, and the processor reads the information in the memory, and completes the steps of the above method in combination with its hardware.
  • the memory in the embodiments of the present application may be a volatile memory or a nonvolatile memory, or may include both volatile and nonvolatile memories.
  • the non-volatile memory can be read-only memory (read-only memory, ROM), programmable read-only memory (programmable ROM, PROM), erasable programmable read-only memory (erasable PROM, EPROM), electrically programmable Erases programmable read-only memory (electrically EPROM, EEPROM) or flash memory.
  • Volatile memory can be random access memory (RAM), which acts as external cache memory.
  • RAM random access memory
  • SRAM static random access memory
  • DRAM dynamic random access memory
  • DRAM synchronous dynamic random access memory
  • SDRAM double data rate synchronous dynamic random access memory
  • ESDRAM enhanced synchronous dynamic random access memory
  • SLDRAM direct memory bus random access memory
  • direct rambus RAM direct rambus RAM
  • the present application also provides a computer program product, the computer program product including: computer program code, when the computer program code is run on the computer, the computer is made to execute the embodiment shown in FIG. 3 Methods.
  • the present application also provides a computer-readable medium, the computer-readable medium stores program code, and when the program code is run on the computer, the computer executes the embodiment shown in FIG. 3 Methods.
  • the disclosed systems, devices and methods may be implemented in other ways.
  • the device embodiments described above are only illustrative.
  • the division of the units is only a logical function division. In actual implementation, there may be other division methods.
  • multiple units or components can be combined or May be integrated into another system, or some features may be ignored, or not implemented.
  • the mutual coupling or direct coupling or communication connection shown or discussed may be through some interfaces, and the indirect coupling or communication connection of devices or units may be in electrical, mechanical or other forms.
  • the units described as separate components may or may not be physically separated, and the components shown as units may or may not be physical units, that is, they may be located in one place, or may be distributed to multiple network units. Part or all of the units can be selected according to actual needs to achieve the purpose of the solution of this embodiment.
  • each functional unit in each embodiment of the present application may be integrated into one processing unit, each unit may exist separately physically, or two or more units may be integrated into one unit.
  • the functions described above are realized in the form of software function units and sold or used as independent products, they can be stored in a computer-readable storage medium.
  • the technical solution of the present application is essentially or the part that contributes to the prior art or the part of the technical solution can be embodied in the form of a software product, and the computer software product is stored in a storage medium, including Several instructions are used to make a computer device (which may be a personal computer, a server, or a network device, etc.) execute all or part of the steps of the methods described in the various embodiments of the present application.
  • the aforementioned storage medium includes: U disk, mobile hard disk, read-only memory (read-only memory, ROM), random access memory (random access memory, RAM), magnetic disk or optical disc and other media that can store program codes. .

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Software Systems (AREA)
  • Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Artificial Intelligence (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Medical Informatics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

一种联邦学习的方法和装置,该方法包括:根据目标联邦学习任务从算子库中获取N个算子(S110),算子库包括K个与联邦学习相关联的算子,N和K为大于1的整数,N小于或等于K;将N个算子中的多个算子进行复合得到第一复合算子(S120);将第一数据集通过第一复合算子进行处理得到第二数据集(S130),第一数据集为用于目标联邦学习任务的训练集;利用第二数据集进行联邦学习得到第一联邦学习模型(S140)。该方法通过一种级联元学习结构,使得输出的方法组合结构能够适应于优化联邦学习实际场景的问题,从而提高联邦学习的效率和质量。

Description

联邦学习的方法和装置
本申请要求于2021年08月30日提交中国专利局、申请号为202111001636.4、申请名称为“联邦学习的方法和装置”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。
技术领域
本申请涉及人工智能领域,并且更加具体地,涉及一种联邦学习的方法和装置。
背景技术
随着大数据时代的到来,每台设备每天都会以各种形式产生巨量的原始数据,这些数据将以“孤岛”的形式诞生并存在于世界的各个角落。传统的集中式学习要求各个边缘设备将本地数据统一传输到中心端的服务器上,其后再利用收集到的数据进行模型的训练与学习,然而这一架构随着时代的发展逐渐受到如下因素的限制:(1)边缘设备广泛地分布于世界上各个地区和角落,这些设备将以飞快的速度源源不断地产生和积累巨大量级的原始数据。若中心端需要收集来自全部边缘设备的原始数据,势必会带来巨大的通信损耗和算力需求;(2)随着现实生活中实际场景的复杂化,越来越多的学习任务要求边缘设备能够做出及时而有效的决策与反馈。传统的集中式学习由于涉及到大量数据的上传势必会导致较大程度的时延,致使其无法满足实际任务场景的实时需求;(3)考虑到行业竞争、用户隐私安全、行政手续复杂等问题,将数据进行集中整合将面临越来越大的阻力制约。因而系统部署将越来越倾向于在本地存储数据,同时由边缘设备自身完成模型的本地计算。
因此,如何在满足数据隐私、安全和监管要求的前提下,设计一个机器学习框架,让人工智能(artificial intelligence,AI)系统能够更加高效、准确地共同使用各自的数据,成为了当前人工智能发展的一个重要议题。联邦学习(federated learning,FL)这一概念的提出有效地解决了当前人工智能发展所面临的困境,其在充分保障用户数据隐私和安全的前提下,通过促使各个边缘设备和中心端服务器协同合作来高效地完成模型的学习任务。然而,FL的提出虽然在一定程度上解决了人工智能领域当前发展所面临的问题,但其本身仍存在多种多样的缺陷使其在实际应用中仍然具有很大的局限性。
发明内容
本申请提供一种联邦学习的方法和装置,能够提高联邦学习的效率和质量。
本申请实施例提供了一种联邦学习的方法,包括:根据目标联邦学习任务从算子库中获取N个算子,所述算子库包括K个与联邦学习相关联的算子,所述N和所述K为大于1的整数,所述N小于或等于所述K;将所述N个算子中的多个算子进行复合得到第一复合算子;通过所述第一复合算子将第一数据集进行处理得到第二数据集,所述第一数据集 为用于所述目标联邦学习任务的训练集;利用所述第二数据集进行联邦学习得到与该目标联邦学习任务对应的第一联邦学习模型。
因此,本申请实施例提供的联邦学习的方法,通过将多个算子进行复合得到用于解决多目标问题的复合算子,通过该复合算子对本次联邦学习的训练集进行处理,可以实现对多个目标问题的同时优化,解决联邦学习实际场景中多问题叠加的挑战。
应理解,这里的K个与联邦学习相关联的算子可以理解为,该K个算子可用于处理联邦学习中的本地训练集,或者说,该K个算子是各种联邦学习相关的单一技术对应的算子。
该目标联邦学习任务可以包括客户端需求信息。这里的客户端需求信息可以是客户端在模型训练之前协商确定的综合需求。
结合第一方面,在第一方面的一种可能的实现方式中,根据目标联邦学习任务从算子库中获取N个算子可以包括:获取目标联邦学习任务,然后根据该目标联邦学习任务从算子库中选择N个算子,该N个算子与该目标联邦学习任务相关联,该目标联邦学习任务包括客户需求信息。
通过该方案,使得用于本次联邦学习的算子(该N个算子)是用于解决客户需求的算子,也就是用于执行本次目标联邦学习任务的算子,因此通过对该N个算子中的部分或者全部算子进行复合得到的复合算子可以应对联邦学习中多问题叠加的挑战。
结合第一方面,在第一方面的一种可能的实现方式中,所述将所述N个算子中的多个算子进行复合得到第一复合算子,包括:根据所述N个算子中任意两个算子之间的算子复合关系,将所述N个算子中的多个算子进行复合得到所述第一复合算子。
通过该方案,使得复合算子是基于算子复合关系的指导下来获取的,即根据算子复合关系指导复合算子的结构的调整方向,从而可以提高效率,同时提高复合算子的实用性和可用性。
这里的算子复合关系可以采用算子复合关系矩阵来表示。例如,该算子复合关系矩阵为N阶矩阵,该矩阵的行表示输出,列表示输入(也可以颠倒过来),对于矩阵中的各元素,1表示可复合,0表示不可复合。
这里的第一复合算子指的是将N个算子中的多个算子进行串联复合组成一定的算子复合结构之后,对该复合结构内的多个算子的一个统称,在其他场景中该复合算子还可以叫做其他名称
结合第一方面,在第一方面的一种可能的实现方式中,所述方法还包括:确定所述N个算子中任意两个算子之间的输入输出空间的匹配关系;根据所述N个算子中任意两个算子之间的输入输出空间的匹配关系确定所述N个算子中任意两个算子之间的算子复合关系。
通过该方案,可以根据算子之间的输入输出空间的匹配关系获取算子之间的复合关系,从而可以更加准确得获取任意算子之间的可复合性,提高算子复合的效率。
结合第一方面,在第一方面的一种可能的实现方式中,在利用所述第二数据集进行联邦学习得到第一联邦学习模型之后,所述方法还包括:确定是否满足第一条件;在不满足第一条件的情况下,对所述第一复合算子的结构进行调整得到第二复合算子。
通过该方案,可以对第一复合算子的结构进行调整,以便获取更加符合用户需求或者 性能更好的复合算子。
结合第一方面,在第一方面的一种可能的实现方式中,所述确定是否满足第一条件,包括:获取所述第一联邦学习模型的测量参数;根据所述测量参数计算评估值,所述测量参数与所述目标联邦学习任务相关联;在所述评估值小于第一阈值的情况下,确定满足所述第一条件。
结合第一方面,在第一发面的一种可能的实现方式中,所述根据所述测量参数计算评估值,包括:根据所述测量参数和成本计算公式计算所述评估值,所述成本计算公式与所述目标联邦学习任务相关联,或者说,所述成本计算公式的形式与客户需求相关。该成本计算公式的形式可以是测量参数的线性和,该测量参数的系数与目标联邦学习任务相关联。
该第一阈值可以是根据目标联邦学习任务设置的,例如,根据目标联邦学习模型的目标参数设置第一阈值。
通过该方案,根据成本计算公式的计算结果为性能指标,通过对复合算子的结构不断调整使得联邦学习模型的性能指标满足设定的阈值,从而可以保证本次联邦学习的质量,提高联邦学习模型的实用性。
结合第一方面,在第一方面的一种可能的实现方式中,所述确定是否满足第一条件,包括:确定调整复合算子的结构的次数;在所述调整复合算子的结构的次数大于或等于第二阈值的情况下,确定满足第一条件。
该方案中,以复合算子的结构的调整次数为终止条件,可以提高算子复合的效率,节省资源。
结合第一方面,在第一方面的一种可能的实现方式中,所述对所述第一复合算子的结构进行调整得到第二复合算子,包括:根据所述测量参数和目标联邦模型参数之间的差异计算损失值,将所述损失值进行反向传播,根据所述损失值对所述第一复合算子的结构进行调整得到第二复合算子。
在另一种可能的实现方式中,也可以是根据评估值和目标评估值之间的差异计算损失值(或者称为损失函数值,本申请对损失函数的具体形式不作限定)。
通过该方案,使得复合算子的结构的调整过程可以以损失值为指导有方向地进行,从而可以提高算子复合结构的调整效率,节省开销。
结合第一方面,在第一方面的一种可能的实现方式中,所述对所述第一复合算子的结构进行调整得到第二复合算子,包括:当所述第一联邦学习模型中的第一参数小于第三阈值时,将所述N个算子中的第一算子与所述第一复合算子进行复合得到所述第二复合算子,所述第一算子与所述第一参数相关联。
结合第一方面,在第一方面的一种可能的实现方式中,所述对所述第一复合算子的结构进行调整得到第二复合算子,包括:将所述第一复合算子的结构进行自动化调参得到所述第二复合算子。
结合第一方面,在第一方面的一种可能的实现方式中,所述算子库包括数据集蒸馏算子、客户端分组算子、虚拟样本平衡算子、元学习算子中的至少一种。
第二方面,本申请实施例提供一种联邦学习的装置,该装置包括用于实现上述第一方面的方法设计中任意一种可能的实现方式的单元。
第三方面,本申请实施例提供一种联邦学习的装置,包括收发器和处理器。可选地, 该装置还包括存储器。该处理器用于控制收发器收发信号,该存储器用于存储计算机程序,该处理器用于从存储器中调用并运行该计算机程序,使得该计算机设备执行上述第一方面的方法设计中任意一种可能的实现方式中的方法。
第四方面,本申请实施例提供一种电子装置,该电子装置可以为用于实现上述第一方面方法设计中的联邦学习的装置,或者为设置在该装置中的芯片。该电子装置包括:处理器,与存储器耦合,可用于执行存储器中的指令和/或程序代码,以实现上述第一方面的方法设计中任意一种可能的实现方式中的方法。可选地,该电子装置还包括存储器。可选地,该电子装置还包括通信接口,处理器与通信接口耦合。
当该电子装置为上述联邦学习的装置时,该通信接口可以是收发器,或,输入/输出接口。
当该电子装置为配置于上述联邦学习的装置的芯片时,该通信接口可以是输入/输出接口。
可选地,该收发器可以为收发电路。可选地,该输入/输出接口可以为输入/输出电路。
第五方面,本申请实施例提供一种计算机程序产品,所述计算机程序产品包括:计算机程序代码,当所述计算机程序代码在计算机上运行时,使得计算机执行上述第一方面的方法设计中任意一种可能的实现方式中的方法。
第六方面,本申请实施例提供一种计算机可读介质,所述计算机可读介质存储有程序代码,当所述计算机程序代码在计算机上运行时,使得计算机执行上述第一方面的方法设计中任意一种可能的实现方式中的方法。
附图说明
图1是一种人工智能主体框架示意图。
图2是一种联邦学习的系统架构的示意图。
图3是数据集蒸馏联邦学习技术流程图。
图4是客户端分组联邦学习技术流程图。
图5是联邦元学习技术流程图。
图6是数据生成联邦学习技术流程图。
图7是本申请实施例联邦学习的方法的示意性框图。
图8是本申请实施例提供的联邦学习的一种系统框架图。
图9是本申请实施例提供的算子复合关系模型执行方法的流程图。
图10是本申请实施例提供的算子复合决策模块执行方法的流程图。
图11是根据本申请实施例提供的一种联邦学习的装置的示意性结构框图。
图12是根据本申请实施例提供的一种联邦学习的装置的示意性结构框图。
具体实施方式
下面将结合本申请实施例中的附图,对本申请实施例中的技术方案进行描述,显然,所描述的实施例仅仅是本申请一部分实施例,而不是全部的实施例。基于本申请中的实施例,本领域普通技术人员在没有做出创造性劳动前提下所获得的所有其他实施例,都属于本申请保护的范围。
图1是一种人工智能主体框架示意图,该主体框架描述了人工智能系统总体工作流程,适用于通用的人工智能领域需求。
下面从“智能信息链”(水平轴)和“信息技术(information technology,IT)价值链”(垂直轴)两个维度对上述人工智能主题框架进行详细的阐述。
“智能信息链”反映从数据的获取到处理的一列过程。举例来说,可以是智能信息感知、智能信息表示与形成、智能推理、智能决策、智能执行与输出的一般过程。在这个过程中,数据经历了“数据—信息—知识—智慧”的凝练过程。
“IT价值链”从人智能的底层基础设施、信息(提供和处理技术实现)到系统的产业生态过程,反映人工智能为信息技术产业带来的价值。
(1)基础设施:
基础设施为人工智能系统提供计算能力支持,实现与外部世界的沟通,并通过基础平台实现支撑。
基础设施可以通过传感器与外部沟通,基础设施的计算能力可以由智能芯片提供。
这里的智能芯片可以是中央处理器(central processing unit,CPU)、神经网络处理器(neural-network processing unit,NPU)、图形处理器(graphics processing unit,GPU)、专门应用的集成电路(application specific integrated circuit,ASIC)以及现场可编程门阵列(field programmable gate array,FPGA)等硬件加速芯片。
基础设施的基础平台可以包括分布式计算框架及网络等相关的平台保障和支持,可以包括云存储和计算、互联互通网络等。
例如,对于基础设施来说,可以通过传感器和外部沟通获取数据,然后将这些数据提供给基础平台提供的分布式计算系统中的智能芯片进行计算。
(2)数据:
基础设施的上一层的数据用于表示人工智能领域的数据来源。该数据涉及到图形、图像、语音、文本、序列,还涉及到传统设备的物联网数据,包括已有系统的业务数据以及力、位移、液位、温度、湿度等感知数据。
应理解,序列可以理解为数据序列,最常见的有时序序列数据,即时序数据,例如一段时间内的天气预报数据(温度、风向等),又例如股市数据、人体血糖变化数据等生理数据的序列等等。
(3)数据处理:
上述数据处理通常包括数据训练,机器学习,深度学习,搜索,推理,决策等处理方式。
其中,机器学习和深度学习可以对数据进行符号化和形式化的智能信息建模、抽取、预处理、训练等。
推理是指在计算机或智能系统中,模拟人类的智能推理方式,依据推理控制策略,利用形式化的信息进行机器思维和求解问题的过程,典型的功能是搜索与匹配。
决策是指智能信息经过推理后进行决策的过程,通常提供分类、排序、预测等功能。
(4)通用能力:
对数据经过上面提到的数据处理后,进一步基于数据处理的结果可以形成一些通用的能力,比如可以是算法或者一个通用系统,例如,翻译,文本的分析,计算机视觉的处理, 语音识别,图像的识别等等。
(5)智能产品及行业应用:
智能产品及行业应用指人工智能系统在各领域的产品和应用,是对人工智能整体解决方案的封装,将智能信息决策产品化、实现落地应用,其应用领域主要包括:智能制造、智能交通、智能家居、智能医疗、智能安防、自动驾驶,平安城市,智能终端等。
本申请实施例可以应用在人工智能中的很多领域,例如,智能制造、智能交通、智能家居、智能医疗、智能安防、自动驾驶,平安城市等领域。
具体地,本申请实施例可以具体应用在各类数据处理的领域,例如图像分类、图像检索、图像语义分割、图像超分辨率等图像类数据相关的处理任务,又例如语音合成、语音识别、语音搜索等语音类数据相关的处理任务,又例如智能翻译、语义提取等文字类数据相关的处理任务,又例如草图绘制、手势控制等涉及图形数据的相关处理任务。图像类数据相关的处理任务还可以是视频类图像数据,例如可以应用于视频搜索、视频分类、关键人物追踪等任务。此外,本申请实施例还可以应用于数据序列的处理任务,例如时序数据。还应理解上述各任务只是按照涉及的处理数据的数据类型来举例说明,但有些任务可以涉及多种数据类型,例如可以是利用语音搜索图像的任务,也就是说,输入一段语音,根据这段语音从图像库中找到相应的图像,该任务就是既涉及图像数据有又涉及语音数据。
图2是适用于本申请实施例的一种联邦学习的系统架构的示意图,该联邦学习的系统仅仅是多种联邦学习架构中的一种,该系统仅仅是一种示例,本申请实施例还可以应用于其他架构形式的联邦学习系统,本申请对此不作限定。
如图2所示,在该场景中包括用户设备#1和用户设备#2以及云端设备#1,为了确保用户数据隐私和安全,用户设备之间不能进行数据交互。需要说明的是,图2中也可能存在3个或3个以上的用户设备,原理是相同的,对此进行了省略。
本地数据#1和本地数据#2分别表示用户设备#1和用户设备#2所能提供的数据,或者可以理解为本地数据#1和本地数据#2分别是用户设备#1和用户设备#2各自的本地真实数据,在联邦学习中,这些数据不希望被其他用户设备获取。这些数据例如可以是图像数据、语音数据、文本数据、序列数据等等。用户设备#1利用本地数据#1进行模型训练,得到本地模型#1;用户设备#2利用本地数据#2进行模型训练得到本地模型#2。之后,用户设备#1将本地模型#1或本地模型#1的参数发送给云端设备#1,用户设备#2将本地模型#2或本地模型#2的参数发送给云端设备#1。
云端设备#1在获得所有本地模型(包括本地模型#1和本地模型#2)后,将所有获得的本地模型汇聚成一个联邦模型,联邦模型也可以理解为统一模型,也就是说所有用户设备能够通用的模型。
之后云端设备#1将得到的联邦模型发送给所有用户设备,用户设备在接收到联邦模型后,将本地模型替换成联邦模型,再重复下一轮的训练:在联邦学习模型的基础上训练得到新的本地模型、上传新的本地模型、模型汇聚出新的联邦模型,如此循环,直到达到一定的预设要求。在图2所示场景中,整个的训练过程可以看作是对联邦模型的模型参数的不断调整,使得联邦模型的参数能够同时适用于所有用户设备,想要获得满足需求的参数,就需要很多次的训练,也就是说,需要循环执行上述过程:训练得到本地模型、上传本地模型、模型汇聚、下发联邦模型、更新本地模型,直到达到能够适用于所有用户设备 的要求,从图2中可以看出,需要用户设备与云端设备之间分别多次的数据交互,即多次本地模型和联邦模型的传输。
下面以图像分类任务为例进行介绍。假设该联邦学习系统是用于图像分类的,具体而言可以假设用于识别动物,也就是说能够区分不同的动物类别,用户设备#1和用户设备#2均能够提供包括有动物的图像。本地数据#1为用户设备#1的图像样本数据(带有动物类别的标签的图像),本地数据#2为用户设备#2的图像样本数据(带有动物类别的标签的图像),也就是说,本地数据#1和本地数据#2的样本维度是可以对齐的,但本地数据#1和本地数据#2的特征维度是不同的。则在训练过程中,用户设备#1利用本地数据#1训练获得本地模型#1,用户设备#2利用本地数据#2训练获得本地模型#2。之后用户设备#1将本地模型#1的模型参数上传到云端设备#1,用户设备#2将本地模型#2的模型参数上传到云端设备#1,云端设备#1将本地模型#1和本地模型#2汇聚成了一个联邦模型,云端设备将该联邦模型分发给用户设备#1和用户设备#2,用户设备#1和用户设备#2在联邦模型的基础上利用本地数据继续训练得到本地模型,重复上述过程直到联邦模型的参数既能够适用于用户设备#1,又能够适用于用户设备#2,云端设备#1将该联邦模型分发给用户设备#1和用户设备#2,用户设备#1和用户设备#2接收该联邦模型以及将其取代原本的本地模型。
在图2所示的系统中,主要存在以下问题。
首先,由于联邦学习中的多个客户端不会共享本地模型数据,即各客户端数据不互通,因此会出现异构数据不平衡的问题,而要想不同客户端之间的数据完全对齐、模型结构完全一致,又会使得应用的场景非常局限。
其次,由于用户之间的模型参数差异和数据差异,如果希望得到一个准确度相对较高的联邦模型,通常需要多次训练,这无疑带来大量的运算资源的消耗,且需要很长的训练时间;同时由于每一轮次的训练过程都需要多次传输模型,因此该架构需要更多次数的数据传输,这会带来很高的传输带宽成本。同时,跨地区跨国等参与方联邦学习必然会涉及较远距离的传输,这也会导致较大的资源消耗。
其他地,由于客户端之间数据不共享,单一客户端的数据质量较差时可能会影响整体模型质量;由于要进行多轮次参数上传、模型训练等过程,很容易造成数据丢失等问题。
因此,图2所示的系统面临着复杂问题叠加的挑战,问题难以被分析和定位,探索过程往往又需要很多的人工和研究成本。
本申请提供一种联邦学习的方法,希望解决联邦学习中多问题叠加的挑战,提高联邦学习的质量和效率。
为了便于理解,下面先对本申请实施例可能涉及相关术语、概念或者技术进行介绍:
(1)联邦学习
联邦学习(federated learning)是一种加密的分布式机器学习技术,又可以称之为联合学习或联盟学习等。联邦学习能够充分地利用参与方的数据和计算能力,使多方可以在不需要共享数据的前提下协作地构建通用、健壮的机器学习模型,因此联邦学习能有效帮助多个机构在满足用户隐私保护、数据安全和政府法规的要求下,进行数据使用和学习建模。联邦学习包括横向联邦学习(horizontal federated learning)、纵向联邦学习(vertical federated learning)。可以说,联邦学习就是为了能够使得既能够共享知识和参数,又不需要交换 任何自身数据。在数据监管越来越严格的大环境下,联邦学习能够解决数据所有权,数据隐私,数据访问权以及异构数据的访问等关键问题。
数据矩阵的横向的一行表示一条训练样本,纵向的一列表示一个数据特征。横向联邦学习是联合多个参与者的具有相同特征的多行样本进行联邦学习,即各个参与者的训练数据是横向划分的。横向联邦学习也称为特征对齐的联邦学习(feature-aligned federated learning),即参与者的数据特征是对齐的,横向联邦学习可以增加训练样本总量。纵向联邦学习是联合多个参与者的共同样本的不同数据特征进行联邦学习,即各个参与者的训练数据是纵向划分的。纵向联邦学习也称为样本对齐的联邦学习(sample-aligned federated learning),即参与者的训练样本是对齐的,纵向联邦学习可以增加训练数据特征维度。
联邦学习在工业界已经得到应用,例如谷歌将其应用于自身的GBoard项目,又例如微众银行提出的FATE联邦学习框架等等,均是上述参数服务器架构的联邦学习方法的实际应用的典型例子,它们都需要参与学习的不同设备之间具有非常高的一致性,模型结构一致,数据能够对齐。这就使得,对于一些不满足上述一致性要求的场景,现有的联邦学习方法并不适用。
(2)数据集蒸馏(dataset distillation)
区别于一般意义上的模型蒸馏,数据集蒸馏不是将知识从复杂模型迁移到小模型,而是在数据层面上的蒸馏。数据集蒸馏是将知识从大的训练数据集压缩到少量的合成数据中。在合成数据上可以通过几步梯度下降,快速训练模型以达到较高的准确率。
数据集蒸馏处理后获得的数据能够达到与原始数据相似的训练效果,数据集蒸馏相当于是减少数据,所以蒸馏后的数据虽然尽可能包含了足够多的数据知识,但会受到模型初始化的方式的影响。
(3)分组联邦学习
分组联邦学习指的是一种对客户端进行分组以减少各边端应用效果差异性较大的问题的联邦学习方式。由于单一模型联邦学习难以满足云模型适用于众多客户端的条件,在各边端应用效果差异性较大,因此考虑对客户端进行分组,例如根据相似度对客户端进行分组,然后组内进行联邦学习形成多个模型以适应于多个客户端组的需求。
(4)元学习(meta learning)
元学习是一种学习如何进行机器学习的方法,其核心思想是学习学习训练的过程。元学习从大量任务训练模型,并通过少量数据在新任务中更快地学习。元学习旨在寻找可以很容易适应新任务的内部表示,在深度学习的背景下,内部表示可以被视为神经网络的一组参数,训练网络只需要使用少量的新数据来对网络进行微调,即可得到新任务的有效模型。
(5)超参数
在机器学习的上下文中,超参数是在开始学习过程之前设置值的参数,而不是通过训练得到的参数数据。通常情况下,需要对超参数进行优化,给学习机选择一组最优超参数,以提高学习的性能和效果。
(6)数据集蒸馏联邦学习技术
该技术通过将传统联邦学习中多轮边云传输模型参数的过程替换为一次蒸馏数据的传输,来有效地减少了联邦学习过程中的通信代价。该技术实现过程示意图如图1所示。
(7)基于客户端分组的多模型联邦学习技术
该技术通过根据客户端在联邦学习过程中产生的参数更新的相似度来划分组,在各组内进行联邦优化可以大大提高模型在各组内的测试准确率,并且提高效率。该技术实现过程示意图如图2所示。
(8)基于元学习的联邦学习技术
该技术通过结合元学习思想来实现模型个性化,各客户端元学习一个好的初始化权重,得到一个共同初始模型作为起点,目的是使得模型可以在新任务上快速迁移适应。随后各客户端基于初始模型与少量的本地数据上更新模型,最终得到个性化模型。该技术实现过程示意图如图3所示。
(9)针对标签异构的数据生成联邦学习技术
该项技术针对客户端本地数据中对应数据量较少的标签,通过在服务器端训练GAN生成器用以生成目标数据,客户端下载GAN生成器补充本地目标标签数据直到各客户端数据集为独立同分布(即IID)。该技术实现过程示意图如图4所示。
图7是本申请实施例提供的一种联邦学习的方法100的示例性流程图,下面对图7所述的各个步骤进行介绍。应理解,图7的例子仅仅是为了帮助本领域技术人员理解本申请实施例,而非要将本申请实施例限制于图7的具体数值或具体场景。本领域技术人员根据所给出的例子,显然可以进行各种等价的修改或变化,这样的修改和变化也落入本申请实施例的范围内。
S110,根据目标联邦学习任务从算子库中获取N个算子。
示例性地,该算子库包括K个与联邦学习相关联的算子,K为大于1的整数。应理解,这里的算子库可以是由人工定期维护的算子库。这里的K个与联邦学习相关联的算子,可以理解为,该K个算子可用于处理联邦学习中的本地训练集,或者说,该K个算子是各种联邦学习相关的单一技术对应的算子,因此该K个算子可以称为K个联邦算子。应理解,该算子库中保存的算子可以包括该算子对应的函数并规范的输入和输出。例如,该该K个算子包括:数据集蒸馏算子,输入为原始数据,输出为蒸馏数据;客户端分组算子,输入为原始客户端集合以及数据,输出为新的客户端集合。还应理解,该算子库中可以仅包括该K个算子,即该算子库可以是一个联邦算子库;或者,该算子库中除了该K个算子还包括其他不同类型的算子,即该算子库中除了包括多个联邦算子,还可能包括其他类型的算子,不同类型的算子可以通过不同标签分类保存。本申请对原始算子库的结构不做限定。
示例性地,该目标联邦学习任务可以包括客户端需求信息。这里的客户端需求信息可以是客户端在模型训练之前协商确定的综合需求。例如,该需求包括数据通信量小而无需性能的提升,或者该需求包括对性能的需求高于对通信量的需求。
根据目标联邦学习任务从算子库中获取N个算子可以包括:获取目标联邦学习任务,然后根据该目标联邦学习任务从算子库中选择N个算子,该N个算子与该目标联邦学习任务相关联。例如,根据该目标联邦学习任务确定客户端需求为通信量小及数据平衡,则在算子库中选择N个用于该目标联邦学习任务的算子,或者说在算子库中选择N个用于实现客户需求的算子,例如,该N个算子为数据集蒸馏算子和客户端分组算子。其中N为大于1的整数且N小于或等于K,即该N个算子应当是该K个算子的子集。
可以将该N个算子的集合称为用于本次联邦学习的算子库。
S120,将N个算子中的多个算子进行复合得到第一复合算子。
示例性地,获取到N个算子之后,将该N个算子中的部分或者全部算子进行复合,以得到第一复合算子。应理解,这里的第一复合算子指的是将N个算子中的多个算子进行串联复合组成一定的算子复合结构之后,对该复合结构内的多个算子的一个统称,在其他场景中该复合算子还可以叫做其他名称,本申请对此不作限定。
可选地,可以根据N个算子的算子复合关系对该N个算子中的多个算子进行复合,这里的算子复合关系可用于表示该N个算子中任意两个算子之间的复合关系,即可以根据该算子复合关系确定该N个算子中任意两个算子之间是否可以复合。
作为一个示例,该算子复合关系可以采用算子复合关系矩阵来表示。
示例性地,生成一个N×N的零矩阵,然后对算子i和算子j的复合关系进行判断,该算子i和算子j为N个算子中任意两个算子,算子i个算子j可以相同也可以不同。例如,通过确定任意两个算子的输入输出空间的匹配关系,确定该该两个算子是否可以复合。本申请对判断两个之间输入输出空间是否匹配的具体方式不做限定,现有技术中任何用于确定两算子之间是否匹配的方案均应在本申请的保护范围内。作为一个具体示例,可以利用算子匹配函数(equal函数)来判断两个算子之间输入输出空间是否匹配。该equal函数也可以有多种设计形式,例如,通过遍历两个算子输入输出dict中的所有元素类型,如果所有元素类型都相同,则认为该两个算子之间的输入输出空间可以匹配。
当算子i的输出空间和算子j的输入空间相匹配,则认为算子i可复合到算子j,可以得到算子组合i+j。应理解,这里的算子i的输出空间和算子j的输入空间相匹配可以理解为,算子i的输入空间可以作为算子j的输出空间。
当算子i和算子j的输入输出空间相匹配,则更新算子复合关系矩阵M,否则不更新。
这里更新算子复合关系矩阵指的是,将矩阵M中的a ij更新为1,这里的a ij表示矩阵M的第i行第j列的元素。这里是以矩阵M的行表示输出、列表示输入为例进行说明的,本申请对矩阵M的其他表现形式不做限定。以该规则遍历N个算子中的任意两个算子(包括每个算子和其本身),可以得到最终的算子复合关系矩阵,该算子复合关系矩阵可表示不同算子之间的可复合性。
作为一个具体示例,假设该N个算子为A、B、C、D,遍历任意两个算子之间的输入输出匹配关系之后,得到的矩阵M例如是:
Figure PCTCN2022112055-appb-000001
其中,该矩阵的列表示输出,该矩阵的行表示输入。例如,a11为1,表示算子A的输出可以作为算子A的输入(即算子A的输出空间与算子A的输入空间相匹配),即算子A可以多次迭代复合本身;a12为0,表示算子A的输出不能作为算子B的输入,即,算子A不能串联复合到算子B。
需要说明的是,两个算子之间的匹配关系是有顺序要求的,即,算子i的输入空间与 算子j的输出空间相匹配,并不代表算子i的输出空间与算子j的输入空间一定可以匹配。算子i的输出空间与算子j的输入空间相匹配代表可以实现算子i+j的组合关系,算子i的输入空间与算子j的输出空间相匹配代表可以实现j+i的组合关系。例如在上面的矩阵M中,a23为0,a32为1,则表示算子B不能复合到算子C,但算子C可以复合到算子B。
应理解,本实施例是以算子复合关系矩阵来表示算子复合关系为例进行说明的,但本申请并不限于此,即在其他可能的实现方式中,可以采用别的可行方式来表示算子之间的复合关系。例如,还可以采用算子复合列表,该列表中包括两个算子之间所有可行的算子复合方式,或者包括两个算子之间所有不可行的算子复合方式等,本申请不作限定。
进一步地,根据该算子复合关系矩阵可以得到一系列算子复合的指导信息。例如,对于矩阵M,可以获得的指导信息包括但不限于:算子A可以多次迭代本身,例如A+A+A+…;算子A+C或A+D或C+D可作为一个级联块进行多次迭代复合,例如A+C+A+C…,或A+D+A+D…,或C+D+C+D…;A、B、C、D可以实现A+C+B或A+D+C+B等串联复合关系。因此,算子复合决策模块在根据算子复合关系矩阵生成算子复合结构或者对算子复合结构进行调整的过程中,会涉及到多个算子之间的顺序复合,也会涉及到两个算子复合后视为一个新算子之后,对新算子的迭代复合,等等。通过不同的方式进行迭代、复合等,可以得到多种算子复合结构。例如,根据上述实施例中的矩阵M,可以生成的算子复合结构包括但不限于:A+C,A+C+D,A+A+D,A+D+D,C+B……。
因此可以根据算子复合关系矩阵生成多种算子复合结构。这里的算子复合结构指的是不同算子之间复合之后得到的结构形式,或者可以理解为不同算子的串联结构。根据算子复合关系矩阵生成算子复合结构指的是,根据算子复合关系矩阵确定两两算子之间的可复合性,并将可复合的算子进行串联组合,得到不同的算子复合结构。第一复合算子对应的算子复合结构是该多种算子复合结构中的一种。
这里,可以是根据一定规则生成确定第一复合算子的算子复合结构,也可以在多种可行的算子复合结构中随机确定一种作为第一复合算子的算子复合结构,本申请对此不做限定。
S130,将第一数据集通过第一复合算子进行处理得到第二数据集。
示例性地,该第一数据集为用于目标联邦学习任务的训练集,或者说,该第一数据集为用于本次联邦学习的训练集(或者说数据集)。
通过第一复合算子对用于本次联邦学习的本地训练集进行处理,生成新的训练集,该新的训练集是经过第一复合算子处理后得到的新的数据的集合,或者是经过第一复合算子处理后(例如进行客户端分组)的具有新的参与方组成结构的数据集合。
S140,利用第二数据集进行联邦学习得到第一联邦学习模型。
应理解,本申请对联邦学习的具体方案不作限定,即本申请对联邦学习所使用的基模型或者算法不作限定。
可选地,在利用第二数据集进行联邦学习得到第一联邦学习模型之后,还可以进一步确定是否满足第一条件,在不满足第一条件的情况下,可以对第一复合算子的结构进行调整,以得到第二复合算子。
因此,这里的第一条件可以看作对算子复合结构进行调整的终止条件,该第一条件例如包括:第一联邦学习模型的评估值小于或等于第一阈值,或者,调整复合算子的结构的 次数大于或等于第二阈值。
其中,第一联邦学习模型的评估值是根据第一联邦学习模型的测量参数计算得到的,该测量参数与目标联邦学习任务相关联,或者说,该测量参数是用于评估本次联邦学习的相关参数,这些参数既可以是用户自定义的,也可以是系统默认的。评估参数例如可以是通信量(Mb)、模型精度(%)、模型训练时长(s)、通信时长(s)中的一项或多项。
这里的评估值可以是将评估参数代入评估公式计算得到的。本申请对该评估公式的具体形式不作限定,这里以评估公式为一种参数线性和的形式为例进行说明。例如,该评估公式为:
Figure PCTCN2022112055-appb-000002
其中,acc表示联邦测试的平均准确率,size(i)表示第i次通信所传输的通信量,系数a和系数b与用户需求相关联,用于表示各参数的权重。这里的a和b可以是用户自定义的,也可以是系统根据用户需求确定的。例如,a和b分别为0.7和0.3,该权重可以表示用户更偏向于对模型准确度的维护。
接下来对调整第一复合算子的结构的方式作示例性说明。
在一种可能的实现方式中,假设调整的终止条件为评估值小于10或调整次数达到30次。在利用第二数据集进行联邦学习得到第一联邦学习模型之后,获取测量参数并计算评估值。
当该评估值大于或等于10,则对第一复合算子的结构进行调整。一种示例,可以通过第一复合算子的结构(或者是第一联邦学习模型)和目标联邦学习任务对该第一复合算子的结构进行调整,这里的目标联邦学习任务可以包括用户需求。例如,获取第一联邦学习的测量参数,当该测量参数中的第一参数小于第三阈值时,将N个算子中的第一算子与该第一复合算子进行复合得到该第二复合算子,其中,该第一算子与该第一参数相关联,该第三阈值与目标联邦学习任务相关联。这里的第一算子与该第一参数相关联,可以表示第一算子可以用于调整第一参数。具体例如,该第一参数为第一联邦学习模型的模型精度,当第一参数小于第三阈值,说明第一联邦学习模型的模型精度不能满足用户需求,则可以选择第一算子与第一复合算子进行复合(或者将第一算子在第一复合算子中进行多次迭代),此时该第一算子可以是用于提升模型精度的算子。具体又例如,该第一参数为训练第一联邦学习模型的成本,当第一参数小于第三阈值,说明训练第一联邦学习模型的成本不能满足用户需求,则可以选择第一算子与第一复合算子进行复合(或者将第一算子在第一复合算子中进行多次迭代),此时该第一算子可以是用于节约成本的算子。通过上述实现方式,可以算子复合结构进行调整,以满足用户需求。另一种示例,可以通过自动化调参的方式对该第一复合算子的结构进行调整,该自动化调参例如可以是网格搜索,该网格搜索可以是按照一定顺序对算子复合结构进行调整的穷举搜索方式。例如,根据算子复合关系矩阵确定所有可能的算子复合结构,并将这些算子复合结构按照一定的顺序进行排序,当本次用于联邦学习的评估值大于或等于10,则选择本次联邦学习的算子复合结构的下一个算子复合结构构成下一次联邦学习的复合算子。然后循环该过程,直到满足调整的终止条件。或者,这里的网格搜索也可以是按照随机的顺序对算子复合结构进行调整的穷举搜索方式,本申请不做限定。需要说明的是,当调整次数达到30次,但这30次内的评估值均大于或等于10,则可以选择这30次中评估值最接近10的那一次联邦学习所使用的算子复合结构。又一种示例,可以通过多次迭代的方式对第一复合算子的结构进行调整。例如,可以根据测量参数和目标联邦模型参数之间的差异计算损失值,然后将损失值进行 反向传播,根据损失值对该第一复合算子的结构进行调整得到第二复合算子,这里的损失值可以指示误差大小,这里的目标联邦模型参数可以是根据用户需求确定的,本申请对损失值的计算方式不作限定。应理解,还可以采用其他迭代优化的方法,例如无损失计算函数的无监督算法,甚至是前馈算法等,本申请不做限定。
因此,本申请实施例提供的联邦学习的方法,将多个算子级联,每个算子用于解决不同的问题,将不同算子串联起来相互作用,将针对解决某一联邦学习技术问题的算子的误差由其他算子来减弱,通过对不同算子的动态组合调整可以获得更好的整体效果,从而可以提高联邦学习的效率和质量。即该方法能够将针对不同单一问题的方法进行复合,通过一种级联元学习结构,使得输出的方法组合结构能够适应于优化联邦学习实际场景的问题。
也就是说,现有的用于解决联邦学习单一问题的技术仍存在一些缺陷,无法应对实际联邦学习场景中各种问题。例如,对于数据集蒸馏联邦学习技术,虽然该技术可以有效减少联邦学习过程的通信代价,但是该技术并未涉及联邦学习中的异构数据不平衡问题,缺少对异构数据的讨论和解决;对于基于客户端分组的多模型联邦学习技术,由于是利用模型参数更新的差异来指导分组,需要多次联邦学习过程,会带来更多的通信代价;对于基于元学习的联邦学习技术,单纯的根据全局模型的准确率来优化模型会损失模型后续的个性化能力;对于针对标签异构的数据生成联邦学习技术,该方法针对于标签异构情况,对于客户端数据特征异构的情况无法有效处理,并且即使上传少量原始数据,仍存在隐私问题。而在显示情况下,联邦学习本身就是多问题的叠加,定位问题以及单一问题的方法调用本身就不够有效的情况下,本申请实施例提供的方法就显得非常重要,因为它能够自适应问题去学习适合当前问题的联邦学习结构,有效调用现有方法。算子之间的有效复合能够增大单个算子对于联邦学习的增益效果,实现对多个目标问题的同时优化,而不是仅针对联邦学习中单一问题进行优化。
图8示出了本申请实施例提供的一种联邦学习的系统架构200的示意性框图,该系统可用于执行方法100。该系统主要包括四个模块:第一模块、第二模块、第三模块和第四模块。下面对各个模块作示例性说明。
第一模块,用于生成联邦算子复合关系,以便生成本次联邦学习的复合算子。第一模块也可以称为联邦算子复合关系模块。下面结合图9对该第一模块所执行的流程300作示例性说明。
首先,第一模块获取本次联邦学习的算子库,并检查该算子库中每两个算子之间的输入输出空间是否匹配。
示例性地,第一模块从该联邦学习系统的中央服务器接收本次联邦学习的算子库(例如可对应于方法100中的N个算子)。在一种可能的情况中,中央服务器接收目标联邦学习任务,该目标联邦学习任务例如可以包括客户端需求信息。中央服务器根据该客户端需求信息在算子库中获取N个算子,该N个算子组成了本次联邦学习的算子库。在另一种可能的情况中,用户直接根据需求选取本次联邦学习的算子库,并将该算子库发送或指示给中央服务器。进一步地,中央服务器将本次联邦学习的算子库传递给第一模块。
第一模块生成算子复合关系零矩阵,每当算子库中的两个算子的输入输出空间匹配时,第一模块更新算子复合关系矩阵,否则不更新。遍历N个算子中的任意两个算子(包括每个算子和其本身),可以得到最终的算子复合关系矩阵,该算子复合关系矩阵可表示不同 算子之间的可复合性。具体方案与方法100中的通过算子复合关系矩阵来表示算子复合关系的方案类似,这里不再重复说明。
进一步地,第一模块将最终生成的算子复合关系矩阵输出给第二模块。第二模块根据算子复合关系矩阵调整算子复合结构,得到复合算子。该第二模块可以称为算子复合决策模块,或联邦算子复合决策模块。
应理解,第一模块和第二模块可以是相互独立的模块,也可以同属于一个模块,例如第一模块和第二模块同属于决策器模块,或者说,第一模块和第二模块可以是决策器模块中的不同子模块。本申请对此不做限定。
下面结合图10对该第二模块所执行的流程400作示例性说明。
首先,第二模块获取本次联邦学习的算子库,并接收来自第一模块的算子复合关系矩阵,根据该算子复合关系矩阵确定第一复合算子的结构,并将第一复合算子输出到第三模块,该第三模块是用于联邦学习的模块。
应理解,第二模块可以是从第一模块获取的算子库,也可以是直接从中央服务器获取的算子库,本申请不作限定。
第三模块将第一数据集通过第一复合算子进行处理得到第二数据集,该第一数据集为用于本次联邦学习的训练集。然后第四模块获取本次联邦学习的测量参数,并根据该测量参数计算评估值,第四模块将该评估值通过反馈信息发送给第二模块。第二模块可以根据该反馈信息调整算子复合结构。具体方案可以参见方法100中的相关说明,为了简洁,这里不再赘述。
应理解,第四模块和第三模块可能是相同的模块,也可能是不同的模块,本申请对此不作限定。
本申请实施例提供的联邦学习的系统,通过决策器模块(包含上述实施例中的第一模块和第二模块)寻找针对单一方法复合的最大化价值的目标结构,即寻找到如何针对真实业务场景去应用较佳的复合方法,来满足客户的确定需求。决策器模块在整个系统中负责动态调整系统结构并评价,通过决策器模块来学习得到满足客户需求的级联联邦学习结构。同时,决策器模块以评估值(用户成本公式计算结果)作为性能指标(benchmark),在整个算子库中以算子复合关系矩阵作为指导,有方向地动态调整算子复合结构,从而可以得到满足用户需求的级联联邦学习系统。
下面以控制空调舒适度联邦模型的训练过程为例,详细介绍本申请实施例提供的联邦学习的方法。
首先,需要进行一些初始化设置。
示例性地,用户对一些基本参数及其数据类型进行初始化设置。例如,表1列出了一种情况下用户在初始化阶段需要输入的参数列表及其示例。
表1
Figure PCTCN2022112055-appb-000003
Figure PCTCN2022112055-appb-000004
下面对表1中的各项输入参数作示例性说明:
(1)名称:用户自定义联邦学习项目实例名称;
(2)工作环境:用户自定义工作环境,也可以采用默认工作环境,例如Python3;
(3)元联邦学习算子类型列表:用户从算子库中选择算法,例如可选择数据集蒸馏算法、客户端分组算法、虚拟样本平衡算法、元学习算法等;
(4)基模型:用户自定义联邦学习基模型,例如可选择神经网络模型、Boost模型、Bagging模型等;
(5)数据范例:用户需要提交至少一条范例数据来反映数据的格式,描述它的特征和标签结构;
(6)训练集:用户需提交用于联邦学习的数据集;
(7)评估参数列表:用户可自定义关于联邦学习的评估指标,这里的评估指标用于评估通过联邦学习训练出来的模型。例如,可选的联邦学习过程中间变量为通信量(Mb)、模型精度(%)、模型训练时长(s)、通信时长(s)等;
(8)自定义评估公式:用户可自定义评估公式。可设置系统默认评估公式为线性和的形式,用户可直接设置各参数的权重。需要说明的是,评估公式中的参数应当是上述参数评估列表中的参数中的一项或多项。
(9)终止条件:用户可以设置算子复合结构调整的终止条件,例如用户设置算子复合结构调整的复合次数为30。当用户没有设置终止条件时,可以采用默认的最大调整次数,例如默认为20次。
(10)下载目录:用户需要设置本地保存模型的本地目录;
(11)上载目录:用户需要设置用于保存本地运行的需要上载日志的目录。
在该实施例中,以多联机空调舒适度控制预测公开数据集ATCII(记为联邦模拟数据集)作为联邦学习的训练集。该联邦模拟数据集中,包括69个国家以及每个国家细化到不同城市下的采集数据,一共有约76000条采集数据,共包含特征70个,基于实际场景以及特征实际采集难易程度的考量,最终选择8个通用特征用于测试。这8个通用特征均为温度以及速度的数值形式信息,其中后三个特征为原始特征进行平均值运算而来。其次,ATCII数据集中特征Thermal preference的采集值为‘cooler’、’no change’、’warmer’,故选择特征Thermal preference作为实验的标签,用以训练三分类预测模型。
另外,选用ACTII数据集中的五个城市Berkeley、Darwin、Kalgoorlie、Oxford、Townsvill作为5个客户端来参与整个联邦学习,各个城市保存的数据作为对应客户端的训练集。
在方法500中,假设用户设置的评估参数为模型测试准确率acc和通信量size,评估 的标准设置为(100-acc)和size线性组合成本,人为设置的成本计算公式中两项参数的单位为%和Mb,权重分别为0.7和0.3,这里表示用户更偏向于对模型准确度的维护。但公式仅仅是一个测试,并非该成本计算公式中的单位和成本有合理的计算关系。抽象为结果计算公式的计算结果单位为成本(元),计算公式简记形式为:
Figure PCTCN2022112055-appb-000005
其中Acc表示联邦测试平均准确率,size(i)表示第i次通信所传输的通信量。
假设用户设置的终止条件为调整结构30次/成本计算公式结果小于10。
然后,生成本次联邦学习的算子库信息。
示例性地,系统根据用户输入的信息确定本次联邦学习的算子库信息。表2列出了一种情况下用户输入的参数列表及其示例。
表2
输入参数 例子(数据类型)
元联邦学习算子列表 [“data_distillation”,“client_grouping”,…](list of string)
数据范例 {“feature1”:a,”feature2”:b,…,”label”:y}(dict)
下面对表2中的各项输入参数做示例性说明:
(1)元联邦学习算子列表:用户从原始算子库中选择算子,例如用户根据用户需求选择数据集蒸馏算子和客户端分组算子为本次联邦学习的算子;
(2)数据范例:用户需要提交至少一条范例数据来反映数据的格式、描述它的特征和标签结构。
表3
输出参数 例子(数据类型)
算子函数列表 [“funA”,“funB”,…,“funM”](list of string)
表3列出了一种情况下系统输出的参数列表。下面对表3中的输出参数做示例性说明:
算子函数列表:系统根据用户提供的数据范例的结构形式,对用户选择的元联邦学习算子进行调整得到的标准函数库,通过函数名即可调用对应函数,例如,通过“funA”可以调用算子A。
应理解,表2和表3所列出的输入和输出参数仅仅是一种示例,本申请实施例还可以采用其他方式输出本次联邦学习的算子库。例如,用户可以向系统输入用户需求,例如,用户输入数据加密、数据平衡等需求(例如在系统界面选择数据加密、数据平衡等需求),系统根据用户需求在原始算子库中选择本次联邦学习的算子。应理解,系统选择的本次联邦学习的算子应当与用户需求相对应,或者说,系统选择的本次联邦学习的算子应当是用于实现用户需求的算子。
在确定了本次联邦学习的算子库之后,还需要对本次联邦学习的算子库中的算子进行参数设置。为了方便说明,以本次联邦学习的算子库为{数据集蒸馏算子(E算子)、客户端分组算子(F算子)}为例进行说明。下面对这两个算子进行参数设置的方式作示例性说明:
(1)数据集蒸馏算子:选择包含两个隐藏层的人工神经网络ANN;激活函数选择ReLU函数;蒸馏采用固定初始化的初始模型;训练轮数设置为20轮,每轮30步梯度下 降(10步梯度下降重复3次);蒸馏数据数目设置为每类5个。
(2)客户端分组算子:设置最大划分组数为输入城市数,最小划分组数为1;设置划分方法采用自定义Non-iid度量算法描述任务相似程度;设置迭代终止条件为组间度量和最大,组内度量和最小;设置迭代最大次数10。
需要说明的是,系统输出的算子库信息中包括算子的输入输出的数据(以及该数据的格式)以及参数设置的表格。作为一个示例,对于数据集蒸馏算子,输入为各客户端原始数据集,输出为客户端蒸馏数据集;对于客户端分组算子,输入为原始客户端集合以及数据,输出为新的客户端集合。
进一步地,生成联邦算子复合关系矩阵。
示例性地,联邦算子复合关系模块根据本次联邦学习的算子库信息,将任意两个算子之间的输入输出空间进行匹配以生成联邦算子复合关系矩阵。
下面以E算子和F算子为例进行说明。
联邦算子复合关系模块输入本次联邦学习的算子库信息(例如该算子库信息为{E算子、F算子}、E算子的输入输出并数据格式、F算子的输入输出并数据格式、E算子的参数设置表格、F算子的参数设置表格等)以及超参数(该超参数不能自复合),然后判断E算子和F算子的输入输出空间是否匹配。
一种方式是,通过E算子和F算子输入输出空间的数据形式,来判断E算子和F算子的输入输出空间是否匹配。
首先简单介绍一种获取数据形式的方式:比如说有一个叫做BaseAlgorithm的基类,它有两个基类方法get_input_type()、get_output_type(),输入到系统中的算子都必须继承自这个BaseAlgorithm基类,算子包含一个run()方法,实现对应的蒸馏、客户分组等运算。然后系统通过调用get_input_type()、get_output_type()方法获取到算子的输入输出数据形式,存储到算子库中。数据形式可以是list/set或者用户自定义的对象。
进一步地,可以对E算子和F算子的数据形式进行equal()计算,以判断E算子和F算子的输入输出空间是否能够对接上。
应理解,E算子和F算子的数据形式的获取流程以及判断E算子和F算子的输入输出空间是否能够对接上的流程,既可以是在联邦复合关系模块中执行的,也可以是在其他模块中执行的,本申请不作限定。
假设E算子的输出空间与F算子的输入空间相匹配,F算子的输出空间与E算子的输入空间也相匹配,E算子的输出空间与E算子的输入空间相匹配,则可以得到算子复合关系E+E、E+F和F+E。可以采用邻接矩阵的方式进行建模,输出算子复合关系矩阵N:
Figure PCTCN2022112055-appb-000006
以数据集蒸馏算子+客户端分组算子(即E+F)级联形式为例进行说明:对于(E+F)这个复合算子来说,E算子的输入为客户端集合以及各客户端本地数据集,输出为各客户端蒸馏数据集;F算子的输入为客户端集合以及客户端蒸馏数据集,输出为新的客户端组。
进一步地,生成和调整联邦算子复合结构。
示例性地,算子复合决策模块获取算子复合关系矩阵及联邦算子库信息,生成第一复 合算子。然后将该第一复合算子复合到联邦学习系统,将各客户端的本地训练集经过该联邦学习系统进行联邦学习,得到第一联邦模型。
表4列出了一种情况下联邦学习过程用户需要输入的参数列表及其示例。应理解,这些参数可以是用户在初始化设置阶段即输入到系统的。
表4
输入参数 例子(数据类型)
基模型 “AlexNet_model”(string)
训练集 {“X”:array1,”Y”:array2}(dict)
联邦模型 [“fed_model1”,”fed_model2”,…](list of string)
下面对表4中的各项输入参数做示例性说明:
(1)基模型:用户可选择本次联邦训练所使用的基模型的类型,例如神经网络模型、Boost模型、Bagging模型等;
(2)训练集:用户需要提交用户需提交用于联邦学习的数据集。
(3)联邦模型:用户可选择联邦模型的类型,例如神经网络模型、Boost模型、Bagging模型等。
在联邦训练过程中,或者联邦训练完成之后,可以获取测量参数,并根据测量参数计算评估值,将测量参数和评估值反馈给算子复合决策模块,以便算子复合决策模块对第一复合算子的结构进行调整。
表5列出了一种情况下用户输入到算子复合决策模块的参数列表及其示例。应理解,这些参数可以是用户在初始化设置阶段即输入到系统的。
表5
Figure PCTCN2022112055-appb-000007
对于各输入参数的说明可以参见表1中相关参数的说明,这里不再赘述。
本次联邦学习之后,如果不满足终止条件,则对第一复合算子的结构进行调整。表6列出了一种情况下算子复合结构的搜索过程。为了方便描述,将算子复合结构接入传统联邦系统的过程记为G,需要说明的是,这里的G并非为算子。在该实施例中,联邦学习过程元学习器采用三分类XGBoost进行联邦学习,XGBoost参数控制不变,具体联邦学习过程为联邦蒸馏数据与本地数据结合训练。
假设用户设置的终止条件为调整结构30次/成本计算公式结果小于10。每一次联邦学习之后,如果不满足终止条件,则可以按照表6中的顺序调整算子复合结构。
表6
Figure PCTCN2022112055-appb-000008
最后迭代终止,触发迭代10次的终止条件,未满足评估值小于10的终止条件,最后算子复合决策模块输出E+F+E+F+E+G的算子复合的联邦学习结构,是搜索过程中满足评估公式值最小的结构。将E+F+E+F+E+G统称为一个复合算子。进一步有通过该复合算子对训练集进行处理:
5客户数据集输入数据集蒸馏算子输出蒸馏数据,压缩数据;蒸馏数据传输至云端,云端对蒸馏数据进行Non-iid度量,划分新组别(Berkeley、Kalgoorlie、Oxford)(Darwin、Townsvill);继续对组内进行多云端控制的联邦学习,组内客户进行数据集蒸馏,并传输至云端;组内云对蒸馏数据进行Non-iid度量,重新划分组结构(Berkeley)(Kalgoorlie、Oxford)(Darwin、Townsvill);新组内进行数据集蒸馏,传输至云端;云端汇总组内蒸馏数据,组内用户结合分享的蒸馏数据和本地数据训练Xgboost模型,并测试。
由于通常蒸馏数据集规模和本地训练数据集规模差距较大,本实施例采用自适应地对蒸馏数据集进行实例迁移的方法,来提升蒸馏数据在结合本地数据后的占比,强调蒸馏数据对于客户端分组的指导意义。
图11是根据本申请实施例提供的一种计算机设备的示意性结构框图。如图11所示的计算机设备500包括获取单元501和处理单元502。
获取单元501,用于根据目标联邦学习任务从算子库中获取N个算子,所述算子库包 括K个与联邦学习相关联的算子,所述N和所述K为大于1的整数,所述N小于或等于所述K。
处理单元502,用于将所述N个算子中的多个算子进行复合得到第一复合算子;将第一数据集通过所述第一复合算子进行处理得到第二数据集,所述第一数据集为用于所述目标联邦学习任务的训练集;利用所述第二数据集进行联邦学习得到第一联邦学习模型。
可选的,在一些实施例中,处理单元502,具体用于:根据所述N个算子中任意两个算子之间的算子复合关系,将所述N个算子中的多个算子进行复合得到所述第一复合算子。
可选的,在一些实施例中,所述处理单元502还用于:确定所述N个算子中任意两个算子之间的输入输出空间的匹配关系;根据所述N个算子中任意两个算子之间的输入输出空间的匹配关系确定所述N个算子中任意两个算子之间的算子复合关系。
可选的,在一些实施例中,所述处理单元502还用于:确定是否满足第一条件;在不满足第一条件的情况下,对所述第一复合算子的结构进行调整得到第二复合算子。
可选的,在一些实施例中,所述获取单元501具体用于:获取所述第一联邦学习模型的测量参数;所述处理单元502具体用于:根据所述测量参数计算评估值,所述测量参数与所述目标联邦学习任务相关联;在所述评估值小于第一阈值的情况下,确定满足所述第一条件。
可选的,在一些实施例中,所述处理单元502具体用于:确定调整复合算子的结构的次数;在所述调整复合算子的结构的次数大于或等于第二阈值的情况下,确定满足第一条件。
可选的,在一些实施例中,所述处理单元502具体用于:根据所述测量参数和目标联邦模型参数之间的差异计算损失值,将所述损失值进行反向传播,根据所述损失值对所述第一复合算子的结构进行调整得到第二复合算子。
可选的,在一些实施例中,所述算子库包括数据集蒸馏算子、客户端分组算子、虚拟样本平衡算子、元学习算子中的至少一种。
获取单元501、处理单元502的具体功能和有益效果可以参考图7-图10所示的实施例,为了简洁,在此就不再赘述。
图12是根据本申请实施例提供的一种用于联邦学习的装置的示意性结构框图。装置600包括总线601、处理器602、通信接口603和存储器604。处理器602、存储器604和通信接口603之间通过总线601通信。其中,处理器702可以是编程门阵列(field programmable gate array,FPGA),可以是专用集成芯片(application specific integrated circuit,ASIC),还可以是系统芯片(system on chip,SoC),还可以是中央处理器(central processor unit,CPU),还可以是网络处理器(network processor,NP),还可以是数字信号处理电路(digital signal processor,DSP),还可以是微控制器(micro controller unit,MCU),还可以是可编程控制器(programmable logic device,PLD)、其他可编程逻辑器件、分立门或者晶体管逻辑器件、分立硬件组件,或其他集成芯片。存储器604中存储有病理图像处理系统所包括的可执行代码,处理器602读取存储器604中的该可执行代码以执行如图3所示的方法。存储器604中还可以包括操作系统等其他运行进程所需的软件模块。操作系统可以为LINUX TM,UNIX TM,WINDOWS TM等。
本申请实施例还提供一种芯片系统,其特征在于,包括:逻辑电路,所述逻辑电路用 于与输入/输出接口耦合,通过所述输入/输出接口传输数据,以执行如图3所述的方法。
在实现过程中,上述方法的各步骤可以通过处理器中的硬件的集成逻辑电路或者软件形式的指令完成。结合本申请实施例所公开的方法的步骤可以直接体现为硬件处理器执行完成,或者用处理器中的硬件及软件模块组合执行完成。软件模块可以位于随机存储器,闪存、只读存储器,可编程只读存储器或者电可擦写可编程存储器、寄存器等本领域成熟的存储介质中。该存储介质位于存储器,处理器读取存储器中的信息,结合其硬件完成上述方法的步骤。为避免重复,这里不再详细描述。
应注意,本申请实施例中的处理器可以是一种集成电路芯片,具有信号的处理能力。在实现过程中,上述方法实施例的各步骤可以通过处理器中的硬件的集成逻辑电路或者软件形式的指令完成。通用处理器可以是微处理器或者该处理器也可以是任何常规的处理器等。结合本申请实施例所公开的方法的步骤可以直接体现为硬件译码处理器执行完成,或者用译码处理器中的硬件及软件模块组合执行完成。软件模块可以位于随机存储器,闪存、只读存储器,可编程只读存储器或者电可擦写可编程存储器、寄存器等本领域成熟的存储介质中。该存储介质位于存储器,处理器读取存储器中的信息,结合其硬件完成上述方法的步骤。
可以理解,本申请实施例中的存储器可以是易失性存储器或非易失性存储器,或可包括易失性和非易失性存储器两者。其中,非易失性存储器可以是只读存储器(read-only memory,ROM)、可编程只读存储器(programmable ROM,PROM)、可擦除可编程只读存储器(erasable PROM,EPROM)、电可擦除可编程只读存储器(electrically EPROM,EEPROM)或闪存。易失性存储器可以是随机存取存储器(random access memory,RAM),其用作外部高速缓存。通过示例性但不是限制性说明,许多形式的RAM可用,例如静态随机存取存储器(static RAM,SRAM)、动态随机存取存储器(dynamic RAM,DRAM)、同步动态随机存取存储器(synchronous DRAM,SDRAM)、双倍数据速率同步动态随机存取存储器(double data rate SDRAM,DDR SDRAM)、增强型同步动态随机存取存储器(enhanced SDRAM,ESDRAM)、同步连接动态随机存取存储器(synchlink DRAM,SLDRAM)和直接内存总线随机存取存储器(direct rambus RAM,DR RAM)。应注意,本文描述的系统和方法的存储器旨在包括但不限于这些和任意其它适合类型的存储器。
根据本申请实施例提供的方法,本申请还提供一种计算机程序产品,该计算机程序产品包括:计算机程序代码,当该计算机程序代码在计算机上运行时,使得该计算机执行图3所示实施例的方法。
根据本申请实施例提供的方法,本申请还提供一种计算机可读介质,该计算机可读介质存储有程序代码,当该程序代码在计算机上运行时,使得该计算机执行图3所示实施例的方法。
本领域普通技术人员可以意识到,结合本文中所公开的实施例描述的各示例的单元及算法步骤,能够以电子硬件、或者计算机软件和电子硬件的结合来实现。这些功能究竟以硬件还是软件方式来执行,取决于技术方案的特定应用和设计约束条件。专业技术人员可以对每个特定的应用来使用不同方法来实现所描述的功能,但是这种实现不应认为超出本申请的范围。
所属领域的技术人员可以清楚地了解到,为描述的方便和简洁,上述描述的系统、装 置和单元的具体工作过程,可以参考前述方法实施例中的对应过程,在此不再赘述。
在本申请所提供的几个实施例中,应该理解到,所揭露的系统、装置和方法,可以通过其它的方式实现。例如,以上所描述的装置实施例仅仅是示意性的,例如,所述单元的划分,仅仅为一种逻辑功能划分,实际实现时可以有另外的划分方式,例如多个单元或组件可以结合或者可以集成到另一个系统,或一些特征可以忽略,或不执行。另一点,所显示或讨论的相互之间的耦合或直接耦合或通信连接可以是通过一些接口,装置或单元的间接耦合或通信连接,可以是电性,机械或其它的形式。
所述作为分离部件说明的单元可以是或者也可以不是物理上分开的,作为单元显示的部件可以是或者也可以不是物理单元,即可以位于一个地方,或者也可以分布到多个网络单元上。可以根据实际的需要选择其中的部分或者全部单元来实现本实施例方案的目的。
另外,在本申请各个实施例中的各功能单元可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中。
所述功能如果以软件功能单元的形式实现并作为独立的产品销售或使用时,可以存储在一个计算机可读取存储介质中。基于这样的理解,本申请的技术方案本质上或者说对现有技术做出贡献的部分或者该技术方案的部分可以以软件产品的形式体现出来,该计算机软件产品存储在一个存储介质中,包括若干指令用以使得一台计算机设备(可以是个人计算机,服务器,或者网络设备等)执行本申请各个实施例所述方法的全部或部分步骤。而前述的存储介质包括:U盘、移动硬盘、只读存储器(read-only memory,ROM)、随机存取存储器(random access memory,RAM)、磁碟或者光盘等各种可以存储程序代码的介质。
以上所述,仅为本申请的具体实施方式,但本申请的保护范围并不局限于此,任何熟悉本技术领域的技术人员在本申请揭露的技术范围内,可轻易想到变化或替换,都应涵盖在本申请的保护范围之内。因此,本申请的保护范围应以所述权利要求的保护范围为准。

Claims (18)

  1. 一种联邦学习的方法,其特征在于,包括:
    根据目标联邦学习任务从算子库中获取N个算子,所述算子库包括K个与联邦学习相关联的算子,所述N和所述K为大于1的整数,所述N小于或等于所述K;
    将所述N个算子中的多个算子进行复合得到第一复合算子;
    通过所述第一复合算子将第一数据集进行处理得到第二数据集,所述第一数据集为用于所述目标联邦学习任务的训练集;
    利用所述第二数据集进行联邦学习得到与所述目标联邦学习任务相对应的第一联邦学习模型。
  2. 根据权利要求1中任一项所述的方法,其特征在于,所述将所述N个算子中的多个算子进行复合得到第一复合算子,包括:
    确定所述N个算子中任意两个算子之间的输入输出空间的匹配关系;
    根据所述N个算子中任意两个算子之间的输入输出空间的匹配关系,将所述N个算子中的多个算子进行复合得到所述第一复合算子。
  3. 根据权利要求1或2所述的方法,其特征在于,在利用所述第二数据集进行联邦学习得到第一联邦学习模型之后,所述方法还包括:
    获取所述第一联邦学习模型的测量参数;
    根据所述测量参数计算评估值,所述测量参数与所述目标联邦学习任务相关联;
    在所述评估值大于或等于第一阈值的情况下,对所述第一复合算子的结构进行调整得到所述第二复合算子。
  4. 根据权利要求3所述的方法,其特征在于,在利用所述第二数据集进行联邦学习得到第一联邦学习模型之后,所述方法还包括:
    确定调整复合算子的结构的次数;
    在所述调整复合算子的结构的次数小于第二阈值的情况下,对所述第一复合算子的结构进行调整得到所述第二复合算子。
  5. 根据权利要求3或4所述的方法,其特征在于,所述对所述第一复合算子的结构进行调整得到第二复合算子,包括:
    根据所述测量参数和目标联邦模型参数之间的差异计算损失值,将所述损失值进行反向传播,根据所述损失值对所述第一复合算子的结构进行调整得到第二复合算子。
  6. 根据权利要求3或4所述的方法,其特征在于,所述对所述第一复合算子的结构进行调整得到第二复合算子,包括:
    当所述第一联邦学习模型的测量参数中的第一参数小于第三阈值时,将所述N个算子中的第一算子与所述第一复合算子进行复合得到所述第二复合算子,所述第一算子与所述第一参数相关联。
  7. 根据权利要求3或4所述的方法,其特征在于,所述对所述第一复合算子的结构进行调整得到第二复合算子,包括:
    将所述第一复合算子的结构进行自动化调参得到所述第二复合算子。
  8. 根据权利要求1至7中任一项所述的方法,其特征在于,所述算子库包括数据集蒸馏算子、客户端分组算子、虚拟样本平衡算子、元学习算子中的至少一种。
  9. 一种联邦学习的装置,其特征在于,包括:
    获取单元,用于根据目标联邦学习任务从算子库中获取N个算子,所述算子库包括K个与联邦学习相关联的算子,所述N和所述K为大于1的整数,所述N小于或等于所述K;
    处理单元,用于将所述N个算子中的多个算子进行复合得到第一复合算子;
    所述处理单元还用于将第一数据集通过所述第一复合算子进行处理得到第二数据集,所述第一数据集为用于所述目标联邦学习任务的训练集;
    所述处理单元还用于利用所述第二数据集进行联邦学习得到与所述目标联邦学习任务对应的第一联邦学习模型。
  10. 根据权利要求9中任一项所述的装置,其特征在于,所述处理单元具体用于:
    确定所述N个算子中任意两个算子之间的输入输出空间的匹配关系;
    根据所述N个算子中任意两个算子之间的输入输出空间的匹配关系,将所述N个算子中的多个算子进行复合得到所述第一复合算子。
  11. 根据权利要求9或10所述的装置,其特征在于,所述处理单元还用于:
    获取所述第一联邦学习模型的测量参数;
    所述处理单元具体用于:
    根据所述测量参数计算评估值,所述测量参数与所述目标联邦学习任务相关联;
    在所述评估值大于或等于第一阈值的情况下,对所述第一复合算子的结构进行调整得到所述第二复合算子。
  12. 根据权利要求9至11中任一项所述的装置,其特征在于,所述处理单元还用于:
    确定调整复合算子的结构的次数;
    在所述调整复合算子的结构的次数小于第二阈值的情况下,对所述第一复合算子进行调整得到所述第二复合算子。
  13. 根据权利要求11或12所述的装置,其特征在于,所述处理单元具体用于:
    根据所述测量参数和目标联邦模型参数之间的差异计算损失值,将所述损失值进行反向传播,根据所述损失值对所述第一复合算子的结构进行调整得到第二复合算子。
  14. 根据权利要求11或12所述的装置,其特征在于,所述处理单元具体用于:
    当所述第一联邦学习模型中的第一参数小于第三阈值时,将所述N个算子中的第一算子与所述第一复合算子进行复合得到所述第二复合算子,所述第一算子与所述第一参数相关联。
  15. 根据权利要求11或12所述的装置,其特征在于,所述处理单元具体用于:
    将所述第一复合算子的结构进行自动化调参得到所述第二复合算子。
  16. 根据权利要求9至15中任一项所述的装置,其特征在于,所述算子库包括数据集蒸馏算子、客户端分组算子、虚拟样本平衡算子、元学习算子中的至少一种。
  17. 一种计算机可读存储介质,其特征在于,所述计算机可读存储介质中存储有计算机程序,当所述计算机程序在一个或多个处理器上运行时,使得所述计算机执行如权利要求1至8中任一项所述的方法。
  18. 一种芯片系统,其特征在于,所述芯片系统包括处理器与数据接口,所述处理器通过所述数据接口读取存储器上存储的指令,以执行如权利要求1至8中任一项所述的方法。
PCT/CN2022/112055 2021-08-30 2022-08-12 联邦学习的方法和装置 WO2023029944A1 (zh)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202111001636.4A CN115730631A (zh) 2021-08-30 2021-08-30 联邦学习的方法和装置
CN202111001636.4 2021-08-30

Publications (1)

Publication Number Publication Date
WO2023029944A1 true WO2023029944A1 (zh) 2023-03-09

Family

ID=85290824

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/112055 WO2023029944A1 (zh) 2021-08-30 2022-08-12 联邦学习的方法和装置

Country Status (2)

Country Link
CN (1) CN115730631A (zh)
WO (1) WO2023029944A1 (zh)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116506469A (zh) * 2023-06-25 2023-07-28 华东交通大学 基于振动取能rfid传感器与联邦学习的桥梁状态监测方法

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111310934A (zh) * 2020-02-14 2020-06-19 北京百度网讯科技有限公司 一种模型生成方法、装置、电子设备和存储介质
CN112257876A (zh) * 2020-11-15 2021-01-22 腾讯科技(深圳)有限公司 联邦学习方法、装置、计算机设备及介质
CN112464287A (zh) * 2020-12-12 2021-03-09 同济大学 基于秘密共享与联邦学习的多方XGBoost安全预测模型训练方法
CN112598121A (zh) * 2020-12-21 2021-04-02 北京时代民芯科技有限公司 一种面向深度学习编译器的高效算子优化方法
CN112733967A (zh) * 2021-03-30 2021-04-30 腾讯科技(深圳)有限公司 联邦学习的模型训练方法、装置、设备及存储介质
CN113052334A (zh) * 2021-04-14 2021-06-29 中南大学 一种联邦学习实现方法、系统、终端设备及可读存储介质
CN113128419A (zh) * 2021-04-23 2021-07-16 京东鲲鹏(江苏)科技有限公司 一种障碍物识别方法和装置、电子设备及存储介质
WO2021158313A1 (en) * 2020-02-03 2021-08-12 Intel Corporation Systems and methods for distributed learning for wireless edge dynamics

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2021158313A1 (en) * 2020-02-03 2021-08-12 Intel Corporation Systems and methods for distributed learning for wireless edge dynamics
CN111310934A (zh) * 2020-02-14 2020-06-19 北京百度网讯科技有限公司 一种模型生成方法、装置、电子设备和存储介质
CN112257876A (zh) * 2020-11-15 2021-01-22 腾讯科技(深圳)有限公司 联邦学习方法、装置、计算机设备及介质
CN112464287A (zh) * 2020-12-12 2021-03-09 同济大学 基于秘密共享与联邦学习的多方XGBoost安全预测模型训练方法
CN112598121A (zh) * 2020-12-21 2021-04-02 北京时代民芯科技有限公司 一种面向深度学习编译器的高效算子优化方法
CN112733967A (zh) * 2021-03-30 2021-04-30 腾讯科技(深圳)有限公司 联邦学习的模型训练方法、装置、设备及存储介质
CN113052334A (zh) * 2021-04-14 2021-06-29 中南大学 一种联邦学习实现方法、系统、终端设备及可读存储介质
CN113128419A (zh) * 2021-04-23 2021-07-16 京东鲲鹏(江苏)科技有限公司 一种障碍物识别方法和装置、电子设备及存储介质

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116506469A (zh) * 2023-06-25 2023-07-28 华东交通大学 基于振动取能rfid传感器与联邦学习的桥梁状态监测方法
CN116506469B (zh) * 2023-06-25 2023-09-05 华东交通大学 基于振动取能rfid传感器与联邦学习的桥梁状态监测方法

Also Published As

Publication number Publication date
CN115730631A (zh) 2023-03-03

Similar Documents

Publication Publication Date Title
WO2021063171A1 (zh) 决策树模型的训练方法、系统、存储介质及预测方法
CN109919316A (zh) 获取网络表示学习向量的方法、装置和设备及存储介质
CN111667022A (zh) 用户数据处理方法、装置、计算机设备和存储介质
Gao et al. Federated region-learning for environment sensing in edge computing system
CN114580663A (zh) 面向数据非独立同分布场景的联邦学习方法及系统
Jin et al. Automated dilated spatio-temporal synchronous graph modeling for traffic prediction
CN113191479A (zh) 联合学习的方法、系统、节点及存储介质
WO2023029944A1 (zh) 联邦学习的方法和装置
Xia et al. SW-BiLSTM: a Spark-based weighted BiLSTM model for traffic flow forecasting
CN114584406B (zh) 一种联邦学习的工业大数据隐私保护系统及方法
CN117034100A (zh) 基于分层池化架构自适应图分类方法、系统、设备和介质
Li et al. Evaluation of smart city construction and optimization of city brand model under neural networks
Zhu et al. Cognitive analytics of social media services for edge resource pre-allocation in industrial manufacturing
CN113128667A (zh) 一种跨域自适应的图卷积平衡迁移学习方法与系统
CN113609337A (zh) 图神经网络的预训练方法、训练方法、装置、设备及介质
CN116993185A (zh) 时间序列预测方法、装置、设备及存储介质
WO2021000411A1 (zh) 基于神经网络的文档分类方法、装置、设备及存储介质
CN116244484A (zh) 一种面向不平衡数据的联邦跨模态检索方法及系统
Guo et al. PicPick: a generic data selection framework for mobile crowd photography
Zhai et al. Deep convolutional neural network for facial expression recognition
Zhu et al. DSCPL: A deep cloud manufacturing service clustering method using pseudo-labels
Hu et al. Communication efficient federated learning with heterogeneous structured client models
Tan et al. Fuzzy retrieval algorithm for film and television animation resource database based on deep neural network
Tan et al. Deep learning video action recognition method based on key frame algorithm
Li et al. HFML: heterogeneous hierarchical federated mutual learning on non-IID data

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22863084

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE