WO2022199480A1 - 实现隐私保护的多方协同更新模型的方法、装置及系统 - Google Patents
实现隐私保护的多方协同更新模型的方法、装置及系统 Download PDFInfo
- Publication number
- WO2022199480A1 WO2022199480A1 PCT/CN2022/081672 CN2022081672W WO2022199480A1 WO 2022199480 A1 WO2022199480 A1 WO 2022199480A1 CN 2022081672 W CN2022081672 W CN 2022081672W WO 2022199480 A1 WO2022199480 A1 WO 2022199480A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- model
- round
- server
- participant
- public
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 58
- 230000002776 aggregation Effects 0.000 claims abstract description 16
- 238000004220 aggregation Methods 0.000 claims abstract description 16
- 230000004931 aggregating effect Effects 0.000 claims description 9
- 238000004590 computer program Methods 0.000 claims description 6
- 238000004364 calculation method Methods 0.000 claims description 2
- 239000000523 sample Substances 0.000 description 103
- 238000010586 diagram Methods 0.000 description 11
- 238000004891 communication Methods 0.000 description 9
- 230000003993 interaction Effects 0.000 description 8
- 230000008569 process Effects 0.000 description 8
- 238000012549 training Methods 0.000 description 8
- 230000006870 function Effects 0.000 description 6
- 238000012545 processing Methods 0.000 description 3
- 238000013528 artificial neural network Methods 0.000 description 2
- 238000013527 convolutional neural network Methods 0.000 description 2
- 230000007246 mechanism Effects 0.000 description 2
- 238000012805 post-processing Methods 0.000 description 2
- 238000010187 selection method Methods 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000004422 calculation algorithm Methods 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 230000014509 gene expression Effects 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000000750 progressive effect Effects 0.000 description 1
- 238000012546 transfer Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F21/00—Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
- G06F21/60—Protecting data
- G06F21/62—Protecting access to data via a platform, e.g. using keys or access control rules
- G06F21/6209—Protecting access to data via a platform, e.g. using keys or access control rules to a single file or object, e.g. in a secure envelope, encrypted and accessed using a key, or with access control rules appended to the object itself
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F21/00—Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
- G06F21/60—Protecting data
- G06F21/62—Protecting access to data via a platform, e.g. using keys or access control rules
- G06F21/6218—Protecting access to data via a platform, e.g. using keys or access control rules to a system of files or objects, e.g. local or distributed file system or database
- G06F21/6245—Protecting personal data, e.g. for financial or medical purposes
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N20/00—Machine learning
Definitions
- One or more embodiments of this specification relate to the field of computer technology, and in particular, to a method, apparatus, and system for implementing a multi-party collaborative update model for privacy protection.
- federated learning also known as federated learning
- federated learning has revolutionized traditional centralized machine learning, allowing participants to collaboratively build more accurate models without uploading local data.
- federated learning is often realized by sharing model parameters or gradients among the participants.
- model parameters or gradients are usually high-dimensional private data
- traditional federated learning is accompanied by high communication overhead and privacy to a certain extent. leaks, etc.
- One or more embodiments of this specification describe a method, device and system for implementing a multi-party collaborative updating model for privacy protection, which can effectively reduce the consumption of communication resources caused by multi-party collaborative modeling, and at the same time play a role in privacy protection.
- a first aspect provides a method for implementing a multi-party collaborative update model for privacy protection, including: the server delivering the aggregated result of the t-th round of public samples to each participant i; wherein the aggregated result is all
- the server aggregates the n first prediction results output by the n participants based on their respective local models for the t-th round of public samples; each participant i is based on the t-th round of public samples and the Aggregate the results, and perform a first update on its local i-th model; each participant i performs a second update on the i-th model after the first update based on the fixed first private sample and its sample label in the local sample set; each Each participant i will use the t+1th round of common samples for the next round of iterations, input the second updated i-th model, and send the second output prediction result to the server; the server aggregates the The n second prediction results sent by n participants are used for the next round of iterations; after the multiple rounds of iterations, each
- a method for implementing a multi-party collaborative update model for privacy protection including: receiving an aggregated result of the t-th round of public samples sent by the server; wherein the aggregated result is the result of the server on all
- the n participants are obtained by aggregating the n first prediction results output by the t-th round of public samples based on their respective local models; according to the t-th round of public samples and the aggregation results, their local i-th
- the model is first updated; based on the fixed first private sample and its sample label in the local sample set, the second update is performed on the ith model after the first update; the t+1th round of public samples will be used for the next round of iterations , input the second updated i-th model, and send the output second prediction result to the server, so that the server can aggregate the second prediction result and other prediction results sent by other participants for use in The next round of iteration; after the multiple rounds of iterations, the i-th model after the second update is used as
- a system for implementing a multi-party collaborative update model for privacy protection including: the server, configured to deliver the aggregated result of the t-th round of public samples to each participant i; wherein, the aggregated result , which is obtained by the server aggregating the n first prediction results output by the n participants based on their respective local models for the t-th round of public samples; The round of public samples and the aggregated results, the first update is performed on its local i-th model; each participant i is also used to update the first update based on the fixed first private sample and its sample label in the local sample set.
- the i-th model is updated for the second time; each participant i is also used to input the second-updated i-th model for the t+1th round of common samples used in the next round of iterations, and output the second
- the prediction result is sent to the server; the server is used for aggregating the n second prediction results sent by the n participants for use in the next round of iterations; each participant i is also used in the After multiple rounds of iterations, the i-th model after the second update is used as the model for collaborative updating with other participants.
- an apparatus for implementing a multi-party collaborative update model for privacy protection comprising: a receiving unit configured to receive the aggregation result of the t-th round of public samples sent by the server; wherein the aggregation result is The server obtains by aggregating the n first prediction results output by the n participants based on their respective local models for the t-th round of public samples; the updating unit is configured to obtain according to the t-th round of public samples and all the first prediction results; The aggregation result is to perform the first update on its local i-th model; the updating unit is also used to perform the first update on the i-th model after the first update based on the fixed first private sample and its sample label in the local sample set Second update; the input unit is used to input the t+1th round of public samples used for the next round of iterations into the second updated ith model, and send the second output prediction result to the server for The server aggregates the second prediction result and other prediction results sent by other participants for the next round
- a computer storage medium is provided on which a computer program is stored, and when the computer program is executed in a computer, the computer is made to execute the method of the first aspect or the second aspect.
- a computing device including a memory and a processor, the memory stores executable code, and when the processor executes the executable code, the method of the first aspect or the second aspect is implemented.
- One or more embodiments of this specification provide a method, device, and system for a multi-party collaborative update model for privacy protection. Only the prediction results are transmitted between each participant and the server, because the dimensions of the prediction results are usually far lower than the model parameters or gradient, so this scheme can reduce the consumption of communication resources. Furthermore, only the prediction results are transmitted between the participants and the server, so that the participants can cooperate to build models with different network structures, thus greatly improving the general adaptability of federated learning. Finally, when each participant performs the second update to their local model, they only use the fixed private samples in their local sample set, which can reduce the cost of privacy protection.
- Figure 1 is a schematic diagram of federated learning based on centralized differential privacy
- Figure 2 is a schematic diagram of federated learning based on local differential privacy
- Figure 3 is a schematic diagram of federated learning based on shared prediction results
- FIG. 4 is a schematic diagram of an implementation scenario of an embodiment provided in this specification.
- FIG. 5 is an interaction diagram of a method for implementing a multi-party collaborative update model for privacy protection provided by an embodiment of this specification
- FIG. 6 is a schematic diagram of a system for implementing a multi-party collaborative update model for privacy protection provided by an embodiment of this specification
- FIG. 7 is a schematic diagram of an apparatus for implementing a multi-party collaborative update model for privacy protection according to an embodiment of the present specification.
- CDP Central Differential Privacy
- LDP Local Differential Privacy
- FIG. 1 is a schematic diagram of federated learning based on centralized differential privacy.
- each participating direction uploads its own model gradients ⁇ w 1 , ⁇ w 2 , . . . , ⁇ w n to a trusted third-party server (hereinafter referred to as a server).
- the server aggregates the model gradients uploaded by each participant:
- FIG. 2 is a schematic diagram of federated learning based on local differential privacy.
- each participant before uploading, each participant first performs local differential privacy through the differential privacy mechanism M on their respective model gradients, and then uses the model gradients (M( ⁇ w 1 ), M( ⁇ w 2 ), M( ⁇ w 2 ), ..., M( ⁇ w n )) are uploaded to the server.
- the server aggregates the locally differentially private model gradients of each participant:
- the aggregated model gradient w' is sent to each participant for each participant to update their local model based on it.
- the model built with this method has poor performance.
- each participant first pre-trains their local models based on their own local sample sets, then uses the pre-trained model to predict the common sample x p , and uploads their prediction results for the public sample x p to the server, respectively.
- Y p 1 , Y p 2 , ..., Y p n The server aggregates the prediction results uploaded by each participant:
- Y p aggregate(Y p 1 +Y p 2 +...+Y p n ),
- the aggregated result Yp is issued to each participant, so that each participant can update their own pre-training model based on it. It should be noted that although this method can solve the problem that traditional federated learning can only synergistically build models of the same structure, and can reduce the consumption of communication resources (the dimension of the prediction result is much smaller than the model parameters or gradients). However, since the prediction results of each participant are also output by the model, sensitive information of the local sample set will also be leaked to a certain extent. For example, an attacker can conduct member inference attacks based on the prediction results.
- the prediction result also leaks the sensitive information of the local sample set: Assuming that there are two parties, one of them (hereinafter referred to as the attacker) obtains the aggregated result Yp , according to the aggregated result and its own Prediction results, the prediction results of the other party can be calculated. After that, based on the calculated prediction results and public samples, the attacker can further infer the parameter information of the pre-trained model of another participant. Since the pre-training model of each participant is trained based on its local sample set, that is to say, the pre-training model of each participant incorporates its privatization information, so when an attacker infers the parameter information of the pre-training model, to a certain extent Sensitive information about the participant's local sample set is also leaked.
- one method is to add random noise that satisfies differential privacy to the prediction results of each participant.
- this alleviates the privacy problem, it also introduces another new problem: the guarantee of model performance usually requires a very large privacy budget to trade off.
- the present application proposes a multi-party collaborative model building method for privacy protection.
- Each participant performs two local model updates, one of which is based on the public sample and the aggregated result of each participant's prediction result. In this way, the consensus of all participants on the public sample can be achieved.
- the other time is based on the fixed sampled private samples (ie, the first private samples) and their sample labels in the local sample set, so that individualized training of the respective local models of each participant can be realized.
- the solutions provided by the embodiments of this specification can solve the privacy protection problem in the process of collaboratively constructing different structural models by the participants, and can balance the communication efficiency and model performance at the same time.
- FIG. 4 is a schematic diagram of an implementation scenario of an embodiment provided in this specification.
- the scenario of multi-party collaborative updating of the model involves a server and n participants, where n is a positive integer.
- each participant can be implemented as any device, platform, server or device cluster with computing and processing capabilities.
- the various parties may be institutions with sample sets of different sizes.
- the model here may be a business forecasting model for performing forecasting tasks for business objects.
- the business objects therein may be, for example, pictures, audios, or texts.
- any participant i maintains the i-th model wi locally and owns the i-th local sample set D i , and the private sample X i in the local sample set D i has the sample label Yi .
- the server may maintain the public sample set D p , or it may only maintain the description information (including the index and other information of the samples) of each public sample X p in the public sample set D p , while the real public sample set D p is provided by a third party supply.
- the server may deliver the aggregated result Y p [t] of the t-th round of public samples X p [t] to each participant i.
- the aggregation result Y p [t] is obtained by the server aggregating n first prediction results output by the n participants based on their respective local models for the t-th round of public samples X p [t].
- the n first prediction results here can be represented as Y p 1 [t], Y p 2 [t], . . . , Y p n [t], respectively.
- Each participant i performs the first update to the local i-th model according to the t-th round of public samples X p [t] and the aggregated result Y p [t]. After that, each participant i performs a second update on the i-th model after the first update based on the fixed first private sample X 1 and its sample label Y 1 in the local sample set. Finally, each participant i will use the t+1th round common sample Xp [t+1] for the next iteration, input the second updated ith model, and send the output second prediction result to server, for the server to aggregate n second prediction results corresponding to n participants (i.e.
- each participant i can use its second updated i-th model as the model updated in collaboration with other participants.
- the following takes the implementation scenario shown in FIG. 4 as an example to describe the method for implementing the multi-party collaborative update model for privacy protection provided in this specification.
- FIG. 5 is an interaction diagram of a method for implementing a multi-party collaborative update model for privacy protection provided by an embodiment of the present specification. It should be noted that this method involves multiple rounds of iterations, and FIG. 5 shows the interaction steps included in the t-th (t is a positive integer) round of iteration, and because the interaction process between each participant participating in the t-th round of iteration and the server is similar, Therefore, Fig. 5 mainly shows the interaction steps between any participant participating in the t-th round of iteration (for ease of description, referred to as the first participant) and the server, and the interaction steps between other participants participating in this round of iteration and the server, Please refer to the interaction steps between the first participant and the server.
- Step 502 the server delivers the aggregated result of the t-th round of public samples to each participant i.
- the aggregation result here may be obtained by the server aggregating the n first prediction results output by the n participants based on their respective local models for the t-th round of common samples.
- the above i is a positive integer not exceeding n.
- the t-th round of common samples here can be one of the following: pictures, texts, and audios.
- any one of the above n first prediction results may be the identification result of the picture, for example, the picture corresponds to the score of each predetermined category.
- the above-mentioned t-th round of common samples may be pre-selected by the server before starting multiple rounds of iterations.
- the above-mentioned t-th round of common samples may be selected by the server in advance in the t-1 round of iterations. The following describes the method for the server to select the t-th round of public samples in two cases.
- a public sample set is maintained in the server.
- the server can randomly select several public samples from the public sample set it maintains as the t-th round of public samples, and send them to each Participant i; or, the server may randomly select several public samples from the public sample set maintained by it as the t-th round of public samples, and send an indication of the t-th round of public samples to each participant i. After that, each participant i obtains the t-th round of public samples from the server according to the received instructions.
- the third party maintains a public sample set
- the server maintains description information of each public sample in the public sample set.
- the description information here includes, but is not limited to, information such as the index of each public sample.
- the server may randomly select the description information of several public samples from the description information of each public sample maintained by the server as the description information of the t-th round of public samples, and deliver it to each participant i. After that, each participant i obtains the t-th round of public samples from the third party according to the received description information.
- the server selects the t-th round of public samples, and each participant i obtains the t-th round of public samples, it can output the corresponding first prediction result for the t-th round of public samples based on the local i-th model .
- the network structures of the local models of each participant may be the same or different.
- the difference here can be divided into the following two situations: First, the types of models are different.
- the model maintained by one party may be an Artificial Neural Network (ANN), and the model maintained by the other party may be eXtreme Gradient Boosting, XGBoost).
- the types of models are the same, but the specific expressions are inconsistent.
- the models maintained by each participant are Convolutional Neural Networks (CNN), but the number of network layers is inconsistent.
- CNN Convolutional Neural Networks
- each participant i can obtain the initial model of the corresponding network structure (for different network structures situation); or, the initial model uniformly distributed by the server (for the same network structure) can be received. Then, based on the randomly selected sample subset (hereinafter referred to as the first private sample) and its sample label from the local sample set, the corresponding initial model is updated to obtain its local i-th model.
- the first private sample the randomly selected sample subset
- the corresponding initial model is updated to obtain its local i-th model.
- the local ith model of each participant i may refer to the ith model after the second update in the t-1th iteration.
- the two update processes of the i-th model after the second update will be described later.
- each participant i updates the corresponding initial model based on the first private sample and its sample label randomly selected from the local sample set, so as to avoid the leakage of sensitive information of each participant.
- the reason is that each participant i uses the first private sample randomly selected from its local sample set to update the corresponding initial model, which is equivalent to adding random noise while personalizing the training model, so that the obtained i-th model With differential privacy function.
- the subsequent data output based on the i-th model meets the requirements of differential privacy.
- differential privacy protection of the private data of each participant can be ensured.
- Step 504 each participant i performs a first update of its local i-th model according to the t-th round of public samples and the aggregated results.
- each participant i can input the t-th round of public samples into its local i-th model to obtain the first local prediction result.
- the aggregated result is used as the sample label of the t-th round of common samples, and the first prediction loss is determined based on the sample label and the first local prediction result.
- the first prediction loss may be determined based on the difference between the sample label and the local prediction result.
- a first update is made to its local i-th model.
- the back-propagation method can be used to first calculate the update gradient corresponding to the model parameters of the ith model based on the first prediction loss, and then perform the first update on the ith model based on the update gradient. More specifically, the model parameters of the ith model are subtracted from the product of the corresponding update gradient and the learning step size (which is a hyperparameter) to obtain the updated model parameters of the ith model, and it is determined as the first The updated i-th model.
- the learning step size which is a hyperparameter
- the first update of the respective local models by each participant based on the public sample can realize the consensus of each participant on the public sample.
- Step 506 each participant i performs a second update on the i-th model after the first update based on the fixed first private sample and its sample label in the local sample set.
- the private samples in the local sample set of each participant i may be one of the following: pictures, texts, and audios. It should be understood that when the i-th round of public samples and the first private samples are both pictures, the second updated i-th model here may be a picture recognition model. When the i-th round of public samples and the first private samples are both texts, the i-th model after the second update here may be a text recognition model. And, when the i-th round of public samples and the first private samples are both audio, the second updated i-th model here may be an audio recognition model.
- the second update in the above step 506 may specifically include: each participant i may input the fixed first private sample in the local sample set into the first updated i-th model to obtain a second local prediction result. Afterwards, a second prediction loss is determined based on the sample label of the first private sample and the second local prediction result. For example, the second prediction loss may be determined based on the difference between the sample label of the first private sample and the second local prediction result. Finally, a second update is performed on the i-th model after the first update according to the second prediction loss. For the step of performing the second update based on the second prediction loss here, reference may be made to the above-mentioned first update, which will not be repeated in this application.
- the first private sample used in step 506 is the first private sample randomly selected locally and used by each participant i when updating the corresponding initial model. That is, after multiple rounds of iterations start, any participant i performs a second update based on the fixed first private sample in each round of iterations.
- any participant i performs the second update based on the fixed first private sample in each round of iteration, which can reduce the cost of privacy protection.
- This step uses a randomly selected private sample, then it is equivalent to adding new random noise in each iteration after the start of multiple iterations, which makes as the number of iterations increases, the added The random noise is getting bigger and bigger.
- the added random noise usually needs to be balanced by the privacy budget. Specifically, the larger the random noise added, the larger the privacy budget required to ensure the availability of data, which will greatly consume the cost of privacy protection.
- any participant i uses a fixed first private sample in each round of iterations, so as to ensure that no new random noise will be added after the start of multiple rounds of iterations, thereby ensuring privacy
- the budget can be fixed. It should be understood that due to the post-processing feature of differential privacy, even in each subsequent iteration, no new random noise will be added. It also meets the requirements of differential privacy.
- each participant performs a second update on their respective local models based on their respective fixed first private samples, so that individualized training of the respective local models of each participant can be realized.
- Step 508 each participant i inputs the second updated i-th model for the t+1th round of common samples used in the next round of iteration, and sends the second output prediction result to the server.
- the t+1th round of common samples here may be selected by the server in the t round of iterations before step 508 is executed.
- the specific selection method is similar to the above-mentioned selection method of the t-th round of public samples, and will not be repeated here.
- the second prediction result here is similar to the above-mentioned first prediction result, for example, it may be an image recognition result, a text recognition result, an audio recognition result, or the like.
- the local i-th model of each participant i has a differential privacy function, so when this step outputs the second prediction result based on it, the outputted second prediction result satisfies the requirement of differential privacy. That is to say, when each participant i sends the corresponding second prediction result to the server, the sensitive information of each participant will not be leaked.
- Step 510 the server aggregates n copies of the second prediction results sent by the n participants for the next round of iteration.
- the server may perform a summation, a weighted summation, or a weighted average calculation on the n second prediction results, so as to obtain an aggregated result of the t+1th round of common samples. Afterwards, after entering the t+1th round of iteration, the server may deliver the aggregated result to each participant, and each participant then performs the first and second updates on their local models again, and so on.
- the above-mentioned steps 502 to 510 are repeatedly performed for multiple times, so that multiple rounds of iterative updating of the respective local models of each participant can be implemented.
- the model parameters used in each iteration are the parameters updated in the previous round.
- the termination condition of the iteration may be that the number of iterations reaches a predetermined round or the model parameters converge.
- each participant i uses its second updated i-th model as the model updated in collaboration with other participants.
- the model updated collaboratively with other participants may be a picture recognition model.
- the model that is updated collaboratively with other participants may be an audio recognition model.
- the model updated in collaboration with other participants may be a text recognition model or the like.
- this solution can reduce the consumption of communication resources.
- only the prediction results are transmitted between the participants and the server, so that the participants can cooperate to build models with different network structures, thus greatly improving the general adaptability of federated learning.
- each participant updates their respective initial models based on the private samples randomly selected from the local sample set to obtain their own local models, which can ensure the differential privacy protection of each participant's data.
- each participant performs the second update to their respective local models, only the fixed private samples in their respective local sample sets are used, thereby reducing the cost of privacy protection.
- an embodiment of this specification also provides a system for implementing a multi-party collaborative update model for privacy protection.
- the system may include servers 602 and n Participants 604 .
- the server 602 is configured to deliver the aggregated result of the t-th round of public samples to each participant i.
- the aggregation result is obtained by the server aggregating the n first prediction results output by the n participants based on their respective local models for the t-th round of common samples.
- Each participant 604 is configured to perform a first update of its local i-th model according to the t-th round of public samples and the aggregated results.
- the network structures of the respective local models of the n participants are different.
- Each participant 604 is further configured to perform a second update on the i-th model after the first update based on the fixed first private sample and its sample label in the local sample set.
- Each participant 604 is further configured to input the t+1th round of common samples used for the next round of iterations into the second updated ith model, and send the second output prediction result to the server 602 .
- the server 602 is configured to aggregate the n second prediction results sent by the n participants for use in the next round of iteration.
- the server 602 is specifically configured to: perform a summation, a weighted summation, or a weighted average of the n second prediction results.
- Each participant 604 is further configured to use its second updated i-th model as the model to be updated collaboratively with other participants after multiple rounds of iterations.
- the samples in the local sample set of any participant i are pictures, and the model updated in collaboration with other participants is the image recognition model; or, the samples in the local sample set of any participant i are audio, which cooperates with other participants
- the updated model is an audio recognition model; or, the samples in the local sample set of any participant i are text, and the model updated in collaboration with other participants is a text recognition model.
- the server 602 maintains a public sample set; the server 602 is further configured to randomly select several public samples from the public sample set maintained by the server 602 as the t+1th round of public samples, and send them to each participant 604 , or; the server 602 is further configured to randomly select a number of public samples from the public sample set maintained by it as the t+1th round of public samples, and send the indication of the t+1th round of public samples to each participant 604; The participating parties 604 are further configured to obtain the t+1th round of public samples from the server 602 according to the received instruction.
- the third party maintains a public sample set
- the server 602 maintains description information of each public sample in the public sample set
- the server 602 is also used to randomly select the description information of several public samples as the t+1th round of public samples.
- the description information is sent to each participant 604; each participant 604 is also used to obtain the t+1th round of public samples from a third party according to the received description information.
- An embodiment of this specification provides a system for implementing a multi-party collaborative update model for privacy protection, which can solve the privacy protection problem of each participant in the process of collaboratively building different structural models, and can balance communication efficiency and model performance.
- an embodiment of the present specification further provides an apparatus for implementing a multi-party collaborative update model for privacy protection.
- the multi-party here includes the server and n participating parties.
- the device is set at any participant i among the n participants, and is used to perform multiple rounds of iterations.
- the apparatus executes any t-th round of iterations through the following units: a receiving unit 702, configured to receive the aggregated results of the t-th round of public samples delivered by the server.
- the aggregation result is obtained by the server aggregating the n first prediction results output by the n participants based on their respective local models for the t-th round of common samples.
- the updating unit 704 is configured to perform a first update of the local i-th model according to the t-th round of public samples and the aggregation result.
- the updating unit 704 is specifically used to: input the t-th round of public samples into its local i-th model to obtain a local prediction result; take the aggregated result as the sample label of the t-th round of public samples, and determine the prediction loss based on it and the local prediction result ; According to the prediction loss, make the first update of its local i-th model.
- the updating unit 704 is further configured to perform a second update on the i-th model after the first update based on the fixed first private sample and its sample label in the local sample set.
- the input unit 706 is used for inputting the t+1th round public samples used for the next round of iterations into the second updated ith model, and sending the output second prediction result to the server for the server to aggregate the second Predictions and other predictions sent by other parties for use in the next iteration.
- the determining unit 708 is configured to, after multiple rounds of iterations, use the second updated i-th model as the model to be updated collaboratively with other participants.
- the receiving unit 702 is further configured to receive the initial model delivered by the server; the updating unit 704 is further configured to update the initial model based on the fixed first private sample and its sample label in the local sample set to obtain the i-th model.
- An apparatus for implementing a multi-party collaborative update model for privacy protection provided by an embodiment of this specification can solve the privacy protection problem of each participant in the process of collaboratively building different structural models, and can balance communication efficiency and model performance.
- a computer-readable storage medium having a computer program stored thereon, which, when executed in a computer, causes the computer to perform the method described in conjunction with FIG. 5 .
- a computing device including a memory and a processor, where executable codes are stored in the memory, and when the processor executes the executable codes, the processing described in conjunction with FIG. 5 is implemented. method.
- the steps of the method or algorithm described in conjunction with the disclosure of this specification may be implemented in a hardware manner, or may be implemented in a manner in which a processor executes software instructions.
- the software instructions can be composed of corresponding software modules, and the software modules can be stored in RAM memory, flash memory, ROM memory, EPROM memory, EEPROM memory, registers, hard disk, removable hard disk, CD-ROM or any other form of storage well known in the art in the medium.
- An exemplary storage medium is coupled to the processor, such that the processor can read information from, and write information to, the storage medium.
- the storage medium can also be an integral part of the processor.
- the processor and storage medium may reside in an ASIC.
- the ASIC may be located in a server.
- the processor and storage medium may also exist in the server as discrete components.
- the functions described in the present invention may be implemented in hardware, software, firmware, or any combination thereof.
- the functions may be stored on or transmitted over as one or more instructions or code on a computer-readable medium.
- Computer-readable media includes both computer storage media and communication media including any medium that facilitates transfer of a computer program from one place to another.
- a storage medium can be any available medium that can be accessed by a general purpose or special purpose computer.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Bioethics (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Software Systems (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Computer Hardware Design (AREA)
- Computer Security & Cryptography (AREA)
- Medical Informatics (AREA)
- Databases & Information Systems (AREA)
- Artificial Intelligence (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Computing Systems (AREA)
- Mathematical Physics (AREA)
- Information Transfer Between Computers (AREA)
- Telephonic Communication Services (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
Description
Claims (20)
- 一种实现隐私保护的多方协同更新模型的方法,所述多方包括服务器和n个参与方;所述方法包括多轮迭代,其中任意的第t轮迭代包括:所述服务器向每个参与方i下发第t轮公共样本的聚合结果;其中,所述聚合结果,是所述服务器对所述n个参与方基于各自本地的模型针对所述第t轮公共样本输出的n份第一预测结果进行聚合得到;每个参与方i根据所述第t轮公共样本和所述聚合结果,对其本地的第i模型进行第一更新;每个参与方i基于本地样本集中固定的第一私有样本及其样本标签,对第一更新后的第i模型进行第二更新;每个参与方i将用于下一轮迭代的第t+1轮公共样本,输入第二更新后的第i模型,并将输出的第二预测结果发送给所述服务器;所述服务器聚合所述n个参与方发送的n份第二预测结果,以用于下一轮迭代;在所述多轮迭代后,每个参与方i将其第二更新后的第i模型,作为其与其它参与方协同更新的模型。
- 根据权利要求1所述的方法,所述服务器中维护有公共样本集;在所述每个参与方i将用于下一轮迭代的第t+1轮公共样本,输入第二更新后的第i模型之前,所述方法还包括:所述服务器从其维护的公共样本集中随机选取若干公共样本作为第t+1轮公共样本,并将其发送给每个参与方i,或者;所述服务器从其维护的公共样本集中随机选取若干公共样本作为第t+1轮公共样本,并将所述第t+1轮公共样本的指示发送给每个参与方i;每个参与方i根据所述指示,从所述服务器获取所述第t+1轮公共样本。
- 根据权利要求1所述的方法,第三方中维护有公共样本集,所述服务器中维护有所述公共样本集中各公共样本的描述信息;在所述每个参与方i将用于下一轮迭代的第t+1轮公共样本,输入第二更新后的第i模型之前,所述方法还包括:所述服务器随机选取若干公共样本的描述信息作为所述第t+1轮公共样本的描述信息,并将其下发至每个参与方i;每个参与方i根据所述描述信息,从第三方获取所述第t+1轮公共样本。
- 根据权利要求1所述的方法,任意的参与方i本地样本集中的样本为图片,其与其它参与方协同更新的模型为图片识别模型;或者,任意的参与方i本地样本集中的样本为音频,其与其它参与方协同更新的模型为音频识别模型;或者,任意的参与方i本地样本集中的样本为文本,其与其它参与方协同更新的模型为文本识别模型。
- 根据权利要求1所述的方法,所述n个参与方各自本地的模型的网络结构不同。
- 根据权利要求1所述的方法,所述服务器聚合所述n个参与方发送的n份第二预测结果,包括:所述服务器对所述n份第二预测结果进行求和、加权求和或者求加权平均。
- 一种实现隐私保护的多方协同更新模型的方法,所述多方包括服务器和n个参与方;所述方法通过所述n个参与方中任意的参与方i执行;所述方法包括多轮迭代,其中任意的第t轮迭代包括:接收所述服务器下发的第t轮公共样本的聚合结果;其中,所述聚合结果,是所述服务器对所述n个参与方基于各自本地的模型针对所述第t轮公共样本输出的n份第一 预测结果进行聚合得到;根据所述第t轮公共样本和所述聚合结果,对其本地的第i模型进行第一更新;基于本地样本集中固定的第一私有样本及其样本标签,对第一更新后的第i模型进行第二更新;将用于下一轮迭代的第t+1轮公共样本,输入第二更新后的第i模型,并将输出的第二预测结果发送给所述服务器,以供所述服务器聚合所述第二预测结果以及其它参与方发送的其它预测结果,以用于下一轮迭代;在所述多轮迭代后,将其第二更新后的第i模型,作为其与其它参与方协同更新的模型。
- 根据权利要求7所述的方法,所述第i模型通过以下步骤获得:接收所述服务器下发的初始模型;基于本地样本集中固定的第一私有样本及其样本标签,更新所述初始模型,得到所述第i模型。
- 根据权利要求7所述的方法,所述根据所述第t轮公共样本和所述聚合结果,对其本地的第i模型进行第一更新,包括:将所述第t轮公共样本输入其本地的第i模型,得到本地预测结果;将所述聚合结果作为所述第t轮公共样本的样本标签,并基于其以及所述本地预测结果,确定预测损失;根据所述预测损失,对其本地的第i模型进行第一更新。
- 一种实现隐私保护的多方协同更新模型的系统,包括服务器和n个参与方;所述服务器,用于向每个参与方i下发第t轮公共样本的聚合结果;其中,所述聚合结果,是所述服务器对所述n个参与方基于各自本地的模型针对所述第t轮公共样本输出的n份第一预测结果进行聚合得到;每个参与方i,用于根据所述第t轮公共样本和所述聚合结果,对其本地的第i模型进行第一更新;每个参与方i,还用于基于本地样本集中固定的第一私有样本及其样本标签,对第一更新后的第i模型进行第二更新;每个参与方i,还用于将用于下一轮迭代的第t+1轮公共样本,输入第二更新后的第i模型,并将输出的第二预测结果发送给所述服务器;所述服务器,用于聚合所述n个参与方发送的n份第二预测结果,以用于下一轮迭代;每个参与方i,还用于在所述多轮迭代后,将其第二更新后的第i模型,作为其与其它参与方协同更新的模型。
- 根据权利要求10所述的系统,所述服务器中维护有公共样本集;所述服务器,还用于从其维护的公共样本集中随机选取若干公共样本作为第t+1轮公共样本,并将其发送给每个参与方i,或者;所述服务器,还用于从其维护的公共样本集中随机选取若干公共样本作为第t+1轮公共样本,并将所述第t+1轮公共样本的指示发送给每个参与方i;每个参与方i,还用于根据所述指示,从所述服务器获取所述第t+1轮公共样本。
- 根据权利要求10所述的系统,第三方中维护有公共样本集,所述服务器中维护有所述公共样本集中各公共样本的描述信息;所述服务器,还用于随机选取若干公共样本的描述信息作为所述第t+1轮公共样本的描述信息,并将其下发至每个参与方i;每个参与方i,还用于根据所述描述信息,从第三方获取所述第t+1轮公共样本。
- 根据权利要求10所述的系统,任意的参与方i本地样本集中的样本为图片,其与其它参与方协同更新的模型为图 片识别模型;或者,任意的参与方i本地样本集中的样本为音频,其与其它参与方协同更新的模型为音频识别模型;或者,任意的参与方i本地样本集中的样本为文本,其与其它参与方协同更新的模型为文本识别模型。
- 根据权利要求10所述的系统,所述n个参与方各自本地的模型的网络结构不同。
- 根据权利要求10所述的系统,所述服务器具体用于:对所述n份第二预测结果进行求和、加权求和或者求加权平均。
- 一种实现隐私保护的多方协同更新模型的装置,所述多方包括服务器和n个参与方;所述装置设置于所述n个参与方中任意的参与方i,用于执行多轮迭代,所述装置通过其包括的以下单元执行其中任意的第t轮迭代:接收单元,用于接收所述服务器下发的第t轮公共样本的聚合结果;其中,所述聚合结果,是所述服务器对所述n个参与方基于各自本地的模型针对所述第t轮公共样本输出的n份第一预测结果进行聚合得到;更新单元,用于根据所述第t轮公共样本和所述聚合结果,对其本地的第i模型进行第一更新;所述更新单元,还用于基于本地样本集中固定的第一私有样本及其样本标签,对第一更新后的第i模型进行第二更新;输入单元,用于将用于下一轮迭代的第t+1轮公共样本,输入第二更新后的第i模型,并将输出的第二预测结果发送给所述服务器,以供所述服务器聚合所述第二预测结果以及其它参与方发送的其它预测结果,以用于下一轮迭代;确定单元,用于在所述多轮迭代后,将其第二更新后的第i模型,作为其与其它参与方协同更新的模型。
- 根据权利要求16所述的装置,所述接收单元,还用于接收所述服务器下发的初始模型;所述更新单元,还用于基于本地样本集中固定的第一私有样本及其样本标签,更新所述初始模型,得到所述第i模型。
- 根据权利要求16所述的装置,所述更新单元具体用于:将所述第t轮公共样本输入其本地的第i模型,得到本地预测结果;将所述聚合结果作为所述第t轮公共样本的样本标签,并基于其以及所述本地预测结果,确定预测损失;根据所述预测损失,对其本地的第i模型进行第一更新。
- 一种计算机可读存储介质,其上存储有计算机程序,其中,当所述计算机程序在计算机中执行时,令计算机执行权利要求1至9中任一项所述的方法。
- 一种计算设备,包括存储器和处理器,其中,所述存储器中存储有可执行代码,所述处理器执行所述可执行代码时,实现权利要求1至9中任一项所述的方法。
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US18/283,343 US20240176906A1 (en) | 2021-03-25 | 2022-03-18 | Methods, apparatuses, and systems for collaboratively updating model by multiple parties for implementing privacy protection |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110320900.4A CN112948885B (zh) | 2021-03-25 | 2021-03-25 | 实现隐私保护的多方协同更新模型的方法、装置及系统 |
CN202110320900.4 | 2021-03-25 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2022199480A1 true WO2022199480A1 (zh) | 2022-09-29 |
Family
ID=76227771
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2022/081672 WO2022199480A1 (zh) | 2021-03-25 | 2022-03-18 | 实现隐私保护的多方协同更新模型的方法、装置及系统 |
Country Status (3)
Country | Link |
---|---|
US (1) | US20240176906A1 (zh) |
CN (1) | CN112948885B (zh) |
WO (1) | WO2022199480A1 (zh) |
Families Citing this family (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112948885B (zh) * | 2021-03-25 | 2023-10-27 | 支付宝(杭州)信息技术有限公司 | 实现隐私保护的多方协同更新模型的方法、装置及系统 |
CN114239064A (zh) * | 2021-12-20 | 2022-03-25 | 杭州趣链科技有限公司 | 隐私计算方法、装置、电子设备及计算机可读存储介质 |
CN116415676A (zh) * | 2021-12-29 | 2023-07-11 | 新智我来网络科技有限公司 | 联合学习中的预测方法及装置 |
CN115081642B (zh) * | 2022-07-19 | 2022-11-15 | 浙江大学 | 一种多方协同更新业务预测模型的方法及系统 |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110647765A (zh) * | 2019-09-19 | 2020-01-03 | 济南大学 | 协同学习框架下基于知识迁移的隐私保护方法及系统 |
CN111062487A (zh) * | 2019-11-28 | 2020-04-24 | 支付宝(杭州)信息技术有限公司 | 基于数据隐私保护的机器学习模型特征筛选方法及装置 |
CN112329940A (zh) * | 2020-11-02 | 2021-02-05 | 北京邮电大学 | 一种结合联邦学习与用户画像的个性化模型训练方法及系统 |
CN112948885A (zh) * | 2021-03-25 | 2021-06-11 | 支付宝(杭州)信息技术有限公司 | 实现隐私保护的多方协同更新模型的方法、装置及系统 |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20200272945A1 (en) * | 2019-02-21 | 2020-08-27 | Hewlett Packard Enterprise Development Lp | System and method of decentralized model building for machine learning and data privacy preserving using blockchain |
-
2021
- 2021-03-25 CN CN202110320900.4A patent/CN112948885B/zh active Active
-
2022
- 2022-03-18 WO PCT/CN2022/081672 patent/WO2022199480A1/zh active Application Filing
- 2022-03-18 US US18/283,343 patent/US20240176906A1/en active Pending
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110647765A (zh) * | 2019-09-19 | 2020-01-03 | 济南大学 | 协同学习框架下基于知识迁移的隐私保护方法及系统 |
CN111062487A (zh) * | 2019-11-28 | 2020-04-24 | 支付宝(杭州)信息技术有限公司 | 基于数据隐私保护的机器学习模型特征筛选方法及装置 |
CN112329940A (zh) * | 2020-11-02 | 2021-02-05 | 北京邮电大学 | 一种结合联邦学习与用户画像的个性化模型训练方法及系统 |
CN112948885A (zh) * | 2021-03-25 | 2021-06-11 | 支付宝(杭州)信息技术有限公司 | 实现隐私保护的多方协同更新模型的方法、装置及系统 |
Also Published As
Publication number | Publication date |
---|---|
US20240176906A1 (en) | 2024-05-30 |
CN112948885A (zh) | 2021-06-11 |
CN112948885B (zh) | 2023-10-27 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2022199480A1 (zh) | 实现隐私保护的多方协同更新模型的方法、装置及系统 | |
CN113191484B (zh) | 基于深度强化学习的联邦学习客户端智能选取方法及系统 | |
CN113408746B (zh) | 一种基于区块链的分布式联邦学习方法、装置及终端设备 | |
WO2022257730A1 (zh) | 实现隐私保护的多方协同更新模型的方法、装置及系统 | |
US11520899B2 (en) | System and method for machine learning architecture with adversarial attack defense | |
Gao et al. | FGFL: A blockchain-based fair incentive governor for Federated Learning | |
WO2019228232A1 (zh) | 对话系统间的知识分享方法、对话方法及装置 | |
CN113222175B (zh) | 信息处理方法及系统 | |
WO2023124296A1 (zh) | 基于知识蒸馏的联合学习训练方法、装置、设备及介质 | |
CN113469373A (zh) | 基于联邦学习的模型训练方法、系统、设备及存储介质 | |
CN111460528A (zh) | 一种基于Adam优化算法的多方联合训练方法及系统 | |
US11100406B2 (en) | Knowledge network platform | |
US20220101189A1 (en) | Federated inference | |
Adams | AI-powered social bots | |
WO2022222795A1 (zh) | 用于隐私保护的模型的训练方法、系统及装置 | |
WO2023174018A1 (zh) | 一种纵向联邦学习方法、装置、系统、设备及存储介质 | |
US20230006980A1 (en) | Systems and methods for securely training a decision tree | |
CN112292696A (zh) | 确定执行设备的动作选择方针 | |
CN116627970A (zh) | 一种基于区块链和联邦学习的数据共享方法及装置 | |
CN112541570A (zh) | 一种多模型训练方法、装置、电子设备及存储介质 | |
Wang et al. | Social-aware clustered federated learning with customized privacy preservation | |
Moudoud et al. | Multi-tasking federated learning meets blockchain to foster trust and security in the Metaverse | |
US20230394583A1 (en) | Customer partner program methods and systems | |
Chen et al. | Advances in Robust Federated Learning: Heterogeneity Considerations | |
Wu et al. | Bayesian stochastic configuration networks for robust data modeling |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 22774141 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 18283343 Country of ref document: US |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWE | Wipo information: entry into national phase |
Ref document number: 11202307137U Country of ref document: SG |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 22774141 Country of ref document: EP Kind code of ref document: A1 |