CN113591115A - Method for batch normalization in logistic regression model for safe federal learning - Google Patents

Method for batch normalization in logistic regression model for safe federal learning Download PDF

Info

Publication number
CN113591115A
CN113591115A CN202110890465.9A CN202110890465A CN113591115A CN 113591115 A CN113591115 A CN 113591115A CN 202110890465 A CN202110890465 A CN 202110890465A CN 113591115 A CN113591115 A CN 113591115A
Authority
CN
China
Prior art keywords
model
data
training
calculation parameters
gradient
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202110890465.9A
Other languages
Chinese (zh)
Inventor
赵培江
祝文伟
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenpu Technology Shanghai Co ltd
Original Assignee
Shenpu Technology Shanghai Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenpu Technology Shanghai Co ltd filed Critical Shenpu Technology Shanghai Co ltd
Priority to CN202110890465.9A priority Critical patent/CN113591115A/en
Publication of CN113591115A publication Critical patent/CN113591115A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F21/00Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F21/60Protecting data
    • G06F21/602Providing cryptographic facilities or services
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning

Abstract

The invention relates to the related field of federal learning, and particularly discloses a method for batch standardization in a logistic regression model for safe federal learning, which comprises the following steps: s1, preparing training data by all parties of the federal learning participants: the federally learned participants may have multiple parties P1, P2.., Pn, each having training data _1, data _ 2.., data _ n, respectively, wherein one party has a tag y, assuming that the party having the tag is P1, and a coordinator for gradient decryption; s2, fusing data sets: in the longitudinal federal study, data should be submitted, namely, a sample ID commonly owned by each participant is screened to form a fused data set fuse _ data; s3, training a logistic regression model in longitudinal federal learning; s4, predicting; on one hand, the method can avoid the model training failure caused by the fact that too large gradient cannot be converged in the logistic regression algorithm modeling of safe federal learning, and the training success rate of the model is improved.

Description

Method for batch normalization in logistic regression model for safe federal learning
Technical Field
The invention relates to the related field of federal learning, in particular to a method for batch standardization in a logistic regression model of safe federal learning.
Background
Machine learning refers to a process of using some algorithms to guide a computer to independently construct a reasonable model by using known data and to judge a new situation by using the model, and plays a very important role in various applications such as network search, online advertisement, commodity recommendation, mechanical failure prediction, insurance pricing, financial risk management and the like. Traditionally, machine learning models are trained on a centralized corpus of data, which may be collected by a single or multiple data providers. Although parallel distributed algorithms have been developed to speed up the training process, the training data itself is still collected centrally and stored in one data center.
In 5.2018, the European Union mentioned a new height to the privacy Protection requirement by the General Data Protection Regulation (GDPR) act. In addition to this, there are many legal regulations that have been published about private data. Therefore, the previous platform mechanism is challenged to share data in any way, and also brings serious privacy problems to the data collection of machine learning. Because data used for machine learning training is often sensitive, it may come from multiple owners with different privacy requirements. This serious privacy problem limits the actual amount of data.
The federal study defines that all data are kept locally, so that privacy is not disclosed and regulations are not violated; a virtual common model is established by combining data of a plurality of participants, and a system which benefits jointly can specifically achieve the purpose that the respective data cannot go out of the local area, and then a virtual common model is established in a parameter exchange mode under an encryption mechanism under the condition of not violating data privacy regulations. Federal learning is used as a modeling method for guaranteeing data safety, and has huge application prospects in industries such as sales, finance and the like. In these industries, data cannot be aggregated directly for machine learning model training due to a number of factors including intellectual property, privacy protection, data security, etc. At this point, a federated model needs to be trained via federated learning.
In a longitudinal federal learning logistic regression model, an approximation function of a sigmoid function adopted when homomorphic encrypted data is calculated cannot limit the upper limit and the lower limit of a calculation result; in the model operation, the sum of the products of the model parameters and the input features is not limited, so that when the sum is too large, the gradient of the model is always increased to become incapable of convergence.
Disclosure of Invention
The invention aims to provide a method for batch standardization in a logistic regression model for safe federal learning, so as to solve the problems in the background technology.
In order to achieve the purpose, the invention provides the following technical scheme:
a method for batch normalization in a logistic regression model for safe federal learning, comprising the steps of:
s1, preparing training data by all parties of the federal learning participants: the federally learned participants may have multiple parties P1, P2.., Pn, each having training data _1, data _ 2.., data _ n, respectively, wherein one party has a tag y, assuming that the party having the tag is P1, and a coordinator for gradient decryption;
s2, fusing data sets: in the longitudinal federal study, data should be submitted, namely, a sample ID commonly owned by each participant is screened to form a fused data set fuse _ data;
s3, training a logistic regression model in longitudinal federal learning: the specific training method comprises the following steps of,
s301, all the participants of the joint training in each training process form a data set X locally according to the sample ID and the characteristics after intersectionPn=[x1,x2,...,xm]∈RmAnd respective model parameters WPn∈Rm
S302, each participant respectively takes out a batch B [ X ] with the data volume B from the owned data]∈Rb×mAccording to the characteristic dimension m, performing matrix multiplication with the model parameter W to obtain the model calculation parameter WX belonging to the R in the first stepb×1
S303, each participant calculates the average value of the parameters according to the model calculation parameters of the first step
Figure BDA0003195805970000021
And standard deviation v2b;
S304, calculating the average value of the parameters of the model of the current batch
Figure BDA0003195805970000031
And standard deviation v2b put into a sliding window queue of length k for recording the mean and standard deviation of previous batches
Figure BDA0003195805970000032
And
Figure BDA0003195805970000033
performing the following steps;
s305, if the amount of the data stored in the sliding window queue recording the average value and the standard deviation of the previous batch is larger than a set value, dequeuing the data at the head of the queue;
s306, calculating the moving average of the mean value and the standard deviation of the model calculation parameters in the sliding window queue to obtain
Figure BDA0003195805970000034
And M (v)2) And standardizing the WX according to the following formula to obtain standardized model calculation parameters:
Figure BDA0003195805970000035
wherein, epsilon is an additional item for preventing from dividing 0, and is a settable decimal, [ alpha, beta ] is a standardized parameter, and the initial value is [1,0 ];
s307, standardizing the model calculation parameters obtained in the step S302 by using the average value and the standard deviation of the model calculation parameters obtained in the step S306 to obtain standardized model calculation parameters;
s308, calculating parameter BN (WX) for the normalized model except P1b) Performing homomorphic encryption to obtain EPn[BN(WXb)]And sends the encrypted result to P1;
s309, summing the normalized model calculation parameters in P1 to obtain:
Figure BDA0003195805970000036
and calculates gradient calculation parameters E [ h (WX) ] of the encrypted model by using local supervision information]=f(E[WX]) Gradient E [ g ] of-y and P1P1]=E[h(WX)]xi
S310, P1 gradient calculation parameters E [ h (WX)]Sending to other participants the gradient E g of P1P1]Sending the data to a coordinator;
s311, the other participants calculate the respective gradients E [ g ] using the gradient calculation parametersPn]And sending the gradient to a coordinator;
s312, the coordinator decrypts the gradient of each participant to obtain gPnAnd distributed to each participant;
s313, each participant uses the decrypted gradient gPnTo update the model parameters [ W, alpha, beta ]];
S314, when the model reaches the maximum iteration times, the model terminates the training termination condition, otherwise, the step S302 is repeated;
s315, after the model training is finished, all the participators simultaneously save the model parameter WPnAnd the mean value of the model calculation parameters in the sliding window queue
Figure BDA0003195805970000041
And standard deviation M (v)2);
S4, prediction: the specific prediction steps are as follows:
s401, the new user carries out intersection;
s402, data characteristic X and model parameter W of new userPnCarrying out matrix multiplication to obtain model calculation parameters WX, and standardizing by using the average value and standard deviation of the model calculation parameters stored in model training;
Figure BDA0003195805970000042
s403, use
Figure BDA0003195805970000043
Obtaining the normalized model calculation parameters, wherein the normalized model calculation parameters are obtained by the formula: y-pred ═ f (wx) to calculate prediction results and prediction probabilities, where
Figure BDA0003195805970000044
Compared with the prior art, the invention has the beneficial effects that: the method can avoid the model training failure caused by the fact that too large gradient cannot be converged in the logistic regression algorithm modeling of safe federal learning on one hand, and improve the training success rate of the model on the other hand, can change the distribution condition of the original data, and enables other participants to be more difficult to obtain the original distribution shape of the original data so as to improve the safety.
Drawings
FIG. 1 is a schematic overall flow chart of the present invention.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
Referring to fig. 1, the present invention provides a technical solution: a method for batch normalization in a logistic regression model for safe federal learning, comprising the steps of:
s1, preparing training data by all parties of the federal learning participants: the federally learned participants may have multiple parties P1, P2.., Pn, each having training data _1, data _ 2.., data _ n, respectively, wherein one party has a tag y, assuming that the party having the tag is P1, and a coordinator for gradient decryption;
s2, fusing data sets: in the longitudinal federal study, data should be submitted, namely, a sample ID commonly owned by each participant is screened to form a fused data set fuse _ data;
s3, training a logistic regression model in longitudinal federal learning: the specific training method comprises the following steps of,
s301, all the participants of the joint training in each training process form a data set X locally according to the sample ID and the characteristics after intersectionPn=[x1,x2,...,xm]∈RmAnd respective model parameters WPn∈Rm
S302, each participant respectively takes out a batch B [ X ] with the data volume B from the owned data]∈Rb×mAccording to the characteristic dimension m, performing matrix multiplication with the model parameter W to obtain the model calculation parameter WX belonging to the R in the first stepb×1
S303, each participant calculates the average value of the parameters according to the model calculation parameters of the first step
Figure BDA0003195805970000051
And standard deviation v2b;
S304, calculating the average value of the parameters of the model of the current batch
Figure BDA0003195805970000061
And standard deviation v2b put into a sliding window queue of length k for recording the mean and standard deviation of previous batches
Figure BDA0003195805970000062
And
Figure BDA0003195805970000063
performing the following steps;
s305, if the amount of the data stored in the sliding window queue recording the average value and the standard deviation of the previous batch is larger than a set value, dequeuing the data at the head of the queue;
s306, calculating the moving average of the mean value and the standard deviation of the model calculation parameters in the sliding window queue to obtain
Figure BDA0003195805970000064
And M (v)2) And standardizing the WX according to the following formula to obtain standardized model calculation parameters:
Figure BDA0003195805970000065
wherein, epsilon is an additional item for preventing from dividing 0, and is a settable decimal, [ alpha, beta ] is a standardized parameter, and the initial value is [1,0 ];
s307, standardizing the model calculation parameters obtained in the step S302 by using the average value and the standard deviation of the model calculation parameters obtained in the step S306 to obtain standardized model calculation parameters;
s308, calculating parameter BN (WX) for the normalized model except P1b) Performing homomorphic encryption to obtain EPn[BN(WXb)]And sends the encrypted result to P1;
s309, summing the normalized model calculation parameters in P1 to obtain:
Figure BDA0003195805970000066
and calculates gradient calculation parameters E [ h (WX) ] of the encrypted model by using local supervision information]=f(E[WX]) Gradient E [ g ] of-y and P1P1]=E[h(WX)]xi
S310, P1 gradient calculation parameters E [ h (WX)]Sending to other participants the gradient E g of P1P1]Sending the data to a coordinator;
s311, the other participants calculate the respective gradients E [ g ] using the gradient calculation parametersPn]And sending the gradient to a coordinator;
s312, the coordinator decrypts the gradient of each participant to obtain gPnAnd distributed to each participant;
s313, each participant uses the decrypted gradient gPnTo update model parametersNumber [ W, alpha, beta ]];
S314, when the model reaches the maximum iteration times, the model terminates the training termination condition, otherwise, the step S302 is repeated;
s315, after the model training is finished, all the participators simultaneously save the model parameter WPnAnd the mean value of the model calculation parameters in the sliding window queue
Figure BDA0003195805970000071
And standard deviation M (v)2);
S4, prediction: the specific prediction steps are as follows:
s401, the new user carries out intersection;
s402, data characteristic X and model parameter W of new userPnCarrying out matrix multiplication to obtain model calculation parameters WX, and standardizing by using the average value and standard deviation of the model calculation parameters stored in model training;
Figure BDA0003195805970000072
s403, use
Figure BDA0003195805970000073
Obtaining the normalized model calculation parameters, wherein the normalized model calculation parameters are obtained by the formula: y-pred ═ f (wx) to calculate prediction results and prediction probabilities, where
Figure BDA0003195805970000074
Although embodiments of the present invention have been shown and described, it will be appreciated by those skilled in the art that changes, modifications, substitutions and alterations can be made in these embodiments without departing from the principles and spirit of the invention, the scope of which is defined in the appended claims and their equivalents.

Claims (1)

1. A method for batch normalization in a logistic regression model for safe federal learning, comprising: the method comprises the following steps:
s1, preparing training data by all parties of the federal learning participants: the federally learned participants may have multiple parties P1, P2.., Pn, each having training data _1, data _ 2.., data _ n, respectively, wherein one party has a tag y, assuming that the party having the tag is P1, and a coordinator for gradient decryption;
s2, fusing data sets: in the longitudinal federal study, data should be submitted, namely, a sample ID commonly owned by each participant is screened to form a fused data set fuse _ data;
s3, training a logistic regression model in longitudinal federal learning: the specific training method comprises the following steps of,
s301, all the participants of the joint training in each training process form a data set X locally according to the sample ID and the characteristics after intersectionPn=[x1,x2,...,xm]∈RmAnd respective model parameters WPn∈Rm
S302, each participant respectively takes out a batch B [ X ] with the data volume B from the owned data]∈Rb×mAccording to the characteristic dimension m, performing matrix multiplication with the model parameter W to obtain the model calculation parameter WX belonging to the R in the first stepb×1
S303, each participant calculates the average value of the parameters according to the model calculation parameters of the first step
Figure FDA0003195805960000011
And standard deviation v2b;
S304, calculating the average value of the parameters of the model of the current batch
Figure FDA0003195805960000012
And standard deviation v2b put into a sliding window queue of length k for recording the mean and standard deviation of previous batches
Figure FDA0003195805960000013
And
Figure FDA0003195805960000014
performing the following steps;
s305, if the amount of the data stored in the sliding window queue recording the average value and the standard deviation of the previous batch is larger than a set value, dequeuing the data at the head of the queue;
s306, calculating the moving average of the mean value and the standard deviation of the model calculation parameters in the sliding window queue to obtain
Figure FDA0003195805960000021
And M (v)2) And standardizing the WX according to the following formula to obtain standardized model calculation parameters:
Figure FDA0003195805960000022
wherein, epsilon is an additional item for preventing from dividing 0, and is a settable decimal, [ alpha, beta ] is a standardized parameter, and the initial value is [1,0 ];
s307, standardizing the model calculation parameters obtained in the step S302 by using the average value and the standard deviation of the model calculation parameters obtained in the step S306 to obtain standardized model calculation parameters;
s308, calculating parameter BN (WX) for the normalized model except P1b) Performing homomorphic encryption to obtain EPn[BN(WXb)]And sends the encrypted result to P1;
s309, summing the normalized model calculation parameters in P1 to obtain:
Figure FDA0003195805960000023
and calculates gradient calculation parameters E [ h (WX) ] of the encrypted model by using local supervision information]=f(E[WX]) Gradient E [ g ] of-y and P1P1]=E[h(WX)]xi
S310, P1 gradient calculation parameters E [ h (WX)]Sending to other participants the gradient E g of P1P1]Sending the data to a coordinator;
s311, the other participants calculate the respective gradients E [ g ] using the gradient calculation parametersPn]And sending the gradient to a coordinator;
s312, the coordinator decrypts the gradient of each participant to obtain gPnAnd distributed to each participant;
s313, each participant uses the decrypted gradient gPnTo update the model parameters [ W, alpha, beta ]];
S314, when the model reaches the maximum iteration times, the model terminates the training termination condition, otherwise, the step S302 is repeated;
s315, after the model training is finished, all the participators simultaneously save the model parameter WPnAnd the mean value of the model calculation parameters in the sliding window queue
Figure FDA0003195805960000031
And standard deviation M (v)2);
S4, prediction: the specific prediction steps are as follows:
s401, the new user carries out intersection;
s402, data characteristic X and model parameter W of new userPnCarrying out matrix multiplication to obtain model calculation parameters WX, and standardizing by using the average value and standard deviation of the model calculation parameters stored in model training;
Figure FDA0003195805960000032
s403, use
Figure FDA0003195805960000033
Obtaining the normalized model calculation parameters, wherein the normalized model calculation parameters are obtained by the formula: y-pred ═ f (wx) to calculate prediction results and prediction probabilities, where
Figure FDA0003195805960000034
CN202110890465.9A 2021-08-04 2021-08-04 Method for batch normalization in logistic regression model for safe federal learning Pending CN113591115A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110890465.9A CN113591115A (en) 2021-08-04 2021-08-04 Method for batch normalization in logistic regression model for safe federal learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110890465.9A CN113591115A (en) 2021-08-04 2021-08-04 Method for batch normalization in logistic regression model for safe federal learning

Publications (1)

Publication Number Publication Date
CN113591115A true CN113591115A (en) 2021-11-02

Family

ID=78254874

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110890465.9A Pending CN113591115A (en) 2021-08-04 2021-08-04 Method for batch normalization in logistic regression model for safe federal learning

Country Status (1)

Country Link
CN (1) CN113591115A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115292738A (en) * 2022-10-08 2022-11-04 豪符密码检测技术(成都)有限责任公司 Method for detecting security and correctness of federated learning model and data

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115292738A (en) * 2022-10-08 2022-11-04 豪符密码检测技术(成都)有限责任公司 Method for detecting security and correctness of federated learning model and data
CN115292738B (en) * 2022-10-08 2023-01-17 豪符密码检测技术(成都)有限责任公司 Method for detecting security and correctness of federated learning model and data

Similar Documents

Publication Publication Date Title
Nabeeh et al. Neutrosophic multi-criteria decision making approach for iot-based enterprises
US20230078061A1 (en) Model training method and apparatus for federated learning, device, and storage medium
Ishizaka et al. Selection of new production facilities with the group analytic hierarchy process ordering method
Khalili-Damghani et al. A hybrid fuzzy multiple criteria group decision making approach for sustainable project selection
Hong et al. Rapid identification of the optimal product configuration and its parameters based on customer-centric product modeling for one-of-a-kind production
CN111737749A (en) Measuring device alarm prediction method and device based on federal learning
CN112039702B (en) Model parameter training method and device based on federal learning and mutual learning
Zaidan et al. Uncertainty decision modeling approach for control engineering tools to support industrial cyber-physical metaverse smart manufacturing systems
Liu Based on intelligent advertising recommendation and abnormal advertising monitoring system in the field of machine learning
CN113591115A (en) Method for batch normalization in logistic regression model for safe federal learning
CN111724175A (en) Citizen credit point evaluation method applying logistic regression modeling
CN112883070A (en) Generation type countermeasure network recommendation method with differential privacy
Wu et al. A supervised learning-driven heuristic for solving the facility location and production planning problem
Wang et al. A Bi-objective AHP-MINLP-GA approach for Flexible Alternative Supplier Selection amid the COVID-19 pandemic
Haryadi et al. Implementation of support vector regression for polkadot cryptocurrency price prediction
CN117235811A (en) Internet finance-based payment big data analysis method and big data analysis system
Liu et al. Three-way decisions with single-valued neutrosophic decision theory rough sets based on grey relational analysis
Śmietanka et al. Federated learning for privacy-preserving data access
Chou et al. Forecasting enterprise resource planning software effort using evolutionary support vector machine inference model
Song et al. Group decision making with hesitant fuzzy linguistic preference relations based on multiplicative DEA cross-efficiency and stochastic acceptability analysis
Süer et al. Cell loading and family scheduling for jobs with individual due dates to minimise maximum tardiness
Chen et al. Fed-EINI: an efficient and interpretable inference framework for decision tree ensembles in federated learning
Li et al. A Temporal–Spatial network embedding model for ICT supply chain market trend forecasting
Han et al. New approach for quality function deployment based on multi-granular unbalanced linguistic information and consensus reaching process
Kubler et al. Information dissemination process for context-aware products

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination