CN110442457A - Model training method, device and server based on federation's study - Google Patents

Model training method, device and server based on federation's study Download PDF

Info

Publication number
CN110442457A
CN110442457A CN201910739086.2A CN201910739086A CN110442457A CN 110442457 A CN110442457 A CN 110442457A CN 201910739086 A CN201910739086 A CN 201910739086A CN 110442457 A CN110442457 A CN 110442457A
Authority
CN
China
Prior art keywords
model
training
local model
training pattern
precision
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201910739086.2A
Other languages
Chinese (zh)
Inventor
雷凯
黄济乐
方俊杰
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Peking University Shenzhen Graduate School
Original Assignee
Peking University Shenzhen Graduate School
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Peking University Shenzhen Graduate School filed Critical Peking University Shenzhen Graduate School
Priority to CN201910739086.2A priority Critical patent/CN110442457A/en
Publication of CN110442457A publication Critical patent/CN110442457A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F21/00Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F21/60Protecting data
    • G06F21/62Protecting access to data via a platform, e.g. using keys or access control rules
    • G06F21/6218Protecting access to data via a platform, e.g. using keys or access control rules to a system of files or objects, e.g. local or distributed file system or database
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F21/00Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F21/60Protecting data
    • G06F21/62Protecting access to data via a platform, e.g. using keys or access control rules
    • G06F21/6218Protecting access to data via a platform, e.g. using keys or access control rules to a system of files or objects, e.g. local or distributed file system or database
    • G06F21/6245Protecting personal data, e.g. for financial or medical purposes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/50Allocation of resources, e.g. of the central processing unit [CPU]
    • G06F9/5005Allocation of resources, e.g. of the central processing unit [CPU] to service a request
    • G06F9/5027Allocation of resources, e.g. of the central processing unit [CPU] to service a request the resource being a machine, e.g. CPUs, Servers, Terminals
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/94Hardware or software architectures specially adapted for image or video understanding
    • G06V10/95Hardware or software architectures specially adapted for image or video understanding structured as a network, e.g. client-server architectures

Abstract

The embodiment of the present invention provides a kind of model training method, device and server based on federation's study.This method comprises: multiple working nodes will be sent to training pattern;Receive the local model of multiple working nodes feedback, local model is that each working node treats training pattern according to the data respectively possessed and is trained acquisition;The precision of each local model is determined according to test data set;The weight coefficient of each local model is determined according to the precision of each local model, weight coefficient and precision are positively correlated;According to multiple local models and corresponding weight coefficient, treats training pattern and be updated.The method of the embodiment of the present invention accelerates the convergence rate to training pattern, improves the precision to training pattern by increasing the weight coefficient of the local model of high-precision and reducing the weight coefficient of low precision local model.

Description

Model training method, device and server based on federation's study
Technical field
The present embodiments relate to field of computer technology, and in particular to a kind of model training side based on federation's study Method, device and server.
Background technique
On the one hand with the continuous reinforcement for constantly improve and monitoring of laws and regulations, the centralized processing of data will face Huge legal risk;On the other hand it is unwilling between each data owning side for factors such as safety, economic interests shared former Beginning data.These factors all can cause data to exist in the form of isolated island.In order to break data silo, federation's study meet the tendency of and It is raw.
Federation's study shares initial data without each data owning side, can be under conditions of guaranteeing safety, fully Model training is carried out using the initial data of each data owning side, efficiently solves the problems, such as the data silo in artificial intelligence epoch. Currently based on the model training process of federation's study are as follows: multiple work sections will be issued to training pattern by server node Point, each working node are trained to what is received to training pattern based on the data respectively possessed, and by trained mould Shape parameter is back to server node, and server node treats training pattern according to the model parameter of each working node received It is updated, then repeats the above process until meeting preset performance indicator to training pattern.
During the existing study progress model training based on federation, server node can not effectively identify working node In low quality node, lead to the introducing of low quality model parameter, to reduce the convergence rate and model of model training Precision.
Summary of the invention
The embodiment of the present invention provides a kind of model training method, device and server based on federation's study, to solve The problem that convergence rate is slow and model accuracy is low present in the existing model training method based on federation's study.
In a first aspect, the embodiment of the present invention provides a kind of model training method based on federation's study, comprising:
Multiple working nodes will be sent to training pattern;
The local model of multiple working node feedbacks is received, local model is each working node according to the number respectively possessed Acquisition is trained according to training pattern is treated;
The precision of each local model is determined according to test data set;
The weight coefficient of each local model, weight coefficient and precision positive are determined according to the precision of each local model It closes;
According to multiple local models and corresponding weight coefficient, treats training pattern and be updated.
In one possible implementation, the method also includes:
The reward that each working node is determined according to the precision of each local model, so that each working node is according to reward Participation is adjusted, reward is positively correlated with precision.
In one possible implementation, it will be sent to training pattern before multiple working nodes, the method is also Include:
It is determined according to training mission to training pattern, and treats training pattern and initialized;
According to the network connection state and training mission between server node and working node, multiple work are determined Node.
In one possible implementation, it treats after training pattern is updated, the method also includes:
Judge updated whether meet pre-set level to training pattern;
If satisfied, model training process is then terminated, conversely, then continuing repetitive exercise.
In one possible implementation, receiving the local model that multiple working nodes are fed back includes:
Receive the local model that multiple working nodes are fed back within a preset period of time.
In one possible implementation, the precision for determining each local model according to test data set includes:
Test verifying is carried out to each local model according to test data set, determines each local model in test data set On accuracy rate and/or recall rate;
According to accuracy rate and/or recall rate of each local model in test data set, each local model is determined Precision.
In one possible implementation, according to multiple local models and corresponding weight coefficient, trained mould is treated Type is updated, comprising:
Training pattern is treated according to the following formula to be updated:
Wherein, M ' expression is updated to training pattern, and M indicates that, to training pattern, α is the weight system to training pattern Number, LiIndicate i-th of local model, αiFor the weight coefficient of i-th of local model, N indicates the quantity of local model.
Second aspect, the embodiment of the present invention provide a kind of model training apparatus based on federation's study, comprising:
Sending module, for multiple working nodes will to be sent to training pattern;
Receiving module, for receiving the local model of multiple working nodes feedback, local model be each working node according to The data respectively possessed treat training pattern and are trained acquisition;
Test module, for determining the precision of each local model according to test data set;
Processing module, for determining the weight coefficient of each local model, weight system according to the precision of each local model It is several to be positively correlated with precision;
Update module, for treating training pattern and being updated according to multiple local models and corresponding weight coefficient.
The third aspect, the embodiment of the present invention provide a kind of server, comprising:
At least one processor and memory;
Memory stores computer executed instructions;
At least one processor executes the computer executed instructions of memory storage, so that at least one processor executes such as The described in any item model training methods based on federation's study of first aspect.
Fourth aspect, the embodiment of the present invention provide a kind of computer readable storage medium, the computer-readable storage medium Computer executed instructions are stored in matter, for realizing any one of such as first aspect when computer executed instructions are executed by processor The model training method based on federation's study.
Model training method, device and server provided in an embodiment of the present invention based on federation's study, by will be wait instruct Practice model and be sent to multiple working nodes, and receive the local model of multiple working node feedbacks, then according to test data set The precision for determining each local model is each local model setting and the positively related weight coefficient of precision, finally according to multiple Local model and corresponding weight coefficient, treat training pattern and are updated, and realize multiple working nodes for wait train The coorinated training of model.Due to being not necessarily to shared data between each node, the peace of data be effectively ensured during model training Full property and data-privacy;By the weight coefficient of increase high quality working node, the weight coefficient of low quality working node is reduced, The convergence to training pattern is not only accelerated, but also can be improved the precision to training pattern;Further, model training process Acceleration, also synchronize the workload for reducing each working node, reduce the communication overhead between node.
Detailed description of the invention
Fig. 1 is system architecture schematic diagram of the invention;
Fig. 2 is the flow chart of model training method one embodiment provided by the invention based on federation's study;
Fig. 3 is the flow chart of the model training method another embodiment provided by the invention based on federation's study;
Fig. 4 is the structural schematic diagram of model training apparatus one embodiment provided by the invention based on federation's study;
Fig. 5 is the structural schematic diagram of one embodiment of server provided by the invention.
Specific embodiment
Below by specific embodiment combination attached drawing, invention is further described in detail.Wherein different embodiments Middle similar component uses associated similar element numbers.In the following embodiments, many datail descriptions be in order to The application is better understood.However, those skilled in the art can recognize without lifting an eyebrow, part of feature It is dispensed, or can be substituted by other elements, material, method in varied situations.In some cases, this Shen Please it is relevant it is some operation there is no in the description show or describe, this is the core in order to avoid the application by mistake More descriptions are flooded, and to those skilled in the art, these relevant operations, which are described in detail, not to be necessary, they Relevant operation can be completely understood according to the general technology knowledge of description and this field in specification.
It is formed respectively in addition, feature described in this description, operation or feature can combine in any suitable way Kind embodiment.Meanwhile each step in method description or movement can also can be aobvious and easy according to those skilled in the art institute The mode carry out sequence exchange or adjustment seen.Therefore, the various sequences in the description and the appended drawings are intended merely to clearly describe a certain A embodiment is not meant to be necessary sequence, and wherein some sequentially must comply with unless otherwise indicated.
It is herein component institute serialization number itself, such as " first ", " second " etc., is only used for distinguishing described object, Without any sequence or art-recognized meanings.And " connection ", " connection " described in the application, unless otherwise instructed, include directly and It is indirectly connected with (connection).
Fig. 1 is system architecture schematic diagram of the invention.As shown in Figure 1, server node 101 and working node 102 can be with It is communicatively coupled by way of wiredly and/or wirelessly.Wherein, server node 101 includes but is not limited to single network clothes Be engaged in device, multiple network servers composition server group or based on the consisting of a large number of computers or network servers of cloud computing Cloud;Working node 102 includes but is not limited to computer, smart phone, tablet computer, personal digital assistant messaging device Deng.
Fig. 1 embodies a kind of completely new distributed artificial intelligence network architecture system.Server node 101 and working node It is not necessarily to shared initial data between 102 and between each working node, ensure that Information Security and data-privacy.Server Node 101 is issued to working node 102 to training pattern, and each working node is instructed based on the local data respectively possessed Practice, and return trained local model, server node 101 merges the local model of multiple working nodes and is polymerized to one World model continues to issue, and is iterated training until world model restrains.
Below in conjunction with system architecture shown in FIG. 1, the model provided in an embodiment of the present invention based on federation's study is instructed The some possible application scenarios for practicing method are illustrated.For example, there are first, second, the third three companies for managing video website, The video recommendations model for thinking acquisition high quality, for recommending video to user, to improve user satisfaction.Each company is possessed User data it is limited, the video recommendations model of high quality can not be trained based on the user data respectively possessed;In order to protect The safety of user data is demonstrate,proved, privacy of user is protected, each company can not share the user data respectively possessed.This When, each company can be by constructing respective working node, and by means of server node coorinated training video recommendations model. It says, in order to obtain high-precision human face recognition model, can be possessed by server node Issuance model training mission for another example The terminal device of face image data can be used as working node and participate in model training.Server node can obtain high quality Human face recognition model, and terminal device can also calculate power by contribution and data obtain corresponding remuneration, realize two-win.
Since the data between server node 101 and working node 102 and between each working node are not shared, information Asymmetry, and the parameter in the local model returned does not have actual physical meaning yet, therefore server node 101 can not ensure The quality for the local model that working node 102 returns.It is difficult to avoid that and introduces low-quality local model, low quality local model The convergence rate and precision that will affect world model are introduced, meanwhile, the reduction of convergence rate also will increase communication overhead, for whole Very big negative effect can be generated for a system.
It will elaborate how the present invention solves the above problems by specific embodiment below.
Fig. 2 is the flow chart of model training method one embodiment provided by the invention based on federation's study.The present embodiment The model training method based on federation's study provided can be applied to server node.As shown in Fig. 2, provided in this embodiment May include: based on the federal model training method learnt
S201, multiple working nodes will be sent to training pattern.
It can include but is not limited to identification model, disaggregated model, detection model, prediction to training pattern in the present embodiment Model etc., in specific implementation such as can be using artificial nerve network model, support vector machines, convolutional neural networks model, The present embodiment to this with no restriction.Working node is the node for possessing training data and participating in model interoperability training, the present embodiment The quantity of the middle working node for participating in model interoperability training is more than or equal to two.
It is understood that server node can be with root before it will be sent to multiple working nodes to training pattern It determines according to training mission to training pattern.For example, if training mission is to carry out recognition of face, mould to be trained can be determined Type is identification model;If training mission is to carry out risk profile, it can determine that training pattern be prediction model.It can be preparatory It establishes training mission and to the mapping relations between training pattern, mould to be trained is determined according to training mission and mapping relations Type.Server node, can be by building after determining to training pattern in advance between server node and each working node Vertical network connection will be sent to each working node to training pattern.
S202, the local model for receiving multiple working nodes feedback, local model are each working nodes according to respectively being gathered around Some data treat training pattern and are trained acquisition.
In the present embodiment when working node receive that server node issues to training pattern after, work can be used Make the local data stored in node to be trained to what is received to training pattern, generates local model.It is received in local model It holds back, alternatively, the local model with trained parameter to be passed through to the network pre-established and is connected when the training time reaches preset value It takes back and reaches server node.The local data stored in each working node in the present embodiment is protected without sharing among the nodes Safety and the privacy for having data in each working node by oneself are demonstrate,proved.
Optionally, in order to further increase the training speed of model, the local model of multiple working nodes feedbacks is received A kind of possible implementation are as follows: receive the local model that multiple working nodes are fed back within a preset period of time.
In order to avoid server node waits working node to feed back for a long time in the present embodiment, the training speed of model is reduced Degree, server node only receive the local model fed back within a preset period of time.Such as server node can issued wait instruct Opening timing device can close receiving channel refusal and be further continued for receiving local mould after preset time arrival after practicing model Type, or can directly be discarded in the local model received after preset time reaches.As an example it is assumed that there is A, B and C tri- A local model, wherein A and B are reached within a preset period of time, and C time-out reaches, then server node is carrying out mould to be trained When type updates, local model A and B are only considered.Preset time period in the present embodiment for example can be according to answering to training pattern The factors such as miscellaneous degree, traffic rate between the average calculation power of each working node, node determine.
S203, the precision that each local model is determined according to test data set.
It, can be according to service in the present embodiment after server node receives the local model of working node feedback Test data set in device node carries out test verifying to the local model received, with the precision of each local model of determination. The precision of local model can embody the quality of local model, it is to be understood that the corresponding high-precision of the local model of high quality. Test data set in the present embodiment includes the test data marked in advance, and each working node can not know test data set. Unified test verifying is carried out by server node, it is possible to prevente effectively from each working node is practised fraud, guarantees test verifying Accuracy and fairness.Specifically, such as the data of test data concentration can be inputted into local model, determines local model Reality output, concentrate the desired output marked in advance to be compared the reality output of local model and test data, according to Comparison result determines the precision of local model.
Optionally, a kind of possible implementation of the precision of each local model is determined according to test data set are as follows: root Test verifying is carried out to each local model according to test data set, determines accuracy rate of each local model in test data set And/or recall rate;According to accuracy rate and/or recall rate of each local model in test data set, each local mould is determined The precision of type.Wherein, accuracy rate can be used for the correctness of measurement model, the correct sample number of accuracy rate=prediction/total sample Number;Recall rate can be used for the covering surface of measurement model, the recall rate=prediction of positive class is positive the positive class number of the quantity of class/total Amount.The precision of local model can perhaps the weighted average of recall rate or accuracy rate and recall rate be true according to accuracy rate It is fixed.
Optionally, the test data set in the present embodiment can be determined according to training pattern.Such as it can pre-establish It, can be according to the mapping after having determined to training pattern to the mapping relations between training pattern and test data set Relationship determines corresponding test data set.
S204, the weight coefficient that each local model is determined according to the precision of each local model, weight coefficient and precision It is positively correlated.
In the present embodiment after the precision for determining each local model, each local mould can be determined according to the precision The weight coefficient of type.Wherein, the size of weight coefficient embodies the size that local model contributes entire model training process. It is understood that introducing the local model of high quality, the convergence to training pattern can speed up, improve the essence to training pattern Degree, is beneficial to entire model training process;And low-quality local model is introduced, have no not only for entire model training process It helps, will increase the workload of other working nodes instead, increase the communication overhead of whole system.Therefore, each in the present embodiment The weight coefficient of local model and its precision are positively correlated, i.e. the weight coefficient of the higher local model acquisition of precision is bigger.It is optional , the local model of default precision threshold is lower than for precision, its weight coefficient 0 can be set as, may be brought to avoid it Adverse effect.
By introducing attention mechanism in the present embodiment, increases the weight coefficient of high quality working node, reduce low quality The weight coefficient of working node is conducive to acceleration model convergence, improves training speed.
S205, according to multiple local models and corresponding weight coefficient, treat training pattern and be updated.
Server node can treat training after the weight coefficient for determining each local model accordingly in the present embodiment Model is updated.For example, can treat training pattern using the weighted average of multiple local models and be updated.
Model training method provided in this embodiment based on federation's study, by the way that multiple works will be sent to training pattern Make node, and receive the local model of multiple working node feedbacks, each local model is then determined according to test data set Precision, for each local model setting and the positively related weight coefficient of precision, finally according to multiple local models and corresponding Weight coefficient is treated training pattern and is updated, and realizes multiple working nodes for the coorinated training to training pattern.Due to During model training, it is not necessarily to shared data between each node, safety and the data-privacy of data has been effectively ensured;Pass through Increase the weight coefficient of high quality working node, reduces the weight coefficient of low quality working node, not only accelerate mould to be trained The convergence of type, and can be improved the precision to training pattern;Further, the acceleration of model training process, it is also synchronous to reduce The workload of each working node, reduces the communication overhead between node.
In order to transfer the enthusiasm of each working node, so that it can play an active part in model training, it will usually to each work Node gives excitation appropriate.Currently, being broadly divided into two classes to the incentive mechanism of working node: the incentive mechanism based on game theory With the incentive mechanism based on contract theory.Prize is distributed based on the contribution of the incentive mechanism Equilibrium Game working node of game theory It encourages, is mostly that total system is contributed to stimulate working node.In the mechanism, the work quality or work of evaluation work node The contribution that node is done by system has following two mode: being overall model institute according to after a working node participation model training The decline of overall model income, is commented caused by the promotion of bring income or a working node exit model training Estimate.Specifically, can be according to indexs such as operation time, communication overhead and model accuracies come comprehensive descision, and then determine work The work quality of node.Incentive mechanism based on contract theory be substantially the incentive mechanism based on game theory a kind of derivative and Variation can preset the specification that a plurality of types of contracts sign reward, and the distributed of information asymmetry that be more adaptive to motivates Situation.Server node can set up a revenue function, with lift scheme precision and receipts as far as possible under the paying that contract limits Speed is held back to reduce expense, strives for the maximum return of oneself;Working node can set up utility function, in data quality and Strive for high-grade contract as far as possible under conditions of computing capability to improve the income of oneself.Both sides are in the state of a game Under, mutual game is to reach balance.
Above-mentioned incentive mechanism is difficult to avoid that malicious node gains reward by cheating, and the node that can not effectively prevent doing evil is to model training mistake The influence of journey.By taking the incentive mechanism based on contract theory as an example, it is assumed that server node has been signed high-grade with working node Contract, server node not can guarantee the data of working node and calculate power and do not play tricks.For working node, if intentionally It does evil or in order to gain excitation by cheating, model can be returned without model training and directly completely, or do not have in local model Model is returned under convergent state, or directly makes up parameter passback model.Such passback model is low-quality model, right What model convergence not only help without, will increase the workload of other working nodes instead.
In order to avoid there is the phenomenon that node gains reward by cheating, while in order to transfer the enthusiasm of each node, in above-mentioned implementation On the basis of example, method provided in this embodiment can also include: to determine each work section according to the precision of each local model The reward of point, so that each working node adjusts participation according to reward, reward is positively correlated with precision.
Reward in the present embodiment for example can be Permission Levels, calculating money of each working node using trained model Source, storage resource etc..Can according to specific application scenarios determine, the present embodiment for reward concrete form with no restriction.
Server node, can be according to the precision using clothes after the precision for determining each local model in the present embodiment The income of the utility function evaluating server node of business device node, is distributed according to the income of server node to each working node Reward.Each working node is assessed the income of each working node according to respective utility function, is adjusted accordingly after being rewarded Participation, specifically, data volume, computing resource, storage resource, communication used in adjustable participation model training process Resource etc..
It is understood that high-precision local model can make server node obtain higher income, therefore in order to The working node of high quality is attracted to participate in model training process, it can be to the work of the local model of offer high-precision in the present embodiment The more rewards of node distribution, i.e., precision is higher, and the reward that can be obtained is also more.The reward of great number can promote working node product The training of the participation model of pole, contributes higher calculation power and more data, provides high-precision local.Optionally, for essence Lower than the local model for presetting precision threshold, server node can be refused to distribute to the working node for providing the local model degree Reward.Therefore, it for the working node for making up parameter passback model, possibly can not be rewarded, or be only capable of obtaining very Small reward can effectively reduce the phenomenon that working node gains reward by cheating.
It is provided in this embodiment based on federation study model training method, on the basis of the above embodiments, pass through to The positively related reward of precision for the local model that each working node distribution is provided with it, the working node for promoting high quality are positive The training process for participating in model, alleviates the phenomenon that malicious node gains reward by cheating, has further speeded up the convergence to training pattern And improve precision to training pattern.
Fig. 3 is the flow chart of the model training method another embodiment provided by the invention based on federation's study.Such as Fig. 3 institute Show, method provided in this embodiment may include:
S301, it is determined according to training mission to training pattern, and treats training pattern and initialized.
Determine that the specific implementation process to training pattern can refer to above-described embodiment according to training mission, it is no longer superfluous herein It states.After determining to training pattern, server node can treat training pattern and be initialized.Such as training can be treated Model carries out random initializtion, is random value by the parameter initialization to training pattern;Alternatively, can will be initial to training pattern Turn to pre-stored model in server node.
It should be noted that server node is only treated training pattern and is carried out once during being iterated training Initialization.
S302, according to the network connection state and training mission between server node and working node, determine multiple Working node.
It, can be according to server node and working node in order to improve the quality for the working node for participating in model training process Between network connection state, training mission, the precision of local model that provides in history iterative process etc. to working node into Row screening.It for example, can be from if during being determined according to training mission 3 working nodes being needed to participate in model trainings 3 working nodes are randomly selected in good 4 working nodes of network connection state and participate in model training, or can be according to each The history of working node shows, and higher 3 working nodes of quality are chosen from good 4 working nodes of network connection state Participate in model training.
It should be noted that can be chosen to working node during every wheel repetitive exercise.Namely It says, the node that model training is participated in during repetitive exercise can not be identical.For example, in first round training process, work The network connection state made between node W and server node is good, the selected participation model training of working node W, however During second wheel training, the network connection state between working node W and server node deteriorates, and can not efficiently be led to Letter, working node W will be removed.It is possible to prevente effectively from the introducing of low quality working node.
S303, multiple working nodes will be sent to training pattern.
Specific implementation process can be with reference to the step S201 in above-described embodiment, and details are not described herein again.
S304, the local model that multiple working nodes are fed back within a preset period of time is received.
Specific implementation process can refer to above-described embodiment, and details are not described herein again.
S305, the precision that each local model is determined according to test data set.
Specific implementation process can be with reference to the step S203 in above-described embodiment, and details are not described herein again.
It should be noted that improving the generalization ability to training pattern in order to avoid there is over-fitting, changing During generation training, test data set can be some or all of not identical.For example, if being stored in server node 10000 test datas marked in advance can be selected at random during every wheel repetitive exercise from 10000 test datas It takes, or chooses 100 formation test data sets according to preset rules, for being verified to local model.
S306, the weight coefficient that each local model is determined according to the precision of each local model, weight coefficient and precision It is positively correlated.
Specific implementation process can be with reference to the step S204 in above-described embodiment, and details are not described herein again.
S307, according to multiple local models and corresponding weight coefficient, treat training pattern and be updated.
Optionally, according to multiple local models and corresponding weight coefficient, one kind that training pattern is updated is treated Possible implementation are as follows:
Training pattern is treated according to the following formula to be updated:
Wherein, M ' expression is updated to training pattern, and M indicates that, to training pattern, α is the weight system to training pattern Number, α ∈ [0,1], LiIndicate i-th of local model, αiFor the weight coefficient of i-th of local model, αi∈ [0,1], N indicate this The quantity of ground model,
It should be noted that at repetitive exercise initial stage, due to unstable to training pattern M, for the training of acceleration model Process can reduce α, increase αi, such as using random fashion initialized to training pattern, α can be set as 0;And in the repetitive exercise later period, since M has tended to convergence state, in order to improve the stability of training process, α can be increased, Reduce αi
S308, judge it is updated whether meet pre-set level to training pattern, if satisfied, then terminating model training mistake Journey continues repetitive exercise if not satisfied, thening follow the steps S302.
Whether whether the pre-set level in the present embodiment can be for one or more of following index: restraining, reach Whether maximum number of iterations reaches the maximum training time etc..Various indexs can be combined and be judged, such as can sentenced first Break it is updated whether restrained to training pattern, if convergence if terminate model training process, otherwise continue to determine whether to have reached To maximum number of iterations, model training process is terminated if reaching, and otherwise continues to determine whether to have reached the maximum the training time, Model training process is terminated if reaching, no person executes step S302, continues repetitive exercise.
Model training method provided in this embodiment based on federation's study, on the basis of the above embodiments, further Improve training speed and model accuracy.
The embodiment of the present invention also provides a kind of model training apparatus based on federation's study, shown in Figure 4, the present invention Embodiment is only illustrated by taking Fig. 4 as an example, is not offered as that present invention is limited only to this.Fig. 4 is provided by the invention based on federal The structural schematic diagram of one embodiment of model training apparatus of habit.As shown in figure 4, the mould provided in this embodiment based on federation's study Type training device 40 may include: sending module 401, receiving module 402, test module 403, processing module 404 and update mould Block 405.
Sending module 401, for multiple working nodes will to be sent to training pattern;
Receiving module 402, for receiving the local model of multiple working node feedbacks, local model is each working node root Training pattern, which is treated, according to the data respectively possessed is trained acquisition;
Test module 403, for determining the precision of each local model according to test data set;
Processing module 404, for determining the weight coefficient of each local model, weight according to the precision of each local model Coefficient and precision are positively correlated;
Update module 405, for treating training pattern and carrying out more according to multiple local models and corresponding weight coefficient Newly.
The device of the present embodiment can be used for executing the technical solution of embodiment of the method shown in Fig. 2, realization principle and skill Art effect is similar, and details are not described herein again.
Optionally, the model training apparatus 40 based on federation's study can also include distribution module (not shown), tool Body is used to determine the reward of each working node according to the precision of each local model, so that each working node is adjusted according to reward Whole participation, reward are positively correlated with precision.
Optionally, the model training apparatus 40 based on federation's study can also include initialization module (not shown), Specifically for before it will be sent to multiple working nodes to training pattern, determined according to training mission to training pattern, and right It is initialized to training pattern;According to the network connection state and training mission between server node and working node, Determine multiple working nodes.
Optionally, the model training apparatus 40 based on federation's study can also include judgment module (not shown), tool Body is used for after treating training pattern and being updated, and judges updated whether meet pre-set level to training pattern;If full Foot, then terminate model training process, conversely, then continuing repetitive exercise.
Optionally, receiving module 402 specifically can be used for, and receive the sheet that multiple working nodes are fed back within a preset period of time Ground model.
Optionally, test module 403 specifically can be used for, and carries out test to each local model according to test data set and tests Card determines accuracy rate and/or recall rate of each local model in test data set;According to each local model in test number According to the accuracy rate and/or recall rate on collection, the precision of each local model is determined.
Optionally, update module 405 specifically can be used for, and treats training pattern according to the following formula and is updated:
Wherein, M ' expression is updated to training pattern, and M indicates that, to training pattern, α is the weight system to training pattern Number, LiIndicate i-th of local model, αiFor the weight coefficient of i-th of local model, N indicates the quantity of local model.
The embodiment of the present invention also provides a kind of server, shown in Figure 5, the embodiment of the present invention only by taking Fig. 5 as an example into Row explanation, is not offered as that present invention is limited only to this.Fig. 5 is the structural schematic diagram of one embodiment of server provided by the invention.Such as Shown in Fig. 5, server 50 provided in this embodiment may include: memory 501, processor 502 and bus 503.Wherein, bus 503 for realizing the connection between each element.
Computer program is stored in memory 501, computer program may be implemented above-mentioned when being executed by processor 502 The technical solution that one embodiment of the method provides.
Wherein, be directly or indirectly electrically connected between memory 501 and processor 502, with realize data transmission or Interaction.It is electrically connected for example, these elements can be realized between each other by one or more of communication bus or signal wire, such as It can be connected by bus 503.The computer journey for realizing the model training method based on federation's study is stored in memory 501 Sequence, the software function module that can be stored in the form of software or firmware including at least one in memory 501, processor 502 By running the software program and module that are stored in memory 501, thereby executing various function application and data processing.
Memory 501 may be, but not limited to, random access memory (Random Access Memory, referred to as: RAM), read-only memory (Read Only Memory, referred to as: ROM), programmable read only memory (Programmable Read-Only Memory, referred to as: PROM), erasable read-only memory (Erasable Programmable Read-Only Memory, referred to as: EPROM), electricallyerasable ROM (EEROM) (Electric Erasable Programmable Read- Only Memory, referred to as: EEPROM) etc..Wherein, memory 501 is for storing program, and processor 502 refers to receiving execution After order, program is executed.Further, the software program in above-mentioned memory 501 and module may also include operating system, can Including the various component softwares for management system task (such as memory management, storage equipment control, power management etc.) and/or Driving, and can be in communication with each other with various hardware or component software, to provide the running environment of other software component.
Processor 502 can be a kind of IC chip, the processing capacity with signal.Above-mentioned processor 502 can To be general processor, including central processing unit (Central Processing Unit, referred to as: CPU), network processing unit (Network Processor, referred to as: NP) etc..It may be implemented or execute disclosed each method, the step in the embodiment of the present invention Rapid and logic diagram.General processor can be microprocessor or the processor is also possible to any conventional processor etc.. It is appreciated that Fig. 5 structure be only illustrate, can also include than shown in Fig. 5 more perhaps less component or have with Different configuration shown in Fig. 5.Each component shown in Fig. 5 can use hardware and/or software realization.
It should be noted that server provided in this embodiment includes but is not limited to single network server, multiple networks The server group that server forms or the cloud consisting of a large number of computers or network servers based on cloud computing, wherein cloud meter It is one kind of distributed computing, a super virtual computer being made of the computer of a group loose couplings.
The embodiment of the present invention also provides a kind of computer readable storage medium, is stored thereon with computer program, computer The model training method based on federation's study that any of the above-described embodiment of the method provides may be implemented when program is executed by processor. Computer readable storage medium in the present embodiment can be any usable medium that computer can access, or include one Data storage devices, the usable mediums such as a or multiple usable mediums integrated server, data center can be magnetic medium, (for example, floppy disk, hard disk, tape), optical medium (for example, DVD) or semiconductor medium (such as SSD) etc..
It will be understood by those skilled in the art that all or part of function of various methods can pass through in above embodiment The mode of hardware is realized, can also be realized by way of computer program.When function all or part of in above embodiment When being realized by way of computer program, which be can be stored in a computer readable storage medium, and storage medium can To include: read-only memory, random access memory, disk, CD, hard disk etc., it is above-mentioned to realize which is executed by computer Function.For example, program is stored in the memory of equipment, when executing program in memory by processor, can be realized State all or part of function.In addition, when function all or part of in above embodiment is realized by way of computer program When, which also can store in storage mediums such as server, another computer, disk, CD, flash disk or mobile hard disks In, through downloading or copying and saving into the memory of local device, or version updating is carried out to the system of local device, when logical When crossing the program in processor execution memory, all or part of function in above embodiment can be realized.
Use above specific case is illustrated the present invention, is merely used to help understand the present invention, not to limit The system present invention.For those skilled in the art, according to the thought of the present invention, can also make several simple It deduces, deform or replaces.

Claims (10)

1. a kind of model training method based on federation's study, is applied to server node characterized by comprising
Multiple working nodes will be sent to training pattern;
Receive the local model of multiple working nodes feedbacks, the local model is each working node according to respectively being possessed Data be trained acquisition to training pattern to described;
The precision of each local model is determined according to test data set;
The weight coefficient of each local model, the weight coefficient and institute are determined according to the precision of each local model State precision positive correlation;
According to multiple local models and corresponding weight coefficient, it is updated to described to training pattern.
2. the method as described in claim 1, which is characterized in that the method also includes:
The reward that each working node is determined according to the precision of each local model, so that each working node Participation is adjusted according to the reward, the reward is positively correlated with the precision.
3. the method as described in claim 1, which is characterized in that it is described will be sent to training pattern multiple working nodes it Before, the method also includes:
It is initialized to training pattern, and to described to training pattern according to training mission determination is described;
According to the network connection state and the training mission between the server node and working node, determine described in Multiple working nodes.
4. the method as described in claim 1, which is characterized in that it is described to it is described be updated to training pattern after, it is described Method further include:
Judge updated whether meet pre-set level to training pattern;
If satisfied, model training process is then terminated, conversely, then continuing repetitive exercise.
5. the method as described in claim 1, which is characterized in that the local model for receiving multiple working node feedbacks Include:
Receive the local model that multiple working nodes are fed back within a preset period of time.
6. the method according to claim 1 to 5, which is characterized in that it is described determined according to test data set it is each described The precision of local model includes:
Test verifying is carried out to each local model according to test data set, determines each local model in the survey Try the accuracy rate and/or recall rate on data set;
According to accuracy rate and/or recall rate of each local model in the test data set, each described is determined The precision of ground model.
7. the method according to claim 1 to 5, which is characterized in that described according to multiple local models and right The weight coefficient answered is updated to described to training pattern, comprising:
It is updated according to the following formula to described to training pattern:
Wherein, M ' expression is updated to training pattern, and M indicates described to training pattern, and α is the weight to training pattern Coefficient, LiIndicate i-th of local model, αiFor the weight coefficient of i-th of local model, N indicates the quantity of local model.
8. a kind of model training apparatus based on federation's study characterized by comprising
Sending module, for multiple working nodes will to be sent to training pattern;
Receiving module, for receiving the local model of multiple working node feedbacks, the local model is each working node Acquisition is trained to training pattern to described according to the data respectively possessed;
Test module, for determining the precision of each local model according to test data set;
Processing module, for determining the weight coefficient of each local model, institute according to the precision of each local model It states weight coefficient and the precision is positively correlated;
Update module, for being carried out to training pattern to described according to multiple local models and corresponding weight coefficient It updates.
9. a kind of server characterized by comprising at least one processor and memory;
The memory stores computer executed instructions;
At least one described processor executes the computer executed instructions of the memory storage, so that at least one described processing Device is executed such as the described in any item model training methods based on federation's study of claim 1-7.
10. a kind of computer readable storage medium, which is characterized in that be stored with computer in the computer readable storage medium It executes instruction, for realizing such as described in any item bases of claim 1-7 when the computer executed instructions are executed by processor In the model training method of federation's study.
CN201910739086.2A 2019-08-12 2019-08-12 Model training method, device and server based on federation's study Pending CN110442457A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910739086.2A CN110442457A (en) 2019-08-12 2019-08-12 Model training method, device and server based on federation's study

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910739086.2A CN110442457A (en) 2019-08-12 2019-08-12 Model training method, device and server based on federation's study

Publications (1)

Publication Number Publication Date
CN110442457A true CN110442457A (en) 2019-11-12

Family

ID=68434570

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910739086.2A Pending CN110442457A (en) 2019-08-12 2019-08-12 Model training method, device and server based on federation's study

Country Status (1)

Country Link
CN (1) CN110442457A (en)

Cited By (54)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110956275A (en) * 2019-11-27 2020-04-03 支付宝(杭州)信息技术有限公司 Risk prediction and risk prediction model training method and device and electronic equipment
CN110990870A (en) * 2019-11-29 2020-04-10 上海能塔智能科技有限公司 Operation and maintenance, processing method, device, equipment and medium using model library
CN111027715A (en) * 2019-12-11 2020-04-17 支付宝(杭州)信息技术有限公司 Monte Carlo-based federated learning model training method and device
CN111046433A (en) * 2019-12-13 2020-04-21 支付宝(杭州)信息技术有限公司 Model training method based on federal learning
CN111091200A (en) * 2019-12-20 2020-05-01 深圳前海微众银行股份有限公司 Updating method, system, agent, server and storage medium of training model
CN111260061A (en) * 2020-03-09 2020-06-09 厦门大学 Differential noise adding method and system in federated learning gradient exchange
CN111369042A (en) * 2020-02-27 2020-07-03 山东大学 Wireless service flow prediction method based on weighted federal learning
CN111383096A (en) * 2020-03-23 2020-07-07 中国建设银行股份有限公司 Fraud detection and model training method and device thereof, electronic equipment and storage medium
CN111460528A (en) * 2020-04-01 2020-07-28 支付宝(杭州)信息技术有限公司 Multi-party combined training method and system based on Adam optimization algorithm
CN111488995A (en) * 2020-04-08 2020-08-04 北京字节跳动网络技术有限公司 Method and apparatus for evaluating a joint training model
CN111678696A (en) * 2020-06-17 2020-09-18 南昌航空大学 Intelligent mechanical fault diagnosis method based on federal learning
CN111695701A (en) * 2020-06-12 2020-09-22 上海富数科技有限公司 System for realizing data set construction processing based on federal learning and construction generation method thereof
CN111814985A (en) * 2020-06-30 2020-10-23 平安科技(深圳)有限公司 Model training method under federated learning network and related equipment thereof
CN111882133A (en) * 2020-08-03 2020-11-03 重庆大学 Prediction-based federated learning communication optimization method and system
CN111881966A (en) * 2020-07-20 2020-11-03 北京市商汤科技开发有限公司 Neural network training method, device, equipment and storage medium
CN111898764A (en) * 2020-06-23 2020-11-06 华为技术有限公司 Method, device and chip for federal learning
CN111931876A (en) * 2020-10-12 2020-11-13 支付宝(杭州)信息技术有限公司 Target data side screening method and system for distributed model training
CN111931242A (en) * 2020-09-30 2020-11-13 国网浙江省电力有限公司电力科学研究院 Data sharing method, computer equipment applying same and readable storage medium
CN112001500A (en) * 2020-08-13 2020-11-27 星环信息科技(上海)有限公司 Model training method, device and storage medium based on longitudinal federated learning system
CN112101528A (en) * 2020-09-17 2020-12-18 上海交通大学 Terminal contribution measurement method based on back propagation
CN112232519A (en) * 2020-10-15 2021-01-15 成都数融科技有限公司 Joint modeling method based on federal learning
KR20210065069A (en) * 2019-11-22 2021-06-03 베이징 바이두 넷컴 사이언스 앤 테크놀로지 코., 엘티디. How to create a shared encoder, device and electronics
CN112949837A (en) * 2021-04-13 2021-06-11 中国人民武装警察部队警官学院 Target recognition federal deep learning method based on trusted network
WO2021120677A1 (en) * 2020-07-07 2021-06-24 平安科技(深圳)有限公司 Warehousing model training method and device, computer device and storage medium
WO2021138877A1 (en) * 2020-01-09 2021-07-15 深圳前海微众银行股份有限公司 Vertical federated learning model training optimization method and apparatus, device, and medium
CN113128528A (en) * 2019-12-27 2021-07-16 无锡祥生医疗科技股份有限公司 Ultrasonic image deep learning distributed training system and training method
CN113158550A (en) * 2021-03-24 2021-07-23 北京邮电大学 Method and device for federated learning, electronic equipment and storage medium
CN113191479A (en) * 2020-01-14 2021-07-30 华为技术有限公司 Method, system, node and storage medium for joint learning
CN113222211A (en) * 2021-03-31 2021-08-06 中国科学技术大学先进技术研究院 Multi-region diesel vehicle pollutant emission factor prediction method and system
CN113239351A (en) * 2020-12-08 2021-08-10 武汉大学 Novel data pollution attack defense method for Internet of things system
WO2021164404A1 (en) * 2020-02-20 2021-08-26 中国银联股份有限公司 Inspection method and apparatus
CN113312169A (en) * 2020-02-27 2021-08-27 香港理工大学深圳研究院 Method and device for distributing computing resources
CN113377797A (en) * 2021-07-02 2021-09-10 支付宝(杭州)信息技术有限公司 Method, device and system for jointly updating model
WO2021179196A1 (en) * 2020-03-11 2021-09-16 Oppo广东移动通信有限公司 Federated learning-based model training method, electronic device, and storage medium
CN113516250A (en) * 2021-07-13 2021-10-19 北京百度网讯科技有限公司 Method, device and equipment for federated learning and storage medium
CN113537633A (en) * 2021-08-09 2021-10-22 中国电信股份有限公司 Prediction method, device, equipment, medium and system based on longitudinal federal learning
CN113688855A (en) * 2020-05-19 2021-11-23 华为技术有限公司 Data processing method, federal learning training method, related device and equipment
CN113705823A (en) * 2020-05-22 2021-11-26 华为技术有限公司 Model training method based on federal learning and electronic equipment
CN113792632A (en) * 2021-09-02 2021-12-14 广州广电运通金融电子股份有限公司 Finger vein identification method, system and storage medium based on multi-party cooperation
CN113807157A (en) * 2020-11-27 2021-12-17 京东科技控股股份有限公司 Method, device and system for training neural network model based on federal learning
CN113837766A (en) * 2021-10-08 2021-12-24 支付宝(杭州)信息技术有限公司 Risk identification method and device and electronic equipment
WO2022057433A1 (en) * 2020-09-18 2022-03-24 华为技术有限公司 Machine learning model training method and related device
WO2022096919A1 (en) * 2020-11-05 2022-05-12 Telefonaktiebolaget Lm Ericsson (Publ) Managing training of a machine learning model
CN114548426A (en) * 2022-02-17 2022-05-27 北京百度网讯科技有限公司 Asynchronous federal learning method, business service prediction method, device and system
WO2022110248A1 (en) * 2020-11-30 2022-06-02 华为技术有限公司 Federated learning method, device and system
WO2022121032A1 (en) * 2020-12-10 2022-06-16 广州广电运通金融电子股份有限公司 Data set division method and system in federated learning scene
CN114662340A (en) * 2022-04-29 2022-06-24 烟台创迹软件有限公司 Weighing model scheme determination method and device, computer equipment and storage medium
CN114679332A (en) * 2022-04-14 2022-06-28 浙江工业大学 APT detection method of distributed system
CN114741611A (en) * 2022-06-08 2022-07-12 杭州金智塔科技有限公司 Federal recommendation model training method and system
CN114844653A (en) * 2022-07-04 2022-08-02 湖南密码工程研究中心有限公司 Credible federal learning method based on alliance chain
CN115196730A (en) * 2022-07-19 2022-10-18 南通派菲克水务技术有限公司 Intelligent sodium hypochlorite adding system for water plant
WO2023030221A1 (en) * 2021-08-30 2023-03-09 华为技术有限公司 Communication method and apparatus
CN116127417A (en) * 2023-04-04 2023-05-16 山东浪潮科学研究院有限公司 Code defect detection model construction method, device, equipment and storage medium
CN116245172A (en) * 2023-03-14 2023-06-09 南京航空航天大学 Coalition building game method facing individual model performance optimization in cross-island federal learning

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108898219A (en) * 2018-06-07 2018-11-27 广东工业大学 A kind of neural network training method based on block chain, device and medium
CN109189825A (en) * 2018-08-10 2019-01-11 深圳前海微众银行股份有限公司 Lateral data cutting federation learning model building method, server and medium
CN109635462A (en) * 2018-12-17 2019-04-16 深圳前海微众银行股份有限公司 Model parameter training method, device, equipment and medium based on federation's study
CN109871702A (en) * 2019-02-18 2019-06-11 深圳前海微众银行股份有限公司 Federal model training method, system, equipment and computer readable storage medium
US20190227980A1 (en) * 2018-01-22 2019-07-25 Google Llc Training User-Level Differentially Private Machine-Learned Models

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20190227980A1 (en) * 2018-01-22 2019-07-25 Google Llc Training User-Level Differentially Private Machine-Learned Models
CN108898219A (en) * 2018-06-07 2018-11-27 广东工业大学 A kind of neural network training method based on block chain, device and medium
CN109189825A (en) * 2018-08-10 2019-01-11 深圳前海微众银行股份有限公司 Lateral data cutting federation learning model building method, server and medium
CN109635462A (en) * 2018-12-17 2019-04-16 深圳前海微众银行股份有限公司 Model parameter training method, device, equipment and medium based on federation's study
CN109871702A (en) * 2019-02-18 2019-06-11 深圳前海微众银行股份有限公司 Federal model training method, system, equipment and computer readable storage medium

Cited By (84)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2022512053A (en) * 2019-11-22 2022-02-02 ベイジン バイドゥ ネットコム サイエンス テクノロジー カンパニー リミテッド Shared encoder generation methods, devices and electronic devices
JP7159460B2 (en) 2019-11-22 2022-10-24 ベイジン バイドゥ ネットコム サイエンス テクノロジー カンパニー リミテッド Shared encoder generation method, apparatus and electronic equipment
KR102532368B1 (en) * 2019-11-22 2023-05-15 베이징 바이두 넷컴 사이언스 앤 테크놀로지 코., 엘티디. Shared encoder generation method, device and electronic device
KR20210065069A (en) * 2019-11-22 2021-06-03 베이징 바이두 넷컴 사이언스 앤 테크놀로지 코., 엘티디. How to create a shared encoder, device and electronics
CN110956275B (en) * 2019-11-27 2021-04-02 支付宝(杭州)信息技术有限公司 Risk prediction and risk prediction model training method and device and electronic equipment
CN110956275A (en) * 2019-11-27 2020-04-03 支付宝(杭州)信息技术有限公司 Risk prediction and risk prediction model training method and device and electronic equipment
CN110990870A (en) * 2019-11-29 2020-04-10 上海能塔智能科技有限公司 Operation and maintenance, processing method, device, equipment and medium using model library
CN111027715B (en) * 2019-12-11 2021-04-02 支付宝(杭州)信息技术有限公司 Monte Carlo-based federated learning model training method and device
CN111027715A (en) * 2019-12-11 2020-04-17 支付宝(杭州)信息技术有限公司 Monte Carlo-based federated learning model training method and device
CN111046433A (en) * 2019-12-13 2020-04-21 支付宝(杭州)信息技术有限公司 Model training method based on federal learning
WO2021121029A1 (en) * 2019-12-20 2021-06-24 深圳前海微众银行股份有限公司 Training model updating method and system, and agent, server and computer-readable storage medium
CN111091200A (en) * 2019-12-20 2020-05-01 深圳前海微众银行股份有限公司 Updating method, system, agent, server and storage medium of training model
CN113128528A (en) * 2019-12-27 2021-07-16 无锡祥生医疗科技股份有限公司 Ultrasonic image deep learning distributed training system and training method
WO2021138877A1 (en) * 2020-01-09 2021-07-15 深圳前海微众银行股份有限公司 Vertical federated learning model training optimization method and apparatus, device, and medium
CN113191479A (en) * 2020-01-14 2021-07-30 华为技术有限公司 Method, system, node and storage medium for joint learning
TWI770749B (en) * 2020-02-20 2022-07-11 大陸商中國銀聯股份有限公司 Inspection method and device
WO2021164404A1 (en) * 2020-02-20 2021-08-26 中国银联股份有限公司 Inspection method and apparatus
CN111369042B (en) * 2020-02-27 2021-09-24 山东大学 Wireless service flow prediction method based on weighted federal learning
CN113312169B (en) * 2020-02-27 2023-12-19 香港理工大学深圳研究院 Computing resource allocation method and device
WO2021169577A1 (en) * 2020-02-27 2021-09-02 山东大学 Wireless service traffic prediction method based on weighted federated learning
CN113312169A (en) * 2020-02-27 2021-08-27 香港理工大学深圳研究院 Method and device for distributing computing resources
CN111369042A (en) * 2020-02-27 2020-07-03 山东大学 Wireless service flow prediction method based on weighted federal learning
CN111260061A (en) * 2020-03-09 2020-06-09 厦门大学 Differential noise adding method and system in federated learning gradient exchange
CN111260061B (en) * 2020-03-09 2022-07-19 厦门大学 Differential noise adding method and system in federated learning gradient exchange
WO2021179196A1 (en) * 2020-03-11 2021-09-16 Oppo广东移动通信有限公司 Federated learning-based model training method, electronic device, and storage medium
CN111383096A (en) * 2020-03-23 2020-07-07 中国建设银行股份有限公司 Fraud detection and model training method and device thereof, electronic equipment and storage medium
CN111460528B (en) * 2020-04-01 2022-06-14 支付宝(杭州)信息技术有限公司 Multi-party combined training method and system based on Adam optimization algorithm
CN111460528A (en) * 2020-04-01 2020-07-28 支付宝(杭州)信息技术有限公司 Multi-party combined training method and system based on Adam optimization algorithm
WO2021203919A1 (en) * 2020-04-08 2021-10-14 北京字节跳动网络技术有限公司 Method and apparatus for evaluating joint training model
CN111488995A (en) * 2020-04-08 2020-08-04 北京字节跳动网络技术有限公司 Method and apparatus for evaluating a joint training model
CN113688855A (en) * 2020-05-19 2021-11-23 华为技术有限公司 Data processing method, federal learning training method, related device and equipment
CN113688855B (en) * 2020-05-19 2023-07-28 华为技术有限公司 Data processing method, federal learning training method, related device and equipment
CN113705823A (en) * 2020-05-22 2021-11-26 华为技术有限公司 Model training method based on federal learning and electronic equipment
CN111695701B (en) * 2020-06-12 2021-08-13 上海富数科技有限公司 System for realizing data set construction processing based on federal learning and construction generation method thereof
CN111695701A (en) * 2020-06-12 2020-09-22 上海富数科技有限公司 System for realizing data set construction processing based on federal learning and construction generation method thereof
CN111678696A (en) * 2020-06-17 2020-09-18 南昌航空大学 Intelligent mechanical fault diagnosis method based on federal learning
CN111898764A (en) * 2020-06-23 2020-11-06 华为技术有限公司 Method, device and chip for federal learning
CN111814985A (en) * 2020-06-30 2020-10-23 平安科技(深圳)有限公司 Model training method under federated learning network and related equipment thereof
WO2021120676A1 (en) * 2020-06-30 2021-06-24 平安科技(深圳)有限公司 Model training method for federated learning network, and related device
CN111814985B (en) * 2020-06-30 2023-08-29 平安科技(深圳)有限公司 Model training method under federal learning network and related equipment thereof
WO2021120677A1 (en) * 2020-07-07 2021-06-24 平安科技(深圳)有限公司 Warehousing model training method and device, computer device and storage medium
CN111881966A (en) * 2020-07-20 2020-11-03 北京市商汤科技开发有限公司 Neural network training method, device, equipment and storage medium
CN111882133A (en) * 2020-08-03 2020-11-03 重庆大学 Prediction-based federated learning communication optimization method and system
CN112001500A (en) * 2020-08-13 2020-11-27 星环信息科技(上海)有限公司 Model training method, device and storage medium based on longitudinal federated learning system
CN112101528B (en) * 2020-09-17 2023-10-24 上海交通大学 Terminal contribution measurement method based on back propagation
CN112101528A (en) * 2020-09-17 2020-12-18 上海交通大学 Terminal contribution measurement method based on back propagation
WO2022057433A1 (en) * 2020-09-18 2022-03-24 华为技术有限公司 Machine learning model training method and related device
CN111931242B (en) * 2020-09-30 2021-02-19 国网浙江省电力有限公司电力科学研究院 Data sharing method, computer equipment applying same and readable storage medium
CN111931242A (en) * 2020-09-30 2020-11-13 国网浙江省电力有限公司电力科学研究院 Data sharing method, computer equipment applying same and readable storage medium
CN111931876A (en) * 2020-10-12 2020-11-13 支付宝(杭州)信息技术有限公司 Target data side screening method and system for distributed model training
US11449805B2 (en) 2020-10-12 2022-09-20 Alipay (Hangzhou) Information Technology Co., Ltd. Target data party selection methods and systems for distributed model training
CN112232519B (en) * 2020-10-15 2024-01-09 成都数融科技有限公司 Joint modeling method based on federal learning
CN112232519A (en) * 2020-10-15 2021-01-15 成都数融科技有限公司 Joint modeling method based on federal learning
WO2022096919A1 (en) * 2020-11-05 2022-05-12 Telefonaktiebolaget Lm Ericsson (Publ) Managing training of a machine learning model
CN113807157A (en) * 2020-11-27 2021-12-17 京东科技控股股份有限公司 Method, device and system for training neural network model based on federal learning
WO2022110248A1 (en) * 2020-11-30 2022-06-02 华为技术有限公司 Federated learning method, device and system
CN113239351B (en) * 2020-12-08 2022-05-13 武汉大学 Novel data pollution attack defense method for Internet of things system
CN113239351A (en) * 2020-12-08 2021-08-10 武汉大学 Novel data pollution attack defense method for Internet of things system
WO2022121032A1 (en) * 2020-12-10 2022-06-16 广州广电运通金融电子股份有限公司 Data set division method and system in federated learning scene
CN113158550A (en) * 2021-03-24 2021-07-23 北京邮电大学 Method and device for federated learning, electronic equipment and storage medium
CN113222211A (en) * 2021-03-31 2021-08-06 中国科学技术大学先进技术研究院 Multi-region diesel vehicle pollutant emission factor prediction method and system
CN113222211B (en) * 2021-03-31 2023-12-12 中国科学技术大学先进技术研究院 Method and system for predicting pollutant emission factors of multi-region diesel vehicle
CN112949837A (en) * 2021-04-13 2021-06-11 中国人民武装警察部队警官学院 Target recognition federal deep learning method based on trusted network
CN112949837B (en) * 2021-04-13 2022-11-11 中国人民武装警察部队警官学院 Target recognition federal deep learning method based on trusted network
CN113377797A (en) * 2021-07-02 2021-09-10 支付宝(杭州)信息技术有限公司 Method, device and system for jointly updating model
CN113516250A (en) * 2021-07-13 2021-10-19 北京百度网讯科技有限公司 Method, device and equipment for federated learning and storage medium
CN113516250B (en) * 2021-07-13 2023-11-03 北京百度网讯科技有限公司 Federal learning method, device, equipment and storage medium
CN113537633A (en) * 2021-08-09 2021-10-22 中国电信股份有限公司 Prediction method, device, equipment, medium and system based on longitudinal federal learning
CN113537633B (en) * 2021-08-09 2023-04-18 中国电信股份有限公司 Prediction method, device, equipment, medium and system based on longitudinal federal learning
WO2023030221A1 (en) * 2021-08-30 2023-03-09 华为技术有限公司 Communication method and apparatus
CN113792632A (en) * 2021-09-02 2021-12-14 广州广电运通金融电子股份有限公司 Finger vein identification method, system and storage medium based on multi-party cooperation
CN113837766A (en) * 2021-10-08 2021-12-24 支付宝(杭州)信息技术有限公司 Risk identification method and device and electronic equipment
CN114548426A (en) * 2022-02-17 2022-05-27 北京百度网讯科技有限公司 Asynchronous federal learning method, business service prediction method, device and system
CN114548426B (en) * 2022-02-17 2023-11-24 北京百度网讯科技有限公司 Asynchronous federal learning method, business service prediction method, device and system
CN114679332A (en) * 2022-04-14 2022-06-28 浙江工业大学 APT detection method of distributed system
CN114662340A (en) * 2022-04-29 2022-06-24 烟台创迹软件有限公司 Weighing model scheme determination method and device, computer equipment and storage medium
CN114662340B (en) * 2022-04-29 2023-02-28 烟台创迹软件有限公司 Weighing model scheme determination method and device, computer equipment and storage medium
CN114741611B (en) * 2022-06-08 2022-10-14 杭州金智塔科技有限公司 Federal recommendation model training method and system
CN114741611A (en) * 2022-06-08 2022-07-12 杭州金智塔科技有限公司 Federal recommendation model training method and system
CN114844653A (en) * 2022-07-04 2022-08-02 湖南密码工程研究中心有限公司 Credible federal learning method based on alliance chain
CN115196730A (en) * 2022-07-19 2022-10-18 南通派菲克水务技术有限公司 Intelligent sodium hypochlorite adding system for water plant
CN116245172A (en) * 2023-03-14 2023-06-09 南京航空航天大学 Coalition building game method facing individual model performance optimization in cross-island federal learning
CN116245172B (en) * 2023-03-14 2023-10-17 南京航空航天大学 Federation construction method for optimizing individual model performance in cross-island federation learning
CN116127417A (en) * 2023-04-04 2023-05-16 山东浪潮科学研究院有限公司 Code defect detection model construction method, device, equipment and storage medium

Similar Documents

Publication Publication Date Title
CN110442457A (en) Model training method, device and server based on federation's study
JP7296978B2 (en) Providing incentives to players to engage in competitive gameplay
US7537523B2 (en) Dynamic player groups for interest management in multi-character virtual environments
CN110189174A (en) A kind of mobile intelligent perception motivational techniques based on quality of data perception
CN110610242A (en) Method and device for setting participant weight in federated learning
CN111282267B (en) Information processing method, information processing apparatus, information processing medium, and electronic device
CN107483986A (en) A kind of method and system of gifts
Zhan et al. Incentive mechanism design for federated learning: Challenges and opportunities
EP3852014A1 (en) Method and apparatus for training learning model, and computing device
Ho et al. Towards social norm design for crowdsourcing markets
CN107229966A (en) A kind of model data update method, apparatus and system
CN106529189A (en) User classifying method, application server and application client-side
CN112990987B (en) Information popularization method and device, electronic equipment and storage medium
CN107944460A (en) One kind is applied to class imbalance sorting technique in bioinformatics
CN111957047A (en) Checkpoint configuration data adjusting method, computer equipment and storage medium
Xiao et al. Incentive mechanism design for federated learning: A two-stage stackelberg game approach
CN110533178A (en) A kind of neural network model training method, apparatus and system
CN114372589A (en) Federated learning method and related device
CN114021735A (en) Method and device for processing data in federated learning
CN104079472A (en) Method for mobile game products to achieve real-time social contact functions
CN116866345A (en) Computer enhancement to increase service growth rate
CN116339973A (en) Digital twin cloud platform computing resource scheduling method based on particle swarm optimization algorithm
CN109543725A (en) A kind of method and device obtaining model parameter
CN114530073A (en) Training method and device based on virtual reality
CN108984479B (en) Method for improving operating efficiency of crowdsourcing platform

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination