WO2023284387A1 - Model training method, apparatus, and system based on federated learning, and device and medium - Google Patents

Model training method, apparatus, and system based on federated learning, and device and medium Download PDF

Info

Publication number
WO2023284387A1
WO2023284387A1 PCT/CN2022/091868 CN2022091868W WO2023284387A1 WO 2023284387 A1 WO2023284387 A1 WO 2023284387A1 CN 2022091868 W CN2022091868 W CN 2022091868W WO 2023284387 A1 WO2023284387 A1 WO 2023284387A1
Authority
WO
WIPO (PCT)
Prior art keywords
model
algorithm parameters
local
updated
joint
Prior art date
Application number
PCT/CN2022/091868
Other languages
French (fr)
Chinese (zh)
Inventor
陈录城
诸葛慧玲
张成龙
孙明
贾淇超
李晓璐
Original Assignee
卡奥斯工业智能研究院(青岛)有限公司
海尔卡奥斯物联生态科技有限公司
海尔数字科技(青岛)有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 卡奥斯工业智能研究院(青岛)有限公司, 海尔卡奥斯物联生态科技有限公司, 海尔数字科技(青岛)有限公司 filed Critical 卡奥斯工业智能研究院(青岛)有限公司
Publication of WO2023284387A1 publication Critical patent/WO2023284387A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F21/00Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F21/60Protecting data
    • G06F21/62Protecting access to data via a platform, e.g. using keys or access control rules
    • G06F21/6218Protecting access to data via a platform, e.g. using keys or access control rules to a system of files or objects, e.g. local or distributed file system or database
    • G06F21/6245Protecting personal data, e.g. for financial or medical purposes

Definitions

  • the embodiments of the present application relate to the technical field of artificial intelligence, for example, to a model training method, device, system, electronic device, and storage medium based on federated learning.
  • Industrial data is the core of industrial informatization, especially manufacturing enterprises have a high degree of dependence on industrial data in production operations, such as process parameters, equipment operation data, production data and a series of industrial data are key data affecting manufacturing production, these Data security is directly related to the stable operation of the manufacturing production line. Data loss, malicious tampering or errors will lead to the shutdown of the entire production line and cause huge losses to industrial production. In addition, the leakage of industrial data of key enterprises related to the national economy and people's livelihood will also affect national security. Out of the protection of industrial data, manufacturing companies do not share or transmit data externally due to data security considerations, resulting in the formation of data islands, which has become a challenge for the implementation and continuous optimization of artificial intelligence technology in industrial scenarios. In the process of industrial intelligence, how to effectively protect data security and use sufficient data volume for model training and solve the problem of continuous optimization of artificial intelligence models has become the key to the development of manufacturing technology.
  • the embodiments of the present application provide a model training method, device, electronic equipment, and storage medium based on federated learning, so as to ensure the effect of model training while ensuring the security of industrial data.
  • the embodiment of the present application provides a model training method based on federated learning, which is executed by multiple private cloud servers, including: training the local model based on local data, and sending the algorithm parameters of the trained local model to the public cloud server, To enable the public cloud server to verify whether it is necessary to update the algorithm parameters of the joint model using the received algorithm parameters of the trained local model; receive the algorithm parameters of the updated joint model pushed by the public cloud server; verify whether The received algorithm parameters of the updated joint model need to be used to update the algorithm parameters of the trained local model; the received algorithm parameters of the updated joint model need to be used to update the trained local model. In the case of the algorithm parameters of the model, update the algorithm parameters of the trained local model to the received algorithm parameters of the updated joint model.
  • the verifying whether it is necessary to update the algorithm parameters of the trained local model by using the received algorithm parameters of the updated joint model includes: calculating the trained local model by using a priori data set.
  • the effect index of the model obtains the first index value; after replacing the algorithm parameters of the trained local model with the received algorithm parameters of the updated joint model, the algorithm is replaced by using the prior data set calculation
  • the effect index of the parameterized local model obtains a second index value; determine whether to use the received algorithm parameters of the updated joint model to update the Algorithm parameters of the trained local model.
  • the public cloud server verifying whether it is necessary to use the received algorithm parameters of the trained local model to update the algorithm parameters of the joint model includes: calculating the effect of the joint model using a priori data set The index obtains a third index value; after replacing the algorithm parameters of the joint model with the algorithm parameters of the received local model after training, using the prior data set to calculate the value of the joint model after replacing the algorithm parameters The effect index obtains a fourth index value; determine whether to use the received algorithm parameters of the trained local model to update the algorithm parameters of the joint model according to the magnitude of the third index value and the fourth index value.
  • the performance index includes precision and/or recall.
  • the training of the local model based on the local data before the training of the local model based on the local data, it also includes receiving the algorithm parameters of the initial model issued by the public cloud server; the training of the local model based on the local data includes: based on the initial Algorithmic parameters of the model and the local data train the local model.
  • the embodiment of the present application also provides a model training device based on federated learning, which is configured in multiple private cloud servers.
  • the device includes: a local training and parameter uploading unit, which is configured to train the local model based on local data.
  • the algorithm parameters of the trained local model are sent to the public cloud server, so that the public cloud server verifies whether the algorithm parameters of the joint model need to be updated with the received algorithm parameters;
  • the joint model parameter receiving unit is configured to receive the The algorithm parameters of the updated joint model pushed by the public cloud server;
  • the verification and update unit is configured to verify whether the received algorithm parameters of the updated joint model need to be used to update the algorithm parameters of the trained local model, In the case that the algorithm parameters of the trained local model need to be updated with the received algorithm parameters of the joint model, update the algorithm parameters of the trained local model to the received Algorithm parameters for the updated joint model.
  • the verifying and updating unit is configured to verify whether it is necessary to use the received algorithm parameters of the updated joint model to update the algorithm parameters of the trained local model, including: using a priori data set to calculate The effect index of the trained local model obtains a first index value; after replacing the algorithm parameters of the trained local model with the received algorithm parameters of the updated joint model, the prior data is used Set calculating the effect index of the local model after replacing the algorithm parameters to obtain a second index value; determine whether to use the received updated joint model according to the size of the first index value and the second index value Algorithm parameters for updating the algorithm parameters of the trained local model.
  • the verification by the public cloud server in the local training and parameter uploading unit whether the received algorithm parameters need to be used to update the algorithm parameters of the joint model includes: using a priori data set to calculate the algorithm parameters of the joint model The effect index obtains the third index value; after replacing the algorithm parameters of the joint model with the received algorithm parameters of the trained local model, the joint model after replacing the algorithm parameters is calculated by using the prior data set Obtaining the fourth index value of the effect index; determining whether to use the received algorithm parameters of the trained local model to update the algorithm parameters of the joint model according to the size of the third index value and the fourth index value .
  • the performance index includes precision and/or recall.
  • the device further includes an initial model parameter receiving unit configured to receive the algorithm parameters of the initial model issued by the public cloud server before the local model is trained based on the local data; the local training and The parameter uploading unit is configured to train the local model based on the algorithm parameters of the initial model and the local data.
  • a model training system based on federated learning including a public cloud server and multiple private cloud servers; the multiple private cloud servers train the local model based on local data, and send the algorithm parameters of the trained local model to To the public cloud server; the public cloud server verifies whether the received algorithm parameters of the trained local model need to be used to update the algorithm parameters of the joint model, and the received trained local model needs to be used for verification In the case of updating the algorithm parameters of the joint model using the algorithm parameters of the local model, update the algorithm parameters of the joint model by using the received algorithm parameters of the local model, and push the updated algorithm parameters of the joint model to the multiple Private cloud server; when the plurality of private cloud servers receive the algorithm parameters of the updated joint model pushed by the public cloud server, verify whether the received updated joint model needs to be used Algorithm parameters update the model parameters of the trained local model, and in the case that the received algorithm parameters of the updated joint model need to be used to update the trained local model algorithm parameters, the trained The algorithm parameters of the local model are updated to the received algorithm parameters of the
  • the system before the multiple private cloud servers train the local model based on local data, the system further includes: the public cloud server sends the algorithm parameters of the initial model to the multiple private cloud servers, The multiple private cloud servers train the local model based on the algorithm parameters of the initial model and the local data.
  • the electronic device includes: a processor; and a memory configured to store executable instructions, and when the executable instructions are executed by the processor, the electronic device executes the methods of the foregoing embodiments.
  • a computer-readable storage medium is also provided, on which a computer program is stored, and when the computer program is executed by a processor, the methods of the above-mentioned embodiments are implemented.
  • Fig. 1 is a schematic flowchart of a model training method based on federated learning provided according to an embodiment of the present application
  • FIG. 2 is a schematic flowchart of a method of a federated learning-based model training system provided according to an embodiment of the present application
  • FIG. 3A is a schematic diagram of another federated learning-based model training system method provided according to an embodiment of the present application.
  • FIG. 3B is a schematic flowchart of another federated learning-based model training system method provided according to an embodiment of the present application.
  • Fig. 4 is a schematic structural diagram of a model training device based on federated learning provided according to an embodiment of the present application
  • FIG. 5 is a schematic structural diagram of another model training device based on federated learning provided according to an embodiment of the present application.
  • FIG. 6 shows a schematic structural diagram of an electronic device suitable for implementing the embodiments of the present application.
  • system and “network” are often used interchangeably herein, and “and/or” mentioned herein refers to any and all combinations including one or more related listed items.
  • first, second, etc. in are used to distinguish different objects, not to limit a specific order.
  • FIG. 1 shows a schematic flowchart of a model training method based on federated learning provided by an embodiment of the present application. This embodiment is applicable to the situation where multiple private cloud servers train models through federated learning, and the method can be executed by a model training device based on federated learning configured on multiple private cloud servers. As shown in FIG. 1 , the model training method based on federated learning described in this embodiment includes the following steps.
  • step S110 the local model is trained based on local data, and the algorithm parameters of the trained local model are sent to the public cloud server, so that the public cloud server verifies whether the received trained local model needs to be adopted
  • the algorithm parameters of update the algorithm parameters of the joint model are sent to the public cloud server, so that the public cloud server verifies whether the received trained local model needs to be adopted.
  • the public cloud server verifies whether it is necessary to use the received algorithm parameters of the trained local model to update the algorithm parameters of the joint model, it can use a priori data set for verification to determine if the joint model uses the algorithm parameters of the private cloud server.
  • the algorithm parameters of the trained local model, and whether the training effect of the joint model is better for example, whether the model accuracy is higher).
  • the effect index of the joint model can be calculated by using the prior data set to obtain the third index value, and after the algorithm parameters of the joint model are replaced with the received algorithm parameters of the local model after training, the replacement is calculated using the prior data set
  • the fourth index value is obtained from the effect index of the joint model after the algorithm parameters, and it is determined whether it is necessary to use the received algorithm parameters of the trained local model to update the joint model according to the size of the third index value and the fourth index value. Algorithmic parameters for the model.
  • step S120 the algorithm parameters of the updated joint model pushed by the public cloud server are received.
  • step S130 verify whether it is necessary to use the received algorithm parameters of the updated joint model to update the algorithm parameters of the trained local model, if necessary, use the received algorithm parameters of the updated joint model
  • step S140 executes step S140, if it is not necessary to update the algorithm parameters of the trained local model by using the received algorithm parameters of the updated joint model, then return to step S110 .
  • a priori data set may be used to calculate the effect index of the trained local model to obtain a first index value, and the algorithm parameters of the trained local model may be replaced with the received algorithm of the updated joint model After parameterizing, use the prior data set to calculate the effect index of the local model after replacing the algorithm parameters to obtain the second index value, and determine whether to use the received
  • the algorithm parameters of the updated joint model update the algorithm parameters of the trained local model.
  • step S140 the algorithm parameters of the trained local model are updated to the received algorithm parameters of the updated joint model.
  • the effect index may include various types, including but not limited to the accuracy rate of model prediction, recall rate, and the like.
  • multiple private cloud servers before training the local model based on local data, can also receive the algorithm parameters of the initial model issued by the public cloud server, based on the algorithm parameters of the initial model and The local data trains the local model to synchronize the initial state of the local model on multiple private cloud servers.
  • multiple private cloud servers are used to train the local model based on local data, and the algorithm parameters of the trained local model are sent to the public cloud server, so that the public cloud server verifies whether the received algorithm needs to be used Parameters update the algorithm parameters of the joint model; receive the algorithm parameters of the updated joint model pushed by the public cloud server; verify whether the received algorithm parameters of the updated joint model need to be used to update the trained local model ; If the verification needs to use the received algorithm parameters of the updated joint model to update the trained local model, update the algorithm parameters of the trained local model to the received updated joint model
  • the algorithm parameters of the model can ensure the effect of model training while ensuring the security of industrial data.
  • FIG. 2 is a schematic flowchart of a model training method of a federated learning-based model training system provided according to an embodiment of the present application.
  • the model training system based on federated learning described in this embodiment includes a public cloud server and multiple private cloud servers.
  • the model training method of the federated learning-based model training system described in this embodiment includes the following steps.
  • step S210 multiple private cloud servers train local models based on local data, and send algorithm parameters of the trained local models to the public cloud server.
  • step S220 the public cloud server verifies whether it is necessary to use the received algorithm parameters of the trained local model to update the algorithm parameters of the joint model, and if it needs to use the received algorithm parameters of the trained local model Updating the algorithm parameters of the joint model uses the received algorithm parameters of the trained local model to update the algorithm parameters of the joint model, and pushes the updated algorithm parameters of the joint model to multiple private cloud servers.
  • step S230 if multiple private cloud servers receive the algorithm parameters of the updated joint model pushed by the public cloud server, step S240 is executed.
  • step S240 verify whether it is necessary to use the received algorithm parameters of the updated joint model to update the algorithm parameters of the trained local model, if necessary, use the received algorithm parameters of the updated joint model
  • step S250 executes step S250, and return to step S210 if the received algorithm parameters of the updated joint model do not need to be used to update the algorithm parameters of the trained local model .
  • step S250 the algorithm parameters of the trained local model are updated to the received algorithm parameters of the updated joint model.
  • the public cloud server may also send the algorithm parameters of the initial model to multiple private cloud servers, and the multiple private cloud servers are based on the algorithm parameters of the initial model and local data to train the local model.
  • the technical solution of this embodiment satisfies the continuous optimization of industrial data models and improves the use effect of artificial intelligence (AI) technology in industrial applications without the need for open data.
  • AI artificial intelligence
  • FIG. 3A is a schematic diagram of a model training method of another federated learning-based model training system provided according to an embodiment of the present application.
  • This embodiment is a safe data model training solution that performs local model training on a private cloud server and integrates and optimizes the models in multiple places into a joint model, and then feeds back to the respective private cloud servers in multiple places.
  • this embodiment mainly uses technologies such as federated learning, distributed computing, and algorithm model integration and optimization.
  • Federated learning is a paradigm of distributed collaborative training of machine learning models, which can be used for collaborative training of machine learning models on a large number of edge devices (clients) without centralized training data. It is characterized by a large number of decentralized stages linked to centralized server, these participants have zero trust in each other and only have access to local training data.
  • Distributed computing is to split a large computing task into multiple small computing tasks and distribute them to multiple machines for calculation, and then summarize the results.
  • Distributed computing in the federated learning process is to perform calculations in multiple places to form a data model, and then Upload the model to summarize the results.
  • the integration and optimization technology of the algorithm model is based on the definition of the same parameter, through federated learning and distributed computing, and integrates local computing and models into a joint model.
  • the global model is initialized and maintained by a central parameter server, which is then shared to edge devices.
  • the client uses local private data to calculate and update the model, and then uploads the updated model to the server, while keeping the privacy-sensitive training data on its own device, through distributed security aggregation Over multiple iterations, the federated learning system trains the integrated model.
  • FIG. 3B shows a schematic flowchart of another model training method based on federated learning provided by the embodiment of the present application. This embodiment is based on the foregoing embodiments, and an improved description is made. As shown in FIG. 3B , the model training method based on federated learning described in this embodiment includes the following steps.
  • step S301 the factory private cloud starts model training.
  • step S302 the factory private cloud shares the algorithm parameters of the model with the joint model.
  • step S303 whether the joint model verification needs to update its algorithm parameters, if the joint model verification needs to update the algorithm parameters, then perform step S305, if the joint model verification does not need to update the algorithm parameters, then perform step S304.
  • step S304 the algorithm parameters of the joint model are not updated, and step S306 is executed.
  • step S305 the joint parameters of the joint model are updated.
  • step S306 the joint model pushes the algorithm parameters of the joint model to the factory private cloud.
  • step S307 the factory private cloud judges whether the joint model is optimal, if the joint model is optimal, execute step S310, and if the joint model is not optimal, execute step S308.
  • step S308 the model is not updated, and step S309 is executed.
  • step S309 continue the local training and end.
  • step S310 the local model is updated iteratively.
  • the private cloud of the factory starts the local algorithm model training and shares the trained model with the joint model.
  • the joint model needs to perform model verification before receiving the model shared by multiple factories. If the verification does not need to update the algorithm parameters of the joint model, the algorithm will not be updated Parameters, the factory private cloud uses the information shared by multiple factories to optimize and update the model, and when it corresponds to the sharing of multiple factories, it will use the shared information of multiple factories to optimize the model.
  • the joint model will regularly push the joint model to (multiple) factories, and the factory will verify the model after receiving the push. If the pushed joint model is better than the local model, the local model will be iteratively updated, otherwise the local model will be updated continuously. local training.
  • the push process is that multiple local models push the updated algorithm parameters after learning to the joint model in real time during the continuous learning process to update the algorithm parameters.
  • the joint model is updated according to the real-time algorithm parameters, the updated joint model algorithm
  • the parameters are reversely pushed to multiple factories. These parameters are based on unified definition rules, and the algorithm model of the factory is continuously pushed iteratively with the update of the joint model.
  • the technical solution of this embodiment mainly uses technologies such as federated learning, distributed computing, and algorithm model joint integration.
  • the joint model can be placed on the public cloud. There is no need to share data in multiple places, but only the data model.
  • the joint model is continuously optimized through the public cloud. Shared with multiple regions for continuous model optimization. The entire process data does not leave the factory, and the training results of the local data model are shared to improve the use effect of industrial AI technology. While ensuring the security of industrial data, it satisfies the continuous optimization of industrial data models and improves the use effect of AI technology in industrial applications without the need for open data.
  • this application provides an embodiment of a model training device based on federated learning.
  • Figure 4 shows the structure of a model training device based on federated learning provided in this embodiment Schematic diagram, the device embodiment corresponds to the method embodiment shown in FIG. 1 , and the device can be applied to various electronic devices in multiple private cloud servers.
  • the model training device based on federated learning described in this embodiment includes a local training and parameter uploading unit 410 , a joint model parameter receiving unit 420 and a verification and updating unit 430 .
  • the local training and parameter uploading unit 410 is configured to train the local model based on local data, and send the algorithm parameters of the trained local model to the public cloud server, so that the public cloud server verifies whether the received The algorithm parameters of the update joint model algorithm parameters.
  • the joint model parameter receiving unit 420 is configured to receive the updated algorithm parameters of the joint model pushed by the public cloud server.
  • the verification and updating unit 430 is configured to verify whether it is necessary to use the received algorithm parameters of the updated joint model to update the algorithm parameters of the trained local model, and if the verification needs to use the received update If the algorithm parameters of the updated joint model update the algorithm parameters of the trained local model, the algorithm parameters of the trained local model are updated to the received algorithm parameters of the updated joint model.
  • the verification and updating unit 430 is configured to: use the prior data set to calculate the effect index of the trained local model to obtain a first index value; After the algorithm parameters of the updated local model are replaced with the received algorithm parameters of the updated joint model, the effect index of the local model after replacing the algorithm parameters is calculated using the prior data set to obtain a second index value; according to the The values of the first index value and the second index value determine whether to use the received algorithm parameters of the updated joint model to update the algorithm parameters of the trained local model.
  • the verification by the public cloud server in the local training and parameter uploading unit 410 whether it is necessary to use the received algorithm parameters to update the algorithm parameters of the joint model includes: using a priori data set Calculating the effect index of the joint model to obtain the third index value; after replacing the algorithm parameters of the joint model with the received algorithm parameters of the trained local model, using the prior data set to calculate the joint model after replacing the algorithm parameters Obtaining a fourth index value based on the effect index of the result index; determining whether to use the received algorithm parameters of the trained local model to update the algorithm parameters of the joint model according to the magnitude of the third index value and the fourth index value.
  • the performance index includes precision and/or recall.
  • the device further includes an initial model parameter receiving unit configured to receive the algorithm parameters of the initial model issued by the public cloud server before training the local model based on the local data;
  • the local training and parameter uploading unit being configured to train the local model based on local data includes: training the local model based on algorithm parameters of the initial model and local data.
  • the model training device based on federated learning provided in this embodiment can execute the model training method based on federated learning provided in the method embodiment of the present disclosure, and has corresponding functional modules for executing the method.
  • FIG. 5 shows a schematic structural diagram of another model training device based on federated learning provided by an embodiment of the present application.
  • the model training device based on federated learning in this embodiment includes an initial model parameter receiving unit 510 , a local training and parameter uploading unit 520 , a joint model parameter receiving unit 530 and a verification and updating unit 540 .
  • the initial model parameter receiving unit 510 is configured to receive the algorithm parameters of the initial model issued by the public cloud server.
  • the local training and parameter uploading unit 520 is configured such that training the local model based on local data includes: training the local model based on the algorithm parameters of the initial model and local data, and sending the algorithm parameters of the trained local model to to the public cloud server, so that the public cloud server verifies whether the received algorithm parameters need to be used to update the algorithm parameters of the joint model.
  • the joint model parameter receiving unit 530 is configured to receive the updated algorithm parameters of the joint model pushed by the public cloud server.
  • the verification and updating unit 540 is configured to, if the verification needs to use the received algorithm parameters of the updated joint model to update the algorithm parameters of the trained local model, then update the algorithm parameters of the trained local model to The algorithm parameters are updated to the received algorithm parameters of the updated joint model.
  • the verification and updating unit 540 is configured to verify whether it is necessary to use the received algorithm parameters of the updated joint model to update the algorithm parameters of the trained local model including : Using the prior data set to calculate the effect index of the trained local model to obtain the first index value; after replacing the algorithm parameters of the trained local model with the received algorithm parameters of the updated joint model , using the prior data set to calculate the effect index of the local model after replacing the algorithm parameters to obtain a second index value; determine whether to use the received The updated algorithm parameters of the joint model update the algorithm parameters of the trained local model.
  • the verification of whether the public cloud server in the local training and parameter uploading sheet 520 needs to use the received algorithm parameters to update the algorithm parameters of the joint model includes: using a priori data set Calculating the effect index of the joint model to obtain the third index value; after replacing the algorithm parameters of the joint model with the received algorithm parameters of the trained local model, using the prior data set to calculate the joint model after replacing the algorithm parameters Obtaining a fourth index value based on the effect index of the method; determining whether to use the received algorithm parameters of the trained local model to update the algorithm parameters of the joint model according to the magnitude of the third index value and the fourth index value.
  • the performance index includes precision and/or recall.
  • the model training device based on federated learning provided in this embodiment can execute the model training method based on federated learning provided in the method embodiment of the present disclosure, and has corresponding functional modules for executing the method.
  • FIG. 6 it shows a schematic structural diagram of an electronic device 600 suitable for implementing the embodiment of the present application.
  • the above-mentioned terminal device in the embodiment of the present application is, for example, a mobile device, a computer, or a vehicle-mounted device built in a floating car, or any combination thereof.
  • the mobile device may include, for example, a mobile phone, a smart home device, a wearable device, a smart mobile device, a virtual reality device, etc., or any combination thereof.
  • the electronic device shown in FIG. 6 is only an example, and should not limit the functions and scope of use of this embodiment of the present application.
  • an electronic device 600 may include a processing device (such as a central processing unit, a graphics processing unit, etc.) Various appropriate actions and processes are executed by a program loaded into a random access memory (Random Access Memory, RAM) 603 by 608 . In the RAM 603, various programs and data necessary for the operation of the electronic device 600 are also stored.
  • the processing device 601, ROM 602, and RAM 603 are connected to each other through a bus 604.
  • An input/output (Input/Output, I/O) interface 605 is also connected to the bus 604 .
  • the following devices can be connected to the I/O interface 605: input devices 606 including, for example, a touch screen, touchpad, keyboard, mouse, camera, microphone, accelerometer, gyroscope, etc.; including, for example, a liquid crystal display (LCD), speaker, vibration an output device 607 such as a computer; a storage device 608 including, for example, a magnetic tape, a hard disk, etc.; and a communication device 609.
  • the communication means 609 may allow the electronic device 600 to communicate with other devices wirelessly or by wire to exchange data. While FIG. 6 shows electronic device 600 having various means, it should be understood that implementing or possessing all of the means shown is not a requirement. More or fewer means may alternatively be implemented or provided.
  • the processes described above with reference to the flowcharts may be implemented as computer software programs.
  • the embodiments of the present application include a computer program product, which includes a computer program carried on a computer-readable medium, where the computer program includes program code for executing the method shown in the flowchart.
  • the computer program may be downloaded and installed from a network via communication means 609, or from storage means 608, or from ROM 602.
  • the processing device 601 When the computer program is executed by the processing device 601, the above-mentioned functions defined in the method of the embodiment of the present application are performed.
  • the above-mentioned computer-readable medium in the embodiment of the present application may be a computer-readable signal medium or a computer-readable storage medium, or any combination of the two.
  • a computer readable storage medium may be, for example, but not limited to, an electrical, magnetic, optical, electromagnetic, infrared, or semiconductor system, device, or device, or any combination thereof.
  • Examples of computer readable storage media may include, but are not limited to: electrical connections with one or more wires, portable computer disks, hard disks, RAM, ROM, Erasable Programmable Read-Only Memory, EPROM) or flash memory, optical fiber, portable compact disk read-only memory (Compact Disc Read-Only Memory, CD-ROM), optical storage device, magnetic storage device, or any suitable combination of the above.
  • a computer-readable storage medium may be any tangible medium containing or storing a program, and the program may be used by or in combination with an instruction execution system, device or device.
  • the computer-readable signal medium may include a data signal propagated in the baseband or as a part of the carrier wave, and the computer-readable program code is carried therein.
  • propagated data signals may take many forms, including but not limited to electromagnetic signals, optical signals, or any suitable combination of the foregoing.
  • a computer-readable signal medium may also be any computer-readable medium other than a computer-readable storage medium, which can transmit, propagate, or transmit a program for use by or in conjunction with an instruction execution system, apparatus, or device .
  • the program code contained on the computer readable medium can be transmitted by any appropriate medium, including but not limited to: electric wire, optical cable, radio frequency (Radio Frequency, RF), etc., or any suitable combination of the above.
  • the above-mentioned computer-readable medium may be included in the above-mentioned electronic device, or may exist independently without being incorporated into the electronic device.
  • the above-mentioned computer-readable medium carries one or more programs, and when the above-mentioned one or more programs are executed by the electronic device, the electronic device: trains the local model based on local data, and converts the algorithm parameters of the trained local model to Send to the public cloud server, so that the public cloud server verifies whether the algorithm parameters of the joint model need to be updated by using the received algorithm parameters of the trained local model; receive the updated joint model pushed by the public cloud server Algorithm parameters; verify whether it is necessary to use the received algorithm parameters of the updated joint model to update the algorithm parameters of the trained local model, if the verification needs to use the received algorithm parameters of the updated joint model Updating the algorithm parameters of the trained local model is updating the algorithm parameters of the trained local model to the received algorithm parameters of the updated joint model.
  • Computer program codes for performing the operations of the embodiments of the present application may be written in one or more programming languages or a combination thereof, and the above-mentioned programming languages include object-oriented programming languages—such as Java, Smalltalk, C++, and A conventional procedural programming language - such as the "C" language or a similar programming language.
  • the program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server.
  • the remote computer may be connected to the user computer via any kind of network, including a Local Area Network (LAN) or a Wide Area Network (WAN), or, alternatively, may be connected to an external computer (e.g., via the Internet) service provider via Internet connection).
  • LAN Local Area Network
  • WAN Wide Area Network
  • an external computer e.g., via the Internet
  • each block in a flowchart or block diagram may represent a module, program segment, or portion of code that contains one or more logical functions for implementing specified executable instructions.
  • the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or they may sometimes be executed in the reverse order, depending upon the functionality involved.
  • each block of the block diagrams and/or flowchart illustrations, and combinations of blocks in the block diagrams and/or flowchart illustrations can be implemented by a dedicated hardware-based system that performs the specified functions or operations , or may be implemented by a combination of dedicated hardware and computer instructions.
  • the units involved in the embodiments described in the present application may be implemented by means of software or by means of hardware.
  • the name of the unit does not limit the unit itself in some cases, for example, the first obtaining unit may also be described as "a unit for obtaining at least two Internet Protocol addresses".

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Software Systems (AREA)
  • General Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • Medical Informatics (AREA)
  • General Health & Medical Sciences (AREA)
  • Physics & Mathematics (AREA)
  • Bioethics (AREA)
  • General Engineering & Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • Mathematical Physics (AREA)
  • Evolutionary Computation (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Computing Systems (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • Computer Hardware Design (AREA)
  • Computer Security & Cryptography (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The embodiments of the present application relate to a model training method, apparatus, and system based on federated learning, and a device and a medium. The method comprises: implementing training of a local model on the basis of local data, and sending algorithm parameters of the trained local model to a public cloud server, such that the public cloud server verifies whether the algorithm parameters of a joint model need to be updated using the received algorithm parameters of the trained local model; receiving updated algorithm parameters of the joint model pushed by the public cloud server; verifying whether the algorithm parameters of the trained local model need to be updated using the received updated algorithm parameters of the joint model; and, when verifying that the algorithm parameters of the trained local model need to be updated using the received updated algorithm parameters of the joint model, updating the algorithm parameters of the trained local model to the received updated algorithm parameters of the joint model.

Description

基于联邦学习的模型训练方法、装置、系统、设备和介质Model training method, device, system, equipment and medium based on federated learning
本申请要求在2021年07月15日提交中国专利局、申请号为202110799024.8的中国专利申请的优先权,该申请的全部内容通过引用结合在本申请中。This application claims the priority of the Chinese patent application with application number 202110799024.8 submitted to the China Patent Office on July 15, 2021, the entire content of which is incorporated herein by reference.
技术领域technical field
本申请实施例涉及人工智能技术领域,例如涉及一种基于联邦学习的模型训练方法、装置、系统、电子设备、及存储介质。The embodiments of the present application relate to the technical field of artificial intelligence, for example, to a model training method, device, system, electronic device, and storage medium based on federated learning.
背景技术Background technique
工业数据是工业信息化的核心,尤其是制造业企业在生产运营中对工业数据有着高度的依赖,例如工艺参数、设备运行数据、生产数据等一系列工业数据是影响制造生产的关键数据,这些数据的安全直接关系到制造生产线的稳定运行,数据的丢失、恶意篡改或者错误都会导致整条产线的停产,对工业生产造成巨大的损失。此外,关乎国计民生的关键企业,工业数据的外泄也会影响到国家安全。出于对工业数据的保护,制造企业出于数据安全的考虑,对外不共享也不传输数据,导致数据孤岛形成,这成为当前人工智能技术在工业场景落地和持续优化的挑战。在工业智能化过程中,如何有效的既保护数据安全也能够使用充足的数据量进行模型训练,解决人工智能模型持续优化的问题成为制造业技术发展的关键所在。Industrial data is the core of industrial informatization, especially manufacturing enterprises have a high degree of dependence on industrial data in production operations, such as process parameters, equipment operation data, production data and a series of industrial data are key data affecting manufacturing production, these Data security is directly related to the stable operation of the manufacturing production line. Data loss, malicious tampering or errors will lead to the shutdown of the entire production line and cause huge losses to industrial production. In addition, the leakage of industrial data of key enterprises related to the national economy and people's livelihood will also affect national security. Out of the protection of industrial data, manufacturing companies do not share or transmit data externally due to data security considerations, resulting in the formation of data islands, which has become a challenge for the implementation and continuous optimization of artificial intelligence technology in industrial scenarios. In the process of industrial intelligence, how to effectively protect data security and use sufficient data volume for model training and solve the problem of continuous optimization of artificial intelligence models has become the key to the development of manufacturing technology.
为保障工业企业数据安全,大多采用私有云方案,数据不出厂,数据安全可靠,但由于数据不开放,不共享,导致工业数据分散,优质数据少,影响了数据模型训练和优化,不足以支撑人工智能技术落地实现。In order to ensure the data security of industrial enterprises, most of them adopt private cloud solutions. The data does not leave the factory, and the data is safe and reliable. However, because the data is not open and shared, the industrial data is scattered and the high-quality data is scarce, which affects the data model training and optimization, which is not enough to support The implementation of artificial intelligence technology.
发明内容Contents of the invention
本申请实施例提供一种基于联邦学习的模型训练方法、装置、电子设备、及存储介质,以在保证工业数据安全的同时,保证了模型训练的效果。The embodiments of the present application provide a model training method, device, electronic equipment, and storage medium based on federated learning, so as to ensure the effect of model training while ensuring the security of industrial data.
本申请实施例提供了一种基于联邦学习的模型训练方法,由多个私有云端服务器执行,包括:基于本地数据对本地模型进行训练,将训练后的本地模型的算法参数发送给公有云端服务器,以使所述公有云端服务器验证是否需要采用所接收的所述训练后的本地模型的算法参数更新联合模型的算法参数;接收所述公有云端服务器推送的更新后的联合模型的算法参数;验证是否需要采用 所接收的所述更新后的联合模型的算法参数更新所述训练后的本地模型的算法参数;在验证需要采用所接收的所述更新的联合模型的算法参数更新所述训练后的本地模型的算法参数的情况下,将所述训练后的本地模型的算法参数更新为所接收的所述更新后的联合模型的算法参数。The embodiment of the present application provides a model training method based on federated learning, which is executed by multiple private cloud servers, including: training the local model based on local data, and sending the algorithm parameters of the trained local model to the public cloud server, To enable the public cloud server to verify whether it is necessary to update the algorithm parameters of the joint model using the received algorithm parameters of the trained local model; receive the algorithm parameters of the updated joint model pushed by the public cloud server; verify whether The received algorithm parameters of the updated joint model need to be used to update the algorithm parameters of the trained local model; the received algorithm parameters of the updated joint model need to be used to update the trained local model. In the case of the algorithm parameters of the model, update the algorithm parameters of the trained local model to the received algorithm parameters of the updated joint model.
于一实施例中,所述验证是否需要采用所接收的所述更新后的联合模型的算法参数更新所述训练后的本地模型的算法参数包括:采用先验数据集计算所述训练后的本地模型的效果指标得到第一指标值;将所述训练后的本地模型的算法参数替换为所接收的所述更新后的联合模型的算法参数后,采用所述先验数据集计算替换所述算法参数后的本地模型的效果指标得到第二指标值;根据所述第一指标值和所述第二指标值的大小确定是否需要采用所接收的所述更新后的联合模型的算法参数更新所述训练后的本地模型的算法参数。In an embodiment, the verifying whether it is necessary to update the algorithm parameters of the trained local model by using the received algorithm parameters of the updated joint model includes: calculating the trained local model by using a priori data set. The effect index of the model obtains the first index value; after replacing the algorithm parameters of the trained local model with the received algorithm parameters of the updated joint model, the algorithm is replaced by using the prior data set calculation The effect index of the parameterized local model obtains a second index value; determine whether to use the received algorithm parameters of the updated joint model to update the Algorithm parameters of the trained local model.
于一实施例中,所述公有云端服务器验证是否需要采用所接收的所述所述训练后的本地模型的算法参数更新联合模型的算法参数包括:采用先验数据集计算所述联合模型的效果指标得到第三指标值;将所述联合模型的算法参数替换为所接收的所述训练后的本地模型的算法参数后,采用所述先验数据集计算替换所述算法参数后的联合模型的效果指标得到第四指标值;根据所述第三指标值和所述第四指标值的大小确定是否需要采用所接收的所述训练后的本地模型的算法参数更新所述联合模型的算法参数。In one embodiment, the public cloud server verifying whether it is necessary to use the received algorithm parameters of the trained local model to update the algorithm parameters of the joint model includes: calculating the effect of the joint model using a priori data set The index obtains a third index value; after replacing the algorithm parameters of the joint model with the algorithm parameters of the received local model after training, using the prior data set to calculate the value of the joint model after replacing the algorithm parameters The effect index obtains a fourth index value; determine whether to use the received algorithm parameters of the trained local model to update the algorithm parameters of the joint model according to the magnitude of the third index value and the fourth index value.
于一实施例中,所述效果指标包括准确率和/或召回率。In one embodiment, the performance index includes precision and/or recall.
于一实施例中,在所述基于本地数据对本地模型进行训练之前还包括,接收公有云端服务器下发的初始模型的算法参数;所述基于本地数据对本地模型进行训练包括:基于所述初始模型的算法参数和所述本地数据对所述本地模型进行训练。In one embodiment, before the training of the local model based on the local data, it also includes receiving the algorithm parameters of the initial model issued by the public cloud server; the training of the local model based on the local data includes: based on the initial Algorithmic parameters of the model and the local data train the local model.
本申请实施例还提供了一种基于联邦学习的模型训练装置,配置于多个私有云端服务器中,所述装置包括:本地训练与参数上传单元,设置为基于本地数据对本地模型进行训练,将训练后的本地模型的算法参数发送给公有云端服务器,以使所述公有云端服务器验证是否需要采用所接收的所述算法参数更新联合模型的算法参数;联合模型参数接收单元,设置为接收所述公有云端服务器推送的更新后的联合模型的算法参数;验证与更新单元,设置为验证是否需要采用所接收的所述更新后的联合模型的算法参数更新所述训练后的本地模型的算法参数,在验证需要采用所接收的所述更新后的联合模型的算法参数更新所述训练后的本地模型的算法参数的情况下,将所述训练后的本地模型的算法参数更新为所接收的所述更新后的联合模型的算法参数。The embodiment of the present application also provides a model training device based on federated learning, which is configured in multiple private cloud servers. The device includes: a local training and parameter uploading unit, which is configured to train the local model based on local data. The algorithm parameters of the trained local model are sent to the public cloud server, so that the public cloud server verifies whether the algorithm parameters of the joint model need to be updated with the received algorithm parameters; the joint model parameter receiving unit is configured to receive the The algorithm parameters of the updated joint model pushed by the public cloud server; the verification and update unit is configured to verify whether the received algorithm parameters of the updated joint model need to be used to update the algorithm parameters of the trained local model, In the case that the algorithm parameters of the trained local model need to be updated with the received algorithm parameters of the joint model, update the algorithm parameters of the trained local model to the received Algorithm parameters for the updated joint model.
于一实施例中,所述验证与更新单元设置为验证是否需要采用所接收的所 述更新后的联合模型的算法参数更新所述训练后的本地模型的算法参数包括:采用先验数据集计算所述训练后的本地模型的效果指标得到第一指标值;将所述训练后的本地模型的算法参数替换为所接收的所述更新后的联合模型的算法参数后,采用所述先验数据集计算替换所述算法参数后的本地模型的效果指标得到第二指标值;根据所述第一指标值和所述第二指标值的大小确定是否需要采用所接收的所述更新后的联合模型的算法参数更新所述训练后的本地模型的算法参数。In an embodiment, the verifying and updating unit is configured to verify whether it is necessary to use the received algorithm parameters of the updated joint model to update the algorithm parameters of the trained local model, including: using a priori data set to calculate The effect index of the trained local model obtains a first index value; after replacing the algorithm parameters of the trained local model with the received algorithm parameters of the updated joint model, the prior data is used Set calculating the effect index of the local model after replacing the algorithm parameters to obtain a second index value; determine whether to use the received updated joint model according to the size of the first index value and the second index value Algorithm parameters for updating the algorithm parameters of the trained local model.
于一实施例中,所述本地训练与参数上传单元中所述公有云端服务器验证是否需要采用所接收的所述算法参数更新联合模型的算法参数包括:采用先验数据集计算所述联合模型的效果指标得到第三指标值;将所述联合模型的算法参数替换为所接收的所述训练后的本地模型的算法参数后,采用所述先验数据集计算替换所述算法参数后的联合模型的效果指标得到第四指标值;根据所述第三指标值和所述第四指标值的大小确定是否需要采用所接收的所述训练后的本地模型的算法参数更新所述联合模型的算法参数。In one embodiment, the verification by the public cloud server in the local training and parameter uploading unit whether the received algorithm parameters need to be used to update the algorithm parameters of the joint model includes: using a priori data set to calculate the algorithm parameters of the joint model The effect index obtains the third index value; after replacing the algorithm parameters of the joint model with the received algorithm parameters of the trained local model, the joint model after replacing the algorithm parameters is calculated by using the prior data set Obtaining the fourth index value of the effect index; determining whether to use the received algorithm parameters of the trained local model to update the algorithm parameters of the joint model according to the size of the third index value and the fourth index value .
于一实施例中,所述效果指标包括准确率和/或召回率。In one embodiment, the performance index includes precision and/or recall.
于一实施例中,所述装置还包括初始模型参数接收单元,设置为在所述基于本地数据对本地模型进行训练之前,接收公有云端服务器下发的初始模型的算法参数;所述本地训练与参数上传单元是设置为基于所述初始模型的算法参数和所述本地数据对所述本地模型进行训练。In one embodiment, the device further includes an initial model parameter receiving unit configured to receive the algorithm parameters of the initial model issued by the public cloud server before the local model is trained based on the local data; the local training and The parameter uploading unit is configured to train the local model based on the algorithm parameters of the initial model and the local data.
还提供了一种基于联邦学习的模型训练系统,包括公有云端服务器和多个私有云端服务器;所述多个私有云端服务器基于本地数据对本地模型进行训练,将训练后的本地模型的算法参数发送给所述公有云端服务器;所述公有云端服务器验证是否需要采用所接收的所述训练后的本地模型的算法参数更新联合模型的算法参数,在验证需要采用所接收的所述训练后的本地模型的算法参数更新所述联合模型的算法参数的情况下,采用所接收的所述本地模型的算法参数更新所述联合模型的算法参数,将更新后的联合模型的算法参数推送到所述多个私有云端服务器;在所述多个私有云端服务器接收到所述公有云端服务器推送的所述更新后的联合模型的算法参数的情况下,验证是否需要采用所接收的所述更新后的联合模型的算法参数更新所述训练后的本地模型的模型参数,在验证需要采用所接收的所述更新后的联合模型的算法参数更新所述训练后的本地模型算法参数的情况下,将所述训练后的本地模型的算法参数更新为所接收的所述更新后的联合模型的算法参数。Also provided is a model training system based on federated learning, including a public cloud server and multiple private cloud servers; the multiple private cloud servers train the local model based on local data, and send the algorithm parameters of the trained local model to To the public cloud server; the public cloud server verifies whether the received algorithm parameters of the trained local model need to be used to update the algorithm parameters of the joint model, and the received trained local model needs to be used for verification In the case of updating the algorithm parameters of the joint model using the algorithm parameters of the local model, update the algorithm parameters of the joint model by using the received algorithm parameters of the local model, and push the updated algorithm parameters of the joint model to the multiple Private cloud server; when the plurality of private cloud servers receive the algorithm parameters of the updated joint model pushed by the public cloud server, verify whether the received updated joint model needs to be used Algorithm parameters update the model parameters of the trained local model, and in the case that the received algorithm parameters of the updated joint model need to be used to update the trained local model algorithm parameters, the trained The algorithm parameters of the local model are updated to the received algorithm parameters of the updated joint model.
于一实施例中,在所述多个私有云端服务器基于本地数据对本地模型进行训练之前,该系统还包括:所述公有云端服务器将初始模型的算法参数发送到 所述多个私有云端服务器,所述多个私有云端服务器基于所述初始模型的算法参数和所述本地数据对所述本地模型进行训练。In one embodiment, before the multiple private cloud servers train the local model based on local data, the system further includes: the public cloud server sends the algorithm parameters of the initial model to the multiple private cloud servers, The multiple private cloud servers train the local model based on the algorithm parameters of the initial model and the local data.
还提供了一种电子设备。该电子设备包括:处理器;以及存储器,设置为存储可执行指令,所述可执行指令在被所述处理器执行时使得所述电子设备执行上述实施例的方法。An electronic device is also provided. The electronic device includes: a processor; and a memory configured to store executable instructions, and when the executable instructions are executed by the processor, the electronic device executes the methods of the foregoing embodiments.
还提供了一种计算机可读存储介质,其上存储有计算机程序,该计算机程序被处理器执行时实现上述实施例的方法。A computer-readable storage medium is also provided, on which a computer program is stored, and when the computer program is executed by a processor, the methods of the above-mentioned embodiments are implemented.
附图说明Description of drawings
下面将对本申请实施例描述中所需要使用的附图作简单的介绍,显而易见地,下面描述中的附图仅仅是本申请实施例中的一部分实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据本申请实施例的内容和这些附图获得其他的附图。The following will give a brief introduction to the drawings that need to be used in the description of the embodiments of the present application. Obviously, the drawings in the following description are only part of the embodiments of the present application. For those of ordinary skill in the art, On the premise of no creative work, other drawings can also be obtained according to the content of the embodiment of the present application and these drawings.
图1是根据本申请实施例提供的一种基于联邦学习的模型训练方法的流程示意图;Fig. 1 is a schematic flowchart of a model training method based on federated learning provided according to an embodiment of the present application;
图2是根据本申请实施例提供的一种基于联邦学习的模型训练系统的方法流程示意图;FIG. 2 is a schematic flowchart of a method of a federated learning-based model training system provided according to an embodiment of the present application;
图3A是根据本申请实施例提供的另一种基于联邦学习的模型训练系统的方法示意图;FIG. 3A is a schematic diagram of another federated learning-based model training system method provided according to an embodiment of the present application;
图3B是根据本申请实施例提供的另一种基于联邦学习的模型训练系统的方法流程示意图;FIG. 3B is a schematic flowchart of another federated learning-based model training system method provided according to an embodiment of the present application;
图4是根据本申请实施例提供的一种基于联邦学习的模型训练装置的结构示意图;Fig. 4 is a schematic structural diagram of a model training device based on federated learning provided according to an embodiment of the present application;
图5是根据本申请实施例提供的另一种基于联邦学习的模型训练装置的结构示意图;FIG. 5 is a schematic structural diagram of another model training device based on federated learning provided according to an embodiment of the present application;
图6示出了适于用来实现本申请实施例的电子设备的结构示意图。FIG. 6 shows a schematic structural diagram of an electronic device suitable for implementing the embodiments of the present application.
具体实施方式detailed description
下面将结合附图对本申请实施例的技术方案进行描述。所描述的实施例可以是本申请的一部分实施例。基于本申请实施例,本领域技术人员可以在不付出创造性劳动的前提下获得其他实施例,这些实施例都应该属于本申请所保护的范围。The technical solutions of the embodiments of the present application will be described below with reference to the accompanying drawings. The described embodiments may be part of the embodiments of the present application. Based on the embodiments of the present application, those skilled in the art can obtain other embodiments without making creative efforts, and these embodiments should all belong to the protection scope of the present application.
需要说明的是,本文中术语“系统”和“网络”通常可以互换使用,本文中提到的“和/或”是指包括一个或更多个相关所列项目的任何和所有组合,本文中的术语“第一”、“第二”等是用于区别不同对象,而不是用于限定特定顺序。It should be noted that the terms "system" and "network" are often used interchangeably herein, and "and/or" mentioned herein refers to any and all combinations including one or more related listed items. The terms "first", "second", etc. in are used to distinguish different objects, not to limit a specific order.
还需要说明是,下述多个实施例可以单独执行,多个实施例之间也可以相互结合执行,本申请实施例对此不作限制。It should also be noted that the following multiple embodiments can be implemented independently, and multiple embodiments can also be implemented in combination with each other, which is not limited in this embodiment of the present application.
本文中的多个装置之间所交互的消息或者信息的名称仅用于说明性的目的,并不是用于对这些消息或信息的范围进行限制。The names of messages or information exchanged between multiple devices herein are for illustrative purposes only, and are not used to limit the scope of these messages or information.
下面结合附图并通过具体实施方式来说明本申请实施例的技术方案。The technical solutions of the embodiments of the present application will be described below in conjunction with the accompanying drawings and through specific implementation methods.
图1示出了本申请实施例提供的一种基于联邦学习的模型训练方法的流程示意图。本实施例可适用于多个私有云端服务器通过联邦学习训练模型的情况,该方法可以由多个私有云端服务器上配置的基于联邦学习的模型训练装置来执行。如图1所示,本实施例所述的基于联邦学习的模型训练方法包括如下步骤。FIG. 1 shows a schematic flowchart of a model training method based on federated learning provided by an embodiment of the present application. This embodiment is applicable to the situation where multiple private cloud servers train models through federated learning, and the method can be executed by a model training device based on federated learning configured on multiple private cloud servers. As shown in FIG. 1 , the model training method based on federated learning described in this embodiment includes the following steps.
在步骤S110中,基于本地数据对本地模型进行训练,将训练后的本地模型的算法参数发送给公有云端服务器,以使所述公有云端服务器验证是否需要采用所接收的所述训练后的本地模型的算法参数更新联合模型的算法参数。In step S110, the local model is trained based on local data, and the algorithm parameters of the trained local model are sent to the public cloud server, so that the public cloud server verifies whether the received trained local model needs to be adopted The algorithm parameters of update the algorithm parameters of the joint model.
所述公有云端服务器验证是否需要采用所接收的所述训练后的本地模型的算法参数更新联合模型的算法参数时,可采用先验数据集进行验证,以确定联合模型若采用该私有云端服务器所训练的本地模型的算法参数,联合模型的训练效果是否更优(比如模型精度是否更高)。可采用先验数据集计算联合模型的效果指标得到第三指标值,将联合模型的算法参数替换为所接收的所述训练后的本地模型的算法参数后,采用所述先验数据集计算替换算法参数后的联合模型的效果指标得到第四指标值,根据所述第三指标值和所述第四指标值的大小确定是否需要采用所接收的所述训练后的本地模型的算法参数更新联合模型的算法参数。When the public cloud server verifies whether it is necessary to use the received algorithm parameters of the trained local model to update the algorithm parameters of the joint model, it can use a priori data set for verification to determine if the joint model uses the algorithm parameters of the private cloud server. The algorithm parameters of the trained local model, and whether the training effect of the joint model is better (for example, whether the model accuracy is higher). The effect index of the joint model can be calculated by using the prior data set to obtain the third index value, and after the algorithm parameters of the joint model are replaced with the received algorithm parameters of the local model after training, the replacement is calculated using the prior data set The fourth index value is obtained from the effect index of the joint model after the algorithm parameters, and it is determined whether it is necessary to use the received algorithm parameters of the trained local model to update the joint model according to the size of the third index value and the fourth index value. Algorithmic parameters for the model.
在步骤S120中,接收所述公有云端服务器推送的更新后的联合模型的算法参数。In step S120, the algorithm parameters of the updated joint model pushed by the public cloud server are received.
在步骤S130中,验证是否需要采用所接收的所述更新后的联合模型的算法参数更新所述训练后的本地模型的算法参数,若需要采用所接收的所述更新后的联合模型的算法参数更新所述训练后的本地模型的算法参数,则执行步骤S140,若不需要采用所接收的所述更新后的联合模型的算法参数更新所述训练后的本地模型的算法参数,则返回步骤S110。In step S130, verify whether it is necessary to use the received algorithm parameters of the updated joint model to update the algorithm parameters of the trained local model, if necessary, use the received algorithm parameters of the updated joint model To update the algorithm parameters of the trained local model, execute step S140, if it is not necessary to update the algorithm parameters of the trained local model by using the received algorithm parameters of the updated joint model, then return to step S110 .
例如,可采用先验数据集计算所述训练后的本地模型的效果指标得到第一 指标值,将所述训练后的本地模型的算法参数替换为所接收的所述更新后的联合模型的算法参数后,采用所述先验数据集计算替换算法参数后的本地模型的效果指标得到第二指标值,根据所述第一指标值和所述第二指标值的大小确定是否需要采用所接收的所述更新后的联合模型的算法参数更新所述训练后的本地模型的算法参数。For example, a priori data set may be used to calculate the effect index of the trained local model to obtain a first index value, and the algorithm parameters of the trained local model may be replaced with the received algorithm of the updated joint model After parameterizing, use the prior data set to calculate the effect index of the local model after replacing the algorithm parameters to obtain the second index value, and determine whether to use the received The algorithm parameters of the updated joint model update the algorithm parameters of the trained local model.
在步骤S140中,将所述训练后的本地模型的算法参数更新为所接收的所述更新后的联合模型的算法参数。In step S140, the algorithm parameters of the trained local model are updated to the received algorithm parameters of the updated joint model.
所述效果指标可包括多种,包括但不限于模型预测的准确率、召回率等。The effect index may include various types, including but not limited to the accuracy rate of model prediction, recall rate, and the like.
根据本公开的一个或多个实施例,多个私有云端服务器在基于本地数据对本地模型进行训练之前还可接收公有云端服务器下发的初始模型的算法参数,基于所述初始模型的算法参数和本地数据对本地模型进行训练,以同步多个私有云端服务器的本地模型的初始状态。According to one or more embodiments of the present disclosure, before training the local model based on local data, multiple private cloud servers can also receive the algorithm parameters of the initial model issued by the public cloud server, based on the algorithm parameters of the initial model and The local data trains the local model to synchronize the initial state of the local model on multiple private cloud servers.
本实施例通过多个私有云端服务器基于本地数据对本地模型进行训练,将训练后的本地模型的算法参数发送给公有云端服务器,以使所述公有云端服务器验证是否需要采用所接收的所述算法参数更新联合模型的算法参数;接收所述公有云端服务器推送的更新后的联合模型的算法参数;验证是否需要采用所接收的所述更新后的联合模型的算法参数更新所述训练后的本地模型;若验证需要采用所接收的所述更新后的联合模型的算法参数更新所述训练后的本地模型,则将所述训练后的本地模型的算法参数更新为所接收的所述更新后的联合模型的算法参数,能够在保证工业数据安全的同时,保证了模型训练的效果。In this embodiment, multiple private cloud servers are used to train the local model based on local data, and the algorithm parameters of the trained local model are sent to the public cloud server, so that the public cloud server verifies whether the received algorithm needs to be used Parameters update the algorithm parameters of the joint model; receive the algorithm parameters of the updated joint model pushed by the public cloud server; verify whether the received algorithm parameters of the updated joint model need to be used to update the trained local model ; If the verification needs to use the received algorithm parameters of the updated joint model to update the trained local model, update the algorithm parameters of the trained local model to the received updated joint model The algorithm parameters of the model can ensure the effect of model training while ensuring the security of industrial data.
图2是根据本申请实施例提供的一种基于联邦学习的模型训练系统的模型训练方法的流程示意图。本实施例所述的基于联邦学习的模型训练系统包括公有云端服务器和多个私有云端服务器。如图2所示,本实施例所述的基于联邦学习的模型训练系统的模型训练方法包括如下步骤。FIG. 2 is a schematic flowchart of a model training method of a federated learning-based model training system provided according to an embodiment of the present application. The model training system based on federated learning described in this embodiment includes a public cloud server and multiple private cloud servers. As shown in FIG. 2 , the model training method of the federated learning-based model training system described in this embodiment includes the following steps.
在步骤S210中,多个私有云端服务器基于本地数据对本地模型进行训练,将训练后的本地模型的算法参数发送给所述公有云端服务器。In step S210, multiple private cloud servers train local models based on local data, and send algorithm parameters of the trained local models to the public cloud server.
在步骤S220中,所述公有云端服务器验证是否需要采用所接收的所述训练后的本地模型的算法参数更新联合模型的算法参数,若需要采用所接收的所述训练后的本地模型的算法参数更新联合模型的算法参数,则采用所接收的所述训练后的本地模型的算法参数更新联合模型的算法参数,将更新后的联合模型的算法参数推送到多个私有云端服务器。In step S220, the public cloud server verifies whether it is necessary to use the received algorithm parameters of the trained local model to update the algorithm parameters of the joint model, and if it needs to use the received algorithm parameters of the trained local model Updating the algorithm parameters of the joint model uses the received algorithm parameters of the trained local model to update the algorithm parameters of the joint model, and pushes the updated algorithm parameters of the joint model to multiple private cloud servers.
在步骤S230中,若多个私有云端服务器接收到所述公有云端服务器推送的所述更新后的联合模型的算法参数,则执行步骤S240。In step S230, if multiple private cloud servers receive the algorithm parameters of the updated joint model pushed by the public cloud server, step S240 is executed.
在步骤S240中,验证是否需要采用所接收的所述更新后的联合模型的算法参数更新所述训练后的本地模型的算法参数,若需要采用所接收的所述更新后的联合模型的算法参数更新所述训练后的本地模型的算法参数,则执行步骤S250,若不需要采用所接收的所述更新后的联合模型的算法参数更新所述训练后的本地模型的算法参数,则返回步骤S210。In step S240, verify whether it is necessary to use the received algorithm parameters of the updated joint model to update the algorithm parameters of the trained local model, if necessary, use the received algorithm parameters of the updated joint model To update the algorithm parameters of the trained local model, execute step S250, and return to step S210 if the received algorithm parameters of the updated joint model do not need to be used to update the algorithm parameters of the trained local model .
在步骤S250中,将所述训练后是本地模型的算法参数更新为所接收的所述更新后的联合模型的算法参数。In step S250, the algorithm parameters of the trained local model are updated to the received algorithm parameters of the updated joint model.
根据本公开的一个或多个实施例,在步骤S210之前,所述公有云端服务器还可将初始模型的算法参数发送到多个私有云端服务器,多个私有云端服务器基于所述初始模型的算法参数和本地数据对本地模型进行训练。According to one or more embodiments of the present disclosure, before step S210, the public cloud server may also send the algorithm parameters of the initial model to multiple private cloud servers, and the multiple private cloud servers are based on the algorithm parameters of the initial model and local data to train the local model.
本实施例技术方案在保证工业数据安全的同时,在不需要开放数据的前提下,满足了工业数据模型持续优化,提升人工智能(Artificial Intelligence,AI)技术在工业应用中的使用效果。While ensuring the security of industrial data, the technical solution of this embodiment satisfies the continuous optimization of industrial data models and improves the use effect of artificial intelligence (AI) technology in industrial applications without the need for open data.
图3A是根据本申请实施例提供的另一种基于联邦学习的模型训练系统的模型训练方法示意图。本实施例是一种在私有云端服务器进行本地的模型训练并将多地的模型整合优化为联合模型然后反馈给多地各自的私有云端服务器的一种安全的数据模型训练方案。FIG. 3A is a schematic diagram of a model training method of another federated learning-based model training system provided according to an embodiment of the present application. This embodiment is a safe data model training solution that performs local model training on a private cloud server and integrates and optimizes the models in multiple places into a joint model, and then feeds back to the respective private cloud servers in multiple places.
如图3A所示,本实施例主要采用联邦学习、分布式计算、算法模型整合优化等技术。联邦学习是一种分布式协同训练机器学习模型的范式,可以在无需集中训练数据的情况下在大量边缘设备(客户端)上进行协作训练机器学习模型,其特点是大量分散阶段链接到集中式服务器,这些参与者彼此之间的信任为零,并且只能访问本地训练数据。分布式计算是把一个大计算任务拆分成多个小计算任务分布到多台机器上去计算,然后再进行结果汇总,联邦学习过程中的分布式计算是在多地进行计算形成数据模型,然后将模型上传进行结果汇总。算法模型的整合优化技术是在同一参数定义的基础上,通过联邦学习和分布式计算,将本地的计算和模型整合成联合模型。在联合机器学习设定中,全局模型由中央参数服务器初始化并维护,并由该服务器共享给边缘设备。为了分布式协同训练出全局模型,客户端使用本地私有数据计算并更新模型,然后将其更新的模型上传到服务器,同时将隐私敏感的训练数据保留在自己的设备上,通过分布式安全聚合的多次迭代,联邦学习系统训练得到整合的模型。As shown in FIG. 3A , this embodiment mainly uses technologies such as federated learning, distributed computing, and algorithm model integration and optimization. Federated learning is a paradigm of distributed collaborative training of machine learning models, which can be used for collaborative training of machine learning models on a large number of edge devices (clients) without centralized training data. It is characterized by a large number of decentralized stages linked to centralized server, these participants have zero trust in each other and only have access to local training data. Distributed computing is to split a large computing task into multiple small computing tasks and distribute them to multiple machines for calculation, and then summarize the results. Distributed computing in the federated learning process is to perform calculations in multiple places to form a data model, and then Upload the model to summarize the results. The integration and optimization technology of the algorithm model is based on the definition of the same parameter, through federated learning and distributed computing, and integrates local computing and models into a joint model. In a federated machine learning setting, the global model is initialized and maintained by a central parameter server, which is then shared to edge devices. In order to train the global model through distributed collaboration, the client uses local private data to calculate and update the model, and then uploads the updated model to the server, while keeping the privacy-sensitive training data on its own device, through distributed security aggregation Over multiple iterations, the federated learning system trains the integrated model.
图3B示出了本申请实施例提供的又一种基于联邦学习的模型训练方法的流程示意图。本实施例以前述实施例为基础,进行了改进说明。如图3B所示,本实施例所述的基于联邦学习的模型训练方法包括如下步骤。FIG. 3B shows a schematic flowchart of another model training method based on federated learning provided by the embodiment of the present application. This embodiment is based on the foregoing embodiments, and an improved description is made. As shown in FIG. 3B , the model training method based on federated learning described in this embodiment includes the following steps.
在步骤S301中,工厂私有云启动模型训练。In step S301, the factory private cloud starts model training.
在步骤S302中,工厂私有云将模型的算法参数共享给联合模型。In step S302, the factory private cloud shares the algorithm parameters of the model with the joint model.
在步骤S303中,联合模型验证是否需要更新其上的算法参数,若联合模型验证需要更新算法参数,则执行步骤S305,若联合模型验证不需要更新算法参数,则执行步骤S304。In step S303, whether the joint model verification needs to update its algorithm parameters, if the joint model verification needs to update the algorithm parameters, then perform step S305, if the joint model verification does not need to update the algorithm parameters, then perform step S304.
在步骤S304中,不更新联合模型的算法参数,执行步骤S306。In step S304, the algorithm parameters of the joint model are not updated, and step S306 is executed.
在步骤S305中,更新联合模型的联合参数。In step S305, the joint parameters of the joint model are updated.
在步骤S306中,联合模型向工厂私有云推送联合模型的算法参数。In step S306, the joint model pushes the algorithm parameters of the joint model to the factory private cloud.
在步骤S307中,工厂私有云判断联合模型是否为最优,若联合模型是最优,则执行步骤S310,若联合模型不是最优,则执行步骤S308。In step S307, the factory private cloud judges whether the joint model is optimal, if the joint model is optimal, execute step S310, and if the joint model is not optimal, execute step S308.
在步骤S308中,不更新模型,执行步骤S309。In step S308, the model is not updated, and step S309 is executed.
在步骤S309中,持续本地训练,结束。In step S309, continue the local training and end.
在步骤S310中,迭代更新本地模型。In step S310, the local model is updated iteratively.
工厂私有云启动本地算法模型训练,并将训练的模型共享给联合模型,联合模型接收多个工厂共享的模型前需要进行模型验证,如果验证不需要更新联合模型的算法参数,将不更新该算法参数,工厂私有云利用多个工厂共享的信息对模型进行优化更新,在对应多个工厂端的共享的时候会利用多个工厂的共享信息进行模型优化。联合模型会定期向(多个)工厂端推送联合模型,工厂端接到推送会进行模型验证,如果推送的联合模型优于本地模型,对本地模型进行迭代更新,否则更新本地模型,持续的进行本地训练。推送过程是多个本地模型在不断进行学习的过程中将学习后更新的算法参数实时地推送给联合模型进行算法参数的更新,联合模型根据实时算法参数更新后,将更新后的联合模型的算法参数反向推送给多个工厂端,这些参数基于统一的定义规则,工厂端的算法模型不断的伴随联合模型的更新推送迭代更新。The private cloud of the factory starts the local algorithm model training and shares the trained model with the joint model. The joint model needs to perform model verification before receiving the model shared by multiple factories. If the verification does not need to update the algorithm parameters of the joint model, the algorithm will not be updated Parameters, the factory private cloud uses the information shared by multiple factories to optimize and update the model, and when it corresponds to the sharing of multiple factories, it will use the shared information of multiple factories to optimize the model. The joint model will regularly push the joint model to (multiple) factories, and the factory will verify the model after receiving the push. If the pushed joint model is better than the local model, the local model will be iteratively updated, otherwise the local model will be updated continuously. local training. The push process is that multiple local models push the updated algorithm parameters after learning to the joint model in real time during the continuous learning process to update the algorithm parameters. After the joint model is updated according to the real-time algorithm parameters, the updated joint model algorithm The parameters are reversely pushed to multiple factories. These parameters are based on unified definition rules, and the algorithm model of the factory is continuously pushed iteratively with the update of the joint model.
本实施例技术方案主要采用联邦学习、分布式计算、算法模型联合整合等技术,联合模型可置于公有云上,多地无需共享数据,只需共享数据模型,通过公有云不断优化联合模型再共享给多地做持续模型优化。整个流程数据不出厂,共享本地数据模型训练结果提升工业AI技术落地的使用效果。在保证工业数据安全的同时,在不需要开放数据的前提下,满足了工业数据模型持续优化,提升AI技术在工业应用中的使用效果。The technical solution of this embodiment mainly uses technologies such as federated learning, distributed computing, and algorithm model joint integration. The joint model can be placed on the public cloud. There is no need to share data in multiple places, but only the data model. The joint model is continuously optimized through the public cloud. Shared with multiple regions for continuous model optimization. The entire process data does not leave the factory, and the training results of the local data model are shared to improve the use effect of industrial AI technology. While ensuring the security of industrial data, it satisfies the continuous optimization of industrial data models and improves the use effect of AI technology in industrial applications without the need for open data.
作为上述多个图所示方法的实现,本申请提供了一种基于联邦学习的模型训练装置的一个实施例,图4示出了本实施例提供的一种基于联邦学习的模型 训练装置的结构示意图,该装置实施例与图1所示的方法实施例相对应,该装置可以应用于多个私有云端服务器中的多种电子设备中。如图4所示,本实施例所述的基于联邦学习的模型训练装置包括本地训练与参数上传单元410、联合模型参数接收单元420和验证与更新单元430。As an implementation of the methods shown in the above figures, this application provides an embodiment of a model training device based on federated learning. Figure 4 shows the structure of a model training device based on federated learning provided in this embodiment Schematic diagram, the device embodiment corresponds to the method embodiment shown in FIG. 1 , and the device can be applied to various electronic devices in multiple private cloud servers. As shown in FIG. 4 , the model training device based on federated learning described in this embodiment includes a local training and parameter uploading unit 410 , a joint model parameter receiving unit 420 and a verification and updating unit 430 .
所述本地训练与参数上传单元410被配置为,基于本地数据对本地模型进行训练,将训练后的本地模型的算法参数发送给公有云端服务器,以使所述公有云端服务器验证是否需要采用所接收的所述算法参数更新联合模型的算法参数。The local training and parameter uploading unit 410 is configured to train the local model based on local data, and send the algorithm parameters of the trained local model to the public cloud server, so that the public cloud server verifies whether the received The algorithm parameters of the update joint model algorithm parameters.
所述联合模型参数接收单元420被配置为,接收所述公有云端服务器推送的更新后的联合模型的算法参数。The joint model parameter receiving unit 420 is configured to receive the updated algorithm parameters of the joint model pushed by the public cloud server.
所述验证与更新单元430被配置为,验证是否需要采用所接收的所述更新后的联合模型的算法参数更新所述训练后的本地模型的算法参数,若验证需要采用所接收的所述更新后的联合模型的算法参数更新所述训练后的本地模型的算法参数,则将所述训练后的本地模型的算法参数更新为所接收的所述更新后的联合模型的算法参数。The verification and updating unit 430 is configured to verify whether it is necessary to use the received algorithm parameters of the updated joint model to update the algorithm parameters of the trained local model, and if the verification needs to use the received update If the algorithm parameters of the updated joint model update the algorithm parameters of the trained local model, the algorithm parameters of the trained local model are updated to the received algorithm parameters of the updated joint model.
根据本公开的一个或多个实施例,所述验证与更新单元430是被配置为,:采用先验数据集计算所述训练后的本地模型的效果指标得到第一指标值;将所述训练后的本地模型的算法参数替换为所接收的所述更新后的联合模型的算法参数后,采用所述先验数据集计算替换算法参数后的本地模型的效果指标得到第二指标值;根据所述第一指标值和所述第二指标值的大小确定是否需要采用所接收的所述更新后的联合模型的算法参数更新所述训练后的本地模型的算法参数。According to one or more embodiments of the present disclosure, the verification and updating unit 430 is configured to: use the prior data set to calculate the effect index of the trained local model to obtain a first index value; After the algorithm parameters of the updated local model are replaced with the received algorithm parameters of the updated joint model, the effect index of the local model after replacing the algorithm parameters is calculated using the prior data set to obtain a second index value; according to the The values of the first index value and the second index value determine whether to use the received algorithm parameters of the updated joint model to update the algorithm parameters of the trained local model.
根据本公开的一个或多个实施例,所述本地训练与参数上传单元410中所述公有云端服务器验证是否需要采用所接收的所述算法参数更新联合模型的算法参数包括:采用先验数据集计算联合模型的效果指标得到第三指标值;将联合模型的算法参数替换为所接收的所述训练后的本地模型的算法参数后,采用所述先验数据集计算替换算法参数后的联合模型的效果指标得到第四指标值;根据所述第三指标值和所述第四指标值的大小确定是否需要采用所接收的所述训练后的本地模型的算法参数更新联合模型的算法参数。According to one or more embodiments of the present disclosure, the verification by the public cloud server in the local training and parameter uploading unit 410 whether it is necessary to use the received algorithm parameters to update the algorithm parameters of the joint model includes: using a priori data set Calculating the effect index of the joint model to obtain the third index value; after replacing the algorithm parameters of the joint model with the received algorithm parameters of the trained local model, using the prior data set to calculate the joint model after replacing the algorithm parameters Obtaining a fourth index value based on the effect index of the result index; determining whether to use the received algorithm parameters of the trained local model to update the algorithm parameters of the joint model according to the magnitude of the third index value and the fourth index value.
根据本公开的一个或多个实施例,所述效果指标包括准确率和/或召回率。According to one or more embodiments of the present disclosure, the performance index includes precision and/or recall.
根据本公开的一个或多个实施例,所述装置还包括初始模型参数接收单元,配置为在基于本地数据对本地模型进行训练之前,接收公有云端服务器下发的初始模型的算法参数;According to one or more embodiments of the present disclosure, the device further includes an initial model parameter receiving unit configured to receive the algorithm parameters of the initial model issued by the public cloud server before training the local model based on the local data;
所述本地训练与参数上传单元配置为基于本地数据对本地模型进行训练包括:基于所述初始模型的算法参数和本地数据对本地模型进行训练。The local training and parameter uploading unit being configured to train the local model based on local data includes: training the local model based on algorithm parameters of the initial model and local data.
本实施例提供的基于联邦学习的模型训练装置可执行本公开方法实施例所提供的基于联邦学习的模型训练方法,具备执行方法相应的功能模块。The model training device based on federated learning provided in this embodiment can execute the model training method based on federated learning provided in the method embodiment of the present disclosure, and has corresponding functional modules for executing the method.
图5示出了本申请实施例提供的另一种基于联邦学习的模型训练装置的结构示意图。如图5所示,本实施例所述的基于联邦学习的模型训练装置包括初始模型参数接收单元510、本地训练与参数上传单元520、联合模型参数接收单元530和验证与更新单元540。FIG. 5 shows a schematic structural diagram of another model training device based on federated learning provided by an embodiment of the present application. As shown in FIG. 5 , the model training device based on federated learning in this embodiment includes an initial model parameter receiving unit 510 , a local training and parameter uploading unit 520 , a joint model parameter receiving unit 530 and a verification and updating unit 540 .
所述初始模型参数接收单元510被配置为,接收公有云端服务器下发的初始模型的算法参数。The initial model parameter receiving unit 510 is configured to receive the algorithm parameters of the initial model issued by the public cloud server.
所述本地训练与参数上传单元520被配置为,基于本地数据对本地模型进行训练包括:基于所述初始模型的算法参数和本地数据对本地模型进行训练,将训练后的本地模型的算法参数发送给公有云端服务器,以使所述公有云端服务器验证是否需要采用所接收的所述算法参数更新联合模型的算法参数。The local training and parameter uploading unit 520 is configured such that training the local model based on local data includes: training the local model based on the algorithm parameters of the initial model and local data, and sending the algorithm parameters of the trained local model to to the public cloud server, so that the public cloud server verifies whether the received algorithm parameters need to be used to update the algorithm parameters of the joint model.
所述联合模型参数接收单元530被配置为,接收所述公有云端服务器推送的更新后的联合模型的算法参数。The joint model parameter receiving unit 530 is configured to receive the updated algorithm parameters of the joint model pushed by the public cloud server.
所述验证与更新单元540被配置为,若验证需要采用所接收的所述更新后的联合模型的算法参数更新所述训练后的本地模型的算法参数,则将所述训练后的本地模型的算法参数更新为所接收的所述更新后的联合模型的算法参数。The verification and updating unit 540 is configured to, if the verification needs to use the received algorithm parameters of the updated joint model to update the algorithm parameters of the trained local model, then update the algorithm parameters of the trained local model to The algorithm parameters are updated to the received algorithm parameters of the updated joint model.
根据本公开的一个或多个实施例,所述验证与更新单元540被配置为验证是否需要采用所接收的所述更新后的联合模型的算法参数更新所述训练后的本地模型的算法参数包括:采用先验数据集计算所述训练后的本地模型的效果指标得到第一指标值;将所述训练后的本地模型的算法参数替换为所接收的所述更新后的联合模型的算法参数后,采用所述先验数据集计算替换算法参数后的本地模型的效果指标得到第二指标值;根据所述第一指标值和所述第二指标值的大小确定是否需要采用所接收的所述更新后的联合模型的算法参数更新所述训练后的本地模型的算法参数。According to one or more embodiments of the present disclosure, the verification and updating unit 540 is configured to verify whether it is necessary to use the received algorithm parameters of the updated joint model to update the algorithm parameters of the trained local model including : Using the prior data set to calculate the effect index of the trained local model to obtain the first index value; after replacing the algorithm parameters of the trained local model with the received algorithm parameters of the updated joint model , using the prior data set to calculate the effect index of the local model after replacing the algorithm parameters to obtain a second index value; determine whether to use the received The updated algorithm parameters of the joint model update the algorithm parameters of the trained local model.
根据本公开的一个或多个实施例,所述本地训练与参数上传单520中所述公有云端服务器验证是否需要采用所接收的所述算法参数更新联合模型的算法参数包括:采用先验数据集计算联合模型的效果指标得到第三指标值;将联合模型的算法参数替换为所接收的所述训练后的本地模型的算法参数后,采用所述先验数据集计算替换算法参数后的联合模型的效果指标得到第四指标值;根据所述第三指标值和所述第四指标值的大小确定是否需要采用所接收的所述训 练后的本地模型算法参数更新联合模型的算法参数。According to one or more embodiments of the present disclosure, the verification of whether the public cloud server in the local training and parameter uploading sheet 520 needs to use the received algorithm parameters to update the algorithm parameters of the joint model includes: using a priori data set Calculating the effect index of the joint model to obtain the third index value; after replacing the algorithm parameters of the joint model with the received algorithm parameters of the trained local model, using the prior data set to calculate the joint model after replacing the algorithm parameters Obtaining a fourth index value based on the effect index of the method; determining whether to use the received algorithm parameters of the trained local model to update the algorithm parameters of the joint model according to the magnitude of the third index value and the fourth index value.
根据本公开的一个或多个实施例,所述效果指标包括准确率和/或召回率。According to one or more embodiments of the present disclosure, the performance index includes precision and/or recall.
本实施例提供的基于联邦学习的模型训练装置可执行本公开方法实施例所提供的基于联邦学习的模型训练方法,具备执行方法相应的功能模块。The model training device based on federated learning provided in this embodiment can execute the model training method based on federated learning provided in the method embodiment of the present disclosure, and has corresponding functional modules for executing the method.
下面参考图6,其示出了适于用来实现本申请实施例的电子设备600的结构示意图。本申请实施例中的上述终端设备,例如为移动设备、电脑、或浮动车中内置的车载设备等,或其任意组合。在一些实施例中,移动设备例如可以包括手机、智能家居设备、可穿戴设备、智能移动设备、虚拟现实设备等,或其任意组合。图6示出的电子设备仅仅是一个示例,不应对本申请实施例的功能和使用范围带来任何限制。Referring to FIG. 6 , it shows a schematic structural diagram of an electronic device 600 suitable for implementing the embodiment of the present application. The above-mentioned terminal device in the embodiment of the present application is, for example, a mobile device, a computer, or a vehicle-mounted device built in a floating car, or any combination thereof. In some embodiments, the mobile device may include, for example, a mobile phone, a smart home device, a wearable device, a smart mobile device, a virtual reality device, etc., or any combination thereof. The electronic device shown in FIG. 6 is only an example, and should not limit the functions and scope of use of this embodiment of the present application.
如图6所示,电子设备600可以包括处理装置(例如中央处理器、图形处理器等)601,其可以根据存储在只读存储器(Read-Only Memory,ROM)602中的程序或者从存储装置608加载到随机访问存储器(Random Access Memory,RAM)603中的程序而执行多种适当的动作和处理。在RAM 603中,还存储有电子设备600操作所需的多种程序和数据。处理装置601、ROM 602以及RAM 603通过总线604彼此相连。输入/输出(Input/Output,I/O)接口605也连接至总线604。As shown in FIG. 6, an electronic device 600 may include a processing device (such as a central processing unit, a graphics processing unit, etc.) Various appropriate actions and processes are executed by a program loaded into a random access memory (Random Access Memory, RAM) 603 by 608 . In the RAM 603, various programs and data necessary for the operation of the electronic device 600 are also stored. The processing device 601, ROM 602, and RAM 603 are connected to each other through a bus 604. An input/output (Input/Output, I/O) interface 605 is also connected to the bus 604 .
通常,以下装置可以连接至I/O接口605:包括例如触摸屏、触摸板、键盘、鼠标、摄像头、麦克风、加速度计、陀螺仪等的输入装置606;包括例如液晶显示器(LCD)、扬声器、振动器等的输出装置607;包括例如磁带、硬盘等的存储装置608;以及通信装置609。通信装置609可以允许电子设备600与其他设备进行无线或有线通信以交换数据。虽然图6示出了具有多种装置的电子设备600,但是应理解的是,并不要求实施或具备所有示出的装置。可以替代地实施或具备更多或更少的装置。Typically, the following devices can be connected to the I/O interface 605: input devices 606 including, for example, a touch screen, touchpad, keyboard, mouse, camera, microphone, accelerometer, gyroscope, etc.; including, for example, a liquid crystal display (LCD), speaker, vibration an output device 607 such as a computer; a storage device 608 including, for example, a magnetic tape, a hard disk, etc.; and a communication device 609. The communication means 609 may allow the electronic device 600 to communicate with other devices wirelessly or by wire to exchange data. While FIG. 6 shows electronic device 600 having various means, it should be understood that implementing or possessing all of the means shown is not a requirement. More or fewer means may alternatively be implemented or provided.
根据本申请实施例的实施例,上文参考流程图描述的过程可以被实现为计算机软件程序。例如,本申请实施例的实施例包括一种计算机程序产品,其包括承载在计算机可读介质上的计算机程序,该计算机程序包含用于执行流程图所示的方法的程序代码。在这样的实施例中,该计算机程序可以通过通信装置609从网络上被下载和安装,或者从存储装置608被安装,或者从ROM 602被安装。在该计算机程序被处理装置601执行时,执行本申请实施例的方法中限定的上述功能。According to an embodiment of an embodiment of the present application, the processes described above with reference to the flowcharts may be implemented as computer software programs. For example, the embodiments of the present application include a computer program product, which includes a computer program carried on a computer-readable medium, where the computer program includes program code for executing the method shown in the flowchart. In such an embodiment, the computer program may be downloaded and installed from a network via communication means 609, or from storage means 608, or from ROM 602. When the computer program is executed by the processing device 601, the above-mentioned functions defined in the method of the embodiment of the present application are performed.
需要说明的是,本申请实施例上述的计算机可读介质可以是计算机可读信号介质或者计算机可读存储介质或者是上述两者的任意组合。计算机可读存储 介质例如可以是但不限于电、磁、光、电磁、红外线、或半导体的系统、装置或器件,或者任意以上的组合。计算机可读存储介质的的例子可以包括但不限于:具有一个或多个导线的电连接、便携式计算机磁盘、硬盘、RAM、ROM、可擦式可编程只读存储器(Erasable Programmable Read-Only Memory,EPROM)或闪存、光纤、便携式紧凑磁盘只读存储器(Compact Disc Read-Only Memory,CD-ROM)、光存储器件、磁存储器件、或者上述的任意合适的组合。在本申请实施例中,计算机可读存储介质可以是任何包含或存储程序的有形介质,该程序可以被指令执行系统、装置或者器件使用或者与其结合使用。而在本申请实施例中,计算机可读信号介质可以包括在基带中或者作为载波一部分传播的数据信号,其中承载了计算机可读的程序代码。这种传播的数据信号可以采用多种形式,包括但不限于电磁信号、光信号或上述的任意合适的组合。计算机可读信号介质还可以是计算机可读存储介质以外的任何计算机可读介质,该计算机可读信号介质可以发送、传播或者传输用于由指令执行系统、装置或者器件使用或者与其结合使用的程序。计算机可读介质上包含的程序代码可以用任何适当的介质传输,包括但不限于:电线、光缆、射频(Radio Frequency,RF)等,或者上述的任意合适的组合。It should be noted that the above-mentioned computer-readable medium in the embodiment of the present application may be a computer-readable signal medium or a computer-readable storage medium, or any combination of the two. A computer readable storage medium may be, for example, but not limited to, an electrical, magnetic, optical, electromagnetic, infrared, or semiconductor system, device, or device, or any combination thereof. Examples of computer readable storage media may include, but are not limited to: electrical connections with one or more wires, portable computer disks, hard disks, RAM, ROM, Erasable Programmable Read-Only Memory, EPROM) or flash memory, optical fiber, portable compact disk read-only memory (Compact Disc Read-Only Memory, CD-ROM), optical storage device, magnetic storage device, or any suitable combination of the above. In the embodiments of the present application, a computer-readable storage medium may be any tangible medium containing or storing a program, and the program may be used by or in combination with an instruction execution system, device or device. However, in the embodiment of the present application, the computer-readable signal medium may include a data signal propagated in the baseband or as a part of the carrier wave, and the computer-readable program code is carried therein. Such propagated data signals may take many forms, including but not limited to electromagnetic signals, optical signals, or any suitable combination of the foregoing. A computer-readable signal medium may also be any computer-readable medium other than a computer-readable storage medium, which can transmit, propagate, or transmit a program for use by or in conjunction with an instruction execution system, apparatus, or device . The program code contained on the computer readable medium can be transmitted by any appropriate medium, including but not limited to: electric wire, optical cable, radio frequency (Radio Frequency, RF), etc., or any suitable combination of the above.
上述计算机可读介质可以是上述电子设备中所包含的;也可以是单独存在,而未装配入该电子设备中。The above-mentioned computer-readable medium may be included in the above-mentioned electronic device, or may exist independently without being incorporated into the electronic device.
上述计算机可读介质承载有一个或者多个程序,当上述一个或者多个程序被该电子设备执行时,使得该电子设备:基于本地数据对本地模型进行训练,将训练后的本地模型的算法参数发送给公有云端服务器,以使所述公有云端服务器验证是否需要采用所接收的所述训练后的本地模型的算法参数更新联合模型的算法参数;接收所述公有云端服务器推送的更新后的联合模型的算法参数;验证是否需要采用所接收的所述更新后的联合模型的算法参数更新所述训练后的本地模型的算法参数,若验证需要采用所接收的所述更新后的联合模型的算法参数更新所述训练后的本地模型的算法参数,则将所述训练后的本地模型的算法参数更新为所接收的所述更新后的联合模型的算法参数。The above-mentioned computer-readable medium carries one or more programs, and when the above-mentioned one or more programs are executed by the electronic device, the electronic device: trains the local model based on local data, and converts the algorithm parameters of the trained local model to Send to the public cloud server, so that the public cloud server verifies whether the algorithm parameters of the joint model need to be updated by using the received algorithm parameters of the trained local model; receive the updated joint model pushed by the public cloud server Algorithm parameters; verify whether it is necessary to use the received algorithm parameters of the updated joint model to update the algorithm parameters of the trained local model, if the verification needs to use the received algorithm parameters of the updated joint model Updating the algorithm parameters of the trained local model is updating the algorithm parameters of the trained local model to the received algorithm parameters of the updated joint model.
可以以一种或多种程序设计语言或其组合来编写用于执行本申请实施例的操作的计算机程序代码,上述程序设计语言包括面向对象的程序设计语言—诸如Java、Smalltalk、C++,还包括常规的过程式程序设计语言—诸如“C”语言或类似的程序设计语言。程序代码可以完全地在用户计算机上执行、部分地在用户计算机上执行、作为一个独立的软件包执行、部分在用户计算机上部分在远程计算机上执行、或者完全在远程计算机或服务器上执行。在涉及远程计算机的情形中,远程计算机可以通过任意种类的网络包括局域网(Local Area Network,LAN)或广域网(Wide Area Network,WAN)连接到用户计算机,或者, 可以连接到外部计算机(例如利用因特网服务提供商来通过因特网连接)。Computer program codes for performing the operations of the embodiments of the present application may be written in one or more programming languages or a combination thereof, and the above-mentioned programming languages include object-oriented programming languages—such as Java, Smalltalk, C++, and A conventional procedural programming language - such as the "C" language or a similar programming language. The program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In cases involving a remote computer, the remote computer may be connected to the user computer via any kind of network, including a Local Area Network (LAN) or a Wide Area Network (WAN), or, alternatively, may be connected to an external computer (e.g., via the Internet) service provider via Internet connection).
附图中的流程图和框图,图示了按照本申请实施例多种实施例的系统、方法和计算机程序产品的可能实现的体系架构、功能和操作。在这点上,流程图或框图中的每个方框可以代表一个模块、程序段、或代码的一部分,该模块、程序段、或代码的一部分包含一个或多个用于实现规定的逻辑功能的可执行指令。也应当注意,在有些作为替换的实现中,方框中所标注的功能也可以以不同于附图中所标注的顺序发生。例如,两个接连地表示的方框实际上可以基本并行地执行,它们有时也可以按相反的顺序执行,这依所涉及的功能而定。也要注意的是,框图和/或流程图中的每个方框、以及框图和/或流程图中的方框的组合,可以用执行规定的功能或操作的专用的基于硬件的系统来实现,或者可以用专用硬件与计算机指令的组合来实现。The flow charts and block diagrams in the accompanying drawings illustrate the system architecture, functions and operations of possible implementations of systems, methods and computer program products according to various embodiments of the embodiments of the present application. In this regard, each block in a flowchart or block diagram may represent a module, program segment, or portion of code that contains one or more logical functions for implementing specified executable instructions. It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or they may sometimes be executed in the reverse order, depending upon the functionality involved. It should also be noted that each block of the block diagrams and/or flowchart illustrations, and combinations of blocks in the block diagrams and/or flowchart illustrations, can be implemented by a dedicated hardware-based system that performs the specified functions or operations , or may be implemented by a combination of dedicated hardware and computer instructions.
描述于本申请实施例中所涉及到的单元可以通过软件的方式实现,也可以通过硬件的方式来实现。单元的名称在某种情况下并不构成对该单元本身的限定,例如,第一获取单元还可以被描述为“获取至少两个网际协议地址的单元”。The units involved in the embodiments described in the present application may be implemented by means of software or by means of hardware. The name of the unit does not limit the unit itself in some cases, for example, the first obtaining unit may also be described as "a unit for obtaining at least two Internet Protocol addresses".

Claims (10)

  1. 一种基于联邦学习的模型训练方法,由多个私有云端服务器执行,包括:A model training method based on federated learning, executed by multiple private cloud servers, including:
    基于本地数据对本地模型进行训练,将训练后的本地模型的算法参数发送给公有云端服务器,以使所述公有云端服务器验证是否需要采用所接收的所述训练后的本地模型的算法参数更新联合模型的算法参数;The local model is trained based on local data, and the algorithm parameters of the trained local model are sent to the public cloud server, so that the public cloud server verifies whether it is necessary to use the received algorithm parameters of the trained local model to update the joint Algorithmic parameters of the model;
    接收所述公有云端服务器推送的更新后的联合模型的算法参数;receiving the algorithm parameters of the updated joint model pushed by the public cloud server;
    验证是否需要采用所接收的所述更新后的联合模型的算法参数更新所述训练后的本地模型的算法参数;Verifying whether the received algorithm parameters of the updated joint model need to be used to update the algorithm parameters of the trained local model;
    在验证需要采用所接收的所述更新后的联合模型的算法参数更新所述训练后的本地模型的算法参数的情况下,将所述训练后的本地模型的算法参数更新为所接收的所述更新后的联合模型的算法参数。In the case that the algorithm parameters of the trained local model need to be updated with the received algorithm parameters of the joint model, update the algorithm parameters of the trained local model to the received Algorithm parameters for the updated joint model.
  2. 根据权利要求1所述的方法,其中,所述验证是否需要采用所接收的所述更新后的联合模型的算法参数更新所述训练后的本地模型的算法参数包括:The method according to claim 1, wherein the verifying whether the received algorithm parameters of the updated joint model need to be used to update the algorithm parameters of the trained local model comprises:
    采用先验数据集计算所述训练后的本地模型的效果指标得到第一指标值;calculating an effect index of the trained local model by using a priori data set to obtain a first index value;
    将所述训练后的本地模型的算法参数替换为所接收的所述更新后的联合模型的算法参数后,采用所述先验数据集计算替换所述算法参数后的本地模型的效果指标得到第二指标值;After replacing the algorithm parameters of the trained local model with the received algorithm parameters of the updated joint model, using the prior data set to calculate the effect index of the local model after replacing the algorithm parameters to obtain the first Two index values;
    根据所述第一指标值和所述第二指标值的大小确定是否需要采用所接收的所述更新后的联合模型的算法参数更新所述训练后的本地模型的算法参数。Determine whether to use the received algorithm parameters of the updated joint model to update the algorithm parameters of the trained local model according to the magnitudes of the first index value and the second index value.
  3. 根据权利要求1所述的方法,其中,所述公有云端服务器验证是否需要采用所接收的所述训练后的本地模型的算法参数更新联合模型的算法参数包括:The method according to claim 1, wherein the verification by the public cloud server whether it is necessary to use the received algorithm parameters of the trained local model to update the algorithm parameters of the joint model comprises:
    采用先验数据集计算所述联合模型的效果指标得到第三指标值;Using the prior data set to calculate the effect index of the joint model to obtain the third index value;
    将所述联合模型的算法参数替换为所接收的所述训练后的本地模型的算法参数后,采用所述先验数据集计算替换算法参数后的联合模型的效果指标得到第四指标值;After replacing the algorithm parameters of the joint model with the received algorithm parameters of the trained local model, using the prior data set to calculate the effect index of the joint model after replacing the algorithm parameters to obtain a fourth index value;
    根据所述第三指标值和所述第四指标值的大小确定是否需要采用所接收的所述训练后的本地模型的算法参数更新所述联合模型的算法参数。Determine whether to use the received algorithm parameters of the trained local model to update the algorithm parameters of the joint model according to the magnitude of the third index value and the fourth index value.
  4. 根据权利要2或3所述的方法,其中,所述效果指标包括准确率和召回率中的至少之一。The method according to claim 2 or 3, wherein the effect index includes at least one of precision rate and recall rate.
  5. 根据权利要求1所述的方法,在所述基于本地数据对本地模型进行训练之前,还包括,接收公有云端服务器下发的初始模型的算法参数;The method according to claim 1, before said training the local model based on the local data, further comprising, receiving the algorithm parameters of the initial model sent by the public cloud server;
    所述基于本地数据对本地模型进行训练包括:基于所述初始模型的算法参数和所述本地数据对所述本地模型进行训练。The training the local model based on local data includes: training the local model based on algorithm parameters of the initial model and the local data.
  6. 一种基于联邦学习的模型训练装置,配置于多个私有云端服务器中,包括:A model training device based on federated learning, configured in multiple private cloud servers, including:
    本地训练与参数上传单元,设置为基于本地数据对本地模型进行训练,将训练后的本地模型的算法参数发送给公有云端服务器,以使所述公有云端服务器验证是否需要采用所接收的所述训练后的本地模型的算法参数更新联合模型的算法参数;The local training and parameter uploading unit is configured to train the local model based on local data, and send the algorithm parameters of the trained local model to the public cloud server, so that the public cloud server verifies whether it needs to adopt the received training Algorithm parameters of the subsequent local model update the algorithm parameters of the joint model;
    联合模型参数接收单元,设置为接收所述公有云端服务器推送的更新后的联合模型的算法参数;The joint model parameter receiving unit is configured to receive the algorithm parameters of the updated joint model pushed by the public cloud server;
    验证与更新单元,设置为验证是否需要采用所接收的所述更新后的联合模型的算法参数更新所述训练后的本地模型的算法参数,在验证需要采用所接收的所述更新后的联合模型的算法参数更新所述训练后的本地模型的情况下,将所述训练后的本地模型的算法参数更新为所接收的所述更新后的联合模型的算法参数。The verification and update unit is configured to verify whether the received algorithm parameters of the updated joint model need to be used to update the algorithm parameters of the trained local model, and the received updated joint model needs to be used for verification In a case where the algorithm parameters of the trained local model are updated, the algorithm parameters of the trained local model are updated to the received algorithm parameters of the updated joint model.
  7. 一种基于联邦学习的模型训练系统,包括公有云端服务器和多个私有云端服务器;A model training system based on federated learning, including a public cloud server and multiple private cloud servers;
    所述多个私有云端服务器基于本地数据对本地模型进行训练,将训练后的本地模型的算法参数发送给所述公有云端服务器;The plurality of private cloud servers train the local model based on local data, and send the algorithm parameters of the trained local model to the public cloud server;
    所述公有云端服务器验证是否需要采用所接收的所述训练后的本地模型的算法参数更新联合模型的算法参数,在验证需要采用所接收的所述训练后的本地模型的算法参数更新所述联合模型的情况下,采用所述所接收的所述训练后的本地模型的算法参数更新所述联合模型的算法参数,将更新后的联合模型的算法参数推送到所述多个私有云端服务器;The public cloud server verifies whether the algorithm parameters of the joint model need to be updated by using the received algorithm parameters of the trained local model, and the algorithm parameters of the joint model need to be updated by using the received algorithm parameters of the trained local model. In the case of a model, update the algorithm parameters of the joint model by using the received algorithm parameters of the trained local model, and push the updated algorithm parameters of the joint model to the plurality of private cloud servers;
    在所述多个私有云端服务器接收到所述公有云端服务器推送的所述更新后的联合模型的算法参数的情况下,验证是否需要采用所接收的所述更新后的联合模型的算法参数更新所述训练后的本地模型的算法参数,在验证需要采用所接收的所述更新后的联合模型的算法参数更新所述训练后的本地模型的算法参数的情况下,将所述训练后的本地模型的算法参数更新为所接收的所述更新后的联合模型的算法参数。When the plurality of private cloud servers receive the algorithm parameters of the updated joint model pushed by the public cloud server, verify whether it is necessary to use the received algorithm parameters of the updated joint model to update the Algorithm parameters of the trained local model, in the case that the algorithm parameters of the trained local model need to be updated by using the received algorithm parameters of the updated joint model, the trained local model The algorithm parameters of are updated to the received algorithm parameters of the updated joint model.
  8. 根据权利要求1所述的系统,在所述多个私有云端服务器基于本地数据对本地模型进行训练之前还包括:The system according to claim 1, before the plurality of private cloud servers train the local model based on local data, it also includes:
    所述公有云端服务器将初始模型的算法参数发送到所述多个私有云端服务 器,所述多个私有云端服务器基于所述初始模型的算法参数和所述本地数据对所述本地模型进行训练。The public cloud server sends the algorithm parameters of the initial model to the multiple private cloud servers, and the multiple private cloud servers train the local model based on the algorithm parameters of the initial model and the local data.
  9. 一种电子设备,包括:An electronic device comprising:
    至少一个处理器;以及at least one processor; and
    存储器,设置为存储可执行指令,所述可执行指令在被所述至少一个处理器执行时,使得所述电子设备执行如权利要求1-5中任一项所述的方法。A memory configured to store executable instructions, which, when executed by the at least one processor, cause the electronic device to execute the method according to any one of claims 1-5.
  10. 一种计算机可读存储介质,其上存储有计算机程序,其中,所述计算机程序被处理器执行时实现如权利要求1-5任一项所述的方法。A computer-readable storage medium, on which a computer program is stored, wherein, when the computer program is executed by a processor, the method according to any one of claims 1-5 is implemented.
PCT/CN2022/091868 2021-07-15 2022-05-10 Model training method, apparatus, and system based on federated learning, and device and medium WO2023284387A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202110799024.8A CN113537513A (en) 2021-07-15 2021-07-15 Model training method, device, system, equipment and medium based on federal learning
CN202110799024.8 2021-07-15

Publications (1)

Publication Number Publication Date
WO2023284387A1 true WO2023284387A1 (en) 2023-01-19

Family

ID=78099295

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/091868 WO2023284387A1 (en) 2021-07-15 2022-05-10 Model training method, apparatus, and system based on federated learning, and device and medium

Country Status (2)

Country Link
CN (1) CN113537513A (en)
WO (1) WO2023284387A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117010484A (en) * 2023-10-07 2023-11-07 之江实验室 Personalized federal learning generalization method, device and application based on attention mechanism

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113537513A (en) * 2021-07-15 2021-10-22 青岛海尔工业智能研究院有限公司 Model training method, device, system, equipment and medium based on federal learning
CN115196730A (en) * 2022-07-19 2022-10-18 南通派菲克水务技术有限公司 Intelligent sodium hypochlorite adding system for water plant

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20200050951A1 (en) * 2018-08-09 2020-02-13 International Business Machines Corporation Collaborative distributed machine learning
CN111477336A (en) * 2020-04-07 2020-07-31 中南大学 Fusion method, system and storage medium for infectious disease diagnosis data
CN111935156A (en) * 2020-08-12 2020-11-13 科技谷(厦门)信息技术有限公司 Data privacy protection method for federated learning
CN112261137A (en) * 2020-10-22 2021-01-22 江苏禹空间科技有限公司 Model training method and system based on joint learning
CN112329940A (en) * 2020-11-02 2021-02-05 北京邮电大学 Personalized model training method and system combining federal learning and user portrait
CN112749392A (en) * 2021-01-07 2021-05-04 西安电子科技大学 Method and system for detecting abnormal nodes in federated learning
US20210150269A1 (en) * 2019-11-18 2021-05-20 International Business Machines Corporation Anonymizing data for preserving privacy during use for federated machine learning
CN113537513A (en) * 2021-07-15 2021-10-22 青岛海尔工业智能研究院有限公司 Model training method, device, system, equipment and medium based on federal learning

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110874637B (en) * 2020-01-16 2020-04-28 支付宝(杭州)信息技术有限公司 Multi-target fusion learning method, device and system based on privacy data protection
CN112257105B (en) * 2020-10-19 2022-01-11 中山大学 Federal learning method and system based on parameter replacement algorithm
CN112989944A (en) * 2021-02-08 2021-06-18 西安翔迅科技有限责任公司 Intelligent video safety supervision method based on federal learning
CN112926747B (en) * 2021-03-25 2022-05-17 支付宝(杭州)信息技术有限公司 Method and device for optimizing business model

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20200050951A1 (en) * 2018-08-09 2020-02-13 International Business Machines Corporation Collaborative distributed machine learning
US20210150269A1 (en) * 2019-11-18 2021-05-20 International Business Machines Corporation Anonymizing data for preserving privacy during use for federated machine learning
CN111477336A (en) * 2020-04-07 2020-07-31 中南大学 Fusion method, system and storage medium for infectious disease diagnosis data
CN111935156A (en) * 2020-08-12 2020-11-13 科技谷(厦门)信息技术有限公司 Data privacy protection method for federated learning
CN112261137A (en) * 2020-10-22 2021-01-22 江苏禹空间科技有限公司 Model training method and system based on joint learning
CN112329940A (en) * 2020-11-02 2021-02-05 北京邮电大学 Personalized model training method and system combining federal learning and user portrait
CN112749392A (en) * 2021-01-07 2021-05-04 西安电子科技大学 Method and system for detecting abnormal nodes in federated learning
CN113537513A (en) * 2021-07-15 2021-10-22 青岛海尔工业智能研究院有限公司 Model training method, device, system, equipment and medium based on federal learning

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117010484A (en) * 2023-10-07 2023-11-07 之江实验室 Personalized federal learning generalization method, device and application based on attention mechanism
CN117010484B (en) * 2023-10-07 2024-01-26 之江实验室 Personalized federal learning generalization method, device and application based on attention mechanism

Also Published As

Publication number Publication date
CN113537513A (en) 2021-10-22

Similar Documents

Publication Publication Date Title
WO2023284387A1 (en) Model training method, apparatus, and system based on federated learning, and device and medium
WO2023092792A1 (en) Optimization method for modeling based on federated learning, and electronic device, storage medium and program product
CN107273083B (en) Interaction method, device, equipment and storage medium between terminal equipment
WO2020221190A1 (en) Applet state synchronization method, device and computer storage medium
CN109522363B (en) Cloud platform synchronization method, system, equipment and storage medium based on block chain
CN113032412B (en) Data synchronization method, device, electronic equipment and computer readable medium
CN113505520A (en) Method, device and system for supporting heterogeneous federated learning
WO2023078072A1 (en) Byzantine fault tolerance-based asynchronous consensus method and apparatus, server and medium
CN110909521A (en) Synchronous processing method and device for online document information and electronic equipment
CN113537512B (en) Model training method, device, system, equipment and medium based on federal learning
US20230385080A1 (en) Method and apparatus for presenting information on lock screen interface, terminal, and storage medium
WO2020173381A1 (en) Data interworking method and device, terminal and storage medium
CN116489621A (en) Vehicle key sharing method, device, equipment and medium
WO2023124219A1 (en) Joint learning model iterative update method, apparatus, system, and storage medium
US10659385B2 (en) Provisioning insight services in a data provider landscape
CN113420400B (en) Routing relation establishment method, request processing method, device and equipment
CN112507676B (en) Method and device for generating energy report, electronic equipment and computer readable medium
WO2022017458A1 (en) Data synchronization method and apparatus, electronic device, and medium
CN114595474A (en) Federal learning modeling optimization method, electronic device, medium, and program product
CN114915516A (en) Communication method and device
CN111680754B (en) Image classification method, device, electronic equipment and computer readable storage medium
WO2018000621A1 (en) Communication data synchronization method and electronic device
CN113487041A (en) Horizontal federal learning method, device and storage medium
CN112612806A (en) House resource information processing method and device, electronic equipment and computer readable medium
CN116521377B (en) Service computing unloading method, system, device, equipment and medium

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22841027

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE