WO2021051610A1 - Data training method, apparatus and system - Google Patents

Data training method, apparatus and system Download PDF

Info

Publication number
WO2021051610A1
WO2021051610A1 PCT/CN2019/118407 CN2019118407W WO2021051610A1 WO 2021051610 A1 WO2021051610 A1 WO 2021051610A1 CN 2019118407 W CN2019118407 W CN 2019118407W WO 2021051610 A1 WO2021051610 A1 WO 2021051610A1
Authority
WO
WIPO (PCT)
Prior art keywords
model
training
model parameters
parameters
multiple clients
Prior art date
Application number
PCT/CN2019/118407
Other languages
French (fr)
Chinese (zh)
Inventor
何安珣
王健宗
Original Assignee
平安科技(深圳)有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 平安科技(深圳)有限公司 filed Critical 平安科技(深圳)有限公司
Publication of WO2021051610A1 publication Critical patent/WO2021051610A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/25Integrating or interfacing systems involving database management systems
    • G06F16/252Integrating or interfacing systems involving database management systems between a Database Management System and a front-end application
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H10/00ICT specially adapted for the handling or processing of patient-related medical or healthcare data
    • G16H10/60ICT specially adapted for the handling or processing of patient-related medical or healthcare data for patient-specific data, e.g. for electronic patient records

Landscapes

  • Engineering & Computer Science (AREA)
  • Databases & Information Systems (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Health & Medical Sciences (AREA)
  • Epidemiology (AREA)
  • General Health & Medical Sciences (AREA)
  • Medical Informatics (AREA)
  • Primary Health Care (AREA)
  • Public Health (AREA)
  • Measuring And Recording Apparatus For Diagnosis (AREA)
  • Medical Treatment And Welfare Office Work (AREA)

Abstract

A data training method, apparatus and system. The method comprises: sending an initial training model to a plurality of clients, wherein the plurality of clients all independently communicate with a server (S202); receiving a plurality of sets of first model parameters sent by the plurality of clients, wherein the first model parameters are obtained by training, according to first medical data of a local database, the initial training model by the clients (S204); performing weighted averaging on the plurality of sets of first model parameters to obtain second model parameters (S206); and sending the second model parameters to the plurality of clients, wherein the second model parameters are used for respectively constructing the same second training model on the plurality of clients (S208). The method solves the technical problem, in the related technology, of an algorithm model for processing medical data being relatively complicated and being unable to process large-scale medical data that has a relatively high security and is inconvenient to circulate.

Description

数据的训练方法及装置、系统Data training method, device and system 技术领域Technical field
本申请涉及计算机领域,具体而言,涉及一种数据的训练方法及装置、系统。This application relates to the computer field, specifically, to a data training method, device, and system.
背景技术Background technique
相关技术中,医疗影像辅助识别是人工智能图像识别技术在医疗领域的一项较为成熟的应用,国内外有多家机构以此技术为核心,建立起标准化的区域医疗影像数据中心云平台服务,集成辅助诊断、数据集中储存管理、区域重大疾病分析,和区域人群健康画像等功能。目前广为使用的区域云平台,顾名思义,只是区域卫生信息共享系统,本质上这是一个以诊所、医院为单位或者几家医院为单位的私有云。Among related technologies, medical imaging assisted recognition is a relatively mature application of artificial intelligence image recognition technology in the medical field. Many institutions at home and abroad have established standardized regional medical imaging data center cloud platform services based on this technology. Integrated auxiliary diagnosis, data centralized storage management, regional major disease analysis, and regional population health portraits and other functions. At present, the widely used regional cloud platform, as the name suggests, is just a regional health information sharing system. In essence, this is a private cloud with clinics, hospitals as units, or several hospitals as units.
因医疗健康数据的隐私性,无法产生规模化效应,数据孤岛的问题仍然存在,医疗健康模型的训练仍然受有限数据的牵制,一些医疗机构需要花费较高的费用购买第三方机构已经训练好的模型,行业整体信息共享程度低,经济效率不高,医疗大健康生态难以在此基础上进一步发展。Due to the privacy of medical and health data, the scale effect cannot be produced. The problem of data islands still exists. The training of medical and health models is still restricted by limited data. Some medical institutions need to spend a higher cost to purchase the training that has been trained by third-party institutions. Model, the industry as a whole has a low degree of information sharing and low economic efficiency. It is difficult for the healthcare ecosystem to develop further on this basis.
传统的数据结构和机器学习是将数据整合后,基于集成后的数据集进行训练。这类方法要求数据在分布式数据集和中心服务器端进行传输,中心服务器由于整合了海量数据,训练模型所需要的算力要求高,计算成本也相应的高,并且响应时间比较长。同时,一些对于安全性比较高、不便于进行流动的数据,如医疗健康数据,就无法大规模使用此方法进行模型训练。Traditional data structure and machine learning integrate data and then train based on the integrated data set. This type of method requires data to be transmitted between the distributed data set and the central server. Since the central server integrates massive data, the computing power required for training the model is high, the calculation cost is correspondingly high, and the response time is relatively long. At the same time, for some data that is relatively safe and inconvenient to flow, such as medical health data, this method cannot be used for model training on a large scale.
针对相关技术中存在的上述问题,目前尚未发现有效的解决方案。In view of the above-mentioned problems existing in related technologies, no effective solutions have been found so far.
申请内容Application content
本申请实施例提供了一种数据的训练方法及装置、系统,以至少解决相关技术中处理医疗数据的算法模型较为复杂,无法处理不便于流动的大规模医疗数据等技术问题。The embodiments of the present application provide a data training method, device, and system, so as to at least solve the technical problems that the algorithm model for processing medical data in related technologies is relatively complicated and cannot handle large-scale medical data that is inconvenient to flow.
根据本申请的一个实施例,提供了一种数据的训练方法,包括:向多个客户端发送初始训练模型,其中,所述多个客户端均与服务器单独通信;接收所述多个客户端发送的多套第一模型参数,其中,所述第一模型参数是所述客户端根据本地数据库的 第一医疗数据对所述初始训练模型进行训练得到的;对所述多套第一模型参数进行加权平均,得到第二模型参数;将所述第二模型参数发送至所述多个客户端,其中,所述第二模型参数用于分别在所述多个客户端构建相同的第二训练模型。According to an embodiment of the present application, a data training method is provided, including: sending an initial training model to multiple clients, wherein each of the multiple clients communicates with a server separately; and receiving the multiple clients Sent multiple sets of first model parameters, where the first model parameters are obtained by the client training the initial training model according to the first medical data in the local database; and the multiple sets of first model parameters Perform a weighted average to obtain the second model parameters; send the second model parameters to the multiple clients, where the second model parameters are used to construct the same second training on the multiple clients respectively model.
根据本申请的另一个实施例,提供了一种数据的训练方法,包括:接收服务器发送的初始训练模型;根据本地数据库的第一医疗数据对所述初始训练模型进行训练,得到第一训练模型;将所述第一训练模型的第一模型参数发送至所述服务器,其中,所述服务器用于对多个客户端的多套第一模型参数进行加权平均,得到第二训练模型的第二模型参数,并将所述第二模型参数反馈至所述多个客户端;根据所述第二模型参数构建第二训练模型,并使用所述第二训练模型训练所述本地数据库的第二医疗数据。According to another embodiment of the present application, a data training method is provided, including: receiving an initial training model sent by a server; training the initial training model according to first medical data in a local database to obtain the first training model Send the first model parameters of the first training model to the server, where the server is used to perform a weighted average on multiple sets of first model parameters of multiple clients to obtain a second model of the second training model Parameters, and feed back the second model parameters to the multiple clients; construct a second training model according to the second model parameters, and use the second training model to train the second medical data of the local database .
根据本申请的一个实施例,提供了一种数据的训练装置,包括:第一发送模块,用于向多个客户端发送初始训练模型,其中,所述多个客户端均与服务器单独通信;接收模块,用于接收所述多个客户端发送的多套第一模型参数,其中,所述第一模型参数是所述客户端根据本地数据库的第一医疗数据对所述初始训练模型进行训练得到的;计算模块,用于对所述多套第一模型参数进行加权平均,得到第二模型参数;第二发送模块,用于将所述第二模型参数发送至所述多个客户端,其中,所述第二模型参数用于分别在所述多个客户端构建相同的第二训练模型。According to an embodiment of the present application, there is provided a data training device, including: a first sending module, configured to send an initial training model to multiple clients, wherein each of the multiple clients communicates with the server separately; The receiving module is configured to receive multiple sets of first model parameters sent by the multiple clients, where the first model parameters are the clients training the initial training model according to the first medical data in the local database A calculation module for weighted average of the multiple sets of first model parameters to obtain a second model parameter; a second sending module for sending the second model parameters to the multiple clients, Wherein, the second model parameters are used to construct the same second training model on the multiple clients respectively.
根据本申请的另一个实施例,提供了一种数据的训练装置,包括:接收模块,用于接收服务器发送的初始训练模型;第一训练模块,用于根据本地数据库的第一医疗数据对所述初始训练模型进行训练,得到第一训练模型;发送模块,用于将所述第一训练模型的第一模型参数发送至所述服务器,其中,所述服务器用于对多个客户端的多套第一模型参数进行加权平均,得到第二训练模型的第二模型参数,并将所述第二模型参数反馈至所述多个客户端;第二训练模块,用于根据所述第二模型参数构建第二训练模型,并使用所述第二训练模型训练所述本地数据库的第二医疗数据。According to another embodiment of the present application, there is provided a data training device, which includes: a receiving module for receiving an initial training model sent by a server; a first training module for performing data training based on first medical data in a local database; The initial training model is trained to obtain a first training model; a sending module is used to send the first model parameters of the first training model to the server, where the server is used to provide multiple sets of The first model parameters are weighted and averaged to obtain the second model parameters of the second training model, and the second model parameters are fed back to the multiple clients; the second training module is used to obtain the second model parameters according to the second model parameters. Constructing a second training model, and using the second training model to train the second medical data of the local database.
根据本申请的又一个实施例,还提供了一种数据的训练系统,包括:服务器和多个客户端,其中,所述服务器包括:第一发送模块,用于向多个客户端发送初始训练模型;接收模块,用于接收所述多个客户端发送的多套第一模型参数;计算模块,用于对所述多套第一模型参数进行加权平均,得到第二模型参数;第二发送模块,用于 将所述第二模型参数发送至所述多个客户端;所述多个客户端,均与服务器单独通信,包括:接收模块,用于接收所述初始训练模型;第一训练模块,用于根据本地数据库的第一医疗数据对所述初始训练模型进行训练,得到所述第一训练模型;发送模块,用于将所述第一训练模型的第一模型参数发送至所述服务器;第二训练模块,用于根据所述第二模型参数构建第二训练模型,并使用所述第二训练模型训练所述本地数据库的第二医疗数据。According to another embodiment of the present application, there is also provided a data training system, including: a server and a plurality of clients, wherein the server includes: a first sending module, configured to send initial training to the plurality of clients Model; receiving module for receiving multiple sets of first model parameters sent by said multiple clients; calculation module for weighted average of said multiple sets of first model parameters to obtain second model parameters; second sending The module is used to send the second model parameters to the multiple clients; the multiple clients all communicate with the server separately, including: a receiving module, used to receive the initial training model; first training Module, used to train the initial training model according to the first medical data in the local database to obtain the first training model; sending module, used to send the first model parameters of the first training model to the Server; a second training module for constructing a second training model according to the second model parameters, and using the second training model to train the second medical data of the local database.
根据本申请的又一个实施例,还提供了一种存储介质,所述存储介质中存储有计算机程序,其中,所述计算机程序被设置为运行时执行上述任一项装置实施例中的步骤。According to another embodiment of the present application, there is also provided a storage medium in which a computer program is stored, wherein the computer program is configured to execute the steps in any one of the above-mentioned device embodiments when running.
根据本申请的又一个实施例,还提供了一种计算机设备,包括存储器和处理器,所述存储器中存储有计算机程序,所述处理器被设置为运行所述计算机程序以执行上述任一项方法实施例中的步骤。According to another embodiment of the present application, there is also provided a computer device, including a memory and a processor, the memory is stored with a computer program, and the processor is configured to run the computer program to execute any of the above Steps in the method embodiment.
通过本申请,服务器向多个客户端发送初始训练模型,使得客户端在本地训练本地的医疗数据,得到更新后的第一训练模型,仅需把第一训练模型的第一模型参数发送至服务器,无需将本地医疗数据整合汇总到服务器,保障了本地数据的安全性,减少了服务器的工作负荷和存储资源;服务器将得到的模型参数进行加权处理,并返回给多个客户端进行训练,使得多个客户端共享一个相同的训练模型,解决了相关技术中处理医疗数据的算法模型较为复杂,无法处理、不便于流动的大规模医疗数据等技术问题。Through this application, the server sends the initial training model to multiple clients, so that the client trains local medical data locally to obtain the updated first training model, and only needs to send the first model parameters of the first training model to the server , There is no need to integrate local medical data to the server, which ensures the security of local data, reduces the workload and storage resources of the server; the server weights the model parameters obtained and returns them to multiple clients for training, so Multiple clients share the same training model, which solves the technical problems of complicated medical data processing algorithm models in related technologies, and large-scale medical data that cannot be processed and inconvenient to flow.
附图说明Description of the drawings
此处所说明的附图用来提供对本申请的进一步理解,构成本申请的一部分,本申请的示意性实施例及其说明用于解释本申请,并不构成对本申请的不当限定。在附图中:The drawings described here are used to provide a further understanding of the application and constitute a part of the application. The exemplary embodiments and descriptions of the application are used to explain the application, and do not constitute an improper limitation of the application. In the attached picture:
图1是本申请实施例的一种数据的训练方法应用于计算机终端的硬件结构框图;FIG. 1 is a block diagram of the hardware structure of a data training method applied to a computer terminal according to an embodiment of the present application;
图2是根据本申请提供的一种数据的训练方法的流程图;Fig. 2 is a flowchart of a data training method provided according to the present application;
图3是根据本申请实施例的另一种数据的训练方法的结构框图;Fig. 3 is a structural block diagram of another data training method according to an embodiment of the present application;
图4是根据本申请实施例提供的基于联邦学习医疗数据的流程图;Figure 4 is a flow chart based on federated learning medical data provided according to an embodiment of the present application;
图5是根据本申请实施例的一种数据的训练装置的结构框图;Fig. 5 is a structural block diagram of a data training device according to an embodiment of the present application;
图6是根据本申请实施例的另一种数据的训练装置的结构框图;Fig. 6 is a structural block diagram of another data training device according to an embodiment of the present application;
图7是根据本申请实施例的一种数据的训练系统的结构框图。Fig. 7 is a structural block diagram of a data training system according to an embodiment of the present application.
具体实施方式detailed description
下文中将参考附图并结合实施例来详细说明本申请。需要说明的是,在不冲突的情况下,本申请中的实施例及实施例中的特征可以相互组合。Hereinafter, the present application will be described in detail with reference to the drawings and in conjunction with the embodiments. It should be noted that the embodiments in this application and the features in the embodiments can be combined with each other if there is no conflict.
需要说明的是,本申请的说明书和权利要求书及上述附图中的术语“第一”、“第二”等是用于区别类似的对象,而不必用于描述特定的顺序或先后次序。It should be noted that the terms "first" and "second" in the specification and claims of the application and the above-mentioned drawings are used to distinguish similar objects, and not necessarily used to describe a specific sequence or sequence.
实施例1Example 1
本申请实施例一所提供的方法实施例可以在移动终端、服务器、计算机终端或者类似的运算装置中执行。以运行在计算机终端上为例,图1是本申请实施例的一种数据的训练方法应用于计算机终端的硬件结构框图。如图1所示,计算机终端可以包括一个或多个(图1中仅示出一个)处理器102(处理器102可以包括但不限于微处理器MCU或可编程逻辑器件FPGA等的处理装置)和用于存储数据的存储器104,可选地,上述计算机终端还可以包括用于通信功能的传输设备106以及输入输出设备108。本领域普通技术人员可以理解,图1所示的结构仅为示意,其并不对上述计算机终端的结构造成限定。例如,计算机终端还可包括比图1中所示更多或者更少的组件,或者具有与图1所示不同的配置。The method embodiment provided in Embodiment 1 of the present application may be executed in a mobile terminal, a server, a computer terminal, or a similar computing device. Taking running on a computer terminal as an example, FIG. 1 is a hardware structural block diagram of a data training method applied to a computer terminal in an embodiment of the present application. As shown in FIG. 1, the computer terminal may include one or more (only one is shown in FIG. 1) processor 102 (the processor 102 may include, but is not limited to, a processing device such as a microprocessor MCU or a programmable logic device FPGA) And the memory 104 for storing data. Optionally, the above-mentioned computer terminal may also include a transmission device 106 and an input/output device 108 for communication functions. A person of ordinary skill in the art can understand that the structure shown in FIG. 1 is only for illustration, and does not limit the structure of the foregoing computer terminal. For example, the computer terminal may also include more or fewer components than shown in FIG. 1, or have a different configuration from that shown in FIG.
存储器104可用于存储计算机程序,例如,应用软件的软件程序以及模块,如本申请实施例中的数据训练方法对应的计算机程序,处理器102通过运行存储在存储器104内的计算机程序,从而执行各种功能应用以及数据处理,即实现上述的方法。存储器104可包括高速随机存储器,还可包括非易失性存储器,如一个或者多个磁性存储装置、闪存、或者其他非易失性固态存储器。在一些实例中,存储器104可进一步包括相对于处理器102远程设置的存储器,这些远程存储器可以通过网络连接至计算机终端。上述网络的实例包括但不限于互联网、企业内部网、局域网、移动通信网及其组合。The memory 104 can be used to store computer programs, for example, software programs and modules of application software, such as the computer programs corresponding to the data training method in the embodiment of the present application. The processor 102 executes the computer programs stored in the memory 104 by running the computer programs stored in the memory 104. This kind of functional application and data processing realizes the above-mentioned method. The memory 104 may include a high-speed random access memory, and may also include a non-volatile memory, such as one or more magnetic storage devices, flash memory, or other non-volatile solid-state memory. In some examples, the memory 104 may further include a memory remotely provided with respect to the processor 102, and these remote memories may be connected to a computer terminal through a network. Examples of the aforementioned networks include, but are not limited to, the Internet, corporate intranets, local area networks, mobile communication networks, and combinations thereof.
传输装置106用于经由一个网络接收或者发送数据。上述的网络具体实例可包括 计算机终端的通信供应商提供的无线网络。在一个实例中,传输装置106包括一个网络适配器(Network Interface Controller,简称为NIC),其可通过基站与其他网络设备相连从而可与互联网进行通讯。在一个实例中,传输装置106可以为射频(Radio Frequency,简称为RF)模块,其用于通过无线方式与互联网进行通讯。The transmission device 106 is used to receive or send data via a network. The above-mentioned specific examples of the network may include a wireless network provided by a communication provider of a computer terminal. In one example, the transmission device 106 includes a network adapter (Network Interface Controller, NIC for short), which can be connected to other network devices through a base station to communicate with the Internet. In an example, the transmission device 106 may be a radio frequency (RF) module, which is used to communicate with the Internet in a wireless manner.
在本实施例中提供了一种数据的训练方法,图2是根据本申请提供的一种数据的训练方法的流程图。如图2所示,该流程包括如下步骤:In this embodiment, a data training method is provided, and FIG. 2 is a flowchart of a data training method provided in this application. As shown in Figure 2, the process includes the following steps:
步骤S202,向多个客户端发送初始训练模型,其中,多个客户端均与服务器单独通信;Step S202, sending the initial training model to multiple clients, where each of the multiple clients communicates with the server individually;
步骤S204,接收多个客户端发送的多套第一模型参数,其中,第一模型参数是客户端根据本地数据库的第一医疗数据对初始训练模型进行训练得到的;Step S204, receiving multiple sets of first model parameters sent by multiple clients, where the first model parameters are obtained by the client training the initial training model according to the first medical data in the local database;
其中,客户端本地数据库的第一医疗数据可以包括患者的属性信息,患者的诊疗信息等,比如:患者的年龄、性别等个人身份信息,过往病史、处方效果等诊疗记录。Among them, the first medical data in the client's local database may include the patient's attribute information, the patient's diagnosis and treatment information, etc., such as: the patient's age, gender and other personal identification information, past medical history, prescription effects and other diagnosis and treatment records.
步骤S206,对多套第一模型参数进行加权平均,得到第二模型参数;Step S206: Perform weighted average on multiple sets of first model parameters to obtain second model parameters;
本实施例中,服务器依据联邦平均算法对多个客户端发送的模型参数进行加权计算,其中,权重是根据各个客户端的训练效果确定的。In this embodiment, the server performs a weighted calculation on the model parameters sent by multiple clients according to the federated average algorithm, where the weight is determined according to the training effect of each client.
步骤S208,将第二模型参数发送至多个客户端,其中,第二模型参数用于分别在多个客户端构建相同的第二训练模型。Step S208: Send the second model parameters to multiple clients, where the second model parameters are used to construct the same second training model on the multiple clients respectively.
通过本申请,服务器向多个客户端发送初始训练模型,使得客户端在本地训练本地的医疗数据,得到更新后的第一训练模型,仅需把第一训练模型的第一模型参数返回至服务器,无需将本地医疗数据整合汇总到服务器,保障了本地数据的安全性,减少了服务器的工作负荷和存储资源;服务器将得到的模型参数进行加权处理,并返回至多个客户端进行训练,使得多个客户端共享一个相同的训练模型,解决了相关技术中处理医疗数据的算法模型较为复杂,无法处理安全性较高、不便于流动的大规模医疗数据等技术问题。Through this application, the server sends the initial training model to multiple clients, so that the client trains local medical data locally to obtain the updated first training model, and only needs to return the first model parameters of the first training model to the server , There is no need to integrate local medical data to the server, which guarantees the security of local data and reduces the workload and storage resources of the server; the server weights the model parameters obtained and returns them to multiple clients for training, making more Each client shares the same training model, which solves the technical problems that the algorithm model for processing medical data in related technologies is relatively complex and cannot handle large-scale medical data with high security and inconvenient flow.
可选地,在对多套第一模型参数进行加权平均,得到第二模型参数之前,还包括:根据预先设置的私钥对第一模型参数进行解密,其中,私钥与对应多个客户端的公钥形成一组密钥对,公钥用于对第一模型参数进行加密。Optionally, before the weighted average of multiple sets of first model parameters is performed to obtain the second model parameters, the method further includes: decrypting the first model parameters according to a preset private key, where the private key corresponds to multiple clients The public key forms a set of key pairs, and the public key is used to encrypt the first model parameters.
在本实施例中,为了保证各个客户端与服务器之间的信息安全性,对两者之间传输的参数进行加密训练,加密方式如下:(1)服务器向各个客户端发送公钥,用以对需要交互的参数(即上述第一模型参数)进行加密,其中,服务器还设置有与该公钥对应的私钥,即公钥与私钥是一组密钥对;(2)服务器接收到加密后的参数后,根据私钥对参数进行解密。In this embodiment, in order to ensure the security of the information between each client and the server, the parameters transmitted between the two are encrypted and trained. The encryption method is as follows: (1) The server sends the public key to each client for Encrypt the parameters that need to be interacted (that is, the above-mentioned first model parameters), where the server is also provided with a private key corresponding to the public key, that is, the public key and the private key are a set of key pairs; (2) the server receives After the encrypted parameters, the parameters are decrypted according to the private key.
在一个可选的示例中,对多套第一模型参数进行加权平均,得到第二模型参数,包括:在M套第一模型参数中选择
Figure PCTCN2019118407-appb-000001
次第一模型参数,其中,每次选择N套第一模型参数,对每次选择的N套第一模型参数进行加权平均,得到一个一级模型参数,其中,N为小于M的整数;对
Figure PCTCN2019118407-appb-000002
个一级模型参数进行加权平均,得到第二模型参数。
In an optional example, the weighted average of multiple sets of first model parameters to obtain the second model parameters includes: selecting among M sets of first model parameters
Figure PCTCN2019118407-appb-000001
The first model parameters of the second time, where N sets of first model parameters are selected each time, and the N sets of first model parameters selected each time are weighted and averaged to obtain a first-level model parameter, where N is an integer less than M;
Figure PCTCN2019118407-appb-000002
The first-level model parameters are weighted and averaged to obtain the second model parameters.
在一个可选的实施例中,服务器采用联邦平均算法对每一个客户端本地数据库训练后的第一模型参数进行平均加权。以3个客户端为例,(即客户端1,客户端2,客户端3),按照每轮执行计算的客户端的选择比例,假设每轮从3个客户端中选择2个,共有
Figure PCTCN2019118407-appb-000003
种选法(即客户端1和客户端2一组,客户端1和客户端3一组,客户端2和客户端3一组);将客户端1和客户端2发送的第一模型参数进行加权,得到参数1(即上述一级模型参数),将客户端1和客户端3发送的第一模型参数加权得到参数2,将客户端2和客户端3发送的第一模型参数加权得到参数3;最后将参数1,参数2,参数3进行平均加权,从而得到第二模型参数(相当于二级模型参数)。
In an optional embodiment, the server uses a federated average algorithm to averagely weight the first model parameters after each client's local database training. Take 3 clients as an example, (namely client 1, client 2, client 3), according to the selection ratio of the clients in each round of calculation, assuming that 2 of the 3 clients are selected in each round, there are a total of
Figure PCTCN2019118407-appb-000003
A selection method (ie a group of client 1 and client 2, a group of client 1 and client 3, a group of client 2 and client 3); the first model parameter sent by client 1 and client 2 Perform weighting to obtain parameter 1 (that is, the above-mentioned first-level model parameters), weight the first model parameters sent by client 1 and client 3 to obtain parameter 2, and weight the first model parameters sent by client 2 and client 3 to obtain Parameter 3; Finally, parameter 1, parameter 2, and parameter 3 are averagely weighted to obtain the second model parameter (equivalent to the second-level model parameter).
在本实施例中提供了另一种数据的训练方法,应用于客户端,图3是根据本申请实施例的另一种数据的训练方法的结构框图。如图3所示,该流程包括如下步骤:In this embodiment, another data training method is provided, which is applied to the client. FIG. 3 is a structural block diagram of another data training method according to an embodiment of the present application. As shown in Figure 3, the process includes the following steps:
步骤S302,接收服务器发送的初始训练模型;Step S302, receiving the initial training model sent by the server;
步骤S304,根据本地数据库的第一医疗数据对初始训练模型进行训练,得到第一训练模型;Step S304, training the initial training model according to the first medical data in the local database to obtain the first training model;
步骤S306,将第一训练模型的第一模型参数发送至服务器,其中,服务器用于对多个客户端的多套第一模型参数进行加权平均,得到第二训练模型的第二模型参数,并将第二模型参数反馈至多个客户端;Step S306: Send the first model parameters of the first training model to the server, where the server is used to perform a weighted average on multiple sets of first model parameters of multiple clients to obtain the second model parameters of the second training model, and Feedback of the second model parameters to multiple clients;
步骤S308,根据第二模型参数构建第二训练模型,并使用第二训练模型训练本地数据库的第二医疗数据。In step S308, a second training model is constructed according to the second model parameters, and the second training model is used to train the second medical data of the local database.
通过本申请实施例,多个客户端根据服务器提供的初始训练模型,各自本地的医 疗数据进行训练,得到更新后的第一训练模型,仅需把第一训练模型的第一模型参数发送至服务器,无需将本地医疗数据整合汇总到服务器,保障了本地数据的安全性;多个客户端根据服务器返回的第二模型参数进行平均加权,使得多个客户端继续训练本地的医疗数据,从而实现了多个客户端共享一个相同的训练模型的目的,解决了相关技术中处理医疗数据的算法模型较为复杂,无法处理安全性较高、不便于流动的大规模医疗数据等技术问题。Through the embodiments of this application, multiple clients train based on the initial training model provided by the server and their respective local medical data to obtain the updated first training model, and only need to send the first model parameters of the first training model to the server , There is no need to integrate and aggregate local medical data to the server, which ensures the security of local data; multiple clients are averagely weighted according to the second model parameters returned by the server, so that multiple clients continue to train local medical data, thus achieving The purpose of multiple clients sharing the same training model solves the technical problems that the algorithm model for processing medical data in related technologies is relatively complicated and cannot handle large-scale medical data with high security and inconvenient flow.
在一个可选的实施例中,根据本地数据库的第一医疗数据对初始训练模型进行训练,得到第一训练模型,包括:使用本地数据库的第一医疗数据在初始训练模型上执行批量梯度计算,得到多个梯度值;计算多个梯度值的平均梯度;使用平均梯度更新初始训练模型的初始权重值,得到第一模型参数。In an optional embodiment, training the initial training model according to the first medical data of the local database to obtain the first training model includes: using the first medical data of the local database to perform batch gradient calculation on the initial training model, Obtain multiple gradient values; calculate the average gradient of the multiple gradient values; use the average gradient to update the initial weight value of the initial training model to obtain the first model parameter.
在本实施例中,由于各个客户端本地的医疗数据是不停的更新的,为了使得联邦学习训练出来的模型自适应于各个客户端,且本地医疗数据的损失最小,通过在初始训练模型上执行批量梯度计算(SGD算法,全称为Stochastic Gradient Descent,随机梯度下降),按照每轮执行计算的客户端设备的比例,计算多个客户端本地医疗数据的损失梯度,相当于以多条并行数据通道计算随机抽取子集中的客户端的梯度的平均值,根据梯度的平均值更新初始训练模型的权重;然后每个客户端都使用本地医疗数据根据梯度的平均值在当前模型(即上述初始训练模型)上进行一步梯度下降,服务器对得到的模型(即上述第二模型参数)进行加权平均。通过对多个客户端的多个模型求平均梯度,使得本地医疗数据的损失最小,要比单独在两个客户端上训练所获取的模型的效果更好。In this embodiment, since the local medical data of each client is constantly updated, in order to make the model trained by federated learning adaptive to each client, and the loss of local medical data is minimized, the initial training model is Perform batch gradient calculation (SGD algorithm, full name Stochastic Gradient Descent, stochastic gradient descent), according to the proportion of client devices that perform calculations in each round, calculate the loss gradient of multiple client local medical data, which is equivalent to multiple parallel data The channel calculates the average value of the gradient of the client in a randomly selected subset, and updates the weight of the initial training model according to the average value of the gradient; then each client uses the local medical data according to the average value of the gradient in the current model (that is, the initial training model mentioned above) ) Performs a step gradient descent, and the server performs a weighted average on the obtained model (that is, the aforementioned second model parameter). By averaging the gradients of multiple models of multiple clients, the loss of local medical data is minimized, which is better than the model obtained by training on two clients separately.
下面结合一个具体实施例对本申请实施例做进一步的说明:The following further describes the embodiments of the application in combination with a specific embodiment:
图4是根据本申请实施例提供的基于联邦学习医疗数据的流程图,如图4所示,假设分布式数据中心有3个客户端,即图4中的数据集1号,数据集2号,数据集3号,以及中心服务器。中心服务器为分布式数据中心提供一个初始模型(即上述初始训练模型),数据集1根据其本地数据库中记录的自有数据(即上述本地数据库的第一医疗数据),对初始模型进行模型训练,得到模型更新1以及该模型的第一模型参数1;同时,数据集2根据其本地数据中记录的自有数据,对初始模型进行训练,得到模型更新2以及第一模型参数2;同理,对于数据集3,得到模型更新3以及第一模型参数 3。Figure 4 is a flow chart based on federated learning medical data provided according to an embodiment of the application. As shown in Figure 4, it is assumed that the distributed data center has 3 clients, namely data set No. 1 and data set No. 2 in Fig. 4 , Data Set No. 3, and the central server. The central server provides an initial model for the distributed data center (that is, the above-mentioned initial training model), and data set 1 performs model training on the initial model based on its own data recorded in its local database (that is, the first medical data in the above-mentioned local database) , Get model update 1 and the first model parameter 1 of the model; at the same time, data set 2 trains the initial model according to its own data recorded in its local data to obtain model update 2 and first model parameter 2; the same goes for , For data set 3, get model update 3 and first model parameter 3.
将三个模型参数发送至中心服务器,不必将各个分布式数据中心侧的数据集整合到中心服务器中,由此便可减轻中心服务器的工作负荷,进而能够提高中心服务器的处理速度。中心服务器将接收到的三个参数依据联邦平均算法进行加权计算,得到第二模型参数,中心服务器将第二模型参数返回数据集1、2、3。Sending the three model parameters to the central server eliminates the need to integrate the data sets on each distributed data center side into the central server, thereby reducing the workload of the central server and increasing the processing speed of the central server. The central server performs weighting calculation on the three received parameters according to the federated average algorithm to obtain the second model parameter, and the central server returns the second model parameter to the data set 1, 2, 3.
通过上述步骤,分布式数据中心不必将本地保存的医疗数据发送至中心服务器,而是将各个数据中心训练模型的模型参数经过加密处理后发送至中心服务器,从而保证了数据集侧的数据安全以及用户的个人隐私;中心服务器不必整合各个终端侧的数据集,而是将来自各个数据集的模型参数进行平均加权,得到第二模型参数,实现了统一更新各个数据集的模型的目的。减少了中心服务器的计算成本和计算时间,从而提高了服务器的处理效率。解决了相关技术中无法大规模训练安全性较高、不便于流动的用户医疗数据,导致医疗数据共享程度较低等技术问题。Through the above steps, the distributed data center does not have to send the locally stored medical data to the central server, but sends the model parameters of the training models of each data center to the central server after encryption processing, thereby ensuring the data security on the data set side and Personal privacy of the user; the central server does not need to integrate the data sets of each terminal side, but averagely weights the model parameters from each data set to obtain the second model parameter, which achieves the purpose of uniformly updating the model of each data set. The calculation cost and calculation time of the central server are reduced, thereby improving the processing efficiency of the server. It solves the technical problems such as the inability to train large-scale user medical data with high security and inconvenient flow in related technologies, resulting in a low degree of medical data sharing.
实施例2Example 2
在本实施例中还提供了一种数据的训练装置,该装置用于实现上述实施例及优选实施方式,已经进行过说明的不再赘述。如以下所使用的,术语“模块”可以实现预定功能的软件和/或硬件的组合。尽管以下实施例所描述的装置较佳地以软件来实现,但是硬件,或者软件和硬件的组合的实现也是可能并被构想的。In this embodiment, a data training device is also provided, which is used to implement the above-mentioned embodiments and preferred implementations, and those that have been explained will not be repeated. As used below, the term "module" can implement a combination of software and/or hardware with predetermined functions. Although the devices described in the following embodiments are preferably implemented by software, implementation by hardware or a combination of software and hardware is also possible and conceived.
图5是根据本申请实施例的一种数据的训练装置的结构框图,如图5所示,该装置包括:第一发送模块502,用于向多个客户端发送初始训练模型,其中,多个客户端均与服务器单独通信;接收模块504,连接至上述第一发送模块502,用于接收多个客户端发送的多套第一模型参数,其中,第一模型参数是客户端根据本地数据库的第一医疗数据对初始训练模型进行训练得到的;计算模块506,连接至上述接收模块504,用于对多套第一模型参数进行加权平均,得到第二模型参数;第二发送模块508,连接至上述计算模块506,用于将第二模型参数发送至多个客户端,其中,第二模型参数用于分别在多个客户端构建相同的第二训练模型。FIG. 5 is a structural block diagram of a data training device according to an embodiment of the present application. As shown in FIG. 5, the device includes: a first sending module 502, configured to send initial training models to multiple clients, where multiple Each client communicates with the server separately; the receiving module 504, connected to the first sending module 502, is used to receive multiple sets of first model parameters sent by multiple clients, where the first model parameter is the client according to the local database The calculation module 506, connected to the receiving module 504, is used to perform a weighted average of multiple sets of first model parameters to obtain the second model parameters; the second sending module 508, It is connected to the aforementioned calculation module 506 and is used to send the second model parameters to multiple clients, where the second model parameters are used to construct the same second training model on the multiple clients respectively.
可选的,上述装置还包括:解密模块,用于在对多套第一模型参数进行加权平均,得到第二模型参数之前,根据预先设置的私钥对第一模型参数进行解密,其中,私钥 与对应多个客户端的公钥形成一组密钥对,公钥用于对第一模型参数进行加密。Optionally, the above-mentioned device further includes: a decryption module, which is used to decrypt the first model parameters according to a preset private key before performing a weighted average of multiple sets of first model parameters to obtain the second model parameters, wherein the private The key and the public keys corresponding to multiple clients form a set of key pairs, and the public key is used to encrypt the first model parameter.
可选的,计算模块包括:选择单元,用于在M套第一模型参数中选择
Figure PCTCN2019118407-appb-000004
次第一模型参数,其中,每次选择N套第一模型参数,对每次选择的N套第一模型参数进行加权平均,得到一个一级模型参数,其中,N为小于M的整数;计算单元,用于对
Figure PCTCN2019118407-appb-000005
个一级模型参数进行加权平均,得到第二模型参数。
Optionally, the calculation module includes: a selection unit for selecting among M sets of first model parameters
Figure PCTCN2019118407-appb-000004
The second first model parameter, where N sets of first model parameters are selected each time, and the N sets of first model parameters selected each time are weighted and averaged to obtain a first-level model parameter, where N is an integer less than M; calculation Unit for
Figure PCTCN2019118407-appb-000005
The first-level model parameters are weighted and averaged to obtain the second model parameters.
图6是根据本申请实施例的另一种数据的训练装置的结构框图,如图6所示,该装置包括:接收模块602,用于接收服务器发送的初始训练模型;第一训练模块604,连接至上述接收模块602,用于根据本地数据库的第一医疗数据对初始训练模型进行训练,得到第一训练模型;发送模块606,连接至上述第一训练模块604用于将第一训练模型的第一模型参数发送至服务器,其中,服务器用于对多个客户端的多套第一模型参数进行加权平均,得到第二训练模型的第二模型参数,并将第二模型参数反馈至多个客户端;第二训练模块608,连接至上述发送模块606,用于根据第二模型参数构建第二训练模型,并使用第二训练模型训练本地数据库的第二医疗数据。Fig. 6 is a structural block diagram of another data training device according to an embodiment of the present application. As shown in Fig. 6, the device includes: a receiving module 602 for receiving an initial training model sent by a server; and a first training module 604, Connected to the receiving module 602, used to train the initial training model according to the first medical data in the local database, to obtain the first training model; the sending module 606, connected to the first training module 604, used to train the first training model The first model parameters are sent to the server, where the server is used to perform a weighted average of multiple sets of first model parameters of multiple clients to obtain the second model parameters of the second training model, and feed back the second model parameters to the multiple clients The second training module 608, connected to the above-mentioned sending module 606, is used to construct a second training model according to the second model parameters, and use the second training model to train the second medical data of the local database.
可选的,第一训练模块包括:第一计算单元,用于使用本地数据库的第一医疗数据在初始训练模型上执行批量梯度计算,得到多个梯度值;第二计算单元,用于计算多个梯度值的平均梯度;第三计算单元,用于使用平均梯度更新初始训练模型的初始权重值,得到第一模型参数。Optionally, the first training module includes: a first calculation unit for performing batch gradient calculations on the initial training model using the first medical data in a local database to obtain multiple gradient values; and a second calculation unit for calculating multiple gradients. The average gradient of each gradient value; the third calculation unit is used to update the initial weight value of the initial training model using the average gradient to obtain the first model parameter.
根据本申请的又一个实施例,还提供了一种数据的训练系统,图7是根据本申请实施例的一种数据的训练系统的结构框图,包括:服务器和多个客户端,其中,服务器包括:第一发送模块,用于向多个客户端发送初始训练模型;接收模块,用于接收多个客户端发送的多套第一模型参数;计算模块,用于对多套第一模型参数进行加权平均,得到第二模型参数;第二发送模块,用于将第二模型参数发送至多个客户端;多个客户端,均与服务器单独通信,包括:接收模块,用于接收初始训练模型;第一训练模块,用于根据本地数据库的第一医疗数据对初始训练模型进行训练,得到第一训练模型;发送模块,用于将第一训练模型的第一模型参数发送至服务器;第二训练模块,用于根据第二模型参数构建第二训练模型,并使用第二训练模型训练本地数据库的第二医疗数据。According to another embodiment of the present application, a data training system is also provided. FIG. 7 is a structural block diagram of a data training system according to an embodiment of the present application, including: a server and multiple clients, where the server Including: a first sending module, used to send initial training models to multiple clients; a receiving module, used to receive multiple sets of first model parameters sent by multiple clients; a calculation module, used to compare multiple sets of first model parameters Perform a weighted average to obtain the second model parameters; the second sending module is used to send the second model parameters to multiple clients; multiple clients, all communicating with the server separately, including: a receiving module, used to receive the initial training model The first training module is used to train the initial training model according to the first medical data in the local database to obtain the first training model; the sending module is used to send the first model parameters of the first training model to the server; second The training module is used to construct a second training model according to the second model parameters, and use the second training model to train the second medical data of the local database.
需要说明的是,上述各个模块是可以通过软件或硬件来实现的,对于后者,可以 通过以下方式实现,但不限于此:上述模块均位于同一处理器中;或者,上述各个模块以任意组合的形式分别位于不同的处理器中。It should be noted that each of the above modules can be implemented by software or hardware. For the latter, it can be implemented in the following manner, but not limited to this: the above modules are all located in the same processor; or, the above modules can be combined in any combination. The forms are located in different processors.
实施例3Example 3
本申请的实施例还提供了一种存储介质,该存储介质中存储有计算机程序,其中,该计算机程序被设置为运行时执行上述任一项方法实施例中的步骤。The embodiment of the present application also provides a storage medium in which a computer program is stored, wherein the computer program is configured to execute the steps in any one of the foregoing method embodiments when running.
可选地,在本实施例中,上述存储介质可以被设置为存储用于执行以下步骤的计算机程序:Optionally, in this embodiment, the aforementioned storage medium may be configured to store a computer program for executing the following steps:
S1,向多个客户端发送初始训练模型,其中,所述多个客户端均与服务器单独通信;S1. Send an initial training model to multiple clients, where each of the multiple clients communicates with the server individually;
S2,接收所述多个客户端发送的多套第一模型参数,其中,所述第一模型参数是所述客户端根据本地数据库的第一医疗数据对所述初始训练模型进行训练得到的;S2, receiving multiple sets of first model parameters sent by the multiple clients, where the first model parameters are obtained by the client training the initial training model according to the first medical data in a local database;
S3,对所述多套第一模型参数进行加权平均,得到第二模型参数;S3: Perform a weighted average on the multiple sets of first model parameters to obtain second model parameters;
S4,将所述第二模型参数发送至所述多个客户端,其中,所述第二模型参数用于分别在所述多个客户端构建相同的第二训练模型。S4. Send the second model parameters to the multiple clients, where the second model parameters are used to construct the same second training model on the multiple clients respectively.
可选地,在本实施例中,上述存储介质可以被设置为存储用于执行以下步骤的计算机程序:Optionally, in this embodiment, the aforementioned storage medium may be configured to store a computer program for executing the following steps:
S1,接收服务器发送的初始训练模型;S1, receiving the initial training model sent by the server;
S2,根据本地数据库的第一医疗数据对所述初始训练模型进行训练,得到第一训练模型;S2, training the initial training model according to the first medical data in the local database to obtain the first training model;
S3,将所述第一训练模型的第一模型参数发送至所述服务器,其中,所述服务器用于对多个客户端的多套第一模型参数进行加权平均,得到第二训练模型的第二模型参数,并将所述第二模型参数反馈至所述多个客户端;S3. Send the first model parameters of the first training model to the server, where the server is used to perform a weighted average on multiple sets of first model parameters of multiple clients to obtain the second model parameters of the second training model. Model parameters, and feedback the second model parameters to the multiple clients;
S4,根据所述第二模型参数构建第二训练模型,并使用所述第二训练模型训练所述本地数据库的第二医疗数据。S4: Construct a second training model according to the second model parameters, and use the second training model to train second medical data of the local database.
可选地,在本实施例中,上述存储介质可以包括但不限于:U盘、只读存储器(Read-Only Memory,简称为ROM)、随机存取存储器(Random Access Memory,简称为RAM)、移动硬盘、磁碟或者光盘等各种可以存储计算机程序的介质。Optionally, in this embodiment, the foregoing storage medium may include, but is not limited to: U disk, Read-Only Memory (Read-Only Memory, ROM for short), Random Access Memory (Random Access Memory, RAM for short), Various media that can store computer programs, such as mobile hard disks, magnetic disks, or optical disks.
本申请的实施例还提供了一种电子装置,包括存储器和处理器,该存储器中存储有计算机程序,该处理器被设置为运行计算机程序以执行上述任一项方法实施例中的步骤。An embodiment of the present application also provides an electronic device, including a memory and a processor, the memory stores a computer program, and the processor is configured to run the computer program to execute the steps in any one of the foregoing method embodiments.
可选地,上述电子装置还可以包括传输设备以及输入输出设备,其中,该传输设备和上述处理器连接,该输入输出设备和上述处理器连接。Optionally, the aforementioned electronic device may further include a transmission device and an input-output device, wherein the transmission device is connected to the aforementioned processor, and the input-output device is connected to the aforementioned processor.
可选地,在本实施例中,上述处理器可以被设置为通过计算机程序执行以下步骤:Optionally, in this embodiment, the foregoing processor may be configured to execute the following steps through a computer program:
S1,向多个客户端发送初始训练模型,其中,所述多个客户端均与服务器单独通信;S1. Send an initial training model to multiple clients, where each of the multiple clients communicates with the server individually;
S2,接收所述多个客户端发送的多套第一模型参数,其中,所述第一模型参数是所述客户端根据本地数据库的第一医疗数据对所述初始训练模型进行训练得到的;S2, receiving multiple sets of first model parameters sent by the multiple clients, where the first model parameters are obtained by the client training the initial training model according to the first medical data in a local database;
S3,对所述多套第一模型参数进行加权平均,得到第二模型参数;S3: Perform a weighted average on the multiple sets of first model parameters to obtain second model parameters;
S4,将所述第二模型参数发送至所述多个客户端,其中,所述第二模型参数用于分别在所述多个客户端构建相同的第二训练模型。S4. Send the second model parameters to the multiple clients, where the second model parameters are used to construct the same second training model on the multiple clients respectively.
可选地,在本实施例中,上述处理器可以被设置为通过计算机程序执行以下步骤:Optionally, in this embodiment, the foregoing processor may be configured to execute the following steps through a computer program:
S1,接收服务器发送的初始训练模型;S1, receiving the initial training model sent by the server;
S2,根据本地数据库的第一医疗数据对所述初始训练模型进行训练,得到第一训练模型;S2, training the initial training model according to the first medical data in the local database to obtain the first training model;
S3,将所述第一训练模型的第一模型参数发送至所述服务器,其中,所述服务器用于对多个客户端的多套第一模型参数进行加权平均,得到第二训练模型的第二模型参数,并将所述第二模型参数反馈至所述多个客户端;S3. Send the first model parameters of the first training model to the server, where the server is used to perform a weighted average on multiple sets of first model parameters of multiple clients to obtain the second model parameters of the second training model. Model parameters, and feedback the second model parameters to the multiple clients;
S4,根据所述第二模型参数构建第二训练模型,并使用所述第二训练模型训练所述本地数据库的第二医疗数据。S4: Construct a second training model according to the second model parameters, and use the second training model to train second medical data of the local database.
可选地,本实施例中的具体示例可以参考上述实施例及可选实施方式中所描述的示例,本实施例在此不再赘述。Optionally, for specific examples in this embodiment, reference may be made to the examples described in the above-mentioned embodiments and optional implementation manners, and details are not described herein again in this embodiment.
显然,本领域的技术人员应该明白,上述的本申请的各模块或各步骤可以用通用的计算装置来实现,它们可以集中在单个的计算装置上,或者分布在多个计算装置所组成的网络上,可选地,它们可以用计算装置可执行的程序代码来实现,从而,可以 将它们存储在存储装置中由计算装置来执行,并且在某些情况下,可以以不同于此处的顺序执行所示出或描述的步骤,或者将它们分别制作成各个集成电路模块,或者将它们中的多个模块或步骤制作成单个集成电路模块来实现。这样,本申请不限制于任何特定的硬件和软件结合。Obviously, those skilled in the art should understand that the above-mentioned modules or steps of this application can be implemented by a general computing device, and they can be concentrated on a single computing device or distributed in a network composed of multiple computing devices. Above, alternatively, they can be implemented with program codes executable by a computing device, so that they can be stored in a storage device for execution by the computing device, and in some cases, can be executed in a different order than here. Perform the steps shown or described, or fabricate them into individual integrated circuit modules respectively, or fabricate multiple modules or steps of them into a single integrated circuit module for implementation. In this way, this application is not limited to any specific combination of hardware and software.
以上所述仅为本申请的优选实施例而已,并不用于限制本申请,对于本领域的技术人员来说,本申请可以有各种更改和变化。凡在本申请的原则之内,所作的任何修改、等同替换、改进等,均应包含在本申请的保护范围之内。The above descriptions are only preferred embodiments of the application, and are not intended to limit the application. For those skilled in the art, the application can have various modifications and changes. Any modification, equivalent replacement, improvement, etc. made within the principles of this application shall be included in the protection scope of this application.

Claims (20)

  1. 一种数据的训练方法,其中,包括:A data training method, which includes:
    向多个客户端发送初始训练模型,其中,所述多个客户端均与服务器单独通信;Sending the initial training model to multiple clients, where each of the multiple clients communicates with the server separately;
    接收所述多个客户端发送的多套第一模型参数,其中,所述第一模型参数是所述客户端根据本地数据库的第一医疗数据对所述初始训练模型进行训练得到的;Receiving multiple sets of first model parameters sent by the multiple clients, where the first model parameters are obtained by the client training the initial training model according to the first medical data in the local database;
    对所述多套第一模型参数进行加权平均,得到第二模型参数;Performing a weighted average on the multiple sets of first model parameters to obtain a second model parameter;
    将所述第二模型参数发送至所述多个客户端,其中,所述第二模型参数用于分别在所述多个客户端构建相同的第二训练模型。The second model parameter is sent to the multiple clients, where the second model parameter is used to construct the same second training model on the multiple clients respectively.
  2. 根据权利要求1所述的方法,在对所述多套第一模型参数进行加权平均,得到第二模型参数之前,所述方法还包括:The method according to claim 1, before the multiple sets of first model parameters are weighted and averaged to obtain the second model parameters, the method further comprises:
    根据预先设置的私钥对所述第一模型参数进行解密,其中,所述私钥与对应所述多个客户端的公钥形成一组密钥对,所述公钥用于对所述第一模型参数进行加密。The first model parameter is decrypted according to the preset private key, wherein the private key and the public keys corresponding to the multiple clients form a set of key pairs, and the public key is used to decrypt the first model parameter. Model parameters are encrypted.
  3. 根据权利要求1所述的方法,对所述多套第一模型参数进行加权平均,得到第二模型参数,包括:The method according to claim 1, wherein the weighted average of the multiple sets of first model parameters to obtain the second model parameters includes:
    在M套第一模型参数中选择
    Figure PCTCN2019118407-appb-100001
    次第一模型参数,其中,每次选择N套第一模型参数,对每次选择的N套所述第一模型参数进行加权平均,得到一个一级模型参数,其中,N为小于M的整数;
    Choose from M sets of first model parameters
    Figure PCTCN2019118407-appb-100001
    The first model parameter of the second time, where N sets of first model parameters are selected each time, and the N sets of the first model parameters selected each time are weighted and averaged to obtain a first-level model parameter, where N is an integer less than M ;
    Figure PCTCN2019118407-appb-100002
    个所述一级模型参数进行加权平均,得到所述第二模型参数。
    Correct
    Figure PCTCN2019118407-appb-100002
    Performing a weighted average of the first-level model parameters to obtain the second model parameter.
  4. 一种数据的训练方法,其中,包括:A data training method, which includes:
    接收服务器发送的初始训练模型;Receive the initial training model sent by the server;
    根据本地数据库的第一医疗数据对所述初始训练模型进行训练,得到第一训练模型;Training the initial training model according to the first medical data in the local database to obtain the first training model;
    将所述第一训练模型的第一模型参数发送至所述服务器,其中,所述服务器用于对多个客户端的多套第一模型参数进行加权平均,得到第二训练模型的第二模型参数,并将所述第二模型参数反馈至所述多个客户端;The first model parameters of the first training model are sent to the server, where the server is used to perform a weighted average of multiple sets of first model parameters of multiple clients to obtain the second model parameters of the second training model , And feed back the second model parameters to the multiple clients;
    根据所述第二模型参数构建第二训练模型,并使用所述第二训练模型训练所述本地数据库的第二医疗数据。A second training model is constructed according to the second model parameters, and the second training model is used to train the second medical data of the local database.
  5. 根据权利要求4所述的方法,根据本地数据库的第一医疗数据对所述初始训练模型进行训练,得到第一训练模型,包括:The method according to claim 4, training the initial training model according to the first medical data in the local database to obtain the first training model, comprising:
    使用所述本地数据库的第一医疗数据在所述初始训练模型上执行批量梯度计算,得到多个梯度值;Using the first medical data of the local database to perform batch gradient calculation on the initial training model to obtain multiple gradient values;
    计算所述多个梯度值的平均梯度;Calculating the average gradient of the multiple gradient values;
    使用所述平均梯度更新所述初始训练模型的初始权重值,得到所述第一模型参数。The average gradient is used to update the initial weight value of the initial training model to obtain the first model parameter.
  6. 一种数据的处理训练装置,其中,包括:A data processing and training device, which includes:
    第一发送模块,用于向多个客户端发送初始训练模型,其中,所述多个客户端均与服务器单独通信;The first sending module is configured to send the initial training model to multiple clients, where each of the multiple clients communicates with the server separately;
    接收模块,用于接收所述多个客户端发送的多套第一模型参数,其中,所述第一模型参数是所述客户端根据本地数据库的第一医疗数据对所述初始训练模型进行训练得到的;The receiving module is configured to receive multiple sets of first model parameters sent by the multiple clients, where the first model parameters are the clients training the initial training model according to the first medical data in the local database owned;
    计算模块,用于对所述多套第一模型参数进行加权平均,得到第二模型参数;A calculation module, configured to perform a weighted average of the multiple sets of first model parameters to obtain a second model parameter;
    第二发送模块,用于将所述第二模型参数发送至所述多个客户端,其中,所 述第二模型参数用于分别在所述多个客户端构建相同的第二训练模型。The second sending module is configured to send the second model parameters to the multiple clients, where the second model parameters are used to construct the same second training model on the multiple clients respectively.
  7. 根据权利要求6所述的装置,所述装置还包括:The device according to claim 6, the device further comprising:
    解密模块,用于在对所述多套第一模型参数进行加权平均,得到第二模型参数之前,根据预先设置的私钥对所述第一模型参数进行解密,其中,所述私钥与向所述目标终端发送的公钥是一组密钥对,所述公钥用于对所述第一模型参数进行加密。The decryption module is used to decrypt the first model parameters according to a preset private key before the multiple sets of first model parameters are weighted and averaged to obtain the second model parameters. The public key sent by the target terminal is a set of key pairs, and the public key is used to encrypt the first model parameter.
  8. 根据权利要求6所述的装置,所述计算模块包括:The device according to claim 6, wherein the calculation module comprises:
    选择单元,用于在所述M套第一模型参数中选择
    Figure PCTCN2019118407-appb-100003
    次第一模型参数,其中,每次选择N套第一模型参数,对每次选择的N套所述第一模型参数进行加权平均,得到一个一级模型参数,其中,N为小于M的整数;计算单元,用于对
    Figure PCTCN2019118407-appb-100004
    个所述一级模型参数进行加权平均,得到所述第二模型参数。
    The selection unit is used to select among the M sets of first model parameters
    Figure PCTCN2019118407-appb-100003
    The first model parameter of the second time, where N sets of first model parameters are selected each time, and the N sets of the first model parameters selected each time are weighted and averaged to obtain a first-level model parameter, where N is an integer less than M ;Computer unit, used to
    Figure PCTCN2019118407-appb-100004
    Performing a weighted average of the first-level model parameters to obtain the second model parameter.
  9. 一种数据的训练装置,其中,包括:A data training device, which includes:
    接收模块,用于接收服务器发送的初始训练模型;The receiving module is used to receive the initial training model sent by the server;
    第一训练模块,用于根据本地数据库的第一医疗数据对所述初始训练模型进行训练,得到第一训练模型;The first training module is configured to train the initial training model according to the first medical data in the local database to obtain the first training model;
    发送模块,用于将所述第一训练模型的第一模型参数发送至所述服务器,其中,所述服务器用于对多个客户端的多套第一模型参数进行加权平均,得到第二训练模型的第二模型参数,并将所述第二模型参数反馈至所述多个客户端;The sending module is configured to send the first model parameters of the first training model to the server, where the server is configured to perform a weighted average on multiple sets of first model parameters of multiple clients to obtain a second training model And feedback the second model parameters to the multiple clients;
    第二训练模块,用于根据所述第二模型参数构建第二训练模型,并使用所述第二训练模型训练所述本地数据库的第二医疗数据。The second training module is configured to construct a second training model according to the second model parameters, and use the second training model to train the second medical data of the local database.
  10. 根据权利要求9所述的装置,所述第一训练模块包括:The device according to claim 9, wherein the first training module comprises:
    第一计算单元,用于使用所述本地数据库的第一医疗数据在所述初始训练模型上执行批量梯度计算,得到多个梯度值;The first calculation unit is configured to use the first medical data of the local database to perform batch gradient calculation on the initial training model to obtain multiple gradient values;
    第二计算单元,用于计算所述多个梯度值的平均梯度;The second calculation unit is used to calculate the average gradient of the multiple gradient values;
    第三计算单元,用于使用所述平均梯度更新所述初始训练模型的初始权重值,得到所述第一模型参数。The third calculation unit is configured to use the average gradient to update the initial weight value of the initial training model to obtain the first model parameter.
  11. 一种数据的训练系统,包括:服务器和多个客户端,其中,A data training system, including: a server and multiple clients, among which,
    所述服务器,包括:第一发送模块,用于向多个客户端发送初始训练模型;接收模块,用于接收所述多个客户端发送的多套第一模型参数;计算模块,用于对所述多套第一模型参数进行加权平均,得到第二模型参数;第二发送模块,用于将所述第二模型参数发送至所述多个客户端;The server includes: a first sending module, used to send initial training models to multiple clients; a receiving module, used to receive multiple sets of first model parameters sent by the multiple clients; Performing a weighted average of the multiple sets of first model parameters to obtain a second model parameter; a second sending module, configured to send the second model parameter to the multiple clients;
    所述多个客户端,均与服务器单独通信,包括:接收模块,用于接收所述初始训练模型;第一训练模块,用于根据本地数据库的第一医疗数据对所述初始训练模型进行训练,得到所述第一训练模型;发送模块,用于将所述第一训练模型的第一模型参数发送至所述服务器;第二训练模块,用于根据所述第二模型参数构建第二训练模型,并使用所述第二训练模型训练所述本地数据库的第二医疗数据。The multiple clients all communicate with the server separately, and include: a receiving module for receiving the initial training model; a first training module for training the initial training model according to the first medical data in the local database , Obtain the first training model; a sending module, for sending the first model parameters of the first training model to the server; a second training module, for constructing a second training according to the second model parameters Model, and use the second training model to train the second medical data of the local database.
  12. 根据权利要求11所述的系统,所述服务器还包括:The system according to claim 11, the server further comprising:
    解密模块,用于在对所述多套第一模型参数进行加权平均,得到第二模型参数之前,根据预先设置的私钥对所述第一模型参数进行解密,其中,所述私钥与向所述目标终端发送的公钥是一组密钥对,所述公钥用于对所述第一模型参数进行加密。The decryption module is used to decrypt the first model parameters according to a preset private key before the multiple sets of first model parameters are weighted and averaged to obtain the second model parameters. The public key sent by the target terminal is a set of key pairs, and the public key is used to encrypt the first model parameter.
  13. 根据权利要求11所述的系统,所述计算模块包括:The system according to claim 11, said calculation module comprising:
    选择单元,用于在所述M套第一模型参数中选择
    Figure PCTCN2019118407-appb-100005
    次第一模型参数,其中,每次选择N套第一模型参数,对每次选择的N套所述第一模型参数进行加权平均,得到一个一级模型参数,其中,N为小于M的整数;计算单元,用于对
    Figure PCTCN2019118407-appb-100006
    个所述一级模型参数进行加权平均,得到所述第二模型参数。
    The selection unit is used to select among the M sets of first model parameters
    Figure PCTCN2019118407-appb-100005
    The first model parameter of the second time, where N sets of first model parameters are selected each time, and the N sets of the first model parameters selected each time are weighted and averaged to obtain a first-level model parameter, where N is an integer less than M ;Computer unit, used to
    Figure PCTCN2019118407-appb-100006
    Performing a weighted average of the first-level model parameters to obtain the second model parameter.
  14. 根据权利要求11所述的系统,所述第一训练模块包括:The system according to claim 11, the first training module comprises:
    第一计算单元,用于使用所述本地数据库的第一医疗数据在所述初始训练模型上执行批量梯度计算,得到多个梯度值;The first calculation unit is configured to use the first medical data of the local database to perform batch gradient calculation on the initial training model to obtain multiple gradient values;
    第二计算单元,用于计算所述多个梯度值的平均梯度;The second calculation unit is used to calculate the average gradient of the multiple gradient values;
    第三计算单元,用于使用所述平均梯度更新所述初始训练模型的初始权重值,得到所述第一模型参数。The third calculation unit is configured to use the average gradient to update the initial weight value of the initial training model to obtain the first model parameter.
  15. 一种计算机设备,包括存储器和处理器,所述存储器存储有计算机程序,所述处理器执行所述计算机程序时实现一种数据的训练方法的步骤,包括:A computer device includes a memory and a processor, the memory stores a computer program, and the steps of implementing a data training method when the processor executes the computer program include:
    向多个客户端发送初始训练模型,其中,所述多个客户端均与服务器单独通信;Sending the initial training model to multiple clients, where each of the multiple clients communicates with the server separately;
    接收所述多个客户端发送的多套第一模型参数,其中,所述第一模型参数是所述客户端根据本地数据库的第一医疗数据对所述初始训练模型进行训练得到的;Receiving multiple sets of first model parameters sent by the multiple clients, where the first model parameters are obtained by the client training the initial training model according to the first medical data in the local database;
    对所述多套第一模型参数进行加权平均,得到第二模型参数;Performing a weighted average on the multiple sets of first model parameters to obtain a second model parameter;
    将所述第二模型参数发送至所述多个客户端,其中,所述第二模型参数用于分别在所述多个客户端构建相同的第二训练模型。The second model parameter is sent to the multiple clients, where the second model parameter is used to construct the same second training model on the multiple clients respectively.
  16. 根据权利要求15所述的计算机设备,在对所述多套第一模型参数进行加权平均,得到第二模型参数之前,所述方法还包括:The computer device according to claim 15, before the weighted average of the multiple sets of first model parameters is performed to obtain the second model parameters, the method further comprises:
    根据预先设置的私钥对所述第一模型参数进行解密,其中,所述私钥与对应所述多个客户端的公钥形成一组密钥对,所述公钥用于对所述第一模型参数进行加密。The first model parameter is decrypted according to the preset private key, wherein the private key and the public keys corresponding to the multiple clients form a set of key pairs, and the public key is used to decrypt the first model parameter. Model parameters are encrypted.
  17. 一种计算机设备,包括存储器和处理器,所述存储器存储有计算机程序,所述处理器执行所述计算机程序时实现一种数据的训练方法的步骤,包括:A computer device includes a memory and a processor, the memory stores a computer program, and the steps of implementing a data training method when the processor executes the computer program include:
    接收服务器发送的初始训练模型;Receive the initial training model sent by the server;
    根据本地数据库的第一医疗数据对所述初始训练模型进行训练,得到第一训练模型;Training the initial training model according to the first medical data in the local database to obtain the first training model;
    将所述第一训练模型的第一模型参数发送至所述服务器,其中,所述服务器用于对多个客户端的多套第一模型参数进行加权平均,得到第二训练模型的第二模型参数,并将所述第二模型参数反馈至所述多个客户端;The first model parameters of the first training model are sent to the server, where the server is used to perform a weighted average of multiple sets of first model parameters of multiple clients to obtain the second model parameters of the second training model , And feed back the second model parameters to the multiple clients;
    根据所述第二模型参数构建第二训练模型,并使用所述第二训练模型训练所述本地数据库的第二医疗数据。A second training model is constructed according to the second model parameters, and the second training model is used to train the second medical data of the local database.
  18. 根据权利要求17所述的计算机设备,根据本地数据库的第一医疗数据对所述初始训练模型进行训练,得到第一训练模型,包括:The computer device according to claim 17, training the initial training model according to the first medical data in the local database to obtain the first training model, comprising:
    使用所述本地数据库的第一医疗数据在所述初始训练模型上执行批量梯度计算,得到多个梯度值;Using the first medical data of the local database to perform batch gradient calculation on the initial training model to obtain multiple gradient values;
    计算所述多个梯度值的平均梯度;Calculating the average gradient of the multiple gradient values;
    使用所述平均梯度更新所述初始训练模型的初始权重值,得到所述第一模型参数。The average gradient is used to update the initial weight value of the initial training model to obtain the first model parameter.
  19. 一种计算机存储介质,其上存储有计算机程序,所述计算机程序被处理器执行时实现一种数据的训练方法的步骤,包括:A computer storage medium having a computer program stored thereon, and the steps of implementing a data training method when the computer program is executed by a processor include:
    向多个客户端发送初始训练模型,其中,所述多个客户端均与服务器单独通信;Sending the initial training model to multiple clients, where each of the multiple clients communicates with the server separately;
    接收所述多个客户端发送的多套第一模型参数,其中,所述第一模型参数是所述客户端根据本地数据库的第一医疗数据对所述初始训练模型进行训练得到的;Receiving multiple sets of first model parameters sent by the multiple clients, where the first model parameters are obtained by the client training the initial training model according to the first medical data in the local database;
    对所述多套第一模型参数进行加权平均,得到第二模型参数;Performing a weighted average on the multiple sets of first model parameters to obtain a second model parameter;
    将所述第二模型参数发送至所述多个客户端,其中,所述第二模型参数用于分别在所述多个客户端构建相同的第二训练模型。The second model parameter is sent to the multiple clients, where the second model parameter is used to construct the same second training model on the multiple clients respectively.
  20. 一种计算机存储介质,其上存储有计算机程序,所述计算机程序被处理器执行时实现一种数据的训练方法的步骤,包括:A computer storage medium having a computer program stored thereon, and the steps of implementing a data training method when the computer program is executed by a processor include:
    接收服务器发送的初始训练模型;Receive the initial training model sent by the server;
    根据本地数据库的第一医疗数据对所述初始训练模型进行训练,得到第一训练模型;Training the initial training model according to the first medical data in the local database to obtain the first training model;
    将所述第一训练模型的第一模型参数发送至所述服务器,其中,所述服务器用于对多个客户端的多套第一模型参数进行加权平均,得到第二训练模型的第二模型参数,并将所述第二模型参数反馈至所述多个客户端;The first model parameters of the first training model are sent to the server, where the server is used to perform a weighted average of multiple sets of first model parameters of multiple clients to obtain the second model parameters of the second training model , And feed back the second model parameters to the multiple clients;
    根据所述第二模型参数构建第二训练模型,并使用所述第二训练模型训练所述本地数据库的第二医疗数据。A second training model is constructed according to the second model parameters, and the second training model is used to train the second medical data of the local database.
PCT/CN2019/118407 2019-09-20 2019-11-14 Data training method, apparatus and system WO2021051610A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201910894089.3 2019-09-20
CN201910894089.3A CN110795477A (en) 2019-09-20 2019-09-20 Data training method, device and system

Publications (1)

Publication Number Publication Date
WO2021051610A1 true WO2021051610A1 (en) 2021-03-25

Family

ID=69438611

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2019/118407 WO2021051610A1 (en) 2019-09-20 2019-11-14 Data training method, apparatus and system

Country Status (2)

Country Link
CN (1) CN110795477A (en)
WO (1) WO2021051610A1 (en)

Families Citing this family (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113377931A (en) * 2020-03-09 2021-09-10 香港理工大学深圳研究院 Language model collaborative learning method, system and terminal of interactive robot
CN115280338A (en) * 2020-03-11 2022-11-01 Oppo广东移动通信有限公司 Model training method based on federal learning, electronic equipment and storage medium
CN113470806A (en) * 2020-03-31 2021-10-01 中移(成都)信息通信科技有限公司 Method, device and equipment for determining disease detection model and computer storage medium
CN111477336A (en) * 2020-04-07 2020-07-31 中南大学 Fusion method, system and storage medium for infectious disease diagnosis data
CN113935390A (en) * 2020-06-29 2022-01-14 中兴通讯股份有限公司 Data processing method, system, device and storage medium
CN113988254B (en) * 2020-07-27 2023-07-14 腾讯科技(深圳)有限公司 Method and device for determining neural network model for multiple environments
CN111935156B (en) * 2020-08-12 2022-06-14 科技谷(厦门)信息技术有限公司 Data privacy protection method for federated learning
CN112001321A (en) * 2020-08-25 2020-11-27 商汤国际私人有限公司 Network training method, pedestrian re-identification method, network training device, pedestrian re-identification device, electronic equipment and storage medium
CN112036504A (en) * 2020-09-15 2020-12-04 平安国际智慧城市科技股份有限公司 Temperature measurement model training method, device, equipment and storage medium
CN112259238A (en) * 2020-10-20 2021-01-22 平安科技(深圳)有限公司 Electronic device, disease type detection method, apparatus, and medium
CN112446040A (en) * 2020-11-24 2021-03-05 平安科技(深圳)有限公司 Federal modeling method based on selective gradient update and related equipment
CN112434619B (en) * 2020-11-26 2024-03-26 新奥新智科技有限公司 Case information extraction method, apparatus, device and computer readable medium
CN112434620B (en) * 2020-11-26 2024-03-01 新奥新智科技有限公司 Scene text recognition method, device, equipment and computer readable medium
CN114548356A (en) * 2020-11-27 2022-05-27 华为技术有限公司 Machine learning method, device and system
CN112617855A (en) * 2020-12-31 2021-04-09 平安科技(深圳)有限公司 Electrocardiogram analysis method and device based on federal learning and related equipment
CN112711621A (en) * 2021-01-18 2021-04-27 湛江市前程网络有限公司 Universal object interconnection training platform and control method and device
CN113298229A (en) * 2021-04-12 2021-08-24 云从科技集团股份有限公司 Federal learning model training method, client, server and storage medium
CN113449318B (en) * 2021-06-18 2024-03-19 北京明朝万达科技股份有限公司 Data classification model training method and device and data classification method and device
CN113268727A (en) * 2021-07-19 2021-08-17 天聚地合(苏州)数据股份有限公司 Joint training model method, device and computer readable storage medium
CN113792759A (en) * 2021-08-19 2021-12-14 北京爱笔科技有限公司 Recognition model training method and device, computer equipment and storage medium
CN113792324B (en) * 2021-11-16 2022-04-05 聊城高新生物技术有限公司 Agricultural product data interaction method and device based on federal learning and electronic equipment
CN114742235A (en) * 2022-04-15 2022-07-12 中国电信股份有限公司 Federal learning method, data analysis model training method and device, and storage medium

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107871160A (en) * 2016-09-26 2018-04-03 谷歌公司 Communicate efficient joint study
US20190227980A1 (en) * 2018-01-22 2019-07-25 Google Llc Training User-Level Differentially Private Machine-Learned Models

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109716346A (en) * 2016-07-18 2019-05-03 河谷生物组学有限责任公司 Distributed machines learning system, device and method

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107871160A (en) * 2016-09-26 2018-04-03 谷歌公司 Communicate efficient joint study
US20190227980A1 (en) * 2018-01-22 2019-07-25 Google Llc Training User-Level Differentially Private Machine-Learned Models

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
H. BRENDAN MCMAHAN, MOORE EIDER, RAMAGE DANIEL, HAMPSON SETH, ARCAS BLAISE: "Communication-Efficient Learning of Deep Networks from Decentralized Data", 28 February 2017 (2017-02-28), pages 1 - 11, XP055538798, Retrieved from the Internet <URL:https://arxiv.org/pdf/1602.05629.pdf> *
YANG LIU; TIANJIAN CHEN; QIANG YANG: "Secure Federated Transfer Learning", ARXIV.ORG, CORNELL UNIVERSITY LIBRARY, 201 OLIN LIBRARY CORNELL UNIVERSITY ITHACA, NY 14853, 8 December 2018 (2018-12-08), 201 Olin Library Cornell University Ithaca, NY 14853, XP080990694 *
YUYUNSEU: "Fine-tuning: Testing Other Datasets by Using Existing Models", CSDN BLOG, 29 September 2016 (2016-09-29), XP055792803, Retrieved from the Internet <URL:https://blog.csdn.net/yj3254/article/details/52704361> *

Also Published As

Publication number Publication date
CN110795477A (en) 2020-02-14

Similar Documents

Publication Publication Date Title
WO2021051610A1 (en) Data training method, apparatus and system
WO2021120676A1 (en) Model training method for federated learning network, and related device
Wen et al. A survey on federated learning: challenges and applications
Xue et al. A resource-constrained and privacy-preserving edge-computing-enabled clinical decision system: A federated reinforcement learning approach
CN112712182B (en) Model training method and device based on federal learning and storage medium
CN109977694A (en) A kind of data sharing method based on cooperation deep learning
CN111986755A (en) Data sharing system based on block chain and attribute-based encryption
CN112231756B (en) FL-EM-GMM medical user privacy protection method and system
Aouedi et al. Handling privacy-sensitive medical data with federated learning: challenges and future directions
CN111753006B (en) Prediction system and method based on federal learning
EP2195962A1 (en) Health keyset management
WO2018099577A1 (en) System and method for providing a collective decentralized authority for sharing sensitive data
CN112118099A (en) Distributed multi-task learning privacy protection method and system for resisting inference attack
CN113947156A (en) Health crowd-sourcing perception system and federal learning method for cost optimization thereof
US20220415455A1 (en) Systems and methods for virtual clinical trials
Tian et al. Robust and privacy-preserving decentralized deep federated learning training: Focusing on digital healthcare applications
CN114615094A (en) Storage method and device based on Internet of things and security chip
CN112289448A (en) Health risk prediction method and device based on joint learning
CN113988310A (en) Deep learning model selection method and device, computer equipment and medium
CN111581648B (en) Method of federal learning to preserve privacy in irregular users
CN110610098A (en) Data set generation method and device
CN111429288A (en) User portrait construction method and device, computer equipment and storage medium
Singh et al. An IoT and blockchain-based secure medical care framework using deep learning and nature-inspired algorithms
CN110472758A (en) Medical treatment reserving method, device, equipment and readable storage medium storing program for executing
CN114491616A (en) Block chain and homomorphic encryption-based federated learning method and application

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19945764

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 19945764

Country of ref document: EP

Kind code of ref document: A1