WO2021184836A1 - Method and apparatus for training recognition model, device, and readable storage medium - Google Patents

Method and apparatus for training recognition model, device, and readable storage medium Download PDF

Info

Publication number
WO2021184836A1
WO2021184836A1 PCT/CN2020/134029 CN2020134029W WO2021184836A1 WO 2021184836 A1 WO2021184836 A1 WO 2021184836A1 CN 2020134029 W CN2020134029 W CN 2020134029W WO 2021184836 A1 WO2021184836 A1 WO 2021184836A1
Authority
WO
WIPO (PCT)
Prior art keywords
recognition model
recognition
model
training
client
Prior art date
Application number
PCT/CN2020/134029
Other languages
French (fr)
Chinese (zh)
Inventor
姜迪
Original Assignee
深圳前海微众银行股份有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 深圳前海微众银行股份有限公司 filed Critical 深圳前海微众银行股份有限公司
Publication of WO2021184836A1 publication Critical patent/WO2021184836A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/12Computing arrangements based on biological models using genetic models
    • G06N3/126Evolutionary algorithms, e.g. genetic algorithms or genetic programming
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q40/00Finance; Insurance; Tax strategies; Processing of corporate or income taxes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/172Classification, e.g. identification

Definitions

  • This application relates to the field of data processing technology of financial technology (Fintech), and in particular to a training method, device, equipment and readable storage medium of a recognition model.
  • Machine learning technology is widely used in speech recognition. These technologies often rely heavily on massive speech data for training, which can easily leak the speech data, and thus cannot effectively protect the privacy of the users corresponding to the speech data. Moreover, the current speech recognition system can only provide very limited customization capabilities, which leads to low accuracy of speech recognition by the resulting speech recognition model. From this, it can be seen that the recognition accuracy of the currently trained recognition model is low, and the training data is likely to be leaked in the process of training the recognition model.
  • the main purpose of this application is to provide a training method, device, equipment and readable storage medium for a recognition model, aiming to solve the technical problem of low recognition accuracy of the existing recognition model and easy leakage of training data during the training of the recognition model .
  • the present application provides a method for training a recognition model, which is applied to a client, and the method for training a recognition model includes the steps:
  • Receive the federated learning result sent by the server update the first recognition model according to the federated learning result to obtain the corresponding second recognition model, and send the second recognition model to the server for
  • the server obtains the target recognition model according to the second recognition model sent by each client and the preset genetic algorithm.
  • the present application also provides a training device for a recognition model.
  • the training device for the recognition model is applied to a client, and the training device for the recognition model includes:
  • the acquisition module is used to acquire local training data
  • the training module is used to train according to the training data to obtain the first recognition model
  • the sending module is configured to send the model parameters of the first recognition model to the server, so that the server performs federated learning according to the model parameters sent by each client, obtains the federated learning result, and returns the federated learning result ;
  • the receiving module is used to receive the federated learning result sent by the server;
  • An update module configured to update the first recognition model according to the federated learning result to obtain a corresponding second recognition model
  • the sending module is further configured to send the second recognition model to the server, so that the server obtains the target recognition model according to the second recognition model sent by each client and a preset genetic algorithm.
  • the present application also provides a recognition model training device, the recognition model training device is applied to the server, and the recognition model training device includes:
  • the receiving module is used to receive the model parameters corresponding to the first recognition model sent by each client;
  • the federated learning module is used to perform federated learning according to the model parameters to obtain federated learning results
  • a sending module configured to send the federated learning result to each client, so that each client can update the corresponding first recognition model according to the federated learning result to obtain the corresponding second recognition model;
  • the receiving module is also used to receive the second recognition model sent by each client, and obtain the target recognition model according to each second recognition model and a preset genetic algorithm.
  • the present application also provides a training device for a recognition model.
  • the training device for the recognition model includes a memory, a processor, and a recognition model stored in the memory and running on the processor.
  • the training program of the recognition model is executed by the processor, the steps of the training method of the recognition model corresponding to the federated learning server are implemented.
  • the present application also provides a computer-readable storage medium, the computer-readable storage medium stores a training program for a recognition model, and the training program for the recognition model is executed by a processor to achieve the above The steps of the training method of the recognition model described.
  • This application obtains the first recognition model based on the migration learning algorithm on the client side, trains according to the local training data, and sends the model parameters of the first recognition model to the server, receives the federated learning result sent by the server, and updates it according to the federated learning result
  • the first recognition model obtains the corresponding second recognition model, and sends the second recognition model to the server, so that the server obtains the target recognition model according to the second recognition model sent by each client and the preset genetic algorithm .
  • the recognition model is obtained through training of each client in the migration learning scenario, which improves the accuracy of the recognition model for recognizing the relevant information of each user, and this application supports the combination of multiple client recognition models in the federated learning scenario, which is effective In the case of protecting the privacy of the training data of each client corresponding to the user, the accuracy of the recognition data of the recognition model is further improved; this application supports the integration and optimization of multiple client recognition models in the evolutionary learning scenario through the genetic algorithm, and makes full use of each The client corresponds to the data value behind the recognition model, thereby further improving the recognition accuracy of the obtained recognition model.
  • FIG. 1 is a schematic flowchart of a first embodiment of a training method for a recognition model according to the present application
  • FIG. 2 is a schematic flowchart of a second embodiment of a training method for a recognition model according to the present application
  • Fig. 3 is a functional schematic block diagram of the first embodiment of the training device for the recognition model of the present application
  • Fig. 4 is a functional schematic block diagram of the second embodiment of the training device for the recognition model of the present application.
  • Fig. 5 is a schematic structural diagram of a hardware operating environment involved in a solution of an embodiment of the present application.
  • FIG. 1 is a schematic flowchart of a first embodiment of a training method for a recognition model in this application.
  • the embodiment of the application provides an embodiment of the training method of the recognition model. It should be noted that although the logical sequence is shown in the flowchart, in some cases, the sequence shown here may be executed in a different order than here. Or the steps described.
  • the training method of the recognition model includes:
  • Step S10 Obtain local training data, and train according to the training data to obtain a first recognition model.
  • Each client obtains local training data, and inputs the training data into the basic model to obtain the first recognition model according to the local training data.
  • machine learning models such as decision trees, random forests, artificial neural networks, and Bayesian learning can be used as basic models.
  • the local training data of the client is pre-stored.
  • the type of local training data is determined by the recognition model that needs to be trained. For example, when the recognition model to be trained is a voice recognition model, the local training data is the voice data generated by the client corresponding to the user; when the recognition model to be trained is a face When recognizing the model, the local training data is the face image data generated by the client corresponding to the user.
  • the client can trigger the acquisition instruction through a pre-set timing task, and obtain local training data according to the acquisition instruction; the client can also obtain the local training data according to the acquisition instruction when it detects the acquisition instruction triggered by the client corresponding to the user. Training data.
  • step S10 includes:
  • Step g Obtain local training data, and train the first recognition model based on the training data based on the migration learning algorithm.
  • the client After the client obtains the local training data, the client obtains the first recognition model based on the migration learning algorithm and trains according to the obtained local training data.
  • Transfer learning is a method of machine learning, which refers to a pre-trained model being reused in another task.
  • the trained recognition model can be applied to the party with missing features and annotations, thereby greatly improving It expands the application range of joint learning and effectively improves the predictive ability of the training recognition model.
  • the training data obtained through the migration learning is input into the basic model to obtain the first recognition model.
  • the speech recognition model can be ASR (Speech Recognition Technology, speech recognition technology) corresponding model.
  • the client combines at least one other client's local training data.
  • the principles for obtaining the first recognition model for each client are the same, which will not be repeated here.
  • There is a migration learning component in each client and the migration learning algorithm is called through the migration learning component to obtain the first recognition model.
  • Step S20 Send the model parameters of the first recognition model to the server, so that the server performs federated learning according to the model parameters sent by each client, obtains the federated learning result, and returns the federated learning result.
  • the client After each client obtains the first recognition model, the client obtains the model parameters of its own first recognition model, and sends the obtained model parameters to the server. It should be noted that the model parameters of the first recognition model of each client may be the same or different.
  • the server receives the model parameters sent by each client, the server performs federated learning according to the model parameters sent by each client, obtains the federated learning result, and sends the federated learning result to each client.
  • the server side there is a federated learning component, and the federated learning component can perform federated learning on the model parameters sent by each client to obtain the federated learning result.
  • this embodiment explains federated learning.
  • Federated Learning is an emerging basic artificial intelligence technology. Its design goal is to ensure information security during the exchange of big data, protect terminal data and personal privacy data, and under the premise of legal compliance, in the presence of multiple parties or Carry out high-efficiency machine learning among multiple computing nodes.
  • this embodiment uses a scenario containing two data owners (ie, enterprise A and enterprise B) as an example to introduce the system architecture of federated learning.
  • the framework can be extended to scenarios that include multiple data owners.
  • enterprise A and enterprise B want to jointly train a machine learning model, and their business systems each have relevant data of their respective users.
  • company B also has label data that the model needs to predict.
  • enterprise A and enterprise B cannot directly exchange data, and can use a federated learning system to build models.
  • the use of federated learning to build a model includes two parts.
  • the first part is: the alignment of encrypted samples. Since the user groups of the two companies are not completely overlapped, the system uses encryption-based user sample alignment technology to confirm the mutual users of both companies under the premise that company A and company B do not disclose their respective data, and does not expose users that do not overlap with each other, so that Combine the characteristics of these users for modeling.
  • the second part is: encryption model training. After determining the common user group, the data can be used to train machine learning models. In order to ensure the confidentiality of the data during the training process, a third-party collaborator C needs to be used for encryption training. Taking the linear regression model as an example, the training process can be divided into the following 4 steps:
  • Step 1 Collaborator C distributes the public key to enterprise A and enterprise B to encrypt the data that needs to be exchanged during the training process.
  • Step 2 Encrypted interaction between enterprise A and enterprise B is used to calculate the intermediate result of the gradient.
  • Step 3 Enterprise A and Enterprise B respectively calculate based on the encrypted gradient value, and at the same time, Enterprise B calculates the loss based on its label data, and summarizes the results to collaborator C.
  • Collaborator C calculates the total gradient value by summing up the results and decrypts it.
  • Step 4 Collaborator C sends the decrypted gradient value back to enterprise A and enterprise B respectively, and enterprise A and enterprise B update the parameters of their respective models according to the gradient value.
  • federated learning is divided into horizontal federated learning (horizontal federated learning), vertical federated learning and federated transfer learning (FmL).
  • Horizontal federated learning is to split the data set according to the horizontal (ie user dimension) when the user characteristics of the two data sets overlap more and the user overlaps less, and extract the user characteristics of the two parties that are the same but the users are not exactly the same. That part of the data for training. For example, there are two banks in different regions, and their user groups come from their respective regions, and their mutual intersections are small. However, their businesses are very similar, so the recorded user characteristics are the same. At this point, horizontal federated learning can be used to build a joint model.
  • Longitudinal federated learning is to split the data set according to the longitudinal direction (ie feature dimension) when the users of the two data sets overlap more and the user characteristics are less overlapped, and extract the ones where the two users are the same but the user characteristics are not exactly the same.
  • Part of the data for training For example, there are two different institutions, one is a bank in a certain place, and the other is an e-commerce company in the same place. Their user groups are likely to include most of the residents in that place, so the intersection of users is relatively large. However, because banks record the user's income and expenditure behavior and credit rating, while e-commerce stores the user's browsing and purchase history, their user characteristics have a small intersection.
  • Vertical federated learning is to aggregate these different features in an encrypted state to enhance the model's capabilities. At present, many machine learning models such as logistic regression model, tree structure model and neural network model have gradually been confirmed to be able to build on this federal system.
  • Federated transfer learning is to use transfer learning to overcome the lack of data or labels when the user and user characteristics of the two data sets are less overlapped, without segmenting the data.
  • there are two different institutions one is a bank in China, and the other is an e-commerce company in the United States. Due to geographical restrictions, the user groups of these two institutions have a very small intersection. At the same time, due to the different types of institutions, only a small part of the data characteristics of the two overlap. In this case, if you want to carry out effective federated learning, you must introduce transfer learning to solve the problem of small unilateral data and fewer label samples, so as to improve the effect of the model.
  • training data of each client user is required, and these training data are stored locally in the client held by each user as the local training data of the corresponding user of the client.
  • part of the training data belongs to the user's private data.
  • the user's private data cannot be obtained, that is, the user's private data cannot be used as training data to train the recognition model.
  • some applications in the client usually compulsorily insert clauses requiring the user to agree to the use of private data in advance in the user agreement, or obtain the user's private data in other unknowing ways, which significantly reduces the privacy of the user's private data.
  • the federated learning result can be obtained without touching the user's original local training data, and each client uses the federated learning result to update the first recognition model.
  • Step S30 Receive the federated learning result sent by the server, update the first recognition model according to the federated learning result to obtain a corresponding second recognition model, and send the second recognition model to the server , So that the server can obtain the target recognition model according to the second recognition model sent by each client and the preset genetic algorithm.
  • Each client receives the federated learning result sent by the server, and updates the first recognition model according to the federated learning result to obtain the corresponding second recognition model, that is, the updated first recognition model is the second recognition model.
  • each client sends the second recognition model to the server.
  • the server receives the second recognition model sent by each client, the server obtains the target recognition model according to the second recognition model sent by each client and a preset genetic algorithm.
  • Genetic Algorithm is a method of solving optimization problems through search. It first generates a certain amount of population randomly. Genetic Algorithm can include reproduce operators, crossover operators, mutation operators, etc. son. The crossover operator is the process of encoding the chromosomes of the members of the population and the chromosome codes of the two or two group members are crossed; the mutation operator is the process of mutating the chromosome codes after the crossover with a certain probability.
  • the server there is a genetic algorithm component, through which the target recognition model can be obtained according to the second recognition model sent by each client.
  • the step of receiving the federated learning result sent by the server, updating the first recognition model according to the federated learning result, and obtaining the corresponding second recognition model includes:
  • Step a Receive a federated learning result sent by the server, and obtain a preset learning rate and a model parameter change amount in the federated learning result.
  • step b the updated model parameters are calculated according to the preset learning rate and the change amount of the model parameters, and the first recognition model is updated according to the updated model parameters to obtain the corresponding second recognition model.
  • each client receives the federated learning result sent by the server, and each client obtains a preset learning rate and the amount of model parameter changes in the federated learning result.
  • the preset learning rate can be set according to specific needs.
  • the size of the preset learning rate is not specifically limited, and the preset learning rates corresponding to different clients may be the same or different.
  • each client obtains the model parameters of the first recognition model, calculates the product of the preset learning rate and the amount of model parameter change, and subtracts the calculated product from the model parameters of the first recognition model to obtain the updated Model parameters, and update the first recognition model according to the updated model parameters, that is, use the updated model parameters as the model parameters of the recognition model to obtain the second recognition model.
  • the client is trained based on the migration learning algorithm to obtain the first recognition model according to the local training data, and the model parameters of the first recognition model are sent to the server, and the federated learning result sent by the server is received, and according to the federated learning result Update the first recognition model to obtain the corresponding second recognition model, and send the second recognition model to the server, so that the server can obtain target recognition according to the second recognition model sent by each client and the preset genetic algorithm Model.
  • the recognition model obtained by training each client in the migration learning scenario improves the accuracy of the recognition model for recognizing the relevant information of each user, and this embodiment supports the combination of multiple client recognition models in the federated learning scenario.
  • this embodiment supports the integration and optimization of multiple client recognition models in the evolutionary learning scenario through the genetic algorithm, making full use of Each client corresponds to the data value behind the recognition model, thereby further improving the recognition accuracy of the obtained recognition model.
  • the recognition model training method further includes:
  • Step S40 Receive model parameters corresponding to the first recognition model sent by each client, and perform federated learning according to the model parameters to obtain a federated learning result.
  • the server receives the model parameters corresponding to the first recognition model sent by each client, and performs federated learning according to the model parameters to obtain the federated learning result.
  • the federated learning result may be a model parameter change amount.
  • the model parameter change amount may be a gradient value calculated according to the loss function when the corresponding loss function converges during the federated learning process.
  • Step S50 Send the federated learning result to each client, so that each client can update the corresponding first recognition model according to the federated learning result to obtain the corresponding second recognition model.
  • the server After the server obtains the federated learning result, the server sends the federated learning result to each client, so that each client can update the corresponding first recognition model according to the federated learning result to obtain the corresponding second recognition model.
  • Each client updates the first recognition model according to the federated learning result, and the process of obtaining the second recognition model has been described in detail in the foregoing embodiment, and will not be repeated here.
  • Step S60 Receive the second recognition model sent by each client, and obtain the target recognition model according to each second recognition model and a preset genetic algorithm.
  • the server receives the second recognition model sent by each client, and obtains the target recognition model according to each second recognition model and a preset genetic algorithm. It is understandable that the target recognition model is the optimal recognition model required. After the server obtains the target recognition model, the server can send the target recognition model to each client, so that each client can perform a corresponding recognition operation according to the target recognition model after receiving the recognition request. After each client receives the target recognition model, each client stores the received target recognition model.
  • step S50 includes:
  • Step d receiving the second recognition model sent by each client, selecting the parent recognition model from each of the received second recognition models, and selecting the target operator in the preset genetic algorithm.
  • Step e Obtain a child recognition model corresponding to the parent recognition model through the parent recognition model and the target operator.
  • step f if it is detected that the offspring recognition model meets the end condition, the offspring recognition model with the highest recognition accuracy rate among the end conditions is determined as the target recognition model.
  • the server receives the second recognition model sent by each client, randomly selects the second recognition model from each of the received second recognition models to determine it as the parent recognition model, and randomly selects the calculation model in the preset genetic algorithm.
  • the sub is determined as the target operator.
  • the target operator corresponding to each parent recognition model may be the same or different.
  • the server obtains the child corresponding to the parent recognition model through the parent recognition model and the target operator corresponding to the parent recognition model Recognize the model, and check whether the resulting offspring recognition model meets the end condition.
  • the offspring recognition model with the highest recognition accuracy rate among the end conditions is obtained, and the offspring recognition model with the highest accuracy rate is determined as the target recognition model; if the obtained offspring recognition model is detected If the offspring recognition model does not meet the ending conditions, the server will use the obtained offspring recognition model as the parent recognition model for the next iteration, and determine the target operator corresponding to the next parent recognition model, and continue according to the parent recognition model and The target operator obtains the offspring recognition model corresponding to the parent recognition model, and when it is detected that the offspring recognition model satisfies the end condition, the offspring recognition model with the highest recognition accuracy rate among the satisfied conditions is determined as the target recognition model.
  • step e includes:
  • Step e1 Obtain the to-be-processed model parameters corresponding to the parent recognition model, and determine the to-be-processed model parameters as parameter codes.
  • Step e2 processing the parameter encoding through the target operator to obtain a new parameter encoding.
  • Step e3 Correspondingly update the parent recognition model according to the new parameter code to obtain the offspring recognition model.
  • the server obtains the model parameters in each second recognition model, that is, obtains the model parameters of the second recognition model as the parent recognition model, determines the obtained model parameters as the model parameters to be processed, and sets the parameters to be processed.
  • the parameter coding update corresponds to the parent recognition model, and the offspring recognition model is obtained. It is understandable that each client can also send the model parameters in the second recognition model to the server.
  • the specific processing method used is determined by the corresponding target operator. For example, for a mutation operator, the offspring recognition model can only inherit a few specific model parameters of the parent recognition model.
  • the genetic algorithm extracts the specific parameter code of the parent recognition model, and uses the extracted specific parameter code as the new parameter code .
  • the parameter codes corresponding to the two parent recognition models can be integrated to obtain a new parameter code.
  • the specific algorithm of integration is determined by the operator corresponding to the genetic algorithm. For example, the method of averaging the parameters can be used Coding for integration.
  • the server selects the parent recognition model in each second recognition model
  • the number of selected parent recognition models can be determined according to the current available resources in the server.
  • all second recognition models can be determined as parent recognition models. It is understandable that as the available resources in the server increase, the number of selected parent identification models will increase accordingly.
  • the progeny recognition model when detecting whether the progeny recognition model meets the end condition, it can be determined that the progeny recognition model meets the end condition when all the progeny recognition models meet the end condition, and if one of the progeny recognition models does not meet the end condition, It is determined that the offspring recognition model does not meet the end condition. It is also possible to determine that the child recognition model satisfies the end condition when the child recognition model with a preset ratio in the child recognition model is detected to meet the end condition; otherwise, it is determined that the child recognition model does not meet the end condition. At this time, the size of the preset ratio can be set according to specific needs, such as 60%, 75%, or 80%.
  • step f includes:
  • Step f1 If it is detected that the number of iterations corresponding to the offspring identification model is greater than the preset number, it is determined that the offspring identification model meets the termination condition, and the pre-stored test data is used to obtain the identification of each offspring identification model that meets the termination condition Accuracy.
  • step f2 the offspring recognition model with the highest recognition accuracy is selected and determined as the target recognition model.
  • the server detects that the number of iterations corresponding to the offspring identification model is greater than the preset number, it is determined that the offspring identification model meets the end condition.
  • the preset number of times can be set according to specific needs, and this embodiment does not limit the size of the preset number of times. It is understandable that when the server obtains the offspring recognition model through the genetic algorithm for the first time, the corresponding iteration number is 1; when the server obtains the offspring identification model through the genetic algorithm for the second time, the corresponding iteration number is 2. , That is, the number of iterations is equal to the number of offspring identification models obtained through genetic algorithm.
  • the server obtains the pre-stored test data, and inputs the test data into each offspring recognition model to determine the recognition accuracy of each offspring recognition model, and compares each child
  • the recognition accuracy rate of the generation recognition model is selected, and the offspring recognition model with the highest recognition accuracy rate is selected as the target recognition model. Further, if it is detected that the number of iterations corresponding to the offspring identification model is less than or equal to the preset number, the server determines that the offspring identification model does not meet the end condition.
  • step f also includes:
  • Step f3 If it is detected that the recognition accuracy rate corresponding to the offspring recognition model is greater than or equal to the preset accuracy rate, it is determined that the offspring recognition model meets the end condition.
  • Step f4 selecting the child recognition model with the highest recognition accuracy rate among the child recognition models and determining it as the target recognition model.
  • the server every time the server obtains the offspring recognition model, the server obtains the pre-stored test data, inputs the test data into the offspring recognition model, obtains the recognition accuracy rate corresponding to the offspring recognition model, and detects each child Whether the recognition accuracy rate corresponding to the generation recognition model is greater than or equal to the preset accuracy rate.
  • the size of the preset accuracy rate can be set according to specific needs, and this embodiment does not specifically limit the size of the preset accuracy rate.
  • the server determines that the child recognition model meets the end condition, and selects the child recognition with the highest recognition accuracy in the child recognition model
  • the model is determined to be the target recognition model; when the server detects that the recognition accuracy rate corresponding to the offspring recognition model is less than the preset accuracy rate, the server determines that the offspring recognition model does not meet the end condition.
  • the server can also determine that the child recognition model meets the end condition when the proportion of the child recognition model whose detection and recognition accuracy rate is greater than or equal to the preset accuracy rate to all the child recognition models in the current iteration process is greater than a specific ratio, Otherwise, it is determined that the offspring recognition model does not meet the end condition.
  • this embodiment does not limit the size of the specific ratio.
  • the server can also sort the child recognition models from high to low according to the recognition accuracy when determining that the child recognition model meets the end condition, and obtain the sorted child recognition model, and then the sorted child recognition model Obtain a preset number of offspring recognition models from front to back in the recognition model to determine the target offspring recognition model, obtain the model parameters of the target offspring recognition model, and linearly add the model parameters of each target offspring recognition model to obtain the target Model parameters, the recognition model corresponding to the target model parameters is determined as the target model.
  • the size of the preset number can be set according to specific needs, for example, the preset number can be set to 2, 3, or 5.
  • the specific process of linear addition can be to calculate the average value of the model parameters corresponding to each target offspring recognition model, and determine the average value as the target model parameter; the process of linear addition can also be to determine the corresponding value of each target offspring recognition model Weight, multiply the model parameters of each target offspring recognition model by the corresponding weight to get the product, and then add the products corresponding to the same model parameter in each target offspring recognition model to get the corresponding target model parameters, understandable Yes, the higher the recognition accuracy, the greater the weight of the model parameters corresponding to the target offspring recognition model.
  • the server receives the model parameters corresponding to the first recognition model sent by each client, performs federated learning according to the model parameters, sends the result of federated learning to each client, and receives the second recognition sent by each client Model, obtain the target recognition model according to each second recognition model and the preset genetic algorithm.
  • the recognition model obtained by training each client in the migration learning scenario improves the accuracy of the recognition model for recognizing the relevant information of each user, and this embodiment supports the combination of multiple client recognition models in the federated learning scenario.
  • this embodiment supports the integration and optimization of multiple client recognition models in the evolutionary learning scenario through the genetic algorithm, making full use of Each client corresponds to the data value behind the recognition model, thereby further improving the recognition accuracy of the obtained recognition model.
  • the present application also provides a training device for a recognition model.
  • the training device for the recognition model is applied to the client, and the training device for the recognition model includes:
  • the obtaining module 10 is used to obtain local training data
  • the training module 20 is configured to train according to the training data to obtain a first recognition model
  • the sending module 30 is configured to send the model parameters of the first recognition model to the server, so that the server performs federated learning according to the model parameters sent by each client, obtains the federated learning result, and returns the federated learning result;
  • the receiving module 40 is configured to receive the federated learning result sent by the server;
  • the update module 50 is configured to update the first recognition model according to the federated learning result to obtain a corresponding second recognition model
  • the sending module 30 is further configured to send the second recognition model to the server, so that the server obtains the target recognition model according to the second recognition model sent by each client and a preset genetic algorithm.
  • the training module 20 is also used to obtain local training data, and based on a migration learning algorithm, train according to the training data to obtain a first recognition model.
  • the update module 50 includes:
  • An acquiring unit for acquiring a preset learning rate and the model parameter change amount in the federated learning result
  • a calculation unit configured to calculate the updated model parameters according to the preset learning rate and the change amount of the model parameters
  • the update unit updates the first recognition model according to the updated model parameters to obtain the corresponding second recognition model.
  • the specific implementation of the training device for the recognition model of this application is basically the same as the steps in the first embodiment of the training method for the recognition model described above, and will not be repeated here.
  • the present application also provides a training device for a recognition model.
  • the training device for the recognition model is applied to the server, and the training device for the recognition model includes:
  • the receiving module 60 is configured to receive model parameters corresponding to the first recognition model sent by each client;
  • the federated learning module 70 is configured to perform federated learning according to the model parameters to obtain federated learning results;
  • the sending module 80 is configured to send the federated learning result to each client, so that each client can update the corresponding first recognition model according to the federated learning result to obtain the corresponding second recognition model;
  • the receiving module 60 is also configured to receive the second recognition model sent by each client, and obtain the target recognition model according to each second recognition model and a preset genetic algorithm.
  • the receiving module 60 includes:
  • the receiving unit is configured to receive the second recognition model sent by each client;
  • the selection unit is used to select the parent recognition model among the received second recognition models, and select the target operator in the preset genetic algorithm
  • the determining unit is configured to obtain the child recognition model corresponding to the parent recognition model through the parent recognition model and the target operator; if it is detected that the child recognition model satisfies the end condition, the end condition will be met.
  • the child recognition model with the highest recognition accuracy rate is determined as the target recognition model.
  • the determining unit includes:
  • the first obtaining subunit is configured to obtain the to-be-processed model parameters corresponding to the parent recognition model
  • the first determining subunit is used to determine the parameters of the model to be processed as parameter codes
  • a processing subunit configured to process the parameter encoding through the target operator to obtain a new parameter encoding
  • the update subunit is used to update the parent recognition model corresponding to the new parameter encoding to obtain the child recognition model.
  • the determining unit includes:
  • the second determining subunit is configured to determine that the child recognition model meets the ending condition if it is detected that the number of iterations corresponding to the offspring recognition model is greater than the preset number;
  • the second acquisition subunit is used to acquire the recognition accuracy rate of each offspring recognition model that meets the end condition by using the pre-stored test data;
  • the second determining subunit is also used to select the offspring recognition model with the highest recognition accuracy rate and determine it as the target recognition model.
  • the determining unit is further configured to determine that the child recognition model satisfies the end condition if the recognition accuracy corresponding to the child recognition model is detected to be greater than or equal to a preset accuracy rate; In the model, the child recognition model with the highest recognition accuracy rate is selected as the target recognition model.
  • the specific implementation of the training device for the recognition model of the present application is basically the same as the steps in the second embodiment of the training method for the recognition model described above, and will not be repeated here.
  • FIG. 5 is a schematic structural diagram of the hardware operating environment involved in the solution of the embodiment of the present application.
  • FIG. 5 can be a schematic structural diagram of the hardware operating environment of the training device for the recognition model.
  • the training device for the recognition model may be a client or a server.
  • the training device for the recognition model in the embodiment of the present application may be a terminal device such as a PC and a portable computer.
  • the training device of the recognition model may include: a processor 1001, such as a CPU, a memory 1005, a user interface 1003, a network interface 1004, and a communication bus 1002.
  • the communication bus 1002 is used to implement connection and communication between these components.
  • the user interface 1003 may include a display screen (Display) and an input unit such as a keyboard (Keyboard), and the optional user interface 1003 may also include a standard wired interface and a wireless interface.
  • the network interface 1004 may optionally include a standard wired interface and a wireless interface (such as a WI-FI interface).
  • the memory 1005 may be a high-speed RAM memory, or a stable memory (non-volatile memory), such as a magnetic disk memory.
  • the memory 1005 may also be a storage device independent of the aforementioned processor 1001.
  • the structure of the training device for the recognition model shown in FIG. 5 does not constitute a limitation on the training device for the recognition model, and may include more or less components than shown in the figure, or a combination of certain components, Or different component arrangements.
  • the memory 1005 as a computer storage medium may include an operating system, a network communication module, a user interface module, and a training program for a recognition model.
  • the operating system is a program that manages and controls the hardware and software resources of the training equipment of the recognition model, and supports the operation of the training program of the recognition model and other software or programs.
  • the user interface 1003 is mainly used to connect to the server and communicate with the server; when the training device of the recognition model is the server At the time, the user interface 1003 is mainly used to connect various clients and communicate with each client; the network interface 1004 is mainly used to communicate with the back-end server and perform data communication with the back-end server; the processor 1001 can be used to call the storage in the memory 1005
  • the training program of the recognition model, and the steps of the training method of the recognition model as described above are executed.
  • the specific implementation of the training device for the recognition model of the present application is basically the same as the foregoing embodiments of the training method for the recognition model, and will not be repeated here.
  • an embodiment of the present application also proposes a computer-readable storage medium, the computer-readable storage medium stores a training program for a recognition model, and the training program for the recognition model is executed by a processor to realize the recognition as described above. The steps of the model training method.
  • the technical solution of this application essentially or the part that contributes to the existing technology can be embodied in the form of a software product, and the computer software product is stored in a storage medium (such as ROM/RAM, magnetic disk, The optical disc) includes several instructions to enable a terminal device (which may be a mobile phone, a computer, a server, an air conditioner, or a network device, etc.) to execute the method described in each embodiment of the present application.
  • a terminal device which may be a mobile phone, a computer, a server, an air conditioner, or a network device, etc.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • Biophysics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Software Systems (AREA)
  • Artificial Intelligence (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Business, Economics & Management (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Evolutionary Biology (AREA)
  • General Health & Medical Sciences (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Physiology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Human Computer Interaction (AREA)
  • Medical Informatics (AREA)
  • Multimedia (AREA)
  • Genetics & Genomics (AREA)
  • Biomedical Technology (AREA)
  • Computational Linguistics (AREA)
  • Molecular Biology (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Accounting & Taxation (AREA)
  • Development Economics (AREA)
  • Economics (AREA)
  • Finance (AREA)
  • Marketing (AREA)
  • Strategic Management (AREA)
  • Technology Law (AREA)
  • General Business, Economics & Management (AREA)
  • Image Analysis (AREA)

Abstract

A method and apparatus for training a recognition model, a device, and a readable storage medium, which relate to the field of financial technology. The method comprises the steps of: acquiring local training data, and training according to the training data to obtain a first recognition model (S10); sending model parameters of the first recognition model to a server for the server to perform federated learning according to model parameters sent by each client so as to obtain a federated learning result (S20); and receiving the federated learning result sent by the server, updating the first recognition model according to the federated learning result to obtain a corresponding second recognition model, and sending the second recognition model to the server, so that the server obtains a target recognition model according to the second recognition model sent by each client and a preset genetic algorithm (S30). In the method, migration learning, federated learning and a genetic algorithm are combined for training and obtaining a recognition model, which ensures that training data is not leaked during the training of the recognition model, and improves the recognition accuracy of the obtained recognition model.

Description

识别模型的训练方法、装置、设备及可读存储介质Training method, device, equipment and readable storage medium of recognition model
本申请要求2020年3月20日申请的,申请号为202010206241.7,名称为“识别模型的训练方法、装置、设备及可读存储介质”的中国专利申请的优先权,在此将其全文引入作为参考。This application claims the priority of the Chinese patent application filed on March 20, 2020, with the application number 202010206241.7, titled "Recognition Model Training Method, Device, Equipment, and Readable Storage Medium", which is hereby incorporated in its entirety as refer to.
技术领域Technical field
本申请涉及金融科技(Fintech)的数据处理技术领域,尤其涉及一种识别模型的训练方法、装置、设备及可读存储介质。This application relates to the field of data processing technology of financial technology (Fintech), and in particular to a training method, device, equipment and readable storage medium of a recognition model.
背景技术Background technique
随着计算机技术的发展,越来越多的技术应用在金融领域,传统金融业正在逐步向金融科技(Fintech)转变,数据处理技术也不例外,但由于金融行业的安全性、实时性要求,也对数据处理技术提出的更高的要求。With the development of computer technology, more and more technologies are applied in the financial field. The traditional financial industry is gradually changing to Fintech. Data processing technology is no exception. However, due to the security and real-time requirements of the financial industry, It also places higher requirements on data processing technology.
机器学习技术在语音识别的使用非常广泛,这些技术往往极度依赖于海量的语音数据进行训练,容易泄露语音数据,从而不能有效保护语音数据对应用户的隐私。而且目前语音识别系统只能提供非常有限的定制化能力,从而导致所得的语音识别模型进行语音识别的准确率低下。由此可知,目前训练所得的识别模型识别准确率低下,且训练识别模型过程中容易泄露训练数据。Machine learning technology is widely used in speech recognition. These technologies often rely heavily on massive speech data for training, which can easily leak the speech data, and thus cannot effectively protect the privacy of the users corresponding to the speech data. Moreover, the current speech recognition system can only provide very limited customization capabilities, which leads to low accuracy of speech recognition by the resulting speech recognition model. From this, it can be seen that the recognition accuracy of the currently trained recognition model is low, and the training data is likely to be leaked in the process of training the recognition model.
技术解决方案Technical solutions
本申请的主要目的在于提供一种识别模型的训练方法、装置、设备及可读存储介质,旨在解决现有的识别模型识别准确率低下,且训练识别模型过程中容易泄露训练数据的技术问题。The main purpose of this application is to provide a training method, device, equipment and readable storage medium for a recognition model, aiming to solve the technical problem of low recognition accuracy of the existing recognition model and easy leakage of training data during the training of the recognition model .
为实现上述目的,本申请提供一种识别模型的训练方法,应用于客户端,所述识别模型的训练方法包括步骤:In order to achieve the above objective, the present application provides a method for training a recognition model, which is applied to a client, and the method for training a recognition model includes the steps:
获取本地的训练数据,根据所述训练数据训练得到第一识别模型;Acquiring local training data, and training according to the training data to obtain a first recognition model;
将所述第一识别模型的模型参数发送给服务端,以供所述服务端根据各客户端发送的模型参数进行联邦学习,得到联邦学习结果,并返回所述联邦学习结果;Sending the model parameters of the first recognition model to the server, so that the server performs federated learning according to the model parameters sent by each client to obtain a federated learning result, and return the federated learning result;
接收所述服务端发送的联邦学习结果,根据所述联邦学习结果更新所述第一识别模型,得到对应的第二识别模型,并将所述第二识别模型发送给所述服务端,以供所述服务端根据各客户端发送的第二识别模型和预设的遗传算法得到目标识别模型。Receive the federated learning result sent by the server, update the first recognition model according to the federated learning result to obtain the corresponding second recognition model, and send the second recognition model to the server for The server obtains the target recognition model according to the second recognition model sent by each client and the preset genetic algorithm.
此外,为实现上述目的,本申请还提供一种识别模型的训练装置,所述识别模型的训练装置应用于客户端中,所述识别模型的训练装置包括:In addition, in order to achieve the above object, the present application also provides a training device for a recognition model. The training device for the recognition model is applied to a client, and the training device for the recognition model includes:
获取模块,用于获取本地的训练数据;The acquisition module is used to acquire local training data;
训练模块,用于根据所述训练数据训练得到第一识别模型;The training module is used to train according to the training data to obtain the first recognition model;
发送模块,用于将所述第一识别模型的模型参数发送给服务端,以供所述服务端根据各客户端发送的模型参数进行联邦学习,得到联邦学习结果,并返回所述联邦学习结果;The sending module is configured to send the model parameters of the first recognition model to the server, so that the server performs federated learning according to the model parameters sent by each client, obtains the federated learning result, and returns the federated learning result ;
接收模块,用于接收所述服务端发送的联邦学习结果;The receiving module is used to receive the federated learning result sent by the server;
更新模块,用于根据所述联邦学习结果更新所述第一识别模型,得到对应的第二识别模型;An update module, configured to update the first recognition model according to the federated learning result to obtain a corresponding second recognition model;
所述发送模块还用于将所述第二识别模型发送给所述服务端,以供所述服务端根据各客户端发送的第二识别模型和预设的遗传算法得到目标识别模型。The sending module is further configured to send the second recognition model to the server, so that the server obtains the target recognition model according to the second recognition model sent by each client and a preset genetic algorithm.
此外,为实现上述目的,本申请还提供一种识别模型的训练装置,所述识别模型的训练装置应用于服务端,所述识别模型的训练装置包括:In addition, in order to achieve the above object, the present application also provides a recognition model training device, the recognition model training device is applied to the server, and the recognition model training device includes:
接收模块,用于接收各客户端发送的第一识别模型对应的模型参数;The receiving module is used to receive the model parameters corresponding to the first recognition model sent by each client;
联邦学习模块,用于根据所述模型参数进行联邦学习,得到联邦学习结果;The federated learning module is used to perform federated learning according to the model parameters to obtain federated learning results;
发送模块,用于将所述联邦学习结果发送给各客户端,以供各客户端根据所述联邦学习结果更新对应的第一识别模型,得到对应的第二识别模型;A sending module, configured to send the federated learning result to each client, so that each client can update the corresponding first recognition model according to the federated learning result to obtain the corresponding second recognition model;
所述接收模块还用于接收各客户端发送的第二识别模型,根据各第二识别模型和预设的遗传算法得到目标识别模型。The receiving module is also used to receive the second recognition model sent by each client, and obtain the target recognition model according to each second recognition model and a preset genetic algorithm.
此外,为实现上述目的,本申请还提供一种识别模型的训练设备,所述识别模型的训练设备包括存储器、处理器和存储在所述存储器上并可在所述处理器上运行的识别模型的训练程序,所述识别模型的训练程序被所述处理器执行时实现如联邦学习服务端对应的识别模型的训练方法的步骤。In addition, in order to achieve the above object, the present application also provides a training device for a recognition model. The training device for the recognition model includes a memory, a processor, and a recognition model stored in the memory and running on the processor. When the training program of the recognition model is executed by the processor, the steps of the training method of the recognition model corresponding to the federated learning server are implemented.
此外,为实现上述目的,本申请还提供一种计算机可读存储介质,所述计算机可读存储介质上存储有识别模型的训练程序,所述识别模型的训练程序被处理器执行时实现如上所述的识别模型的训练方法的步骤。In addition, in order to achieve the above object, the present application also provides a computer-readable storage medium, the computer-readable storage medium stores a training program for a recognition model, and the training program for the recognition model is executed by a processor to achieve the above The steps of the training method of the recognition model described.
本申请通过客户端基于迁移学习算法,根据本地的训练数据训练得到第一识别模型,并将第一识别模型的模型参数发送给服务端,接收服务端发送的联邦学习结果,根据联邦学习结果更新第一识别模型,得到对应的第二识别模型,并将第二识别模型发送给所述服务端,以供服务端根据各客户端发送的第二识别模型和预设的遗传算法得到目标识别模型。通过各个客户端在迁移学习的场景下训练得到识别模型,提升了识别模型对各个用户相关信息识别的准确率,且本申请支持在联邦学习场景下进行多个客户端识别模型的结合,在有效保护各个客户端对应用户训练数据隐私的情况下,进一步提高识别模型识别数据的准确率;本申请支持通过遗传算法,在进化学习场景下进行多个客户端识别模型的整合和优化,充分利用各个客户端对应识别模型背后的数据价值,从而进一步地提高了所得识别模型的识别准确率。This application obtains the first recognition model based on the migration learning algorithm on the client side, trains according to the local training data, and sends the model parameters of the first recognition model to the server, receives the federated learning result sent by the server, and updates it according to the federated learning result The first recognition model obtains the corresponding second recognition model, and sends the second recognition model to the server, so that the server obtains the target recognition model according to the second recognition model sent by each client and the preset genetic algorithm . The recognition model is obtained through training of each client in the migration learning scenario, which improves the accuracy of the recognition model for recognizing the relevant information of each user, and this application supports the combination of multiple client recognition models in the federated learning scenario, which is effective In the case of protecting the privacy of the training data of each client corresponding to the user, the accuracy of the recognition data of the recognition model is further improved; this application supports the integration and optimization of multiple client recognition models in the evolutionary learning scenario through the genetic algorithm, and makes full use of each The client corresponds to the data value behind the recognition model, thereby further improving the recognition accuracy of the obtained recognition model.
附图说明Description of the drawings
图1是本申请识别模型的训练方法第一实施例的流程示意图;FIG. 1 is a schematic flowchart of a first embodiment of a training method for a recognition model according to the present application;
图2是本申请识别模型的训练方法第二实施例的流程示意图;2 is a schematic flowchart of a second embodiment of a training method for a recognition model according to the present application;
图3是本申请识别模型的训练装置第一实施例的功能示意图模块图;Fig. 3 is a functional schematic block diagram of the first embodiment of the training device for the recognition model of the present application;
图4是本申请识别模型的训练装置第二实施例的功能示意图模块图;Fig. 4 is a functional schematic block diagram of the second embodiment of the training device for the recognition model of the present application;
图5是本申请实施例方案涉及的硬件运行环境的结构示意图。Fig. 5 is a schematic structural diagram of a hardware operating environment involved in a solution of an embodiment of the present application.
本申请目的的实现、功能特点及优点将结合实施例,参照附图做进一步说明。The realization, functional characteristics, and advantages of the purpose of this application will be further described in conjunction with the embodiments and with reference to the accompanying drawings.
本发明的实施方式Embodiments of the present invention
应当理解,此处所描述的具体实施例仅仅用以解释本申请,并不用于限定本申请。It should be understood that the specific embodiments described here are only used to explain the present application, and are not used to limit the present application.
本申请提供一种识别模型的训练方法,参照图1,图1为本申请识别模型的训练方法第一实施例的流程示意图。The present application provides a method for training a recognition model. Refer to FIG. 1, which is a schematic flowchart of a first embodiment of a training method for a recognition model in this application.
本申请实施例提供了识别模型的训练方法的实施例,需要说明的是,虽然在流程图中示出了逻辑顺序,但是在某些情况下,可以以不同于此处的顺序执行所示出或描述的步骤。The embodiment of the application provides an embodiment of the training method of the recognition model. It should be noted that although the logical sequence is shown in the flowchart, in some cases, the sequence shown here may be executed in a different order than here. Or the steps described.
本实施例中,至少存在两个客户端,识别模型的训练方法包括:In this embodiment, there are at least two clients, and the training method of the recognition model includes:
步骤S10,获取本地的训练数据,根据所述训练数据训练得到第一识别模型。Step S10: Obtain local training data, and train according to the training data to obtain a first recognition model.
各客户端获取本地的训练数据,将训练数据输入至基础模型中,以根据本地训练数据得到第一识别模型。本实施例可采用决策树、随机森林、人工神经网络和贝叶斯学习等机器学习模型作为基础模型。客户端本地的训练数据是预先存储好的。本地的训练数据的类型由需要训练的识别模型决定,如当要训练的识别模型为语音识别模型时,本地的训练数据为客户端对应用户所生成的语音数据;当要训练的识别模型为人脸识别模型时,本地的训练数据为客户端对应用户所生成的人脸图像数据。具体地,客户端可通过预先设置好的定时任务触发获取指令,根据获取指令获取本地的训练数据;客户端也可在侦测到客户端对应用户触发的获取指令时,根据获取指令获取本地的训练数据。Each client obtains local training data, and inputs the training data into the basic model to obtain the first recognition model according to the local training data. In this embodiment, machine learning models such as decision trees, random forests, artificial neural networks, and Bayesian learning can be used as basic models. The local training data of the client is pre-stored. The type of local training data is determined by the recognition model that needs to be trained. For example, when the recognition model to be trained is a voice recognition model, the local training data is the voice data generated by the client corresponding to the user; when the recognition model to be trained is a face When recognizing the model, the local training data is the face image data generated by the client corresponding to the user. Specifically, the client can trigger the acquisition instruction through a pre-set timing task, and obtain local training data according to the acquisition instruction; the client can also obtain the local training data according to the acquisition instruction when it detects the acquisition instruction triggered by the client corresponding to the user. Training data.
进一步地,步骤S10包括:Further, step S10 includes:
步骤g,获取本地的训练数据,基于迁移学习算法,根据所述训练数据训练得到第一识别模型。Step g: Obtain local training data, and train the first recognition model based on the training data based on the migration learning algorithm.
当客户端获取到本地的训练数据后,客户端基于迁移学习算法,根据所获取的本地的训练数据训练得到第一识别模型。迁移学习是一种机器学习的方法,指的是一个预训练的模型被重新用在另一个任务中。通过迁移学习,结合其他客户端的训练数据进行联合建模,在保证两方的训练数据隐私不被泄露的前提下,可以将训练的识别模型应用在特征和标注缺失的一方,从而极大程度的拓展了联合学习的应用范围,同时有效提高训练所得识别模型的预测能力。需要说明的是,在通过迁移学习算法得到第一识别模型过程中,将通过迁移学习得到的训练数据输入基础模型中,得到第一识别模型。在本实施例中,语音识别模型可为ASR(Speech Recognition Technology,语音识别技术)对应的模型。After the client obtains the local training data, the client obtains the first recognition model based on the migration learning algorithm and trains according to the obtained local training data. Transfer learning is a method of machine learning, which refers to a pre-trained model being reused in another task. Through migration learning, combined with the training data of other clients for joint modeling, under the premise of ensuring that the privacy of the training data of the two parties is not leaked, the trained recognition model can be applied to the party with missing features and annotations, thereby greatly improving It expands the application range of joint learning and effectively improves the predictive ability of the training recognition model. It should be noted that, in the process of obtaining the first recognition model through the migration learning algorithm, the training data obtained through the migration learning is input into the basic model to obtain the first recognition model. In this embodiment, the speech recognition model can be ASR (Speech Recognition Technology, speech recognition technology) corresponding model.
需要说明的是,在客户端进行迁移学习过程中,客户端至少结合一个其它客户端的本地训练数据进行。当存在多个客户端时,各个客户端得到第一识别模型的原理一致,在此不再重复赘述。每一客户端中都存在迁移学习组件,通过该迁移学习组件来调用迁移学习算法得到第一识别模型。It should be noted that, during the migration learning process of the client, the client combines at least one other client's local training data. When there are multiple clients, the principles for obtaining the first recognition model for each client are the same, which will not be repeated here. There is a migration learning component in each client, and the migration learning algorithm is called through the migration learning component to obtain the first recognition model.
步骤S20,将所述第一识别模型的模型参数发送给服务端,以供所述服务端根据各客户端发送的模型参数进行联邦学习,得到联邦学习结果,并返回所述联邦学习结果。Step S20: Send the model parameters of the first recognition model to the server, so that the server performs federated learning according to the model parameters sent by each client, obtains the federated learning result, and returns the federated learning result.
当各客户端得到第一识别模型后,客户端获取自己第一识别模型的模型参数,将所获取的模型参数发送给服务端。需要说明的是,各个客户端第一识别模型的模型参数可能相同,也可能不相同。当服务端接收到各个客户端发送的模型参数,服务器根据各个客户端发送的模型参数进行联邦学习,得到联邦学习结果,将联邦学习结果发送给各个客户端。在服务端中,存在联邦学习组件,通过联邦学习组件可对各客户端发送的模型参数进行联邦学习,得到联邦学习结果。After each client obtains the first recognition model, the client obtains the model parameters of its own first recognition model, and sends the obtained model parameters to the server. It should be noted that the model parameters of the first recognition model of each client may be the same or different. When the server receives the model parameters sent by each client, the server performs federated learning according to the model parameters sent by each client, obtains the federated learning result, and sends the federated learning result to each client. On the server side, there is a federated learning component, and the federated learning component can perform federated learning on the model parameters sent by each client to obtain the federated learning result.
为了便于理解,本实施例对联邦学习进行解释。联邦学习(Federated Learning)是一种新兴的人工智能基础技术,其设计目标是在保障大数据交换时的信息安全、保护终端数据和个人隐私数据、在合法合规的前提下,在多方参与或多计算节点之间开展高效率的机器学习。在联邦学习的系统构架中,本实施例以包含两个数据拥有方(即企业A和企业B)的场景为例介绍联邦学习的系统构架。该构架可扩展至包含多个数据拥有方的场景。假设企业A和企业B想联合训练一个机器学习模型,它们的业务系统分别拥有各自用户的相关数据。此外,企业B还拥有模型需要预测的标签数据。出于数据隐私保护和安全考虑,企业A和企业B无法直接进行数据交换,可使用联邦学习系统建立模型。For ease of understanding, this embodiment explains federated learning. Federated Learning (Federated Learning) is an emerging basic artificial intelligence technology. Its design goal is to ensure information security during the exchange of big data, protect terminal data and personal privacy data, and under the premise of legal compliance, in the presence of multiple parties or Carry out high-efficiency machine learning among multiple computing nodes. In the system architecture of federated learning, this embodiment uses a scenario containing two data owners (ie, enterprise A and enterprise B) as an example to introduce the system architecture of federated learning. The framework can be extended to scenarios that include multiple data owners. Suppose enterprise A and enterprise B want to jointly train a machine learning model, and their business systems each have relevant data of their respective users. In addition, company B also has label data that the model needs to predict. For data privacy protection and security considerations, enterprise A and enterprise B cannot directly exchange data, and can use a federated learning system to build models.
使用联邦学习建立模型包括两部分内容,第一部分为:加密样本对齐。由于两家企业的用户群体并非完全重合,系统利用基于加密的用户样本对齐技术,在企业A和企业B不公开各自数据的前提下确认双方的共有用户,并且不暴露不互相重叠的用户,以便联合这些用户的特征进行建模。第二部分为:加密模型训练。在确定共有用户群体后,就可以利用这些数据训练机器学习模型。为了保证训练过程中数据的保密性,需要借助第三方协作者C进行加密训练。以线性回归模型为例,训练过程可分为以下4步:The use of federated learning to build a model includes two parts. The first part is: the alignment of encrypted samples. Since the user groups of the two companies are not completely overlapped, the system uses encryption-based user sample alignment technology to confirm the mutual users of both companies under the premise that company A and company B do not disclose their respective data, and does not expose users that do not overlap with each other, so that Combine the characteristics of these users for modeling. The second part is: encryption model training. After determining the common user group, the data can be used to train machine learning models. In order to ensure the confidentiality of the data during the training process, a third-party collaborator C needs to be used for encryption training. Taking the linear regression model as an example, the training process can be divided into the following 4 steps:
第①步:协作者C把公钥分发给企业A和企业B,用以对训练过程中需要交换的数据进行加密。Step ①: Collaborator C distributes the public key to enterprise A and enterprise B to encrypt the data that needs to be exchanged during the training process.
第②步:企业A和企业B之间以加密形式交互用于计算梯度的中间结果。Step ②: Encrypted interaction between enterprise A and enterprise B is used to calculate the intermediate result of the gradient.
第③步:企业A和企业B分别基于加密的梯度值进行计算,同时企业B根据其标签数据计算损失,并把结果汇总给协作者C。协作者C通过汇总结果计算总梯度值并将其解密。Step ③: Enterprise A and Enterprise B respectively calculate based on the encrypted gradient value, and at the same time, Enterprise B calculates the loss based on its label data, and summarizes the results to collaborator C. Collaborator C calculates the total gradient value by summing up the results and decrypts it.
第④步:协作者C将解密后的梯度值分别回传给企业A和企业B,企业A和企业B根据梯度值更新各自模型的参数。Step ④: Collaborator C sends the decrypted gradient value back to enterprise A and enterprise B respectively, and enterprise A and enterprise B update the parameters of their respective models according to the gradient value.
迭代上述步骤直至线性回归模型对应的损失函数收敛,损失函数为预先设置好的函数,这样就完成了整个训练过程。在样本对齐及模型训练过程中,企业A和企业B各自的数据均保留在本地,且训练中的数据交互也不会导致数据隐私泄露。因此,双方在联邦学习的帮助下得以实现合作训练模型。Iterate the above steps until the loss function corresponding to the linear regression model converges, and the loss function is a preset function, thus completing the entire training process. In the process of sample alignment and model training, the respective data of enterprise A and enterprise B are kept locally, and the data interaction during training will not cause data privacy leakage. Therefore, the two parties can realize the cooperative training model with the help of federated learning.
针对不同数据集,联邦学习分为横向联邦学习(horizontal federated learning)、纵向联邦学习(vertical federated learning)与联邦迁移学习(Federated Transfer Learning,FmL)。For different data sets, federated learning is divided into horizontal federated learning (horizontal federated learning), vertical federated learning and federated transfer learning (FmL).
横向联邦学习是在两个数据集的用户特征重叠较多,而用户重叠较少的情况下,把数据集按照横向(即用户维度)切分,并取出双方用户特征相同而用户不完全相同的那部分数据进行训练。比如有两家不同地区的银行,它们的用户群体分别来自各自所在的地区,相互的交集很小。但是,它们的业务很相似,因此,记录的用户特征是相同的。此时,就可以使用横向联邦学习来构建联合模型。Horizontal federated learning is to split the data set according to the horizontal (ie user dimension) when the user characteristics of the two data sets overlap more and the user overlaps less, and extract the user characteristics of the two parties that are the same but the users are not exactly the same. That part of the data for training. For example, there are two banks in different regions, and their user groups come from their respective regions, and their mutual intersections are small. However, their businesses are very similar, so the recorded user characteristics are the same. At this point, horizontal federated learning can be used to build a joint model.
纵向联邦学习是在两个数据集的用户重叠较多而用户特征重叠较少的情况下,把数据集按照纵向(即特征维度)切分,并取出双方用户相同而用户特征不完全相同的那部分数据进行训练。比如有两个不同的机构,家是某地的银行,另一家是同一个地方的电商,它们的用户群体很有可能包含该地的大部分居民因此用户的交集较大。但是,由于银行记录的都是用户的收支行为与信用评级,而电商则保有用户的浏览与购买历史,因此它们的用户特征交集较小。纵向联邦学习就是将这些不同特征在加密的状态下加以聚合,以增强模型能力。目前,逻辑回归模型、树形结构模型和神经网络模型等众多机器学习模型已经逐渐被证实能够建立在此联邦体系上。Longitudinal federated learning is to split the data set according to the longitudinal direction (ie feature dimension) when the users of the two data sets overlap more and the user characteristics are less overlapped, and extract the ones where the two users are the same but the user characteristics are not exactly the same. Part of the data for training. For example, there are two different institutions, one is a bank in a certain place, and the other is an e-commerce company in the same place. Their user groups are likely to include most of the residents in that place, so the intersection of users is relatively large. However, because banks record the user's income and expenditure behavior and credit rating, while e-commerce stores the user's browsing and purchase history, their user characteristics have a small intersection. Vertical federated learning is to aggregate these different features in an encrypted state to enhance the model's capabilities. At present, many machine learning models such as logistic regression model, tree structure model and neural network model have gradually been confirmed to be able to build on this federal system.
联邦迁移学习是在两个数据集的用户与用户特征重叠都较少的情况下,不对数据进行切分,而利用迁移学习来克服数据或标签不足的情况。比如有两个不同机构,一家是位于中国的银行,另一家是位于美国的电商。由于受地域限制,这两家机构的用户群体交集很小。同时,由于机构类型的不同,二者的数据特征也只有小部分重合。在这种情况下,要想进行有效的联邦学习,就必须引入迁移学习,来解决单边数据规模小和标签样本少的问题,从而提升模型的效果。Federated transfer learning is to use transfer learning to overcome the lack of data or labels when the user and user characteristics of the two data sets are less overlapped, without segmenting the data. For example, there are two different institutions, one is a bank in China, and the other is an e-commerce company in the United States. Due to geographical restrictions, the user groups of these two institutions have a very small intersection. At the same time, due to the different types of institutions, only a small part of the data characteristics of the two overlap. In this case, if you want to carry out effective federated learning, you must introduce transfer learning to solve the problem of small unilateral data and fewer label samples, so as to improve the effect of the model.
需要说明的是,在训练识别模型过程中,需要各个客户端用户的训练数据,这些训练数据是存储在各个用户所持客户端的本地,作为客户端对应用户的本地的训练数据。在本地的训练数据中,有部分训练数据属于用户的隐私数据,若没有用户的授权,无法获取到用户的隐私数据,即不可以将用户的隐私数据作为训练数据训练得到识别模型。目前,客户端中的部分应用通常会在用户协议中强制插入预先要求用户同意使用隐私数据的条款,或者用其它不知情的方式获取用户的隐私数据,这明显降低了用户隐私数据的私密性。而采用联邦学习,可以在不触及用户原始本地的训练数据的基础上,得到联邦学习结果,各客户端通过联邦学习结果来更新第一识别模型。It should be noted that in the process of training the recognition model, training data of each client user is required, and these training data are stored locally in the client held by each user as the local training data of the corresponding user of the client. In the local training data, part of the training data belongs to the user's private data. Without the user's authorization, the user's private data cannot be obtained, that is, the user's private data cannot be used as training data to train the recognition model. At present, some applications in the client usually compulsorily insert clauses requiring the user to agree to the use of private data in advance in the user agreement, or obtain the user's private data in other unknowing ways, which significantly reduces the privacy of the user's private data. With federated learning, the federated learning result can be obtained without touching the user's original local training data, and each client uses the federated learning result to update the first recognition model.
步骤S30,接收所述服务端发送的联邦学习结果,根据所述联邦学习结果更新所述第一识别模型,得到对应的第二识别模型,并将所述第二识别模型发送给所述服务端,以供所述服务端根据各客户端发送的第二识别模型和预设的遗传算法得到目标识别模型。Step S30: Receive the federated learning result sent by the server, update the first recognition model according to the federated learning result to obtain a corresponding second recognition model, and send the second recognition model to the server , So that the server can obtain the target recognition model according to the second recognition model sent by each client and the preset genetic algorithm.
各客户端接收服务端发送的联邦学习结果,并根据联邦学习结果更新第一识别模型,得到对应的第二识别模型,即更新后的第一识别模型就是第二识别模型。当各客户端得到第二识别模型后,各客户端将第二识别模型发送给服务端。当服务端接收到各客户端发送的第二识别模型后,服务端根据各客户端发送的第二识别模型和预设的遗传算法得到目标识别模型。Each client receives the federated learning result sent by the server, and updates the first recognition model according to the federated learning result to obtain the corresponding second recognition model, that is, the updated first recognition model is the second recognition model. After each client obtains the second recognition model, each client sends the second recognition model to the server. After the server receives the second recognition model sent by each client, the server obtains the target recognition model according to the second recognition model sent by each client and a preset genetic algorithm.
遗传算法(Genetic Algorithm)是通过搜索解决优化问题的方法,其先随机生成一定量的种群,遗传算法可包括复制(reproduce)算子、交叉(Crossover)算子和变异(Mutation)算子等算子。其中交叉算子为将种群成员的染色体进行编码,将两两种群成员的染色体编码交叉的过程;突变算子为将交叉后的染色体编码以一定概率变异的过程。在服务端中,存在遗传算法组件,通过遗传算法组件可根据各客户端发送的第二识别模型得到目标识别模型。Genetic Algorithm (Genetic Algorithm) is a method of solving optimization problems through search. It first generates a certain amount of population randomly. Genetic Algorithm can include reproduce operators, crossover operators, mutation operators, etc. son. The crossover operator is the process of encoding the chromosomes of the members of the population and the chromosome codes of the two or two group members are crossed; the mutation operator is the process of mutating the chromosome codes after the crossover with a certain probability. In the server, there is a genetic algorithm component, through which the target recognition model can be obtained according to the second recognition model sent by each client.
进一步地,所述接收所述服务端发送的联邦学习结果,根据所述联邦学习结果更新所述第一识别模型,得到对应的第二识别模型的步骤包括:Further, the step of receiving the federated learning result sent by the server, updating the first recognition model according to the federated learning result, and obtaining the corresponding second recognition model includes:
步骤a,接收所述服务端发送的联邦学习结果,获取预设学习率和所述联邦学习结果中的模型参数改变量。Step a: Receive a federated learning result sent by the server, and obtain a preset learning rate and a model parameter change amount in the federated learning result.
步骤b,根据所述预设学习率和所述模型参数改变量计算得到更新后的模型参数,根据更新后的模型参数更新所述第一识别模型,得到对应的第二识别模型。In step b, the updated model parameters are calculated according to the preset learning rate and the change amount of the model parameters, and the first recognition model is updated according to the updated model parameters to obtain the corresponding second recognition model.
进一步地,各客户端接收服务端发送的联邦学习结果,各客户端获取预设学习率和联邦学习结果中模型参数改变量,其中,预设学习率可根据具体需要而设置,本实施例对预设学习率的大小不做具体限制,不同客户端对应的预设学习率可以相同,也可以不相同。具体地,各客户端获取第一识别模型的模型参数,并计算预设学习率与模型参数改变量之间的乘积,将第一识别模型的模型参数减去计算所得的乘积,得到更新后的模型参数,并根据更新后的模型参数更新第一识别模型,即将更新后的模型参数作为识别模型的模型参数,从而得到第二识别模型。Further, each client receives the federated learning result sent by the server, and each client obtains a preset learning rate and the amount of model parameter changes in the federated learning result. The preset learning rate can be set according to specific needs. The size of the preset learning rate is not specifically limited, and the preset learning rates corresponding to different clients may be the same or different. Specifically, each client obtains the model parameters of the first recognition model, calculates the product of the preset learning rate and the amount of model parameter change, and subtracts the calculated product from the model parameters of the first recognition model to obtain the updated Model parameters, and update the first recognition model according to the updated model parameters, that is, use the updated model parameters as the model parameters of the recognition model to obtain the second recognition model.
本实施例通过客户端基于迁移学习算法,根据本地的训练数据训练得到第一识别模型,并将第一识别模型的模型参数发送给服务端,接收服务端发送的联邦学习结果,根据联邦学习结果更新第一识别模型,得到对应的第二识别模型,并将第二识别模型发送给所述服务端,以供服务端根据各客户端发送的第二识别模型和预设的遗传算法得到目标识别模型。通过各个客户端在迁移学习的场景下训练得到识别模型,提升了识别模型对各个用户相关信息识别的准确率,且本实施例支持在联邦学习场景下进行多个客户端识别模型的结合,在有效保护各个客户端对应用户训练数据隐私的情况下,进一步提高识别模型识别数据的准确率;本实施例支持通过遗传算法,在进化学习场景进行多个客户端识别模型的整合和优化,充分利用各个客户端对应识别模型背后的数据价值,从而进一步地提高了所得识别模型的识别准确率。In this embodiment, the client is trained based on the migration learning algorithm to obtain the first recognition model according to the local training data, and the model parameters of the first recognition model are sent to the server, and the federated learning result sent by the server is received, and according to the federated learning result Update the first recognition model to obtain the corresponding second recognition model, and send the second recognition model to the server, so that the server can obtain target recognition according to the second recognition model sent by each client and the preset genetic algorithm Model. The recognition model obtained by training each client in the migration learning scenario improves the accuracy of the recognition model for recognizing the relevant information of each user, and this embodiment supports the combination of multiple client recognition models in the federated learning scenario. In the case of effectively protecting the privacy of the training data of each client corresponding to the user, the accuracy of the recognition data of the recognition model is further improved; this embodiment supports the integration and optimization of multiple client recognition models in the evolutionary learning scenario through the genetic algorithm, making full use of Each client corresponds to the data value behind the recognition model, thereby further improving the recognition accuracy of the obtained recognition model.
进一步地,提出本申请识别模型的训练方法第二实施例。所述识别模型的训练方法第二实施例与所述识别模型的训练方法第一实施例的区别在于,参照图2,所述识别模型的训练方法还包括:Further, a second embodiment of the training method of the recognition model of the present application is proposed. The difference between the second embodiment of the recognition model training method and the first embodiment of the recognition model training method is that, referring to FIG. 2, the recognition model training method further includes:
步骤S40,接收各客户端发送的第一识别模型对应的模型参数,根据所述模型参数进行联邦学习,得到联邦学习结果。Step S40: Receive model parameters corresponding to the first recognition model sent by each client, and perform federated learning according to the model parameters to obtain a federated learning result.
服务端接收各客户端发送的第一识别模型对应的模型参数,并根据模型参数进行联邦学习,得到联邦学习结果。需要说明的是,联邦学习的过程已在上述实施例中详细阐述,在此不再重复赘述。在本实施例中,联邦学习结果可为模型参数改变量,具体地,模型参数改变量可为在进行联邦学习过程中,对应损失函数收敛时,根据损失函数计算得到的梯度值。The server receives the model parameters corresponding to the first recognition model sent by each client, and performs federated learning according to the model parameters to obtain the federated learning result. It should be noted that the process of federated learning has been described in detail in the foregoing embodiment, and will not be repeated here. In this embodiment, the federated learning result may be a model parameter change amount. Specifically, the model parameter change amount may be a gradient value calculated according to the loss function when the corresponding loss function converges during the federated learning process.
步骤S50,将所述联邦学习结果发送给各客户端,以供各客户端根据所述联邦学习结果更新对应的第一识别模型,得到对应的第二识别模型。Step S50: Send the federated learning result to each client, so that each client can update the corresponding first recognition model according to the federated learning result to obtain the corresponding second recognition model.
当服务端得到联邦学习结果后,服务端将联邦学习结果发送给各客户端,以供各客户端根据联邦学习结果更新各自对应的第一识别模型,得到对应的第二识别模型。各客户端根据联邦学习结果更新第一识别模型,得到第二识别模型的过程已在上述实施例中详细描述,在此不再重复赘述。After the server obtains the federated learning result, the server sends the federated learning result to each client, so that each client can update the corresponding first recognition model according to the federated learning result to obtain the corresponding second recognition model. Each client updates the first recognition model according to the federated learning result, and the process of obtaining the second recognition model has been described in detail in the foregoing embodiment, and will not be repeated here.
步骤S60,接收各客户端发送的第二识别模型,根据各第二识别模型和预设的遗传算法得到目标识别模型。Step S60: Receive the second recognition model sent by each client, and obtain the target recognition model according to each second recognition model and a preset genetic algorithm.
服务端接收各客户端发送的第二识别模型,根据各第二识别模型和预设的遗传算法得到目标识别模型。可以理解的是,该目标识别模型即为所需的最优识别模型。当服务端得到目标识别模型后,服务端可将目标识别模型发送给各客户端,以供各客户端在接收到识别请求后,根据目标识别模型进行相应的识别操作。当各客户端接收到目标识别模型后,各客户端存储所接收的目标识别模型。The server receives the second recognition model sent by each client, and obtains the target recognition model according to each second recognition model and a preset genetic algorithm. It is understandable that the target recognition model is the optimal recognition model required. After the server obtains the target recognition model, the server can send the target recognition model to each client, so that each client can perform a corresponding recognition operation according to the target recognition model after receiving the recognition request. After each client receives the target recognition model, each client stores the received target recognition model.
进一步地,步骤S50包括:Further, step S50 includes:
步骤d,接收各客户端发送的第二识别模型,在所接收的各第二识别模型中选择父代识别模型,并在预设的遗传算法中选择目标算子。Step d, receiving the second recognition model sent by each client, selecting the parent recognition model from each of the received second recognition models, and selecting the target operator in the preset genetic algorithm.
步骤e,通过所述父代识别模型和所述目标算子得到所述父代识别模型对应的子代识别模型。Step e: Obtain a child recognition model corresponding to the parent recognition model through the parent recognition model and the target operator.
步骤f,若检测到所述子代识别模型满足结束条件,则将满足结束条件中识别准确率最高的子代识别模型确定为目标识别模型。In step f, if it is detected that the offspring recognition model meets the end condition, the offspring recognition model with the highest recognition accuracy rate among the end conditions is determined as the target recognition model.
具体地,服务端接收各客户端发送的第二识别模型,在所接收的各第二识别模型中随机选择第二识别模型确定为父代识别模型,并在预设的遗传算法中随机选择算子确定为目标算子。当存在至少两个父代识别模型时,每一父代识别模型对应的目标算子可相同,也可不相同。当服务端确定父代识别模型,以及父代识别模型对应的目标算子后,服务端通过父代识别模型和该父代识别模型对应的目标算子,得到该父代识别模型对应的子代识别模型,并检测所得的子代识别模型是否满足结束条件。若检测到所得的子代识别模型满足结束条件,则获取满足结束条件中识别准确率最高的子代识别模型,将该准确率最高的子代识别模型确定为目标识别模型;若检测到所得的子代识别模型未满足结束条件,服务端则将所得的子代识别模型作为下一次迭代的父代识别模型,并确定下一次父代识别模型对应的目标算子,继续根据父代识别模型和目标算子得到父代识别模型对应的子代识别模型,在检测到子代识别模型满足结束条件时,将满足条件中识别准确率最高的子代识别模型确定为目标识别模型。Specifically, the server receives the second recognition model sent by each client, randomly selects the second recognition model from each of the received second recognition models to determine it as the parent recognition model, and randomly selects the calculation model in the preset genetic algorithm. The sub is determined as the target operator. When there are at least two parent recognition models, the target operator corresponding to each parent recognition model may be the same or different. After the server determines the parent recognition model and the target operator corresponding to the parent recognition model, the server obtains the child corresponding to the parent recognition model through the parent recognition model and the target operator corresponding to the parent recognition model Recognize the model, and check whether the resulting offspring recognition model meets the end condition. If the detected offspring recognition model satisfies the end condition, the offspring recognition model with the highest recognition accuracy rate among the end conditions is obtained, and the offspring recognition model with the highest accuracy rate is determined as the target recognition model; if the obtained offspring recognition model is detected If the offspring recognition model does not meet the ending conditions, the server will use the obtained offspring recognition model as the parent recognition model for the next iteration, and determine the target operator corresponding to the next parent recognition model, and continue according to the parent recognition model and The target operator obtains the offspring recognition model corresponding to the parent recognition model, and when it is detected that the offspring recognition model satisfies the end condition, the offspring recognition model with the highest recognition accuracy rate among the satisfied conditions is determined as the target recognition model.
进一步地,步骤e包括:Further, step e includes:
步骤e1,获取所述父代识别模型对应的待处理模型参数,将所述待处理模型参数确定为参数编码。Step e1: Obtain the to-be-processed model parameters corresponding to the parent recognition model, and determine the to-be-processed model parameters as parameter codes.
步骤e2,通过所述目标算子对所述参数编码进行处理,得到新的参数编码。Step e2, processing the parameter encoding through the target operator to obtain a new parameter encoding.
步骤e3,根据所述新的参数编码对应更新所述父代识别模型,得到子代识别模型。Step e3: Correspondingly update the parent recognition model according to the new parameter code to obtain the offspring recognition model.
在本实施例中,服务端获取各第二识别模型中的模型参数,即获取第二识别模型中作为父代识别模型的模型参数,将所获取的模型参数确定为待处理模型参数,将待处理模型参数确定遗传算法的参数编码,通过遗传算法中的算子对参数编码进行处理,即通过父代识别模型对应的目标算子对待处理参数编码进行处理,得到新的参数编码,并根据新的参数编码更新对应父代识别模型,得到子代识别模型。可以理解的是,各客户端也可将第二识别模型中的模型参数发送给服务端。需要说明的是,在对参数编码进行处理过程中,具体采用何种处理手段,是由对应的目标算子决定的。如对于突变算子,子代识别模型可只继承父代识别模型的特定几个模型参数,此时遗传算法提取父代识别模型的特定参数编码,将所提取的特定参数编码作为新的参数编码。对于交叉算子,可将两个父代识别模型对应的参数编码进行整合,得到新的参数编码,整合的具体算法由遗传算法对应的算子决定,如可采用求取平均值的方法对参数编码进行整合。In this embodiment, the server obtains the model parameters in each second recognition model, that is, obtains the model parameters of the second recognition model as the parent recognition model, determines the obtained model parameters as the model parameters to be processed, and sets the parameters to be processed. Process model parameters to determine the parameter encoding of the genetic algorithm, and process the parameter encoding through the operator in the genetic algorithm, that is, process the encoding of the parameters to be processed through the target operator corresponding to the parent identification model, and obtain the new parameter encoding, and according to the new The parameter coding update corresponds to the parent recognition model, and the offspring recognition model is obtained. It is understandable that each client can also send the model parameters in the second recognition model to the server. It should be noted that in the process of processing the parameter encoding, the specific processing method used is determined by the corresponding target operator. For example, for a mutation operator, the offspring recognition model can only inherit a few specific model parameters of the parent recognition model. At this time, the genetic algorithm extracts the specific parameter code of the parent recognition model, and uses the extracted specific parameter code as the new parameter code . For the crossover operator, the parameter codes corresponding to the two parent recognition models can be integrated to obtain a new parameter code. The specific algorithm of integration is determined by the operator corresponding to the genetic algorithm. For example, the method of averaging the parameters can be used Coding for integration.
进一步地,服务端在各第二识别模型中选择父代识别模型过程中,可根据服务端中当前的可利用资源来确定所选择父代识别模型的数量,当服务端中的可利用资源足够多时,可将所有第二识别模型都确定为父代识别模型。可以理解的是,随着服务端中可利用资源的增加,所选择的父代识别模型的数量也会对应增加。Further, when the server selects the parent recognition model in each second recognition model, the number of selected parent recognition models can be determined according to the current available resources in the server. When the available resources in the server are sufficient For a long time, all second recognition models can be determined as parent recognition models. It is understandable that as the available resources in the server increase, the number of selected parent identification models will increase accordingly.
进一步地,在检测子代识别模型是否满足结束条件时,可在所有子代识别模型都满足结束条件时,才确定子代识别模型满足结束条件,若有一个子代识别模型未满足结束条件,则确定子代识别模型不满足结束条件。也可在检测到子代识别模型中,有预设比例的子代识别模型满足结束条件时,确定子代识别模型满足结束条件,否则,确定子代识别模型未满足结束条件。此时,预设比例的大小可根据具体需要而设置,如可设置为60%,75%或者80%等。Further, when detecting whether the progeny recognition model meets the end condition, it can be determined that the progeny recognition model meets the end condition when all the progeny recognition models meet the end condition, and if one of the progeny recognition models does not meet the end condition, It is determined that the offspring recognition model does not meet the end condition. It is also possible to determine that the child recognition model satisfies the end condition when the child recognition model with a preset ratio in the child recognition model is detected to meet the end condition; otherwise, it is determined that the child recognition model does not meet the end condition. At this time, the size of the preset ratio can be set according to specific needs, such as 60%, 75%, or 80%.
进一步地,步骤f包括:Further, step f includes:
步骤f1,若检测到所述子代识别模型对应的迭代次数大于预设次数,则确定所述子代识别模型满足结束条件,并采用预存测试数据获取满足结束条件的各子代识别模型的识别准确率。Step f1: If it is detected that the number of iterations corresponding to the offspring identification model is greater than the preset number, it is determined that the offspring identification model meets the termination condition, and the pre-stored test data is used to obtain the identification of each offspring identification model that meets the termination condition Accuracy.
步骤f2,选择识别准确率最高的子代识别模型确定为目标识别模型。In step f2, the offspring recognition model with the highest recognition accuracy is selected and determined as the target recognition model.
进一步地,若服务端检测到子代识别模型对应迭代次数大于预设次数,则确定子代识别模型满足结束条件。其中,预设次数可根据具体需要而设置,本实施例不限制预设次数的大小。可以理解的是,当服务端第一次通过遗传算法得到子代识别模型时,对应的迭代次数为1;当服务端第二次通过遗传算法得到子代识别模型时,对应的迭代次数为2,即迭代次数等于通过遗传算法得到子代识别模型的次数。当确定子代识别模型满足结束条件时,服务端获取预先存储的测试数据,并将该测试数据输入至各子代识别模型中,以确定各个子代识别模型的识别准确率,并对比各个子代识别模型的识别准确率,选择识别准确率最高的子代识别模型确定为目标识别模型。进一步地,若检测到子代识别模型对应的迭代次数小于或者等于预设次数,服务端则确定子代识别模型未满足结束条件。Further, if the server detects that the number of iterations corresponding to the offspring identification model is greater than the preset number, it is determined that the offspring identification model meets the end condition. The preset number of times can be set according to specific needs, and this embodiment does not limit the size of the preset number of times. It is understandable that when the server obtains the offspring recognition model through the genetic algorithm for the first time, the corresponding iteration number is 1; when the server obtains the offspring identification model through the genetic algorithm for the second time, the corresponding iteration number is 2. , That is, the number of iterations is equal to the number of offspring identification models obtained through genetic algorithm. When it is determined that the offspring recognition model meets the end conditions, the server obtains the pre-stored test data, and inputs the test data into each offspring recognition model to determine the recognition accuracy of each offspring recognition model, and compares each child The recognition accuracy rate of the generation recognition model is selected, and the offspring recognition model with the highest recognition accuracy rate is selected as the target recognition model. Further, if it is detected that the number of iterations corresponding to the offspring identification model is less than or equal to the preset number, the server determines that the offspring identification model does not meet the end condition.
进一步地,步骤f还包括:Further, step f also includes:
步骤f3,若检测到所述子代识别模型对应的识别准确率大于或者等于预设准确率,则确定所述子代识别模型满足结束条件。Step f3: If it is detected that the recognition accuracy rate corresponding to the offspring recognition model is greater than or equal to the preset accuracy rate, it is determined that the offspring recognition model meets the end condition.
步骤f4,在所述子代识别模型中选择识别准确率最大的子代识别模型确定为目标识别模型。Step f4, selecting the child recognition model with the highest recognition accuracy rate among the child recognition models and determining it as the target recognition model.
进一步地,服务端每得到一次子代识别模型时,服务端都获取预先存储的测试数据,将测试数据输入至子代识别模型中,得到子代识别模型对应的识别准确率,并检测各个子代识别模型对应的识别准确率是否大于或者等于预设准确率。其中,预设准确率的大小可根据具体需要而设置,本实施例对预设准确率的大小不做具体限制。当服务端检测到各个子代识别模型对应识别准确率都大于或者等于预设准确率后,服务端确定子代识别模型满足结束条件,在子代识别模型中选择识别准确率最大的子代识别模型确定为目标识别模型;当服务端检测到存在子代识别模型对应的识别准确率小于预设准确率时,服务端确定子代识别模型不满足结束条件。Further, every time the server obtains the offspring recognition model, the server obtains the pre-stored test data, inputs the test data into the offspring recognition model, obtains the recognition accuracy rate corresponding to the offspring recognition model, and detects each child Whether the recognition accuracy rate corresponding to the generation recognition model is greater than or equal to the preset accuracy rate. The size of the preset accuracy rate can be set according to specific needs, and this embodiment does not specifically limit the size of the preset accuracy rate. When the server detects that the corresponding recognition accuracy of each child recognition model is greater than or equal to the preset accuracy, the server determines that the child recognition model meets the end condition, and selects the child recognition with the highest recognition accuracy in the child recognition model The model is determined to be the target recognition model; when the server detects that the recognition accuracy rate corresponding to the offspring recognition model is less than the preset accuracy rate, the server determines that the offspring recognition model does not meet the end condition.
进一步地,服务端也可在检测识别准确率大于或者等于预设准确率的子代识别模型占当前迭代过程中所有子代识别模型的比例大于特定比例时,确定子代识别模型满足结束条件,否则,确定子代识别模型未满足结束条件。其中,本实施例不限制特定比例的大小。Further, the server can also determine that the child recognition model meets the end condition when the proportion of the child recognition model whose detection and recognition accuracy rate is greater than or equal to the preset accuracy rate to all the child recognition models in the current iteration process is greater than a specific ratio, Otherwise, it is determined that the offspring recognition model does not meet the end condition. Among them, this embodiment does not limit the size of the specific ratio.
进一步地,服务端也可在确定子代识别模型满足结束条件中,按照识别准确率将各个子代识别模型从高到低排序,得到排序后的子代识别模型,然后在排序后的子代识别模型中从前到后获取预设数量的子代识别模型确定为目标子代识别模型,获取目标子代识别模型的模型参数,将各个目标子代识别模型的模型参数进行线性相加,得到目标模型参数,将目标模型参数对应的识别模型确定为目标模型。其中,预设数量的大小可根据具体需要而设置,如可将预设数量设置为2、3或者5。线性相加的具体过程可为计算各个目标子代识别模型对应模型参数的平均值,将该平均值对应确定为目标模型参数;线性相加的过程还可为确定各个目标子代识别模型对应的权重,将各个目标子代识别模型的模型参数乘以对应的权重,得到乘积,然后将各个目标子代识别模型中,同一模型参数对应的乘积相加,得到对应的目标模型参数,可以理解的是,识别准确率越高,对应目标子代识别模型的模型参数的权重越大。Further, the server can also sort the child recognition models from high to low according to the recognition accuracy when determining that the child recognition model meets the end condition, and obtain the sorted child recognition model, and then the sorted child recognition model Obtain a preset number of offspring recognition models from front to back in the recognition model to determine the target offspring recognition model, obtain the model parameters of the target offspring recognition model, and linearly add the model parameters of each target offspring recognition model to obtain the target Model parameters, the recognition model corresponding to the target model parameters is determined as the target model. Among them, the size of the preset number can be set according to specific needs, for example, the preset number can be set to 2, 3, or 5. The specific process of linear addition can be to calculate the average value of the model parameters corresponding to each target offspring recognition model, and determine the average value as the target model parameter; the process of linear addition can also be to determine the corresponding value of each target offspring recognition model Weight, multiply the model parameters of each target offspring recognition model by the corresponding weight to get the product, and then add the products corresponding to the same model parameter in each target offspring recognition model to get the corresponding target model parameters, understandable Yes, the higher the recognition accuracy, the greater the weight of the model parameters corresponding to the target offspring recognition model.
本实施例通过服务端接收各客户端发送的第一识别模型对应的模型参数,根据模型参数进行联邦学习,将所得的联邦学习结果发送给各客户端,并接收各客户端发送的第二识别模型,根据各第二识别模型和预设的遗传算法得到目标识别模型。通过各个客户端在迁移学习的场景下训练得到识别模型,提升了识别模型对各个用户相关信息识别的准确率,且本实施例支持在联邦学习场景下进行多个客户端识别模型的结合,在有效保护各个客户端对应用户训练数据隐私的情况下,进一步提高识别模型识别数据的准确率;本实施例支持通过遗传算法,在进化学习场景进行多个客户端识别模型的整合和优化,充分利用各个客户端对应识别模型背后的数据价值,从而进一步地提高了所得识别模型的识别准确率。In this embodiment, the server receives the model parameters corresponding to the first recognition model sent by each client, performs federated learning according to the model parameters, sends the result of federated learning to each client, and receives the second recognition sent by each client Model, obtain the target recognition model according to each second recognition model and the preset genetic algorithm. The recognition model obtained by training each client in the migration learning scenario improves the accuracy of the recognition model for recognizing the relevant information of each user, and this embodiment supports the combination of multiple client recognition models in the federated learning scenario. In the case of effectively protecting the privacy of the training data of each client corresponding to the user, the accuracy of the recognition data of the recognition model is further improved; this embodiment supports the integration and optimization of multiple client recognition models in the evolutionary learning scenario through the genetic algorithm, making full use of Each client corresponds to the data value behind the recognition model, thereby further improving the recognition accuracy of the obtained recognition model.
此外,本申请还提供一种识别模型的训练装置,参照图3,所述识别模型的训练装置应用于客户端中,所述识别模型的训练装置包括:In addition, the present application also provides a training device for a recognition model. Referring to FIG. 3, the training device for the recognition model is applied to the client, and the training device for the recognition model includes:
获取模块10,用于获取本地的训练数据;The obtaining module 10 is used to obtain local training data;
训练模块20,用于根据所述训练数据训练得到第一识别模型;The training module 20 is configured to train according to the training data to obtain a first recognition model;
发送模块30,用于将所述第一识别模型的模型参数发送给服务端,以供所述服务端根据各客户端发送的模型参数进行联邦学习,得到联邦学习结果,并返回所述联邦学习结果;The sending module 30 is configured to send the model parameters of the first recognition model to the server, so that the server performs federated learning according to the model parameters sent by each client, obtains the federated learning result, and returns the federated learning result;
接收模块40,用于接收所述服务端发送的联邦学习结果;The receiving module 40 is configured to receive the federated learning result sent by the server;
更新模块50,用于根据所述联邦学习结果更新所述第一识别模型,得到对应的第二识别模型;The update module 50 is configured to update the first recognition model according to the federated learning result to obtain a corresponding second recognition model;
所述发送模块30还用于将所述第二识别模型发送给所述服务端,以供所述服务端根据各客户端发送的第二识别模型和预设的遗传算法得到目标识别模型。The sending module 30 is further configured to send the second recognition model to the server, so that the server obtains the target recognition model according to the second recognition model sent by each client and a preset genetic algorithm.
进一步地,所述训练模块20还用于获取本地的训练数据,基于迁移学习算法,根据所述训练数据训练得到第一识别模型。Further, the training module 20 is also used to obtain local training data, and based on a migration learning algorithm, train according to the training data to obtain a first recognition model.
进一步地,所述更新模块50包括:Further, the update module 50 includes:
获取单元,用于获取预设学习率和所述联邦学习结果中的模型参数改变量;An acquiring unit for acquiring a preset learning rate and the model parameter change amount in the federated learning result;
计算单元,用于根据所述预设学习率和所述模型参数改变量计算得到更新后的模型参数;A calculation unit, configured to calculate the updated model parameters according to the preset learning rate and the change amount of the model parameters;
更新单元,根据更新后的模型参数更新所述第一识别模型,得到对应的第二识别模型。The update unit updates the first recognition model according to the updated model parameters to obtain the corresponding second recognition model.
本申请识别模型的训练装置具体实施方式与上述识别模型的训练方法第一实施例中各步骤基本相同,在此不再赘述。The specific implementation of the training device for the recognition model of this application is basically the same as the steps in the first embodiment of the training method for the recognition model described above, and will not be repeated here.
此外,本申请还提供一种识别模型的训练装置,参照图4,所述识别模型的训练装置应用于服务端,所述识别模型的训练装置包括:In addition, the present application also provides a training device for a recognition model. Referring to FIG. 4, the training device for the recognition model is applied to the server, and the training device for the recognition model includes:
接收模块60,用于接收各客户端发送的第一识别模型对应的模型参数;The receiving module 60 is configured to receive model parameters corresponding to the first recognition model sent by each client;
联邦学习模块70,用于根据所述模型参数进行联邦学习,得到联邦学习结果;The federated learning module 70 is configured to perform federated learning according to the model parameters to obtain federated learning results;
发送模块80,用于将所述联邦学习结果发送给各客户端,以供各客户端根据所述联邦学习结果更新对应的第一识别模型,得到对应的第二识别模型;The sending module 80 is configured to send the federated learning result to each client, so that each client can update the corresponding first recognition model according to the federated learning result to obtain the corresponding second recognition model;
所述接收模块60还用于接收各客户端发送的第二识别模型,根据各第二识别模型和预设的遗传算法得到目标识别模型。The receiving module 60 is also configured to receive the second recognition model sent by each client, and obtain the target recognition model according to each second recognition model and a preset genetic algorithm.
进一步地,所述接收模块60包括:Further, the receiving module 60 includes:
接收单元,用于接收各客户端发送的第二识别模型;The receiving unit is configured to receive the second recognition model sent by each client;
选择单元,用于在所接收的各第二识别模型中选择父代识别模型,并在预设的遗传算法中选择目标算子;The selection unit is used to select the parent recognition model among the received second recognition models, and select the target operator in the preset genetic algorithm;
确定单元,用于通过所述父代识别模型和所述目标算子得到所述父代识别模型对应的子代识别模型;若检测到所述子代识别模型满足结束条件,则将满足结束条件中识别准确率最高的子代识别模型确定为目标识别模型。The determining unit is configured to obtain the child recognition model corresponding to the parent recognition model through the parent recognition model and the target operator; if it is detected that the child recognition model satisfies the end condition, the end condition will be met The child recognition model with the highest recognition accuracy rate is determined as the target recognition model.
进一步地,所述确定单元包括:Further, the determining unit includes:
第一获取子单元,用于获取所述父代识别模型对应的待处理模型参数;The first obtaining subunit is configured to obtain the to-be-processed model parameters corresponding to the parent recognition model;
第一确定子单元,用于将所述待处理模型参数确定为参数编码;The first determining subunit is used to determine the parameters of the model to be processed as parameter codes;
处理子单元,用于通过所述目标算子对所述参数编码进行处理,得到新的参数编码;A processing subunit, configured to process the parameter encoding through the target operator to obtain a new parameter encoding;
更新子单元,用于根据所述新的参数编码对应更新所述父代识别模型,得到子代识别模型。The update subunit is used to update the parent recognition model corresponding to the new parameter encoding to obtain the child recognition model.
进一步地,所述确定单元包括:Further, the determining unit includes:
第二确定子单元,用于若检测到所述子代识别模型对应的迭代次数大于预设次数,则确定所述子代识别模型满足结束条件;The second determining subunit is configured to determine that the child recognition model meets the ending condition if it is detected that the number of iterations corresponding to the offspring recognition model is greater than the preset number;
第二获取子单元,用于采用预存测试数据获取满足结束条件的各子代识别模型的识别准确率;The second acquisition subunit is used to acquire the recognition accuracy rate of each offspring recognition model that meets the end condition by using the pre-stored test data;
所述第二确定子单元还用于选择识别准确率最高的子代识别模型确定为目标识别模型。The second determining subunit is also used to select the offspring recognition model with the highest recognition accuracy rate and determine it as the target recognition model.
进一步地,所述确定单元还用于若检测到所述子代识别模型对应的识别准确率大于或者等于预设准确率,则确定所述子代识别模型满足结束条件;在所述子代识别模型中选择识别准确率最大的子代识别模型确定为目标识别模型。Further, the determining unit is further configured to determine that the child recognition model satisfies the end condition if the recognition accuracy corresponding to the child recognition model is detected to be greater than or equal to a preset accuracy rate; In the model, the child recognition model with the highest recognition accuracy rate is selected as the target recognition model.
本申请识别模型的训练装置具体实施方式与上述识别模型的训练方法第二实施例中各步骤基本相同,在此不再赘述。The specific implementation of the training device for the recognition model of the present application is basically the same as the steps in the second embodiment of the training method for the recognition model described above, and will not be repeated here.
此外,本申请还提供一种识别模型的训练设备。如图5所示,图5是本申请实施例方案涉及的硬件运行环境的结构示意图。In addition, this application also provides a training device for identifying models. As shown in FIG. 5, FIG. 5 is a schematic structural diagram of the hardware operating environment involved in the solution of the embodiment of the present application.
需要说明的是,图5即可为识别模型的训练设备的硬件运行环境的结构示意图,识别模型的训练设备可为客户端,也可为服务端。本申请实施例识别模型的训练设备可以是PC,便携计算机等终端设备。It should be noted that FIG. 5 can be a schematic structural diagram of the hardware operating environment of the training device for the recognition model. The training device for the recognition model may be a client or a server. The training device for the recognition model in the embodiment of the present application may be a terminal device such as a PC and a portable computer.
如图5所示,该识别模型的训练设备可以包括:处理器1001,例如CPU,存储器1005,用户接口1003,网络接口1004,通信总线1002。其中,通信总线1002用于实现这些组件之间的连接通信。用户接口1003可以包括显示屏(Display)、输入单元比如键盘(Keyboard),可选用户接口1003还可以包括标准的有线接口、无线接口。网络接口1004可选的可以包括标准的有线接口、无线接口(如WI-FI接口)。存储器1005可以是高速RAM存储器,也可以是稳定的存储器(non-volatile memory),例如磁盘存储器。存储器1005可选的还可以是独立于前述处理器1001的存储装置。As shown in FIG. 5, the training device of the recognition model may include: a processor 1001, such as a CPU, a memory 1005, a user interface 1003, a network interface 1004, and a communication bus 1002. Among them, the communication bus 1002 is used to implement connection and communication between these components. The user interface 1003 may include a display screen (Display) and an input unit such as a keyboard (Keyboard), and the optional user interface 1003 may also include a standard wired interface and a wireless interface. The network interface 1004 may optionally include a standard wired interface and a wireless interface (such as a WI-FI interface). The memory 1005 may be a high-speed RAM memory, or a stable memory (non-volatile memory), such as a magnetic disk memory. Optionally, the memory 1005 may also be a storage device independent of the aforementioned processor 1001.
本领域技术人员可以理解,图5中示出的识别模型的训练设备结构并不构成对识别模型的训练设备的限定,可以包括比图示更多或更少的部件,或者组合某些部件,或者不同的部件布置。Those skilled in the art can understand that the structure of the training device for the recognition model shown in FIG. 5 does not constitute a limitation on the training device for the recognition model, and may include more or less components than shown in the figure, or a combination of certain components, Or different component arrangements.
如图5所示,作为一种计算机存储介质的存储器1005中可以包括操作系统、网络通信模块、用户接口模块以及识别模型的训练程序。其中,操作系统是管理和控制识别模型的训练设备硬件和软件资源的程序,支持识别模型的训练程序以及其它软件或程序的运行。As shown in FIG. 5, the memory 1005 as a computer storage medium may include an operating system, a network communication module, a user interface module, and a training program for a recognition model. Among them, the operating system is a program that manages and controls the hardware and software resources of the training equipment of the recognition model, and supports the operation of the training program of the recognition model and other software or programs.
在图5所示的识别模型的训练设备中,当识别模型的训练设备为客户端时,用户接口1003主要用于连接服务端,与服务端进行数据通信;当识别模型的训练设备为服务端时,用户接口1003主要用于连接各客户端,与各客户端进行数据通信;网络接口1004主要用于后台服务端,与后台服务端进行数据通信;处理器1001可以用于调用存储器1005中存储的识别模型的训练程序,并执行如上所述的识别模型的训练方法的步骤。In the training device of the recognition model shown in FIG. 5, when the training device of the recognition model is the client, the user interface 1003 is mainly used to connect to the server and communicate with the server; when the training device of the recognition model is the server At the time, the user interface 1003 is mainly used to connect various clients and communicate with each client; the network interface 1004 is mainly used to communicate with the back-end server and perform data communication with the back-end server; the processor 1001 can be used to call the storage in the memory 1005 The training program of the recognition model, and the steps of the training method of the recognition model as described above are executed.
本申请识别模型的训练设备具体实施方式与上述识别模型的训练方法各实施例基本相同,在此不再赘述。The specific implementation of the training device for the recognition model of the present application is basically the same as the foregoing embodiments of the training method for the recognition model, and will not be repeated here.
此外,本申请实施例还提出一种计算机可读存储介质,所述计算机可读存储介质上存储有识别模型的训练程序,所述识别模型的训练程序被处理器执行时实现如上所述的识别模型的训练方法的步骤。In addition, an embodiment of the present application also proposes a computer-readable storage medium, the computer-readable storage medium stores a training program for a recognition model, and the training program for the recognition model is executed by a processor to realize the recognition as described above. The steps of the model training method.
本申请计算机可读存储介质具体实施方式与上述识别模型的训练方法各实施例基本相同,在此不再赘述。The specific implementation of the computer-readable storage medium of the present application is basically the same as the foregoing embodiments of the training method of the recognition model, and will not be repeated here.
需要说明的是,在本文中,术语“包括”、“包含”或者其任何其他变体意在涵盖非排他性的包含,从而使得包括一系列要素的过程、方法、物品或者装置不仅包括那些要素,而且还包括没有明确列出的其他要素,或者是还包括为这种过程、方法、物品或者装置所固有的要素。在没有更多限制的情况下,由语句“包括一个……”限定的要素,并不排除在包括该要素的过程、方法、物品或者装置中还存在另外的相同要素。It should be noted that in this article, the terms "include", "include" or any other variants thereof are intended to cover non-exclusive inclusion, so that a process, method, article or device including a series of elements not only includes those elements, It also includes other elements that are not explicitly listed, or elements inherent to the process, method, article, or device. If there are no more restrictions, the element defined by the sentence "including a..." does not exclude the existence of other identical elements in the process, method, article, or device that includes the element.
上述本申请实施例序号仅仅为了描述,不代表实施例的优劣。The serial numbers of the foregoing embodiments of the present application are for description only, and do not represent the superiority or inferiority of the embodiments.
通过以上的实施方式的描述,本领域的技术人员可以清楚地了解到上述实施例方法可借助软件加必需的通用硬件平台的方式来实现,当然也可以通过硬件,但很多情况下前者是更佳的实施方式。基于这样的理解,本申请的技术方案本质上或者说对现有技术做出贡献的部分可以以软件产品的形式体现出来,该计算机软件产品存储在一个存储介质(如ROM/RAM、磁碟、光盘)中,包括若干指令用以使得一台终端设备(可以是手机,计算机,服务端,空调器,或者网络设备等)执行本申请各个实施例所述的方法。Through the description of the above implementation manners, those skilled in the art can clearly understand that the above-mentioned embodiment method can be implemented by means of software plus the necessary general hardware platform, of course, it can also be implemented by hardware, but in many cases the former is better.的实施方式。 Based on this understanding, the technical solution of this application essentially or the part that contributes to the existing technology can be embodied in the form of a software product, and the computer software product is stored in a storage medium (such as ROM/RAM, magnetic disk, The optical disc) includes several instructions to enable a terminal device (which may be a mobile phone, a computer, a server, an air conditioner, or a network device, etc.) to execute the method described in each embodiment of the present application.
以上仅为本申请的优选实施例,并非因此限制本申请的专利范围,凡是利用本申请说明书及附图内容所作的等效结构或等效流程变换,或直接或间接运用在其他相关的技术领域,均同理包括在本申请的专利保护范围内。The above are only the preferred embodiments of the application, and do not limit the scope of the patent for this application. Any equivalent structure or equivalent process transformation made using the content of the description and drawings of the application, or directly or indirectly applied to other related technical fields , The same reason is included in the scope of patent protection of this application.

Claims (20)

  1. 一种识别模型的训练方法,应用于客户端,其中,所述识别模型的训练方法包括以下步骤:A training method for a recognition model is applied to a client, wherein the training method for the recognition model includes the following steps:
    获取本地的训练数据,根据所述训练数据训练得到第一识别模型;Acquiring local training data, and training according to the training data to obtain a first recognition model;
    将所述第一识别模型的模型参数发送给服务端,以供所述服务端根据各客户端发送的模型参数进行联邦学习,得到联邦学习结果,并返回所述联邦学习结果;以及Sending the model parameters of the first recognition model to the server, so that the server performs federated learning according to the model parameters sent by each client, obtains the federated learning result, and returns the federated learning result; and
    接收所述服务端发送的联邦学习结果,根据所述联邦学习结果更新所述第一识别模型,得到对应的第二识别模型,并将所述第二识别模型发送给所述服务端,以供所述服务端根据各客户端发送的第二识别模型和预设的遗传算法得到目标识别模型。Receive the federated learning result sent by the server, update the first recognition model according to the federated learning result to obtain the corresponding second recognition model, and send the second recognition model to the server for The server obtains the target recognition model according to the second recognition model sent by each client and the preset genetic algorithm.
  2. 如权利要求1所述的识别模型的训练方法,其中,所述获取本地的训练数据包括:The method for training a recognition model according to claim 1, wherein said obtaining local training data comprises:
    通过预先设置好的定时任务触发获取指令,根据获取指令获取本地的训练数据;或Trigger the acquisition instruction through a pre-set timed task, and acquire local training data according to the acquisition instruction; or
    在侦测到客户端对应用户触发的获取指令时,根据获取指令获取本地的训练数据。When detecting that the client corresponds to the acquisition instruction triggered by the user, the local training data is acquired according to the acquisition instruction.
  3. 如权利要求1所述的识别模型的训练方法,其中,所述获取本地的训练数据,根据所述训练数据训练得到第一识别模型的步骤包括:The method for training a recognition model according to claim 1, wherein the step of obtaining local training data and training according to the training data to obtain the first recognition model comprises:
    获取本地的训练数据,基于迁移学习算法,根据所述训练数据训练得到第一识别模型。Obtain local training data, and train the first recognition model based on the training data based on the migration learning algorithm.
  4. 如权利要求1至3任一项所述的识别模型的训练方法,其中,所述接收所述服务端发送的联邦学习结果,根据所述联邦学习结果更新所述第一识别模型,得到对应的第二识别模型的步骤包括:The method for training a recognition model according to any one of claims 1 to 3, wherein the receiving the federated learning result sent by the server, and updating the first recognition model according to the federated learning result to obtain the corresponding The steps of the second recognition model include:
    接收所述服务端发送的联邦学习结果,获取预设学习率和所述联邦学习结果中的模型参数改变量;以及Receiving the federated learning result sent by the server, and obtaining the preset learning rate and the model parameter change amount in the federated learning result; and
    根据所述预设学习率和所述模型参数改变量计算得到更新后的模型参数,根据更新后的模型参数更新所述第一识别模型,得到对应的第二识别模型。The updated model parameters are calculated according to the preset learning rate and the change amount of the model parameters, and the first recognition model is updated according to the updated model parameters to obtain the corresponding second recognition model.
  5. 如权利要求4所述的识别模型的训练方法,其中,所述根据所述预设学习率和所述模型参数改变量计算得到更新后的模型参数,根据更新后的模型参数更新所述第一识别模型,得到对应的第二识别模型包括:The method for training a recognition model according to claim 4, wherein the updated model parameters are calculated according to the preset learning rate and the amount of change in the model parameters, and the first model parameters are updated according to the updated model parameters. The recognition model to obtain the corresponding second recognition model includes:
    获取所述第一识别模型的模型参数,并计算所述预设学习率与模型参数改变量之间的乘积,将所述第一识别模型的模型参数减去计算所得的乘积,得到更新后的模型参数,并根据更新后的模型参数更新第一识别模型,得到第二识别模型。Obtain the model parameters of the first recognition model, calculate the product of the preset learning rate and the amount of model parameter change, and subtract the calculated product from the model parameters of the first recognition model to obtain the updated Model parameters, and update the first recognition model according to the updated model parameters to obtain the second recognition model.
  6. 一种识别模型的训练方法,其中,所述识别模型的训练方法应用于服务端,所述识别模型的训练方法包括以下步骤:A training method for a recognition model, wherein the training method for the recognition model is applied to a server, and the training method for the recognition model includes the following steps:
    接收各客户端发送的第一识别模型对应的模型参数,根据所述模型参数进行联邦学习,得到联邦学习结果;Receiving model parameters corresponding to the first recognition model sent by each client, and performing federated learning according to the model parameters to obtain a federated learning result;
    将所述联邦学习结果发送给各客户端,以供各客户端根据所述联邦学习结果更新对应的第一识别模型,得到对应的第二识别模型;以及Sending the federated learning result to each client, so that each client can update the corresponding first recognition model according to the federated learning result to obtain the corresponding second recognition model; and
    接收各客户端发送的第二识别模型,根据各第二识别模型和预设的遗传算法得到目标识别模型。The second recognition model sent by each client is received, and the target recognition model is obtained according to each second recognition model and a preset genetic algorithm.
  7. 如权利要求6所述的识别模型的训练方法,其中,所述联邦学习结果为模型参数改变量,所述模型参数改变量为在进行联邦学习过程中,对应损失函数收敛时,根据损失函数计算得到的梯度值。The method for training a recognition model according to claim 6, wherein the federated learning result is a model parameter change amount, and the model parameter change amount is calculated according to the loss function when the corresponding loss function converges during the federated learning process The obtained gradient value.
  8. 如权利要求6所述的识别模型的训练方法,其中,所述目标识别模型为所需的最优识别模型。8. The method for training a recognition model according to claim 6, wherein the target recognition model is a desired optimal recognition model.
  9. 如权利要求6所述的识别模型的训练方法,其中,所述接收各客户端发送的第二识别模型,根据各第二识别模型和预设的遗传算法得到目标识别模型的步骤之后,还包括:The method for training a recognition model according to claim 6, wherein after the step of receiving the second recognition model sent by each client and obtaining the target recognition model according to each second recognition model and a preset genetic algorithm, the method further comprises :
    将所述目标识别模型发送给各客户端,以供各客户端在接收到识别请求后,根据目标识别模型进行相应的识别操作。The target recognition model is sent to each client, so that each client can perform a corresponding recognition operation according to the target recognition model after receiving the recognition request.
  10. 如权利要求6所述的识别模型的训练方法,其中,所述接收各客户端发送的第二识别模型,根据各第二识别模型和预设的遗传算法得到目标识别模型的步骤包括:7. The method for training a recognition model according to claim 6, wherein the step of receiving the second recognition model sent by each client, and obtaining the target recognition model according to each second recognition model and a preset genetic algorithm comprises:
    接收各客户端发送的第二识别模型,在所接收的各第二识别模型中选择父代识别模型,并在预设的遗传算法中选择目标算子;Receiving the second recognition model sent by each client, selecting the parent recognition model from each of the received second recognition models, and selecting the target operator in the preset genetic algorithm;
    通过所述父代识别模型和所述目标算子得到所述父代识别模型对应的子代识别模型;以及Obtain a child recognition model corresponding to the parent recognition model through the parent recognition model and the target operator; and
    若检测到所述子代识别模型满足结束条件,则将满足结束条件中识别准确率最高的子代识别模型确定为目标识别模型。If it is detected that the offspring recognition model meets the end condition, the offspring recognition model with the highest recognition accuracy rate among the end conditions is determined as the target recognition model.
  11. 如权利要求10所述的识别模型的训练方法,其中,所述通过所述父代识别模型和所述目标算子得到所述父代识别模型对应的子代识别模型的步骤包括:10. The method for training a recognition model according to claim 10, wherein the step of obtaining a child recognition model corresponding to the parent recognition model through the parent recognition model and the target operator comprises:
    获取所述父代识别模型对应的待处理模型参数,将所述待处理模型参数确定为参数编码;Acquiring the to-be-processed model parameters corresponding to the parent recognition model, and determining the to-be-processed model parameters as parameter codes;
    通过所述目标算子对所述参数编码进行处理,得到新的参数编码;以及Processing the parameter encoding through the target operator to obtain a new parameter encoding; and
    根据所述新的参数编码对应更新所述父代识别模型,得到子代识别模型。Correspondingly update the parent recognition model according to the new parameter encoding to obtain the offspring recognition model.
  12. 如权利要求11所述的识别模型的训练方法,其中,所述在所接收的各第二识别模型中选择父代识别模型的步骤包括:11. The method for training a recognition model according to claim 11, wherein the step of selecting a parent recognition model among the received second recognition models comprises:
    根据服务端中当前的可利用资源来确定所选择父代识别模型的数量。The number of selected parent recognition models is determined according to the current available resources in the server.
  13. 如权利要求10所述的识别模型的训练方法,其中,所述检测到所述子代识别模型满足结束条件的步骤包括:10. The method for training a recognition model according to claim 10, wherein the step of detecting that the offspring recognition model satisfies an end condition comprises:
    当所有所述子代识别模型都满足结束条件时,确定所述子代识别模型满足结束条件。When all the offspring recognition models meet the end condition, it is determined that the offspring recognition model meets the end condition.
  14. 如权利要求10至13任一项所述的识别模型的训练方法,其中,所述若检测到所述子代识别模型满足结束条件,则将满足结束条件中识别准确率最高的子代识别模型确定为目标识别模型的步骤包括:The method for training a recognition model according to any one of claims 10 to 13, wherein, if it is detected that the offspring recognition model satisfies the end condition, the offspring recognition model with the highest recognition accuracy rate among the end conditions will be satisfied The steps to determine the target recognition model include:
    若检测到所述子代识别模型对应的迭代次数大于预设次数,则确定所述子代识别模型满足结束条件,并采用预存测试数据获取满足结束条件的各子代识别模型的识别准确率;以及If it is detected that the number of iterations corresponding to the offspring recognition model is greater than the preset number, it is determined that the offspring recognition model meets the end condition, and the pre-stored test data is used to obtain the recognition accuracy of each offspring recognition model that meets the end condition; as well as
    选择识别准确率最高的子代识别模型确定为目标识别模型。The offspring recognition model with the highest recognition accuracy is selected as the target recognition model.
  15. 如权利要求10至13任一项所述的识别模型的训练方法,其中,所述若检测到所述子代识别模型满足结束条件,则将满足结束条件中识别准确率最高的子代识别模型确定为目标识别模型的步骤包括:The method for training a recognition model according to any one of claims 10 to 13, wherein, if it is detected that the offspring recognition model satisfies the end condition, the offspring recognition model with the highest recognition accuracy rate among the end conditions will be satisfied The steps to determine the target recognition model include:
    若检测到所述子代识别模型对应的识别准确率大于或者等于预设准确率,则确定所述子代识别模型满足结束条件;以及If it is detected that the recognition accuracy rate corresponding to the offspring recognition model is greater than or equal to the preset accuracy rate, it is determined that the offspring recognition model meets the end condition; and
    在所述子代识别模型中选择识别准确率最大的子代识别模型确定为目标识别模型。Among the offspring recognition models, the offspring recognition model with the highest recognition accuracy is selected and determined as the target recognition model.
  16. 如权利要求10至13任一项所述的识别模型的训练方法,其中,所述若检测到所述子代识别模型满足结束条件,则将满足结束条件中识别准确率最高的子代识别模型确定为目标识别模型的步骤包括:The method for training a recognition model according to any one of claims 10 to 13, wherein, if it is detected that the offspring recognition model satisfies the end condition, the offspring recognition model with the highest recognition accuracy rate among the end conditions will be satisfied The steps to determine the target recognition model include:
    若检测到识别准确率大于或者等于预设准确率的子代识别模型占当前迭代过程中所有子代识别模型的比例大于特定比例时,确定子代识别模型满足结束条件;以及If it is detected that the proportion of the child recognition model with a recognition accuracy greater than or equal to the preset accuracy to all child recognition models in the current iteration process is greater than a certain ratio, it is determined that the child recognition model meets the end condition; and
    在所述子代识别模型中选择识别准确率最大的子代识别模型确定为目标识别模型。Among the offspring recognition models, the offspring recognition model with the highest recognition accuracy is selected and determined as the target recognition model.
  17. 一种识别模型的训练装置,其中,所述识别模型的训练装置应用于客户端中,所述识别模型的训练装置包括:A training device for a recognition model, wherein the training device for the recognition model is applied to a client, and the training device for the recognition model includes:
    获取模块,用于获取本地的训练数据;The acquisition module is used to acquire local training data;
    训练模块,用于根据所述训练数据训练得到第一识别模型;The training module is used to train according to the training data to obtain the first recognition model;
    发送模块,用于将所述第一识别模型的模型参数发送给服务端,以供所述服务端根据各客户端发送的模型参数进行联邦学习,得到联邦学习结果,并返回所述联邦学习结果;The sending module is configured to send the model parameters of the first recognition model to the server, so that the server performs federated learning according to the model parameters sent by each client, obtains the federated learning result, and returns the federated learning result ;
    接收模块,用于接收所述服务端发送的联邦学习结果;以及The receiving module is used to receive the federated learning result sent by the server; and
    更新模块,用于根据所述联邦学习结果更新所述第一识别模型,得到对应的第二识别模型;An update module, configured to update the first recognition model according to the federated learning result to obtain a corresponding second recognition model;
    所述发送模块还用于将所述第二识别模型发送给所述服务端,以供所述服务端根据各客户端发送的第二识别模型和预设的遗传算法得到目标识别模型。The sending module is further configured to send the second recognition model to the server, so that the server obtains the target recognition model according to the second recognition model sent by each client and a preset genetic algorithm.
  18. 一种识别模型的训练装置,其中,所述识别模型的训练装置应用于服务端,所述识别模型的训练装置包括:A training device for a recognition model, wherein the training device for the recognition model is applied to a server, and the training device for the recognition model includes:
    接收模块,用于接收各客户端发送的第一识别模型对应的模型参数;The receiving module is used to receive the model parameters corresponding to the first recognition model sent by each client;
    联邦学习模块,用于根据所述模型参数进行联邦学习,得到联邦学习结果;以及The federated learning module is used to perform federated learning according to the model parameters to obtain federated learning results; and
    发送模块,用于将所述联邦学习结果发送给各客户端,以供各客户端根据所述联邦学习结果更新对应的第一识别模型,得到对应的第二识别模型;A sending module, configured to send the federated learning result to each client, so that each client can update the corresponding first recognition model according to the federated learning result to obtain the corresponding second recognition model;
    所述接收模块还用于接收各客户端发送的第二识别模型,根据各第二识别模型和预设的遗传算法得到目标识别模型。The receiving module is also used to receive the second recognition model sent by each client, and obtain the target recognition model according to each second recognition model and a preset genetic algorithm.
  19. 一种识别模型的训练设备,其中,所述识别模型的训练设备包括存储器、处理器和存储在所述存储器上并可在所述处理器上运行的识别模型的训练程序,所述识别模型的训练程序被所述处理器执行时实现如权利要求1至16任一项中所述的识别模型的训练方法的步骤。A training device for a recognition model, wherein the training device for the recognition model includes a memory, a processor, and a training program for a recognition model that is stored on the memory and can run on the processor. When the training program is executed by the processor, the steps of the training method of the recognition model as described in any one of claims 1 to 16 are realized.
  20. 一种计算机可读存储介质,其中,所述计算机可读存储介质上存储有识别模型的训练程序,所述识别模型的训练程序被处理器执行时实现如权利要求1至16任一项所述的识别模型的训练方法的步骤。A computer-readable storage medium, wherein a training program for a recognition model is stored on the computer-readable storage medium, and when the training program for the recognition model is executed by a processor, the training program of any one of claims 1 to 16 is realized The steps of the training method of the recognition model.
PCT/CN2020/134029 2020-03-20 2020-12-04 Method and apparatus for training recognition model, device, and readable storage medium WO2021184836A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202010206241.7A CN111428881B (en) 2020-03-20 2020-03-20 Recognition model training method, device, equipment and readable storage medium
CN202010206241.7 2020-03-20

Publications (1)

Publication Number Publication Date
WO2021184836A1 true WO2021184836A1 (en) 2021-09-23

Family

ID=71548570

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2020/134029 WO2021184836A1 (en) 2020-03-20 2020-12-04 Method and apparatus for training recognition model, device, and readable storage medium

Country Status (2)

Country Link
CN (1) CN111428881B (en)
WO (1) WO2021184836A1 (en)

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114003949A (en) * 2021-09-30 2022-02-01 清华大学 Model training method and device based on private data set
CN114168988A (en) * 2021-12-16 2022-03-11 大连理工大学 Method for aggregating federated learning models and electronic device
CN114429223A (en) * 2022-01-26 2022-05-03 上海富数科技有限公司 Heterogeneous model establishing method and device
CN114492849A (en) * 2022-01-24 2022-05-13 光大科技有限公司 Model updating method and device based on federal learning
CN114785605A (en) * 2022-04-28 2022-07-22 中国电信股份有限公司 Method, device and equipment for determining network anomaly detection model and storage medium
CN114785810A (en) * 2022-03-31 2022-07-22 海南师范大学 Tree broadcast data synchronization algorithm suitable for federal learning
CN115002031A (en) * 2022-05-11 2022-09-02 大连理工大学 Unbalanced data distribution-based federated learning network traffic classification model training method, model and classification method
CN115659212A (en) * 2022-09-27 2023-01-31 南京邮电大学 Federal learning efficiency evaluation method based on TDD communication under cross-domain heterogeneous scene
US11588621B2 (en) * 2019-12-06 2023-02-21 International Business Machines Corporation Efficient private vertical federated learning
CN116150753A (en) * 2022-12-21 2023-05-23 上海交通大学 Mobile end malicious software detection system based on federal learning
CN116361663A (en) * 2023-06-01 2023-06-30 湖南工商大学 Digital evidence classification model training method, classification method, device and related equipment
WO2023169167A1 (en) * 2022-03-09 2023-09-14 华为技术有限公司 Model training method and apparatus, and device and storage medium
WO2023185485A1 (en) * 2022-04-02 2023-10-05 西安电子科技大学 Model training method and face recognition method based on adaptive split learning-federated learning
WO2024060906A1 (en) * 2022-09-20 2024-03-28 腾讯科技(深圳)有限公司 Data processing method and apparatus for federated learning system, computer, and readable storage medium

Families Citing this family (29)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111428881B (en) * 2020-03-20 2021-12-07 深圳前海微众银行股份有限公司 Recognition model training method, device, equipment and readable storage medium
CN112102939B (en) * 2020-07-24 2023-08-04 西安电子科技大学 Cardiovascular and cerebrovascular disease reference information prediction system, method and device and electronic equipment
CN112085159B (en) * 2020-07-24 2023-08-15 西安电子科技大学 User tag data prediction system, method and device and electronic equipment
CN111839495B (en) * 2020-07-30 2023-04-07 深圳前海微众银行股份有限公司 Detection method, device and storage medium
CN112052959B (en) * 2020-09-04 2023-08-25 深圳前海微众银行股份有限公司 Automatic driving training method, equipment and medium based on federal learning
CN114257386B (en) * 2020-09-10 2023-03-21 华为技术有限公司 Training method, system, equipment and storage medium for detection model
CN112201342B (en) * 2020-09-27 2024-04-26 博雅正链(北京)科技有限公司 Medical auxiliary diagnosis method, device, equipment and storage medium based on federal learning
CN112164224A (en) * 2020-09-29 2021-01-01 杭州锘崴信息科技有限公司 Traffic information processing system, method, device and storage medium for information security
CN112232518B (en) * 2020-10-15 2024-01-09 成都数融科技有限公司 Lightweight distributed federal learning system and method
CN112256874B (en) * 2020-10-21 2023-08-08 平安科技(深圳)有限公司 Model training method, text classification method, device, computer equipment and medium
CN112447298A (en) * 2020-11-24 2021-03-05 平安科技(深圳)有限公司 Neurodegenerative disease modeling device based on federal transfer learning and related equipment
CN113807157A (en) * 2020-11-27 2021-12-17 京东科技控股股份有限公司 Method, device and system for training neural network model based on federal learning
CN112465786A (en) * 2020-12-01 2021-03-09 平安科技(深圳)有限公司 Model training method, data processing method, device, client and storage medium
CN112611080A (en) * 2020-12-10 2021-04-06 浙江大学 Intelligent air conditioner control system and method based on federal learning
CN113762862A (en) * 2020-12-28 2021-12-07 京东城市(北京)数字科技有限公司 Cargo abnormity monitoring method and device, electronic equipment and storage medium
CN113822311B (en) * 2020-12-31 2023-09-01 京东科技控股股份有限公司 Training method and device of federal learning model and electronic equipment
CN113807380B (en) * 2020-12-31 2023-09-01 京东科技信息技术有限公司 Training method and device of federal learning model and electronic equipment
CN112990476B (en) * 2021-02-08 2022-04-12 支付宝(杭州)信息技术有限公司 Method, system and device for updating model parameters based on federal learning
WO2022193173A1 (en) * 2021-03-17 2022-09-22 深圳技术大学 Blockchain-based financial data information federated transfer learning system and method
CN113052323B (en) * 2021-03-23 2023-05-23 支付宝(杭州)信息技术有限公司 Model training method and device based on federal learning and electronic equipment
CN113158842A (en) * 2021-03-31 2021-07-23 中国工商银行股份有限公司 Identification method, system, device and medium
CN113259363B (en) * 2021-05-26 2022-09-02 中国人民解放军战略支援部队信息工程大学 Covert communication method and device
CN113505894B (en) * 2021-06-02 2023-12-15 北京航空航天大学 Longitudinal federal learning linear regression and logistic regression model training method and device
CN113673476B (en) * 2021-09-02 2023-11-07 京东科技控股股份有限公司 Face recognition model training method and device, storage medium and electronic equipment
CN113919508B (en) * 2021-10-15 2022-07-19 河南工业大学 Mobile server-based federal learning system and method
CN114118882B (en) * 2022-01-27 2022-05-27 太平金融科技服务(上海)有限公司 Service data processing method, device, equipment and medium based on combined model
CN114708109B (en) * 2022-03-01 2022-11-11 上海钐昆网络科技有限公司 Risk recognition model training method, device, equipment and storage medium
CN114638998A (en) * 2022-03-07 2022-06-17 支付宝(杭州)信息技术有限公司 Model updating method, device, system and equipment
CN115860116A (en) * 2022-12-02 2023-03-28 广州图灵科技有限公司 Federal learning method based on generative model and deep transfer learning

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105336323A (en) * 2015-10-14 2016-02-17 清华大学 Uighur language voice recognition method and device
CN109165725A (en) * 2018-08-10 2019-01-08 深圳前海微众银行股份有限公司 Neural network federation modeling method, equipment and storage medium based on transfer learning
CN110162939A (en) * 2018-10-25 2019-08-23 腾讯科技(深圳)有限公司 Man-machine recognition methods, equipment and medium
CN110766169A (en) * 2019-10-31 2020-02-07 深圳前海微众银行股份有限公司 Transfer training optimization method and device for reinforcement learning, terminal and storage medium
CN110874638A (en) * 2020-01-19 2020-03-10 同盾控股有限公司 Behavior analysis-oriented meta-knowledge federation method, device, electronic equipment and system
CN111428881A (en) * 2020-03-20 2020-07-17 深圳前海微众银行股份有限公司 Recognition model training method, device, equipment and readable storage medium

Family Cites Families (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11257259B2 (en) * 2017-08-15 2022-02-22 Siemens Healthcare Gmbh Topogram prediction from surface data in medical imaging
US11861461B2 (en) * 2018-08-09 2024-01-02 Autodesk, Inc. Techniques for generating designs that reflect stylistic preferences
CN109698822A (en) * 2018-11-28 2019-04-30 众安信息技术服务有限公司 Combination learning method and system based on publicly-owned block chain and encryption neural network
CN109783959B (en) * 2019-01-23 2023-10-10 保定市博宏高科控制技术有限公司 Valve-based combined multi-model simulation platform and method
CN109977694A (en) * 2019-03-11 2019-07-05 暨南大学 A kind of data sharing method based on cooperation deep learning
CN110163093B (en) * 2019-04-15 2021-03-05 浙江工业大学 Guideboard recognition confrontation defense method based on genetic algorithm
CN110266771B (en) * 2019-05-30 2022-11-22 王静逸 Distributed intelligent node and distributed group intelligent system deployment method
CN110443063B (en) * 2019-06-26 2023-03-28 电子科技大学 Adaptive privacy-protecting federal deep learning method
CN110874648A (en) * 2020-01-16 2020-03-10 支付宝(杭州)信息技术有限公司 Federal model training method and system and electronic equipment

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105336323A (en) * 2015-10-14 2016-02-17 清华大学 Uighur language voice recognition method and device
CN109165725A (en) * 2018-08-10 2019-01-08 深圳前海微众银行股份有限公司 Neural network federation modeling method, equipment and storage medium based on transfer learning
CN110162939A (en) * 2018-10-25 2019-08-23 腾讯科技(深圳)有限公司 Man-machine recognition methods, equipment and medium
CN110766169A (en) * 2019-10-31 2020-02-07 深圳前海微众银行股份有限公司 Transfer training optimization method and device for reinforcement learning, terminal and storage medium
CN110874638A (en) * 2020-01-19 2020-03-10 同盾控股有限公司 Behavior analysis-oriented meta-knowledge federation method, device, electronic equipment and system
CN111428881A (en) * 2020-03-20 2020-07-17 深圳前海微众银行股份有限公司 Recognition model training method, device, equipment and readable storage medium

Cited By (23)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11588621B2 (en) * 2019-12-06 2023-02-21 International Business Machines Corporation Efficient private vertical federated learning
CN114003949A (en) * 2021-09-30 2022-02-01 清华大学 Model training method and device based on private data set
CN114003949B (en) * 2021-09-30 2022-08-30 清华大学 Model training method and device based on private data set
CN114168988A (en) * 2021-12-16 2022-03-11 大连理工大学 Method for aggregating federated learning models and electronic device
CN114168988B (en) * 2021-12-16 2024-05-03 大连理工大学 Federal learning model aggregation method and electronic device
CN114492849B (en) * 2022-01-24 2023-09-08 光大科技有限公司 Model updating method and device based on federal learning
CN114492849A (en) * 2022-01-24 2022-05-13 光大科技有限公司 Model updating method and device based on federal learning
CN114429223A (en) * 2022-01-26 2022-05-03 上海富数科技有限公司 Heterogeneous model establishing method and device
CN114429223B (en) * 2022-01-26 2023-11-07 上海富数科技有限公司 Heterogeneous model building method and device
WO2023169167A1 (en) * 2022-03-09 2023-09-14 华为技术有限公司 Model training method and apparatus, and device and storage medium
CN114785810A (en) * 2022-03-31 2022-07-22 海南师范大学 Tree broadcast data synchronization algorithm suitable for federal learning
CN114785810B (en) * 2022-03-31 2023-05-16 海南师范大学 Tree-like broadcast data synchronization method suitable for federal learning
WO2023185485A1 (en) * 2022-04-02 2023-10-05 西安电子科技大学 Model training method and face recognition method based on adaptive split learning-federated learning
CN114785605B (en) * 2022-04-28 2023-12-12 中国电信股份有限公司 Determination method, device, equipment and storage medium of network anomaly detection model
CN114785605A (en) * 2022-04-28 2022-07-22 中国电信股份有限公司 Method, device and equipment for determining network anomaly detection model and storage medium
CN115002031B (en) * 2022-05-11 2023-09-29 大连理工大学 Federal learning network flow classification model training method, model and classification method based on unbalanced data distribution
CN115002031A (en) * 2022-05-11 2022-09-02 大连理工大学 Unbalanced data distribution-based federated learning network traffic classification model training method, model and classification method
WO2024060906A1 (en) * 2022-09-20 2024-03-28 腾讯科技(深圳)有限公司 Data processing method and apparatus for federated learning system, computer, and readable storage medium
CN115659212A (en) * 2022-09-27 2023-01-31 南京邮电大学 Federal learning efficiency evaluation method based on TDD communication under cross-domain heterogeneous scene
CN115659212B (en) * 2022-09-27 2024-04-09 南京邮电大学 Federal learning efficiency evaluation method based on TDD communication under cross-domain heterogeneous scene
CN116150753A (en) * 2022-12-21 2023-05-23 上海交通大学 Mobile end malicious software detection system based on federal learning
CN116361663A (en) * 2023-06-01 2023-06-30 湖南工商大学 Digital evidence classification model training method, classification method, device and related equipment
CN116361663B (en) * 2023-06-01 2023-08-15 湖南工商大学 Digital evidence classification model training method, classification method, device and related equipment

Also Published As

Publication number Publication date
CN111428881A (en) 2020-07-17
CN111428881B (en) 2021-12-07

Similar Documents

Publication Publication Date Title
WO2021184836A1 (en) Method and apparatus for training recognition model, device, and readable storage medium
US11947680B2 (en) Model parameter training method, terminal, and system based on federation learning, and medium
CN111695697B (en) Multiparty joint decision tree construction method, equipment and readable storage medium
WO2021114822A1 (en) Private data protection-based risk decision making method, apparatus and system, and device
JP7095140B2 (en) Multi-model training methods and equipment based on feature extraction, electronic devices and media
CN110633805A (en) Longitudinal federated learning system optimization method, device, equipment and readable storage medium
CN110309587B (en) Decision model construction method, decision method and decision model
WO2022095352A1 (en) Abnormal user identification method and apparatus based on intelligent decision, and computer device
WO2021159798A1 (en) Method for optimizing longitudinal federated learning system, device and readable storage medium
CN103067918B (en) Method, device and system of privacy data anonymization in communication network
CN111914274A (en) Full-process block chain system based on multiple information sources
Li et al. Retracted: Design of multimedia blockchain privacy protection system based on distributed trusted communication
CN111400504A (en) Method and device for identifying enterprise key people
CN113011632B (en) Enterprise risk assessment method, device, equipment and computer readable storage medium
WO2022156594A1 (en) Federated model training method and apparatus, electronic device, computer program product, and computer-readable storage medium
CN105786581A (en) Multistage server and method for network data operation
CN110837653A (en) Label prediction method, device and computer readable storage medium
CN110675263B (en) Risk identification method and device for transaction data
CN109146683A (en) A kind of information management and Verification System using block chain
CN112529102B (en) Feature expansion method, device, medium and computer program product
WO2021169477A1 (en) Cross feature-based model building and prediction methods, devices and apparatuses, and storage medium
US9336408B2 (en) Solution for continuous control and protection of enterprise data based on authorization projection
CN115203365A (en) Social event processing method applied to comprehensive treatment field
Kim A quantitative and knowledge–based approach to choosing security architectural tactics
CN103634326B (en) A kind of method and device for processing application system request message

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20926145

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20926145

Country of ref document: EP

Kind code of ref document: A1