CN115879467A - Federal learning-based Chinese address word segmentation method and device - Google Patents

Federal learning-based Chinese address word segmentation method and device Download PDF

Info

Publication number
CN115879467A
CN115879467A CN202211626861.1A CN202211626861A CN115879467A CN 115879467 A CN115879467 A CN 115879467A CN 202211626861 A CN202211626861 A CN 202211626861A CN 115879467 A CN115879467 A CN 115879467A
Authority
CN
China
Prior art keywords
model
word segmentation
federal
learning
knowledge
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202211626861.1A
Other languages
Chinese (zh)
Other versions
CN115879467B (en
Inventor
李莹
李文龙
金路
鲍迪恩
彭聪
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang Bangsheng Technology Co ltd
Original Assignee
Zhejiang Bangsheng Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang Bangsheng Technology Co ltd filed Critical Zhejiang Bangsheng Technology Co ltd
Priority to CN202211626861.1A priority Critical patent/CN115879467B/en
Publication of CN115879467A publication Critical patent/CN115879467A/en
Application granted granted Critical
Publication of CN115879467B publication Critical patent/CN115879467B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Landscapes

  • Machine Translation (AREA)

Abstract

The invention discloses a Chinese address word segmentation method and a device based on federal learning, wherein a Chinese address word segmentation task is divided into an address word segmentation model and a place name classification model to be trained independently, the address word segmentation model is trained by using local address data, and the place name classification model is used for realizing multi-data-source combined training by using the federal learning technology; during decision making, the two models jointly complete the Chinese address word segmentation task. The invention firstly proposes a dual-model to complete Chinese address word segmentation task together, the address word segmentation model is trained based on local data, and the model is used for segmenting nouns of complete addresses; the place name classification model is trained by using federal learning to learn data distribution of multi-source data, and is used for grading place name nouns segmented by the address word segmentation model. Meanwhile, the federal learning process of the place name classification model is improved, and a knowledge distillation technology is introduced to accelerate training efficiency. And finally, improving the loss function of the knowledge distillation method and enhancing the effect of knowledge distillation.

Description

Federal learning-based Chinese address word segmentation method and device
Technical Field
The invention relates to the technical field of Chinese address word segmentation, in particular to a method and a device for Chinese address word segmentation based on federal learning.
Background
With the development of artificial intelligence technology and distributed computing, the situation of data islanding is increasingly serious. Modern machine learning algorithms rely on large amounts of data, especially when training deep neural models from high dimensional data such as text and images. Most data naturally comes from various organizations, the data are stored in different machines, and the data all relate to private information and do not allow private sharing and dissemination. Therefore, it is necessary to learn a well-behaved machine learning model while protecting user privacy. Federal learning, a technique for protecting data privacy, has become a new machine learning paradigm that allows models to be trained on multiple distributed organizations or servers that have local data samples without exchanging information. Meanwhile, the knowledge distillation technology is used as a model compression method, information is concentrated into a smaller model, and the method is applied to the federal learning, so that the information transmission efficiency can be improved, and the performance of the federal learning can be improved.
In a real-world scene, different banks, enterprises and other organizations hold address data of users, and the banks want to train an address word segmentation model for everyone to use, but cannot require the banks to upload own data to the cloud in a centralized manner due to laws and other reasons. Even within the same business, different departments often maintain address data locally. It has been shown that training of most language models can present legal and ethical issues, as they can reveal the user's address information in an unexpected way.
Disclosure of Invention
The invention aims to provide a Chinese address word segmentation method and device based on federal learning, aiming at overcoming the defects in the prior art, and the method and device jointly finish the training of an address word segmentation model on the premise of ensuring that the privacy of address data is not revealed by using the federal learning technology, and simultaneously ensure the training efficiency and the model precision.
The purpose of the invention is realized by the following technical scheme: in a first aspect, the invention provides a method for segmenting Chinese addresses based on federal learning, which comprises the following steps:
(1) The local address data is used for training an address word segmentation model and is responsible for segmenting an address into a plurality of place nouns, the whole training process is only carried out locally, and model parameter updating does not participate in federal learning parameter transmission and sharing;
(2) The method comprises the following steps of performing federal learning training by using a place name classification model, wherein the place name classification model is trained by using place nouns and place level label data, and the training process is realized on the basis of a transverse federal learning technology; the training of the place name classification model is completed by a local complex model and a federal simple model; each round of model training, wherein the local complex model learns the aggregation parameters of the federated simple model and the knowledge in the local training data, the federated simple model learns the knowledge of the local complex model, and the federated simple model participates in transverse federated learning and is iteratively updated until a preset termination condition is met;
(3) The trained address word segmentation model and the place name classification model jointly complete the Chinese address word segmentation task; and sending a complete Chinese address into the address word segmentation model, segmenting the address word segmentation model into a plurality of Chinese place nouns, and carrying out hierarchical decision on the place nouns by the local complex model.
Further, in the step (2), the training process of the place name classification model includes three steps:
1) The local complex model learns knowledge of the federated simple model and the local data set through knowledge distillation;
2) After learning the knowledge of the local complex model through knowledge distillation, the federated simple model is used as a participant, updated parameters are uploaded to a cooperative party for parameter aggregation, and the federated simple model participates in federated learning training;
3) And returning the aggregation parameters by the collaborators for updating the federal simple model.
Further, the loss function of the knowledge distillation method consists of a mission loss function and a distillation loss function; the mission loss function of the knowledge distillation method is
Figure BDA0004003799390000021
Wherein y is i Represents a data annotation, <' > or>
Figure BDA0004003799390000022
Representing the prediction result of the student model; the distillation loss function of the knowledge distillation method is +>
Figure BDA0004003799390000023
Wherein->
Figure BDA0004003799390000024
Representing the prediction result of the teacher model, wherein T is a hyperparameter used for controlling the smoothness of the distillation result; the integrity loss function of the knowledge distillation method is
Figure BDA0004003799390000025
Figure BDA0004003799390000026
Wherein alpha is the learning emphasis of the hyper-parameter for controlling knowledge distillation, the fitting to the teacher model is more emphasized when alpha is larger, and the fitting to the data is emphasized when alpha is smaller.
Further, when local data knowledge is transmitted to a federal simple model by knowledge distillation aiming at the local complex model, the knowledge distillation method takes the local complex model as a teacher model and the federal simple model as a student model; and when the federal learning knowledge is transferred to the local complex model by knowledge distillation aiming at the federal simple model, the federal simple model is used as a teacher model, and the local complex model is used as a student model.
Further, in the step (1), the address word segmentation model adopts a model with information memory capacity to complete a location segmentation task, the selected model is a recurrent neural network model, a Bert model or a graph neural network, and model training is performed by using only a locally stored address segmentation labeling data set until a preset termination condition is met.
Further, in the step (2), the place name classification model uses a federal simple model to participate in federal learning, and the structure of the federal simple model of all participants in the federal learning is required to be consistent; for the local complex model to learn local data knowledge, each participant is allowed to participate in training using a different model.
Further, the federal simple model is a bidirectional cyclic neural network model, a convolutional neural network model or a hidden Markov model within 10 layers;
further, the local complex model is a Bert model or a bidirectional recurrent neural network model with a depth of more than 20 layers.
In a second aspect, the invention provides a federally-learning-based Chinese address word segmentation device, which comprises a memory and one or more processors, wherein executable codes are stored in the memory, and when the processors execute the executable codes, the federally-learning-based Chinese address word segmentation method is realized.
In a third aspect, the present invention provides a computer-readable storage medium, on which a program is stored, where the program, when executed by a processor, implements the method for chinese address word segmentation based on federal learning.
The invention has the beneficial effects that: on one hand, the traditional Chinese address word segmentation task is divided into an address word segmentation model for site word segmentation and a place name classification model for site classification, and the address word segmentation accuracy can be better improved through the model after work division. Meanwhile, in order to solve the problem that location nouns are distributed in equipment of each organization and cannot share and enhance model precision, the invention provides a federated learning method to carry out joint training on the location nouns, and address grading knowledge of each organization is extracted on the premise of protecting address data from being leaked. On the other hand, in order to ensure the performance of the federal learning and prevent the transmission bottleneck from being reached because too many updated parameters need to be transmitted, the invention provides a knowledge distillation method for refining the transmitted knowledge, and compared with the traditional federal learning method, the method has the advantages of lower transmission cost and better performance.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below, it is obvious that the drawings in the following description are only some embodiments of the present invention, and for those skilled in the art, other drawings can be obtained according to these drawings without creative efforts.
FIG. 1 is a flow chart of a method for segmenting words of Chinese addresses based on federal learning provided by the invention;
FIG. 2 is a flowchart of a place name classification model training in a Federal learning-based Chinese address word segmentation method according to the present invention;
FIG. 3 is a view of the federal learning architecture of a place name classification model in a Chinese address word segmentation method based on federal learning provided by the present invention;
FIG. 4 is a flow chart of an address word segmentation task in a Federal learning-based Chinese address word segmentation method according to the present invention;
fig. 5 is a structural diagram of a chinese address word segmentation apparatus based on federal learning according to the present invention.
Detailed Description
The invention is described in further detail below with reference to the figures and specific examples.
As shown in fig. 1, the method for segmenting the words of the chinese address based on the federal learning provided by the present invention comprises three steps:
(1) Firstly, local address data is used for training an address word segmentation model, the address word segmentation model is responsible for segmenting an address into a plurality of place nouns, the whole training process is only carried out locally, and model parameters are updated without participating in federal learning parameter transmission and sharing; the training data set is a complete address, and the model parameters are updated by using local data until the model converges, for example, "ancient border street text garden cells" in the western lake region of Hangzhou, zhejiang, data labels are used for specially marking the beginning and the end of each noun in the address, so that the model can identify the boundary of the place name noun. The address word segmentation model completes a place segmentation task by adopting a model with information memory capacity, the address word segmentation model can select a bidirectional cyclic neural network, a conditional random field, a Bert, a graph neural network and the like, the training task can be designed into a normal sequence-to-sequence training task, and model training is only carried out by using a locally stored address segmentation labeling data set until a preset termination condition is met.
(2) Secondly, performing federal learning training by using a place name classification model, wherein the place name classification model is trained by using place nouns and place level label data, and the training process is realized on the basis of a transverse federal learning technology; and the training of the place name classification model is completed by a local complex model and a federal simple model together. Each round of model training, wherein the local complex model learns the aggregation parameters of the federated simple model and the knowledge in the local training set, the federated simple model learns the knowledge of the local complex model, and then the federated simple model participates in the transverse federated learning and is iteratively updated until a preset termination condition is met; the specific process is as follows: 1) The local complex model learns knowledge of the federated simple model and the local data set through knowledge distillation; 2) After learning the knowledge of the local complex model through knowledge distillation, the federated simple model is used as a participant, the updated parameters are uploaded to a cooperative party for parameter aggregation, and the federated simple model participates in federated learning training; 3) And returning the aggregation parameters by the collaborators for updating the federal simple model. The place name classification model combines a plurality of data sources to learn classification of place name nouns, and address levels of the place name nouns are identified, wherein the address levels comprise province, city, district, county, street and the like.
(3) Finally, the trained address word segmentation model and the place name classification model jointly complete the Chinese address word segmentation task; and sending a complete Chinese address into the address word segmentation model, segmenting the address word segmentation model into a plurality of Chinese place nouns, and carrying out grading decision on the place nouns by the local complex model to judge the grades of provinces, cities, districts, streets and the like.
Referring to fig. 2, describing the training process of the place name classification model in step (2), first, the local complex model is trained using local address noun data, where the local address noun data includes "hang state city", "south kyo university", "beijing city", "high new district", "qinghua university", and the like. The training model is a local complex model, a model with more model parameters and more complex structure is used, and the model can select Bert, a bidirectional cyclic neural network with the depth of more than 20 layers and the like. The training process of the local complex model comprises two parts: 1. learning knowledge of the local data; 2. knowledge distillation is used to learn the knowledge of the federal simple model. When the federal learning knowledge is transferred to the local complex model through knowledge distillation aiming at the federal simple model, the federal simple model is used as a teacher model, and the local complex model is used as a student model. The integrity loss function is as follows:
Figure BDA0004003799390000041
Figure BDA0004003799390000042
Figure BDA0004003799390000043
wherein alpha is a hyper-parameter used for controlling the smoothness degree of the teacher model and the student model; y is i The actual data annotation is represented by a label,
Figure BDA0004003799390000044
representing the prediction results of the student model, here using the local complex model prediction results as the output of the student model,
Figure BDA0004003799390000045
representing the prediction result of the teacher model, wherein the prediction result of the federal simple model is used as the output of the teacher model, and T is superThe parameters are used to control the smoothness of the distillation result. And when the training of the local complex model reaches a preset stopping condition, stopping the training and entering the next process, wherein the preset stopping condition can be that 10 epochs are trained or a limit value of the convergence rate of the model is reached. And the next process is to train the simple federated model by knowledge distillation, and transmit the knowledge of the local complex model to the simple federated model, wherein the simple federated model adopts a model with short-term memory and a small number of model parameters, and the selected model can be a bidirectional recurrent neural network model, a convolutional neural network model, a hidden Markov model and the like within 10 layers. The knowledge distillation method is characterized in that local data knowledge is transmitted to a federal simple model by knowledge distillation aiming at a local complex model, the local complex model is used as a teacher model, and the federal simple model is used as a student model; the integrity loss function is as follows:
Figure BDA0004003799390000046
Figure BDA0004003799390000047
Figure BDA0004003799390000048
wherein, alpha is a hyper-parameter used for controlling the smoothness degree of the teacher model and the student model, and y is i The label of the real data is represented,
Figure BDA0004003799390000049
representing the prediction results of the student model, here using the federal simple model prediction results as the output of the student model,
Figure BDA00040037993900000410
and (4) representing the prediction result of the teacher model, wherein the prediction result of the local complex model is used as the output of the teacher model, and T is a hyperparameter used for controlling the smoothness of the distillation result. Of simple federal modelsAnd stopping training when the training reaches a preset stopping condition, wherein the preset stopping condition can be that 10 epochs are trained or a limit value of the convergence rate of the model is reached. The difference between the federal simple model training and the local complex model training is that the setting of the hyper-parameters alpha and T is different, the local complex model training tends to learn the knowledge of local data, and the federal simple model training focuses on learning the knowledge of the local complex model.
Furthermore, a federal simple model is used for participating in the horizontal federal learning training, the parameters of the federal simple model are updated, and horizontal federal learning strategies such as FedAvg and FedSgd can be selected for federal learning. And carrying out iterative updating in such a way until a preset termination condition is met to obtain the trained place name classification model.
As shown in fig. 3, a block diagram of a federal learning framework is presented in which the participants represent a separate entity that can provide knowledge of local data. Each participant can use a self-defined local complex model as required, federal learning does not make uniform requirements on the local complex model, and the model structure is required to be consistent only for the federal simple model. Taking the participant 1 as an example, the local complex model of the participant 1 only participates in local training and transmits knowledge to the federal simple model, the federal simple model encrypts updated parameters and uploads the updated parameters to the collaborator, and the collaborator performs parameter aggregation by using a weighted average equal algorithm and distributes the aggregated parameters to each participant. The encryption algorithm may be a homomorphic encryption algorithm, a secure aggregation algorithm, or the like.
As shown in fig. 4, the procedure of performing the chinese address word segmentation task by the federal learning-based chinese address word segmentation method is to input a complete address, for example, "ancient dangling street and flatstem cell in west lake region of hangzhou city, zhejiang province"; then, address noun word segmentation is carried out by using an address word segmentation model to obtain Zhejiang province, hangzhou city, western lake region, ancient street and flatstem milkvetch cell; then, using a place name classification model to classify the address nouns in grades, for example, classifying Zhejiang province to obtain the level of the address nouns belonging to province, and classifying West lake areas to obtain the level of the address nouns belonging to areas; and finally, outputting a Chinese address word segmentation result, wherein Zhejiang province is classified into a province level, hangzhou city is classified into a city level, west lake region is classified into a region level, gudang streets are classified into a street level, and aster cells are classified into residential regions.
Corresponding to the embodiment of the Chinese address word segmentation method based on the federal learning, the invention also provides an embodiment of a Chinese address word segmentation device based on the federal learning.
Referring to fig. 5, the device for segmenting a chinese address based on federal learning according to an embodiment of the present invention includes a memory and one or more processors, where the memory stores executable codes, and the processors execute the executable codes to implement the method for segmenting a chinese address based on federal learning in the foregoing embodiments.
The embodiment of the Chinese address word segmentation device based on federal learning can be applied to any equipment with data processing capability, such as computers and other equipment or devices. The device embodiments may be implemented by software, or by hardware, or by a combination of hardware and software. The software implementation is taken as an example, and as a logical device, the device is formed by reading corresponding computer program instructions in the nonvolatile memory into the memory for running through the processor of any device with data processing capability. In terms of hardware, as shown in fig. 5, a hardware structure diagram of any device with data processing capability where the federally learned chinese address word segmentation apparatus is located according to the present invention is shown, and in addition to the processor, the memory, the network interface, and the nonvolatile memory shown in fig. 5, any device with data processing capability where the apparatus is located in the embodiment may generally include other hardware according to the actual function of the any device with data processing capability, which is not described again.
The implementation process of the functions and actions of each unit in the above device is specifically described in the implementation process of the corresponding step in the above method, and is not described herein again.
For the device embodiments, since they substantially correspond to the method embodiments, reference may be made to the partial description of the method embodiments for relevant points. The above-described embodiments of the apparatus are merely illustrative, and the units described as separate parts may or may not be physically separate, and parts displayed as units may or may not be physical units, may be located in one position, or may be distributed on multiple network units. Some or all of the modules can be selected according to actual needs to achieve the purpose of the scheme of the invention. One of ordinary skill in the art can understand and implement it without inventive effort.
An embodiment of the present invention further provides a computer-readable storage medium, on which a program is stored, and when the program is executed by a processor, the method for segmenting the chinese address based on the federal learning in the foregoing embodiments is implemented.
The computer readable storage medium may be an internal storage unit, such as a hard disk or a memory, of any data processing capability device described in any of the foregoing embodiments. The computer readable storage medium may also be any external storage device of a device with data processing capabilities, such as a plug-in hard disk, a Smart Media Card (SMC), an SD Card, a Flash memory Card (Flash Card), etc. provided on the device. Further, the computer readable storage medium may include both an internal storage unit and an external storage device of any data processing capable device. The computer-readable storage medium is used for storing the computer program and other programs and data required by the arbitrary data processing-capable device, and may also be used for temporarily storing data that has been output or is to be output.
The above-described embodiments are intended to illustrate rather than to limit the invention, and any modifications and variations of the present invention are within the spirit of the invention and the scope of the appended claims.

Claims (10)

1. A Chinese address word segmentation method based on federal learning is characterized by comprising the following steps:
(1) The local address data is used for training an address word segmentation model and is responsible for segmenting an address into a plurality of place nouns, the whole training process is only carried out locally, and model parameter updating does not participate in federal learning parameter transmission and sharing;
(2) The method comprises the following steps of performing federal learning training by using a place name classification model, wherein the place name classification model is trained by using place nouns and place level label data, and the training process is realized on the basis of a transverse federal learning technology; the training of the place name classification model is completed by a local complex model and a federal simple model; each round of model training, wherein the local complex model learns the aggregation parameters of the federated simple model and the knowledge in the local training data, the federated simple model learns the knowledge of the local complex model, and the federated simple model participates in transverse federated learning and is iteratively updated until a preset termination condition is met;
(3) The trained address word segmentation model and the place name classification model jointly complete the Chinese address word segmentation task; and sending a complete Chinese address into the address word segmentation model, segmenting the address word segmentation model into a plurality of Chinese place nouns, and carrying out hierarchical decision on the place nouns by the local complex model.
2. The method for Chinese address segmentation based on federal learning as claimed in claim 1, wherein in the step (2), the training process of the place name classification model comprises three steps:
1) The local complex model learns knowledge of the federated simple model and the local data set through knowledge distillation;
2) After learning the knowledge of the local complex model through knowledge distillation, the federated simple model is used as a participant, the updated parameters are uploaded to a cooperative party for parameter aggregation, and the federated simple model participates in federated learning training;
3) And returning the aggregation parameters by the collaborators for updating the federal simple model.
3. The Federal learning-based Chinese address word segmentation method as claimed in claim 2, wherein the knowledge distilling partyThe loss function of the method consists of a mission loss function and a distillation loss function; the mission loss function of the knowledge distillation method is
Figure QLYQS_1
Wherein y is i Represents a data annotation, <' > or>
Figure QLYQS_2
Representing the prediction result of the student model; the distillation loss function of the known distillation method is ≥>
Figure QLYQS_3
Wherein->
Figure QLYQS_4
Representing the prediction result of the teacher model, wherein T is a hyperparameter used for controlling the smoothness of the distillation result; the integrity loss function of the knowledge distillation method is ≥>
Figure QLYQS_5
Wherein alpha is the learning emphasis of the hyper-parameter for controlling knowledge distillation, the fitting to the teacher model is more emphasized when alpha is larger, and the fitting to the data is emphasized when alpha is smaller.
4. The method for Chinese address segmentation based on federal learning as claimed in claim 2, wherein when the local data knowledge is transmitted to the federal simple model by knowledge distillation aiming at the local complex model, the knowledge distillation method takes the local complex model as a teacher model and the federal simple model as a student model; and when the federal learning knowledge is transferred to the local complex model by knowledge distillation aiming at the federal simple model, the federal simple model is used as a teacher model, and the local complex model is used as a student model.
5. The method for Chinese address segmentation based on federal learning according to claim 1, wherein in the step (1), the address segmentation model adopts a model with information memory capacity to complete a location segmentation task, the selected model is a recurrent neural network model, a Bert model or a graph neural network, and model training is performed only by using a locally stored address segmentation tagging dataset until a preset termination condition is met.
6. The method for segmenting words into Chinese addresses based on federated learning as defined in claim 1, wherein in the step (2), the place name classification model uses a federated simple model to participate in federated learning, and the federated simple model structure of all participants in federated learning is required to be consistent; for the local complex model to learn local data knowledge, each participant is allowed to participate in training using a different model.
7. The method of claim 1, wherein the federated learning-based Chinese address word segmentation method is a bidirectional recurrent neural network model, a convolutional neural network model, or a hidden Markov model with 10 layers or less.
8. The Federal learning-based Chinese address word segmentation method according to claim 1, wherein the local complex model is a Bert model or a bidirectional recurrent neural network model with a depth exceeding 20 layers.
9. A federal learning-based chinese address participle device, comprising a memory and one or more processors, wherein the memory stores executable codes, and the processors execute the executable codes to realize the federal learning-based chinese address participle method according to any one of claims 1 to 8.
10. A computer-readable storage medium having a program stored thereon, wherein the program, when executed by a processor, implements the federal learning based chinese address tokenization method as claimed in any one of claims 1 to 8.
CN202211626861.1A 2022-12-16 2022-12-16 Chinese address word segmentation method and device based on federal learning Active CN115879467B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202211626861.1A CN115879467B (en) 2022-12-16 2022-12-16 Chinese address word segmentation method and device based on federal learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202211626861.1A CN115879467B (en) 2022-12-16 2022-12-16 Chinese address word segmentation method and device based on federal learning

Publications (2)

Publication Number Publication Date
CN115879467A true CN115879467A (en) 2023-03-31
CN115879467B CN115879467B (en) 2024-04-30

Family

ID=85753897

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202211626861.1A Active CN115879467B (en) 2022-12-16 2022-12-16 Chinese address word segmentation method and device based on federal learning

Country Status (1)

Country Link
CN (1) CN115879467B (en)

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113657607A (en) * 2021-08-05 2021-11-16 浙江大学 Continuous learning method for federal learning
CN114091667A (en) * 2021-11-22 2022-02-25 北京理工大学 Federal mutual learning model training method oriented to non-independent same distribution data
WO2022105714A1 (en) * 2020-11-23 2022-05-27 华为技术有限公司 Data processing method, machine learning training method and related apparatus, and device
CN115374954A (en) * 2022-07-08 2022-11-22 广东华兴银行股份有限公司 Model training method based on federal learning, terminal and storage medium

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2022105714A1 (en) * 2020-11-23 2022-05-27 华为技术有限公司 Data processing method, machine learning training method and related apparatus, and device
CN113657607A (en) * 2021-08-05 2021-11-16 浙江大学 Continuous learning method for federal learning
CN114091667A (en) * 2021-11-22 2022-02-25 北京理工大学 Federal mutual learning model training method oriented to non-independent same distribution data
CN115374954A (en) * 2022-07-08 2022-11-22 广东华兴银行股份有限公司 Model training method based on federal learning, terminal and storage medium

Also Published As

Publication number Publication date
CN115879467B (en) 2024-04-30

Similar Documents

Publication Publication Date Title
CN112085159B (en) User tag data prediction system, method and device and electronic equipment
García-Peñalvo Entrepreneurial and problem solving skills in software engineers
US20180096219A1 (en) Neural network combined image and text evaluator and classifier
CN109919209B (en) Domain self-adaptive deep learning method and readable storage medium
Goh et al. Food-image Classification Using Neural Network Model
WO2022174491A1 (en) Artificial intelligence-based method and apparatus for medical record quality control, computer device, and storage medium
CN110874618B (en) OCR template learning method and device based on small sample, electronic equipment and medium
US20220269927A1 (en) Optimizing machine learning
CN111931814A (en) Unsupervised anti-domain adaptation method based on intra-class structure compactness constraint
Wu et al. Modeling visual and word-conditional semantic attention for image captioning
CN111191041A (en) Characteristic data acquisition method, data storage method, device, equipment and medium
Mu et al. A climate downscaling deep learning model considering the multiscale spatial correlations and chaos of meteorological events
CN115879467A (en) Federal learning-based Chinese address word segmentation method and device
WO2022227169A1 (en) Image classification method and apparatus, and electronic device and storage medium
CN115168609A (en) Text matching method and device, computer equipment and storage medium
Yang et al. Split learning based on self-supervised learning
CN111091198A (en) Data processing method and device
Zhang Sharing of teaching resources for English majors based on ubiquitous learning resource sharing platform and neural network
Bentaleb et al. A survey of federated learning approach for the Sustainable Development aspect: eLearning
Sokolovska Directions of development of digital society in Ukraine
Ma et al. A survey of Chinese character style transfer
CN117077798A (en) Method and device for generating joint learning model based on semi-supervised learning
Huang et al. Incremental Unsupervised Adversarial Domain Adaptation for Federated Learning in IoT Networks
CN116842960B (en) Feature extraction model training and extracting method and device based on large model
Pemasiri et al. Accurate 3D hand mesh recovery from a single RGB image

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant