US12373729B2 - System and method for federated learning with local differential privacy - Google Patents
System and method for federated learning with local differential privacyInfo
- Publication number
- US12373729B2 US12373729B2 US17/085,699 US202017085699A US12373729B2 US 12373729 B2 US12373729 B2 US 12373729B2 US 202017085699 A US202017085699 A US 202017085699A US 12373729 B2 US12373729 B2 US 12373729B2
- Authority
- US
- United States
- Prior art keywords
- gradients
- gradient
- perturbed
- perturbation
- electronic device
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active, expires
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F21/00—Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
- G06F21/60—Protecting data
- G06F21/62—Protecting access to data via a platform, e.g. using keys or access control rules
- G06F21/6218—Protecting access to data via a platform, e.g. using keys or access control rules to a system of files or objects, e.g. local or distributed file system or database
- G06F21/6245—Protecting personal data, e.g. for financial or medical purposes
- G06F21/6263—Protecting personal data, e.g. for financial or medical purposes during internet communication, e.g. revealing personal data from cookies
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/044—Recurrent networks, e.g. Hopfield networks
- G06N3/0442—Recurrent networks, e.g. Hopfield networks characterised by memory or gating, e.g. long short-term memory [LSTM] or gated recurrent units [GRU]
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/0464—Convolutional networks [CNN, ConvNet]
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/09—Supervised learning
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/098—Distributed learning, e.g. federated learning
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N20/00—Machine learning
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/044—Recurrent networks, e.g. Hopfield networks
Definitions
- This disclosure relates generally to database and file management within network environments, and in particular relates to machine learning for databases and file management.
- Machine learning is the study of computer algorithms that improve automatically through experience. It is seen as a subset of artificial intelligence. Machine learning algorithms build a mathematical model based on sample data, known as training data, in order to make predictions or decisions without being explicitly programmed to do so. Usually, when training a machine learning model, one needs to collect a large, representative sample of data from a training set. Data from the training set can be as varied as a corpus of text, a collection of images, and data collected from individual users of a service. Machine learning algorithms are used in a wide variety of applications, such as email filtering and computer vision, where it is difficult or infeasible to develop conventional algorithms to perform the needed tasks.
- FIG. 1 illustrates an example prediction system, in accordance with presently disclosed embodiments.
- FIG. 5 B illustrates an example architecture for federated learning based on perturbed user data.
- FIG. 6 illustrates an example workflow of federated learning enhanced by splitting and shuffling.
- FIG. 11 illustrates is a flow diagram of a method for perturbing gradients in federated learning, in accordance with the presently disclosed embodiments.
- FIG. 12 illustrates is a flow diagram of a method for perturbing user data in federated learning, in accordance with the presently disclosed embodiments.
- FIG. 13 illustrates an example computer system.
- FIG. 14 illustrates a diagram of an example artificial intelligence (AI) architecture.
- AI artificial intelligence
- FIG. 1 illustrates an example prediction system 100 , in accordance with presently disclosed embodiments.
- the prediction system 100 may include a programming analytics system 102 , one or more databases 104 , 106 , and a TV programming and advertising content subnetwork 108 .
- the programming analytics system 102 may include a cloud-based cluster computing architecture or other similar computing architecture that may receive one or more user automatic content recognition (ACR) user viewing data 110 , which may be provided by first-party or third-party sources, and provide TV programming content and advertising content to one or more client devices (e.g., a TV, a standalone monitor, a desktop computer, a laptop computer, a tablet computer, a mobile phone, a wearable electronic device, a voice-controlled personal assistant device, an automotive display, a gaming system, an appliance, or other similar multimedia electronic device) suitable for displaying programming and advertising content and/or playing back programming and advertising content.
- ACR user automatic content recognition
- the programming analytics system 102 may be utilized to process and manage various analytics and/or data intelligence such as TV programming analytics, web analytics, user profile data, user payment data, user privacy preferences, and so forth.
- the programming analytics system 102 may include a Platform as a Service (PaaS) architecture, a Software as a Service (SaaS) architecture, and an Infrastructure as a Service (IaaS), or other various cloud-based cluster computing architectures.
- PaaS Platform as a Service
- SaaS Software as a Service
- IaaS Infrastructure as a Service
- the programming analytics system 102 may include a pre-processing functional block 112 , a deep-learning model functional block 114 , and multi-label classification functional block 116 .
- the pre-processing functional block 112 , the deep-learning model functional block 114 , and the multi-label classification functional block 116 may each include, for example, a computing engine.
- the pre-processing functional block 112 may receive the ACR user viewing data 110 , which may include, for example, specific programming content (e.g., TV programming) recently viewed by one or more particular users or subgroups of users.
- the ACR user viewing data 110 may include an identification of the recently viewed programming content (e.g., TV programs), metadata associated with the recently viewed programming content (e.g., TV programs), the particular timeslot (e.g., day-hour) the recently viewed programming content (e.g., TV programs) was viewed within, and the programming channel on which the programming content (e.g., TV programs) was viewed.
- the recently viewed programming content e.g., TV programs
- metadata associated with the recently viewed programming content e.g., TV programs
- the particular timeslot e.g., day-hour
- the recently viewed programming content e.g., TV programs
- the pre-processing functional block 112 may then interface with the content database 104 to associate the recently viewed programming content included in the ACR user viewing data 110 with TV programming content stored by the database 104 .
- the TV programming content stored by the database 104 may include, for example, user or subgroup profile data, programming genre data, programing category data, programming clustering category group data, or other TV programming content or metadata that may be stored by the database 104 .
- the ACR user viewing data 110 may include time-series data expressed in an hour context and/or day context. For instance, in a particular embodiment, time-series ACR user viewing data 110 may be received, for example, every 2-hour timeslot per 24-hour time period (12 timeslots total per 24-hour day).
- different timeslots may be utilized (e.g., 83-hour timeslots per 24-hour time period, 241-hour timeslots per 24-hour time period, 4830-minute timeslots per 24-hour time period, etc.)
- the pre-processing functional block 112 may also perform stratified sampling and data augmentation on the time-series based ACR user viewing data 110 to, for example, augment and up-sample minority classes (e.g., defined as user subgroups with less than 20 examples per unique class).
- the data augmentation may be based on the introduction of Gaussian noise via one or more multiplicative factors.
- the pre-processing functional block 112 may also be utilized, for example, to split the time-series based ACR user viewing data 110 in an N number of datasets before providing to the deep-learning model functional block 114 for training, cross-validating, and testing.
- the pre-processing functional block 112 may perform the stratified multi-label sampling by, for example, accounting for the existence of one or more disjoint groups within a population and generating samples where the proportion of these groups is maintained.
- the pre-processing functional block 112 may perform a multi-label Synthetic Minority Over-sampling Technique (SMOTE) on the time-series based ACR user viewing training dataset.
- SMOTE Synthetic Minority Over-sampling Technique
- a final pre-processing of the time-series based ACR user viewing data 110 may be performed before providing an output to the deep-learning model functional block 114 for training, cross-validating, and testing.
- the deep-learning model functional block 114 may receive an N number of datasets (e.g., N arrays of time-series based ACR user viewing data 110 in 2-hour timeslots) generate an N number of long short term (LSTM) layers based thereon.
- outputs of the LSTM layers of the deep-learning model functional block 114 may be combined into a single array utilizing, for example, a concatenation layer of the deep-learning model functional block 114 . From the concatenation layer, the deep-learning model functional block 114 may then transfer the single array through one or more dense layers of the deep-learning model functional block 114 .
- the deep-learning model functional block 114 may then transfer the single array through a sigmoid output layer of the deep-learning model functional block 114 .
- the sigmoid output layer of the deep-learning model functional block 114 may include, for example, a number of neurons (e.g., the number of neurons may be equal to the number of classes and/or classification labels) that may be utilized to classify the single array into individual classes, in which one or more final probabilities for individual classification labels may be calculated.
- the deep-learning model functional block 114 may also include a loss function that may be utilized to assign a higher weight to positive classification for individual classification labels, assuming that individual users and/or subgroups of users may typically not exceed more than a maximum number of users (e.g., N users).
- a loss function that may be utilized to assign a higher weight to positive classification for individual classification labels, assuming that individual users and/or subgroups of users may typically not exceed more than a maximum number of users (e.g., N users).
- differential privacy To satisfy the increasing demand for preserving privacy, differential privacy (DP) was proposed as a rigorous principle that guarantees provable privacy protection and has been extensively applied.
- ⁇ be a deterministic function that maps the dataset D to the real numbers .
- This deterministic function ⁇ under the context of differential privacy, is called a query function of the dataset D.
- the query function may request the mean of a feature in the dataset, the gender of each sample.
- ⁇ -differential privacy may be defined as follows.
- the privacy guarantee of mechanism is controlled by privacy budget, denoted as ⁇ .
- a smaller value of E may indicate a stronger privacy guarantee.
- a local differentially private algorithm may provide aggregate representations about a set of data items without leaking information of any data item.
- the immunity to post-processing may also work on local differential privacy, which claims no algorithm can compromise the differentially private output and make it less differentially private. Meanwhile, shuffling and swapping may obtain a better local privacy protection.
- the remote server 265 may be an aggregator that collects a set of weights of local client-side models from the local side and averages the weights after each communication round.
- One goal may be to maximize the accuracy of both remote and local client-side models while preserving the privacy of the users.
- the client system may determine, based on one or more privacy policies, that one or more of the plurality of initial user data 310 should be perturbed. Accordingly, the client system may use a data perturbator 325 to enforce privacy policies such as LDP to perturb the initial user data 310 .
- the perturbed user data may be further sent to a model trainer 330 .
- the model trainer 330 may perform the duty of federated learning locally, e.g., calculating the loss using user data 310 .
- the output of the model trainer 330 may comprise the gradients 335 associated with the machine-learning model 305 .
- the client system may determine, based on one or more privacy policies, that one or more of the plurality of initial gradients 335 should be perturbed.
- the first electronic device may use the following example algorithm to perturb gradients.
- the goal of the algorithm may comprise ensuring LDP but preserving the accuracy of average calculation.
- the algorithm may be as follows. Let A be the LDP mechanism, it changes x into one of two values with probability:
- the gradient-perturbation model may be formulated as:
- the aforementioned gradient-perturbation model may have a tradeoff that if a smaller r or bigger E is chosen, variance may be smaller but privacy may be worse.
- Using a tradeoff between a smaller variance or bigger differential privacy may be an effective solution for addressing the technical challenge of poor accuracy caused by a large variance introduced to the estimated average since the embodiments disclosed herein may balance the variance and differential privacy to protect privacy as well as get a respectable accuracy.
- the first electronic device may access, from a data store associated with the first electronic device, a plurality of initial user data for training a machine-learning model.
- the first electronic device may select one or more of the plurality of initial user data for perturbation.
- the first electronic device may then generate, based on a data-perturbation model, one or more perturbed user data for the one or more selected initial user data, respectively.
- the generation for each selected initial user data may comprise the following sub-steps. Firstly, the first electronic device may feed the selected initial user data as an input to the data-perturbation model.
- the selected initial user data may have a value x within a value range. Secondly, the first electronic device may divide the value range into m intervals.
- the first electronic device may change x into a center value a of one of the m intervals with a probability 1 ⁇ p if a distance between x and a is a minimum distance among distances between x and all the center values of the m intervals or a probability p/(m ⁇ 1) if the distance between x and a is not the minimum distance among distances between x and all the center value of the m intervals.
- the first electronic device may determine, based on the one or more perturbed user data, a plurality of gradients associated with the machine-learning model. The first electronic device may further send, from the first electronic device to a second electronic device, the plurality of gradients.
- client system 1205 may add LDP noise 510 to user data 1210
- client system 2215 may add LDP noise 510 to user data 2220
- client system 3225 may add LDP noise 510 to user data 3230
- client system k 235 may add LDP noise 510 to user data k 240 .
- This may allow the whole training process to occur on the noisy user data.
- client systems encode/compress the user data and then apply LDP because large dimensionality of user data may degrade the privacy level.
- FIG. 5 B illustrates an example architecture for federated learning based on perturbed user data.
- the client systems may respectively generate noisy user data 520 , noisy user data 530 , noisy user data 540 , and noisy user data 550 .
- the client systems may then learn the gradients based on the noisy user data.
- the client system may then send the learned gradients to the remote server 265 .
- the first electronic device may receive, at the first electronic device from the second electronic device, a plurality of weights of the machine-learning model. The plurality of weights may be determined based on the one or more perturbed gradients.
- the first electronic device may further determine, by the first electronic device, a plurality of new gradients for the plurality of weights.
- the second electronic device may send the whole machine-learning model back to the first electronic device for it to start the next iteration of learning gradients for the machine-learning model.
- x be a data item within the domain D, in which x may be binary, categorical, ordinal, discrete or continuous.
- the goal for data perturbation may comprise ensuring LDP but preserving data fidelity.
- the first electronic device may use the following example algorithm to perturb user data. Firstly, the domain D may be divided into m subdomain/intervals/buckets and a center ⁇ i in each subdomain D 1 may be selected. Let A be our LDP mechanism, it may change x into ⁇ i with probability:
- the data-perturbation model may be formulated as:
- A(x) may represent a changed value of x
- x may be a data item within domain D divided into m subdomains D i of the value range
- j may represent an index indexing each of the m intervals of the value range
- ⁇ j may represent the center value of each subdomain of the m intervals of the value range
- i represents the index indexing the interval that has the center value that is closest to x
- ⁇ i represents the center value that is closest to x
- ⁇ k may represent any remaining center
- Using the data-perturbation model that chooses one out of two extreme values as the noisy data may be an effective solution for addressing the technical challenge of risk of information exposure due to noisy data being close to its original value with high probability since the data-perturbation model makes it more distinct from its original value.
- the aforementioned data-perturbation model may have a tradeoff that if smaller p or bigger m is chosen, A(x) may be closer to x but the privacy may be worse.
- the embodiments disclosed herein may use shuffling and splitting model updates as an effective solution for addressing the technical challenge of explosion of privacy budget due to high dimensionality of weights in deep learning models since the remote server 265 may be unable to link different gradient/weight values from the same client system after the gradients/weights are split and uploaded anonymously. As a result, the remote server 265 cannot infer more information about a particular client system, which makes it sufficient to protect ⁇ -LDP for each gradient/weight. Likewise, because of the anonymity, the remote server 265 may be unable to link gradients/weights from the same client system at different iterations. Without splitting and shuffling, the privacy budget of LDP may grow to Td ⁇ , where T is the interaction number and d is the number of gradients/weights in the model.
- FIG. 6 illustrates an example workflow of federated learning enhanced by splitting and shuffling.
- the workflow shows a system which relies on federated learning with local differential privacy.
- Each gradient update of a client system may need to protect the data privacy information with perturbed information for privacy protection locally.
- the workflow includes a local process (i.e., on the client systems) and a cloud process (i.e., on the remote server 265 ).
- a client system may learn a model A 610 ; based on local data B 615 , a client system may learn a model B 620 ; and based on local data Z 625 , a client system may learn a model Z 630 .
- client systems may not need to trust each other or the remote server 265 .
- it may protect a user's all data.
- user refers to who generates or owns the data, not necessarily the client system.
- the client system may have multiple users' data.
- a hospital may be associated with a client system, and a sub-dataset may comprise one patient's data. Making the client-side data indistinguishable may require adding a lot of noise and may profoundly affect the performance of federated learning.
- the client system may partition its data into multiple sub-datasets, calculate gradients and add noises on each sub-dataset then send all noisy gradients to the remote server 265 .
- FIG. 8 illustrates example pseudo-code for federated learning with local privacy perturbation according to particular embodiments.
- M is the number of local client systems;
- B is the local mini-batch size,
- E is the number of local epochs, and
- ⁇ is the learning rate.
- the embodiments disclosed herein introduce a federated learning approach with local differential privacy that comprises two steps, as shown in FIG. 8 .
- Randomized Response Mechanism This mechanism is for binary or categorical data only, whereas data are numeric in the scenario of federated learning.
- a modified version of generalized randomized response mechanism was proposed but it introduces asymptotically higher variance to the estimated average than embodiments described herein and is only feasible when E is very large
- Gaussian Mechanism may be frequently used for differential privacy. However, based on the definition of the local differential privacy, currently most works only study ⁇ -LDP that does not include the ⁇ yet. However, the Gaussian mechanism requires relaxing the differential privacy definition and introduces ⁇ , which does not match the scenario of local differential privacy. ⁇ is the probability that indicate those highly unlikely “bad” events. These “bad” events may break ⁇ -differential privacy and usually defined by the size of the dataset. As a result, the Gaussian Mechanism is less secure than the mechanism of the embodiments disclosed herein (without introducing ⁇ ).
- DP differential privacy
- LDP local differential privacy
- the embodiments disclosed herein may apply the same mechanism to each dimension independently and achieve k ⁇ -local DP.
- DP and LDP differ in definition and in whether the noise is added locally or on the remote server. However, they are not mutually exclusive.
- the definition of LDP is a special case of that of DP.
- noises are added locally to achieve DP (e.g., via additive secret sharing) instead of LDP.
- Cony-Small may be used as the cloud-side DNN.
- VGG-Small may be used for CIFAR-10.
- the learning rate and batch size may be set as 0.03 and 10, respectively.
- the numbers of epochs for MNIST, and CIFAR-10 are 10 and 100, respectively. Considering the randomness during perturbation, the test experiments are run ten times independently to obtain an averaged value.
- Table 1 illustrates an exemplary summary of performance analysis percentage.
- MNIST i.e., a public dataset
- CIFAR-10 i.e., another public dataset
- FIG. 10 illustrates example graphs showing effects of data dimension and differential privacy when the model has been trained with pre-assigned perturbation.
- the performance is evaluated with the number of the client systems.
- the embodiments disclosed herein implement a two-layer CNN for image classification.
- the default network is not working. Therefore, the embodiments disclosed herein re-design a small VGG for the task.
- the training data and the testing data are fed into the network directly in each client system, and for each client system, the size of the training data is the total number of the training samples dividing the number of the client systems. In this case, a larger number of client systems implicit the small size of training data of each client system.
- the learning rate ⁇ is set as 0.03 for MNIST and 0.015 for CIFAR-10.
- FIG. 10 shows that the embodiments disclosed herein may achieve a performance with a low privacy cost because of the new design of the communication and the new local noise perturbation. It may be not hard to see that while increasing the number of client systems in the training, the LDP-FL may perform as close as the noise-free federated learning. The privacy budget may also affect the performance of the central model.
- the analysis of privacy budget is provided as follows.
- the privacy budget represents the privacy cost in the framework.
- the scale is chosen from 0.1 to 1 for MNIST and 1 to 10 for CIFAR-10. It may be not hard to see that more complex data and tasks require more privacy cost. The main reason may be that the complex task requires a sophisticated neural network, which contains a large number of model parameters. Meanwhile, the range of each parameter is also wider in the complex task.
- FIG. 11 illustrates is a flow diagram of a method for perturbing gradients in federated learning, in accordance with the presently disclosed embodiments.
- the method 1100 may be performed utilizing one or more processing devices (e.g., of a client system 1 - k 205 - 235 ) that may include hardware (e.g., a general purpose processor, a graphic processing unit (GPU), an application-specific integrated circuit (ASIC), a system-on-chip (SoC), a microcontroller, a field-programmable gate array (FPGA), a central processing unit (CPU), an application processor (AP), a visual processing unit (VPU), a neural processing unit (NPU), a neural decision processor (NDP), or any other processing device(s) that may be suitable for processing 2D and 3D image data, software (e.g., instructions running/executing on one or more processors), firmware (e.g., microcode), or some combination thereof.
- hardware e.g., a general purpose processor, a
- the method 1100 may begin at step 1110 with the one or more processing devices (e.g., of a client system 1 - k 205 - 235 ).
- the first electronic device may access, from a data store associated with the first electronic device, a plurality of initial gradients associated with a machine-learning model.
- the method 1100 may then continue at step 1120 with the one or more processing devices.
- the first electronic device may determine, based on one or more privacy policies, that one or more of the plurality of initial gradients should be perturbed.
- the method 1100 may then continue at step 1130 with the one or more processing devices.
- the first electronic device may select one or more of the plurality of initial gradients for perturbation.
- the method 1100 may then continue at step 1140 with the one or more processing devices.
- the first electronic device may generate, based on a gradient-perturbation model, one or more perturbed gradients for the one or more selected initial gradients, respectively, wherein for each selected initial gradient: an input to the gradient-perturbation model comprises the selected initial gradient having a value x, the gradient-perturbation model changes x into a first continuous value with a first probability or a second continuous value with a second probability, and the first and second probabilities are determined based on x.
- the method 1100 may then continue at step 1150 with the one or more processing devices.
- the first electronic device may shuffle the one or more perturbed gradients to a random order.
- the method 1100 may then continue at step 1160 with the one or more processing devices.
- the first electronic device may send, from the first electronic device to a second electronic device, the one or more perturbed gradients, wherein the one or more perturbed gradients are sent based on the random order.
- the method 1100 may then continue at step 1170 with the one or more processing devices.
- the first electronic device may receive, at the first electronic device from the second electronic device, a plurality of weights of the machine-learning model, wherein the plurality of weights are determined based on the one or more perturbed gradients.
- the method 1100 may then continue at step 1180 with the one or more processing devices.
- the first electronic device may determine, by the first electronic device, a plurality of new gradients for the plurality of weights.
- Particular embodiments may repeat one or more steps of the method of FIG. 11 , where appropriate.
- this disclosure describes and illustrates an example method for perturbing gradients in federated learning including the particular steps of the method of FIG. 11
- this disclosure contemplates any suitable method for perturbing gradients in federated learning including any suitable steps, which may include all, some, or none of the steps of the method of FIG. 11 , where appropriate.
- this disclosure describes and illustrates particular components, devices, or systems carrying out particular steps of the method of FIG. 11
- this disclosure contemplates any suitable combination of any suitable components, devices, or systems carrying out any suitable steps of the method of FIG. 11 .
- FIG. 12 illustrates is a flow diagram of a method for perturbing user data in federated learning, in accordance with the presently disclosed embodiments.
- the method 1200 may be performed utilizing one or more processing devices (e.g., of a client system 1 - k 205 - 235 ) that may include hardware (e.g., a general purpose processor, a graphic processing unit (GPU), an application-specific integrated circuit (ASIC), a system-on-chip (SoC), a microcontroller, a field-programmable gate array (FPGA), a central processing unit (CPU), an application processor (AP), a visual processing unit (VPU), a neural processing unit (NPU), a neural decision processor (NDP), or any other processing device(s) that may be suitable for processing 2D and 3D image data, software (e.g., instructions running/executing on one or more processors), firmware (e.g., microcode), or some combination thereof.
- hardware e.g., a general purpose processor, a
- the method 1200 may begin at step 1210 with the one or more processing devices (e.g., of a client system 1 - k 205 - 235 ).
- the first electronic device may access, from a data store associated with a first electronic device, a plurality of initial user data for training a machine-learning model.
- the method 1200 may then continue at step 1220 with the one or more processing devices.
- the first electronic device may determine, based on one or more privacy policies, that one or more of the plurality of initial user data should be perturbed.
- the method 1200 may then continue at step 1230 with the one or more processing devices.
- the first electronic device may select one or more of the plurality of initial user data for perturbation.
- the method 1200 may then continue at step 1240 with the one or more processing devices.
- the first electronic device may generate, based on a data-perturbation model, one or more perturbed user data for the one or more selected initial user data, respectively, wherein the generation for each selected initial user data comprises: feeding the selected initial user data as an input to the data-perturbation model, wherein the selected initial user data has a value x within a value range, dividing the value range into m intervals, and changing x into a center value a of one of the m intervals with a first probability 1 ⁇ p if a distance between x and a is a minimum distance among distances between x and all the center values of the m intervals or a second probability p/(m ⁇ 1) if the distance between x and a is not the minimum distance among distances between x and all the center value
- One or more memory buses may couple processor 1302 to memory 1304 .
- Bus 1312 may include one or more memory buses, as described below.
- one or more memory management units reside between processor 1302 and memory 1304 and facilitate accesses to memory 1304 requested by processor 1302 .
- memory 1304 includes random access memory (RAM).
- This RAM may be volatile memory, where appropriate. Where appropriate, this RAM may be dynamic RAM (DRAM) or static RAM (SRAM). Moreover, where appropriate, this RAM may be single-ported or multi-ported RAM.
- DRAM dynamic RAM
- SRAM static RAM
- Memory 1304 may include one or more memory devices 1304 , where appropriate.
- this ROM may be mask-programmed ROM, programmable ROM (PROM), erasable PROM (EPROM), electrically erasable PROM (EEPROM), electrically alterable ROM (EAROM), or flash memory or a combination of two or more of these.
- This disclosure contemplates mass storage 1306 taking any suitable physical form.
- Storage 1306 may include one or more storage control units facilitating communication between processor 1302 and storage 1306 , where appropriate.
- storage 1306 may include one or more storages 1306 .
- this disclosure describes and illustrates particular storage, this disclosure contemplates any suitable storage.
- I/O interface 1308 includes hardware, software, or both, providing one or more interfaces for communication between computer system 1300 and one or more I/O devices.
- Computer system 1300 may include one or more of these I/O devices, where appropriate.
- One or more of these I/O devices may enable communication between a person and computer system 1300 .
- an I/O device may include a keyboard, keypad, microphone, monitor, mouse, printer, scanner, speaker, still camera, stylus, tablet, touch screen, trackball, video camera, another suitable I/O device or a combination of two or more of these.
- An I/O device may include one or more sensors. This disclosure contemplates any suitable I/O devices and any suitable I/O interfaces 1306 for them.
- computer system 1300 may communicate with an ad hoc network, a personal area network (PAN), a local area network (LAN), a wide area network (WAN), a metropolitan area network (MAN), or one or more portions of the Internet or a combination of two or more of these.
- PAN personal area network
- LAN local area network
- WAN wide area network
- MAN metropolitan area network
- computer system 1300 may communicate with a wireless PAN (WPAN) (such as, for example, a BLUETOOTH WPAN), a WI-FI network, a WI-MAX network, a cellular telephone network (such as, for example, a Global System for Mobile Communications (GSM) network), or other suitable wireless network or a combination of two or more of these.
- Computer system 1300 may include any suitable communication interface 1310 for any of these networks, where appropriate.
- Communication interface 1310 may include one or more communication interfaces 1310 , where appropriate.
- bus 1312 includes hardware, software, or both coupling components of computer system 1300 to each other.
- bus 1312 may include an Accelerated Graphics Port (AGP) or other graphics bus, an Enhanced Industry Standard Architecture (EISA) bus, a front-side bus (FSB), a HYPERTRANSPORT (HT) interconnect, an Industry Standard Architecture (ISA) bus, an INFINIBAND interconnect, a low-pin-count (LPC) bus, a memory bus, a Micro Channel Architecture (MCA) bus, a Peripheral Component Interconnect (PCI) bus, a PCI-Express (PCIe) bus, a serial advanced technology attachment (SATA) bus, a Video Electronics Standards Association local (VLB) bus, or another suitable bus or a combination of two or more of these.
- Bus 1312 may include one or more buses 1312 , where appropriate.
- the supervised learning algorithms 1420 may include any algorithms that may be utilized to apply, for example, what has been learned in the past to new data using labeled examples for predicting future events. For example, starting from the analysis of a known training dataset, the supervised learning algorithms 1420 may produce an inferred function to make predictions about the output values. The supervised learning algorithms 1420 can also compare its output with the correct and intended output and find errors in order to modify the supervised learning algorithms 1420 accordingly.
- the unsupervised learning algorithms 1422 may include any algorithms that may applied, for example, when the data used to train the unsupervised learning algorithms 1422 are neither classified or labeled. For example, the unsupervised learning algorithms 1422 may study and analyze how systems may infer a function to describe a hidden structure from unlabeled data.
- the classification algorithms or functions 1426 may include any algorithms that may utilize a supervised learning model (e.g., logistic regression, na ⁇ ve Bayes, stochastic gradient descent (SGD), k-nearest neighbors, decision trees, random forests, support vector machine (SVM), and so forth) to learn from the data input to the supervised learning model and to make new observations or classifications based thereon.
- the machine translation algorithms or functions 1428 may include any algorithms or functions that may be suitable for automatically converting source text in one language, for example, into text in another language.
- the QA algorithms or functions 1430 may include any algorithms or functions that may be suitable for automatically answering questions posed by humans in, for example, a natural language, such as that performed by voice-controlled personal assistant devices.
- the text generation algorithms or functions 1432 may include any algorithms or functions that may be suitable for automatically generating natural language texts.
- the speech recognition algorithms and functions 1412 may include any algorithms or functions that may be suitable for recognizing and translating spoken language into text, such as through automatic speech recognition (ASR), computer speech recognition, speech-to-text (STT), or text-to-speech (TTS) in order for the computing to communicate via speech with one or more users, for example.
- the planning algorithms and functions 1438 may include any algorithms or functions that may be suitable for generating a sequence of actions, in which each action may include its own set of preconditions to be satisfied before performing the action. Examples of AI planning may include classical planning, reduction to other problems, temporal planning, probabilistic planning, preference-based planning, conditional planning, and so forth.
- the robotics algorithms and functions 1440 may include any algorithms, functions, or systems that may enable one or more devices to replicate human behavior through, for example, motions, gestures, performance tasks, decision-making, emotions, and so forth.
- a computer-readable non-transitory storage medium or media may include one or more semiconductor-based or other integrated circuits (ICs) (such, as for example, field-programmable gate arrays (FPGAs) or application-specific ICs (ASICs)), hard disk drives (HDDs), hybrid hard drives (HHDs), optical discs, optical disc drives (ODDs), magneto-optical discs, magneto-optical drives, floppy diskettes, floppy disk drives (FDDs), magnetic tapes, solid-state drives (SSDs), RAM-drives, SECURE DIGITAL cards or drives, any other suitable computer-readable non-transitory storage media, or any suitable combination of two or more of these, where appropriate.
- ICs such, as for example, field-programmable gate arrays (FPGAs) or application-specific ICs (ASICs)
- HDDs hard disk drives
- HHDs hybrid hard drives
- ODDs optical disc drives
- magneto-optical discs magneto-optical drives
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Software Systems (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Bioethics (AREA)
- Mathematical Physics (AREA)
- Data Mining & Analysis (AREA)
- Computing Systems (AREA)
- Evolutionary Computation (AREA)
- Artificial Intelligence (AREA)
- Medical Informatics (AREA)
- Computational Linguistics (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Molecular Biology (AREA)
- Databases & Information Systems (AREA)
- Computer Hardware Design (AREA)
- Computer Security & Cryptography (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
- Storage Device Security (AREA)
Abstract
Description
Pr[M(D)=Y]≤e ε P·Pr[M(D′)=Y]+δ (1)
If δ=0, is E-differentially private. The parameter E represents the privacy budget that controls the privacy loss of M. A larger value of E may indicate weaker privacy protection.
Pr[M(x)=Y]≤e ε P·Pr[M(x′)=Y] (2)
where the inputs x and x′ are any two inputs. The privacy guarantee of mechanism is controlled by privacy budget, denoted as ε. A smaller value of E may indicate a stronger privacy guarantee. According to this definition, a local differentially private algorithm may provide aggregate representations about a set of data items without leaking information of any data item. The immunity to post-processing may also work on local differential privacy, which claims no algorithm can compromise the differentially private output and make it less differentially private. Meanwhile, shuffling and swapping may obtain a better local privacy protection.
Alternatively, the client systems may do the above perturbation directly to weights instead of to gradients 335. In particular embodiments, the second electronic device may send the whole machine-learning model back to the first electronic device for it to start the next iteration of learning gradients for the machine-learning model. In alternative embodiments, the first electronic device may receive, at the first electronic device from the second electronic device, a plurality of weights of the machine-learning model. The plurality of weights may be determined based on the one or more perturbed gradients. The first electronic device may further determine, by the first electronic device, a plurality of new gradients for the plurality of weights.
In the above formulation: A(x) may represent a changed value of x, c may represent a center value of a value range, r may represent a distance from the center value to boundaries of the value range, c−r may represent the left boundary of the value range, c+r may represent the right boundary of the value range, each selected initial gradient may be clipped into the value range, and E may be a positive real number determined based on a local differential policy. Using the gradient-perturbation algorithm that chooses one out of two extreme values as the noisy data may be an effective solution for addressing the technical challenge of risk of information exposure due to noisy data being close to its original value with high probability since the gradient-perturbation algorithm makes it more distinct from its original value.
The proof of accuracy of average calculation may be illustrated as:
In the above formulation: A(x) may represent a changed value of x, x may be a data item within domain D divided into m subdomains Di of the value range, j may represent an index indexing each of the m intervals of the value range, αj may represent the center value of each subdomain of the m intervals of the value range, i represents the index indexing the interval that has the center value that is closest to x, αi represents the center value that is closest to x, αk may represent any remaining center value that is not αi, and Distance (x, αj) may be a function measuring a distance between x and αj. Using the data-perturbation model that chooses one out of two extreme values as the noisy data may be an effective solution for addressing the technical challenge of risk of information exposure due to noisy data being close to its original value with high probability since the data-perturbation model makes it more distinct from its original value.
otherwise.
then for any p, p′∈[c−r, c+r]:
the above still holds.
by Bernstein's inequality,
such that |
The variance of estimated average over n client systems is
which is higher than the embodiments described herein, shown in Equation (17), when ∈<2.3 at least. In the best case, Laplace mechanism's variance is always higher than the embodiments described herein for any ∈. Because a small ∈ is important to stronger privacy protection, one may decide to choose the Laplace mechanism for lower variance, i.e., a better estimation of average weight updates. The advantages of the Laplace mechanism may include that it is easier to understand, and the noisy data is still continuously distributed.
| TABLE 1 | |||||||
| ϵ = 0 | ϵ = 0.1 | ϵ = 0.5 | ϵ = 1 | ϵ = 5 | ϵ = 10 | ||
| MNIST | 97.26% | — | — | — | — | — |
| (k = 100) | ||||||
| MNIST-LDP | 97.26% | 13.84% | 95.36% | 96.24% | ||
| (k = 100) | ||||||
| CIFAR | 62.36% | — | — | — | — | — |
| (k = 500) | ||||||
| CIFAR-LDP | 62.36% | — | — | 10.00% | 58.89% | 60.37% |
| (k = 500) | ||||||
Claims (15)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US17/085,699 US12373729B2 (en) | 2020-05-28 | 2020-10-30 | System and method for federated learning with local differential privacy |
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US202063031531P | 2020-05-28 | 2020-05-28 | |
| US17/085,699 US12373729B2 (en) | 2020-05-28 | 2020-10-30 | System and method for federated learning with local differential privacy |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| US20210374605A1 US20210374605A1 (en) | 2021-12-02 |
| US12373729B2 true US12373729B2 (en) | 2025-07-29 |
Family
ID=78705044
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US17/085,699 Active 2043-12-29 US12373729B2 (en) | 2020-05-28 | 2020-10-30 | System and method for federated learning with local differential privacy |
Country Status (1)
| Country | Link |
|---|---|
| US (1) | US12373729B2 (en) |
Families Citing this family (32)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US11651292B2 (en) * | 2020-06-03 | 2023-05-16 | Huawei Technologies Co., Ltd. | Methods and apparatuses for defense against adversarial attacks on federated learning systems |
| US12481891B2 (en) * | 2020-10-26 | 2025-11-25 | Google Llc | Training neural networks with label differential privacy |
| US12262287B2 (en) * | 2020-12-03 | 2025-03-25 | Qualcomm Incorporated | Wireless signaling in federated learning for machine learning components |
| US11734453B2 (en) * | 2021-02-22 | 2023-08-22 | International Business Machines Corporation | Privacy-preserving motion analysis |
| US20220318412A1 (en) * | 2021-04-06 | 2022-10-06 | Qualcomm Incorporated | Privacy-aware pruning in machine learning |
| US11816243B2 (en) * | 2021-08-09 | 2023-11-14 | Adobe Inc. | Preserving user-entity differential privacy in natural language modeling |
| US20230147398A1 (en) * | 2021-11-05 | 2023-05-11 | Fujitsu Limited | Span loss prediction with federated learning |
| CN113961967B (en) * | 2021-12-13 | 2022-03-22 | 支付宝(杭州)信息技术有限公司 | Method and device for jointly training natural language processing model based on privacy protection |
| CN114338144B (en) * | 2021-12-27 | 2024-07-19 | 杭州趣链科技有限公司 | Data leakage prevention method, electronic device and computer-readable storage medium |
| CN114417423B (en) * | 2022-01-25 | 2024-05-17 | 杭州卷积云科技有限公司 | Real-time privacy protection method and system for unlimited data stream based on dynamic budget allocation |
| CN114611128B (en) * | 2022-03-15 | 2023-06-06 | 抖音视界有限公司 | Longitudinal federal learning method, device, system, equipment and storage medium |
| CN114841364B (en) * | 2022-04-14 | 2024-06-14 | 北京理工大学 | A federated learning approach that meets personalized local differential privacy requirements |
| CN114781448B (en) * | 2022-04-19 | 2025-09-05 | 齐鲁工业大学(山东省科学院) | A bearing fault feature extraction method, system, medium and device |
| CN114817989B (en) * | 2022-04-29 | 2024-06-11 | 浪潮(北京)电子信息产业有限公司 | Model generation method, operation control method, device, equipment and storage medium |
| US20230359907A1 (en) * | 2022-05-09 | 2023-11-09 | Google Llc | System(s) and method(s) for jointly learning machine learning model(s) based on server data and client data |
| US12417314B2 (en) | 2022-05-11 | 2025-09-16 | Cornell University | Task-aware privacy preservation for multi-dimensional data |
| CN115131056A (en) * | 2022-05-16 | 2022-09-30 | 中国科学技术大学 | Advertisement inventory estimation method, device and application system |
| US12373598B2 (en) * | 2022-05-27 | 2025-07-29 | The Toronto-Dominion Bank | Identifying and mitigating disparate group impact in differential-privacy machine-learned models |
| JP7735220B2 (en) * | 2022-06-03 | 2025-09-08 | 株式会社東芝 | Machine learning systems and methods |
| IL294292A (en) * | 2022-06-26 | 2024-01-01 | Google Llc | Privacy-sensitive neural network training |
| CN115081532B (en) * | 2022-07-01 | 2024-09-06 | 西安电子科技大学 | Federated continuous learning training method based on memory replay and differential privacy |
| US12443750B2 (en) * | 2022-07-05 | 2025-10-14 | Twilio Inc. | Context-based entropy management |
| CN115496198B (en) * | 2022-08-05 | 2025-07-04 | 广州大学 | Gradient compression framework with adaptive privacy budget allocation based on federated learning |
| US12314433B2 (en) * | 2022-09-16 | 2025-05-27 | Oracle International Corporation | Systems for design and implementation of privacy preserving AI with privacy regulations within intelligence pipelines |
| US12271704B2 (en) | 2022-09-21 | 2025-04-08 | International Business Machines Corporation | Detection of veracity of responses in machine comprehension question and answer models |
| CN115719092A (en) * | 2022-09-29 | 2023-02-28 | 阿里巴巴(中国)有限公司 | Model training method and federated learning system based on federated learning |
| CN116029390B (en) * | 2022-12-29 | 2025-11-14 | 天翼安全科技有限公司 | A model aggregation training method and apparatus |
| CN116436590A (en) * | 2023-04-12 | 2023-07-14 | 湖南大学 | A trajectory privacy protection method and system in a cloud-edge-device collaborative architecture |
| CN117556470B (en) * | 2023-12-18 | 2024-05-03 | 河北大学 | LDP method for carrying out joint disturbance on FL high-dimensional sparse Top-k gradient vector |
| CN118504712B (en) * | 2024-05-31 | 2024-11-26 | 重庆大学 | A federated learning method and system based on local differential privacy |
| CN119740258B (en) * | 2024-10-31 | 2025-11-11 | 北京邮电大学 | Private data publishing method, system and device based on generation countermeasure network |
| CN119808135B (en) * | 2024-11-12 | 2025-09-19 | 北京航空航天大学 | Privacy protection method, device and electronic device for discrete-time dynamic graphs |
Citations (7)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20110139748A1 (en) * | 2009-12-15 | 2011-06-16 | University Of Houston | Atomic layer etching with pulsed plasmas |
| US20120212375A1 (en) * | 2011-02-22 | 2012-08-23 | Depree Iv William Frederick | Quantum broadband antenna |
| US20170134434A1 (en) * | 2014-03-24 | 2017-05-11 | Amazon Technologies, Inc. | Role-based access control assignment |
| US20190220703A1 (en) | 2019-03-28 | 2019-07-18 | Intel Corporation | Technologies for distributing iterative computations in heterogeneous computing environments |
| KR20190103090A (en) | 2019-08-15 | 2019-09-04 | 엘지전자 주식회사 | Method and apparatus for learning a model to generate poi data using federated learning |
| KR20200010480A (en) | 2017-08-11 | 2020-01-30 | 구글 엘엘씨 | On-Device Machine Learning Platform |
| US20200184278A1 (en) * | 2014-03-18 | 2020-06-11 | Z Advanced Computing, Inc. | System and Method for Extremely Efficient Image and Pattern Recognition and Artificial Intelligence Platform |
-
2020
- 2020-10-30 US US17/085,699 patent/US12373729B2/en active Active
Patent Citations (7)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20110139748A1 (en) * | 2009-12-15 | 2011-06-16 | University Of Houston | Atomic layer etching with pulsed plasmas |
| US20120212375A1 (en) * | 2011-02-22 | 2012-08-23 | Depree Iv William Frederick | Quantum broadband antenna |
| US20200184278A1 (en) * | 2014-03-18 | 2020-06-11 | Z Advanced Computing, Inc. | System and Method for Extremely Efficient Image and Pattern Recognition and Artificial Intelligence Platform |
| US20170134434A1 (en) * | 2014-03-24 | 2017-05-11 | Amazon Technologies, Inc. | Role-based access control assignment |
| KR20200010480A (en) | 2017-08-11 | 2020-01-30 | 구글 엘엘씨 | On-Device Machine Learning Platform |
| US20190220703A1 (en) | 2019-03-28 | 2019-07-18 | Intel Corporation | Technologies for distributing iterative computations in heterogeneous computing environments |
| KR20190103090A (en) | 2019-08-15 | 2019-09-04 | 엘지전자 주식회사 | Method and apparatus for learning a model to generate poi data using federated learning |
Non-Patent Citations (8)
| Title |
|---|
| Bebensee, Bjorn. "Local differential privacy: a tutorial." arXiv preprint arXiv:1907.11908, Jul. 27, 2019. |
| Dwork, Cynthia. "Differential privacy: A survey of results." In International conference on theory and applications of models of computation, pp. 1-19. Springer, Berlin, Heidelberg, Apr. 25, 2008. |
| Erlingsson, Ulfar, Vasyl Pihur, and Aleksandra Korolova. "Rappor: Randomized aggregatable privacy-preserving ordinal response." In Proceedings of the 2014 ACM SIGSAC conference on computer and communications security, pp. 1054-1067, 2014. |
| Jayaraman, Bargav, Lingxiao Wang, David Evans, and Quanquan Gu. "Distributed learning without distress: Privacy-preserving empirical risk minimization." In Advances in Neural Information Processing Systems, pp. 6343-6354, 2018. |
| Kairouz, Peter, Sewoong Oh, and Pramod Viswanath. "Extremal mechanisms for local differential privacy." In Advances in neural information processing systems, pp. 1-51, Apr. 16, 2016. |
| Seif, Mohamed, Ravi Tandon, and Ming Li. "Wireless federated learning with local differential privacy." arXiv preprint arXiv:2002.05151, pp. 1-13, Feb. 13, 2020. |
| Shokri, Reza, and Vitaly Shmatikov. "Privacy-preserving deep learning." In Proceedings of the 22nd ACM SIGSAC conference on computer and communications security, pp. 1310-1321, 2015. |
| Truex, Stacey, Ling Liu, Ka-Ho Chow, Mehmet Emre Gursoy, and Wenqi Wei. "LDP-Fed: federated learning with local differential privacy." In Proceedings of the Third ACM International Workshop on Edge Systems, Analytics and Networking, pp. 61-66, Jun. 5, 2020. |
Also Published As
| Publication number | Publication date |
|---|---|
| US20210374605A1 (en) | 2021-12-02 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US12373729B2 (en) | System and method for federated learning with local differential privacy | |
| US11797843B2 (en) | Hashing-based effective user modeling | |
| US11562243B2 (en) | Machine-learning models based on non-local neural networks | |
| US11790264B2 (en) | Systems and methods for performing knowledge distillation | |
| US11544550B2 (en) | Analyzing spatially-sparse data based on submanifold sparse convolutional neural networks | |
| US11132604B2 (en) | Nested machine learning architecture | |
| US10922609B2 (en) | Semi-supervised learning via deep label propagation | |
| US11144812B2 (en) | Mixed machine learning architecture | |
| US20190073580A1 (en) | Sparse Neural Network Modeling Infrastructure | |
| US20190073590A1 (en) | Sparse Neural Network Training Optimization | |
| US20210279525A1 (en) | Hierarchy-preserving learning for multi-label classification | |
| US20200311198A1 (en) | N-ary relation prediction over text spans | |
| EP3977361B1 (en) | Co-informatic generative adversarial networks for efficient data co-clustering | |
| AU2016350555A1 (en) | Identifying content items using a deep-learning model | |
| US11711558B2 (en) | User classification based on user content viewed | |
| US20250286875A1 (en) | Systems and methods for authentication using partitioned authentication tokens | |
| US20220222539A1 (en) | Adversarial learning of privacy preserving representations | |
| Zhang et al. | A generative adversarial network–based method for generating negative financial samples | |
| US10990762B2 (en) | Chat analysis using machine learning | |
| Liu et al. | AIM: Attributing, interpreting, mitigating data unfairness | |
| US12002569B2 (en) | Generating multi-dimensional recommendation data objects based on decentralized crowd sourcing | |
| EP3166025A1 (en) | Identifying content items using a deep-learning model | |
| Nguyen et al. | Meta-learning and personalization layer in federated learning | |
| US20240320507A1 (en) | Device, method, and system for weighted knowledge transfer | |
| US20240144079A1 (en) | Systems and methods for digital image analysis |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| AS | Assignment |
Owner name: SAMSUNG ELECTRONICS COMPANY, LTD., KOREA, REPUBLIC OF Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:QIAN, JIANWEI;SUN, LICHAO;CHEN, XUN;SIGNING DATES FROM 20201027 TO 20201029;REEL/FRAME:054227/0843 |
|
| FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
| STCF | Information on status: patent grant |
Free format text: PATENTED CASE |