CN117313902A - Signal game-based vehicle formation asynchronous federal learning method - Google Patents

Signal game-based vehicle formation asynchronous federal learning method Download PDF

Info

Publication number
CN117313902A
CN117313902A CN202311619800.7A CN202311619800A CN117313902A CN 117313902 A CN117313902 A CN 117313902A CN 202311619800 A CN202311619800 A CN 202311619800A CN 117313902 A CN117313902 A CN 117313902A
Authority
CN
China
Prior art keywords
vehicle formation
server
vehicle
formation
representing
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202311619800.7A
Other languages
Chinese (zh)
Other versions
CN117313902B (en
Inventor
于海洋
梁育豪
赵亚楠
杨阳
任毅龙
崔志勇
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beihang University
Original Assignee
Beihang University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beihang University filed Critical Beihang University
Priority to CN202311619800.7A priority Critical patent/CN117313902B/en
Publication of CN117313902A publication Critical patent/CN117313902A/en
Application granted granted Critical
Publication of CN117313902B publication Critical patent/CN117313902B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F21/00Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F21/50Monitoring users, programs or devices to maintain the integrity of platforms, e.g. of processors, firmware or operating systems
    • G06F21/55Detecting local intrusion or implementing counter-measures
    • G06F21/56Computer malware detection or handling, e.g. anti-virus arrangements
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F21/00Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F21/60Protecting data
    • G06F21/602Providing cryptographic facilities or services
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F21/00Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F21/60Protecting data
    • G06F21/62Protecting access to data via a platform, e.g. using keys or access control rules
    • G06F21/6218Protecting access to data via a platform, e.g. using keys or access control rules to a system of files or objects, e.g. local or distributed file system or database
    • G06F21/6245Protecting personal data, e.g. for financial or medical purposes
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Software Systems (AREA)
  • Computer Security & Cryptography (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Computer Hardware Design (AREA)
  • Health & Medical Sciences (AREA)
  • Bioethics (AREA)
  • Medical Informatics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Mathematical Physics (AREA)
  • Data Mining & Analysis (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Computation (AREA)
  • Virology (AREA)
  • Computing Systems (AREA)
  • Databases & Information Systems (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The invention discloses a vehicle formation asynchronous federal learning method based on signal game, which comprises the following steps: grouping vehicles according to the position information to obtain a vehicle formation; the method comprises the steps that a server sends a learning task to the Internet of vehicles equipment and receives participation signals sent by vehicle formation; the server identifies the participation signals based on a first strategy of the signal game to judge whether the vehicle formation is in a positive state; if the server judges that the vehicle formation is in a positive state, the global model is sent to the vehicle formation; the server receives a local model of the vehicle formation trained using the local data set, and aggregates the global model in an asynchronous manner based on the local model. The invention can realize the accurate selection of active participants under the condition of incomplete information, simultaneously reduces the communication expenditure by an asynchronous aggregation mode, improves the training efficiency, and combines the two modes to improve the model training quality.

Description

Signal game-based vehicle formation asynchronous federal learning method
Technical Field
The present invention relates to the field of electrical digital data processing. And more particularly to a signal game based asynchronous federal learning method for vehicle formation.
Background
With the development of technology, intelligent internet-connected vehicles (Intelligent Connected Vehicle, ICV) have attracted considerable attention. In ICV, a large amount of sensor data needs to be processed with the cooperation of a central server, providing a comfortable driving experience for the driver. Since these sensor data contain sensitive information such as speed, acceleration, etc., which are easily edited, forged, and stolen, there are security and privacy problems in transmitting the sensor data to the central server. Federal learning solves the above problems by locally training the local model with proprietary data via ICV and uploading local model parameters. Meanwhile, the method also provides an efficient communication mode for solving the machine learning problem about the distributed data. Because of the above-described advantages of federal learning, it has been widely used in some ICV applications, such as vehicle co-location and traffic flow prediction.
However, the expensive training overhead and Non-independent distribution (Non-Independent and Identically Distributed, non-IID) data from a single ICV remain a challenge to be addressed by large-scale application of federal learning in the internet of vehicles. In particular, since sensor data involved in training has different distributions, such Non-IID data can lead to bias in global model aggregation, degrading model accuracy. In terms of communication overhead, the traditional synchronous federation learning method needs to wait for all nodes to upload the model and then aggregate, which greatly increases time overhead. Furthermore, each ICV requires multiple rounds of interaction with a central server, which inevitably creates a significant amount of communication overhead in a large-scale vehicle scenario, thereby reducing the efficiency of the training process.
Furthermore, in the real world, there are some malicious participants that may disrupt the overall model training. Therefore, it is necessary to effectively exclude malicious nodes before training. Existing participant selection strategies are mostly based on reputation models, the nature of the nodes is judged by quantifying subjective and objective reputations, considering that the selector can observe complete information of the vehicle under default conditions. However, the central server often has only incomplete information in the real world. Therefore, it is another challenge to design an effective participant selection strategy to exclude unreliable members in incomplete conditions.
Disclosure of Invention
The invention is based on the above-mentioned demand of the prior art, the technical problem to be solved by the invention is to provide a vehicle formation asynchronous federal learning method based on signal game to realize the accurate selection of active participants under the condition of incomplete information, and simultaneously, the communication overhead is reduced in an asynchronous aggregation mode, the training efficiency is improved, and the combination of the two can improve the model training quality.
In order to solve the problems, the invention is realized by adopting the following technical scheme:
provided is a vehicle formation asynchronous federal learning method based on signal gaming, comprising: grouping vehicles according to the position information to obtain a vehicle formation; the method comprises the steps that a server sends a learning task to the Internet of vehicles equipment and receives participation signals sent by vehicle formation; the server identifies the participation signals based on a first strategy of signal gaming to determine whether the vehicle formation is in a positive state, including: a server and vehicle formation are taken as persons in a bureau to construct a game model; solving the game model to obtain a first strategy of perfect Bayesian equilibrium, wherein the first strategy comprises game equilibrium on four solution spaces; determining a state of a vehicle formation according to the first strategy, including: according toAnd->Determining the boundary value of the high reward +.>Boundary value of low reward +.>Wherein, the method comprises the steps of, wherein,m h representing the received first consideration proposal signal,m l representing a received second reward proposal signal, the first reward proposal signal being higher than the second reward proposal signal,Drepresenting the cost of recovering to the previous global aggregate model after receiving the formation of the malicious vehicle,E s representing the impact of receiving a trained local model of a malicious vehicle formation,E p representing the effect of the server receiving a trained local model of the active vehicle fleet; determining a priori beliefsμAnd according to the boundary value of the high reward +.>Boundary value of low reward +.>Determining a priori beliefs pertaining to a solution space, the priori beliefsμIncluding at the time of receivingm h A priori belief that the time server believes that the vehicle formation is of the positive typeμ h And on receipt ofm l A priori belief that the time server believes that the vehicle formation is of the positive typeμ l The method comprises the steps of carrying out a first treatment on the surface of the By comparison in the corresponding solution spacem h -m l And (3) withFm h And (3) withE s -Dm l And (3) withE s -DAnd (3) withγm l And C%θ p )-Fm l And C%θ p )、m h And C%θ p )+Fm l And (3) withE p m h And (3) withE p γAnd (3) withμ h γAnd (3) withμ l To determine an equilibrium solution for the vehicle formation, wherein,Frepresents the camouflage cost of a plague vehicle formation,γprobability value representing that the vehicle formation belongs to the positive type C #θ p ) Training overhead representing active vehicle formation; if the server judges that the vehicle formation is in a positive state, the global model is sent to the vehicle formation; the server receives a local model of the vehicle formation trained using the local data set, and aggregates the global model in an asynchronous manner based on the local model.
Optionally, the game model constructed by the server and the vehicle formation for the persons in the office includes: building a signal gaming model between a server and a vehicle formationThe method comprises the steps of carrying out a first treatment on the surface of the Wherein N represents the collection of persons in the office, +.>,/>,/>,/>,/>Representing a set of types for both parties of the game->Representing a set of types of servers, +.>A set of types representing a formation of vehicles,θ h representing classes of serversThe method comprises the steps of,θ p indicating the positive type of vehicle formation,θ s representing malicious types of vehicle platoon, Pr{θ=θ p -representing natural probabilities of formation belonging to the positive type, Pr{θ=θ s Natural probability of formation belonging to malicious type, 1-γProbability values representing that the formation belongs to a malicious type, m= {m h ,m l M represents a signal set, n= {π a ,π r And, pi represents the set of policies,π a indicating that the server receives a request for vehicle formation,π r representing that the server refuses the request for vehicle formation; u represents a reward set, U= { U r ,U v },U r Representing server profit, U v Representing vehicle formation benefits.
Optionally, the reward set includes: when a=θ p ,B=m h ,C=π a When U r =E p -m h ,U v =m h -C(θ p ) The method comprises the steps of carrying out a first treatment on the surface of the When a=θ p ,B=m l ,C=π a When U r =E p -m l ,U v =m l -C(θ p ) The method comprises the steps of carrying out a first treatment on the surface of the When a=θ p ,B=m h ||m l ,C=π r When U r =0,U v =0; when a=θ s ,B=m h ,C=π a When U r =E s -m h -D,U v =m h -C(θ p )-FThe method comprises the steps of carrying out a first treatment on the surface of the When a=θ s ,B=m h ,C=π r When U r =0,U v =-FThe method comprises the steps of carrying out a first treatment on the surface of the When a=θ s ,B=m l ,C=π a When U r =E s -m l -D,U v =m l -C(θ S ) The method comprises the steps of carrying out a first treatment on the surface of the When a=θ s ,B=m l ,C=π r When U r =0,U v =0, wherein, C%θ S ) The training overhead of the malicious vehicle formation is represented, A is the type of the vehicle formation, B is a reward proposal of the vehicle formation received by the server, and C is a decision of the server.
Optionally, according to the boundary value of the high rewardBoundary value of low reward +.>Determining a solution space to which a priori belief belongs, including +.> Wherein,D 1 a first one of the solution spaces is represented,D 2 a second one of the solution spaces is represented,D 3 a third one of the solutions is represented by a third one of the solutions,D 4 representing a fourth solution space.
Optionally, the first strategy includes game balancing on four solutions, including: if it isThe server is balanced +.>The method comprises the steps of carrying out a first treatment on the surface of the Judging whether the first limit condition is satisfied>,/>,/>,/>If so, determine the equalization solution of the vehicle formation +.>The method comprises the steps of carrying out a first treatment on the surface of the Judging whether the second limiting condition is satisfied,/>,/>If so, determine the equalization solution of the vehicle formation +.>The method comprises the steps of carrying out a first treatment on the surface of the If->The server is balanced +.>The method comprises the steps of carrying out a first treatment on the surface of the Judging whether the third limiting condition is satisfied,/>,/>,/>If so, determine the balance of the vehicle formationSolutionThe method comprises the steps of carrying out a first treatment on the surface of the Judging whether the fourth limitation condition is satisfied>,/>,/>If so, determine the equalization solution of the vehicle formation +.>The method comprises the steps of carrying out a first treatment on the surface of the If->Equalization of the serverThe method comprises the steps of carrying out a first treatment on the surface of the Judging whether the fifth limitation condition is satisfied>,/>,/>If so, determine the equalization solution of the vehicle formation +.>The method comprises the steps of carrying out a first treatment on the surface of the If->The server is balanced +.>The method comprises the steps of carrying out a first treatment on the surface of the Judging whether or not the sixth constraint is satisfied>,/>,/>,/>If so, determine the equalization solution of the vehicle formation +.>The method comprises the steps of carrying out a first treatment on the surface of the Judging whether or not the seventh constraint is satisfied>,/>If so, determine the equalization solution of the vehicle formation +.>
Optionally, the prior beliefs are modified based on gaming equalizationμComprising: if it isWhen the first constraint is satisfied, the server updates the prior beliefs toλ={λ h ,λ l }, whereinλ h =μ h ,/>The method comprises the steps of carrying out a first treatment on the surface of the When the second constraint is satisfied, the server updates the prior beliefs toλ={λ h ,λ l }, whereinλ h =1,λ l =0; if->When the third constraint is satisfied, the server updates the prior beliefs toλ={λ h ,λ l }, whereinλ h =μ h ,/>The method comprises the steps of carrying out a first treatment on the surface of the When the fourth constraint is met, the server updates the a priori beliefs toλ={λ h ,λ l }, whereinλ h =1,λ l =0; if->When the fifth constraint is met, the server updates the a priori beliefs toλ={λ h ,λ l }, whereinλ h =μ h ,/>The method comprises the steps of carrying out a first treatment on the surface of the If->When the sixth constraint is met, the server updates the a priori beliefs toλ={λ h ,λ l }, wherein->λ l =μ l The method comprises the steps of carrying out a first treatment on the surface of the When the seventh constraint is met, the server updates the a priori beliefs toλ={λ h ,λ l }, whereinλ h =1,λ l =0。
Optionally, the local modelBased on training formula->Obtaining; wherein,w r-1 global model representing the round on the server that the vehicle formation received,/for>Is a strong convex function->As a function of the local loss,ρis the parameter of the ultrasonic wave to be used as the ultrasonic wave,D p representing the sum of the vehicle data sets within the formation.
Optionally, the server receives a local model of the vehicle formation trained using the local data set, and aggregates the global model in an asynchronous manner based on the local model, wherein the formula includes:wherein,rthe polymerization run is represented by the number of polymerization runs,w r a global model representing the current turn server,w r-1 a global model representing a round on a server, < +.>Representation modelw r-1 And model->Is used as a similarity function of the (c) for the (c),zrepresenting the super parameter.
Optionally, the vehicle formation trains the global model with a local data set to obtain a local model, including: vehicle formation utilizes a random gradient descent method to find a locally optimal solution, based on a Mini-batch gradient descent method to control the number of training samples.
Optionally, if the server determines that the vehicle formation is not in a positive state, the server disregards the vehicle formation and waits for participation signals of other vehicle formations.
Compared with the prior art, the invention provides a vehicle formation asynchronous federal learning method based on signal game. By building an asynchronous federal learning framework based on vehicle formation, the vehicle formation is made a client for federal learning that provides diversified and sufficient data through secure data sharing within the formation while aggregating computing resources to reduce training overhead. Then, asynchronous optimization is performed, and the server immediately aggregates the global model after uploading the trained model for each formation. The invention further establishes a first strategy based on the signal game to judge the participants, in the strategy, the server is a receiver of the signal, the vehicle formation is a sender of the signal, the participants can be selected under the condition of incomplete information, the server can obtain the optimal signals of the vehicle formation under different conditions through perfect Bayesian equilibrium, the privacy information of the user is ensured, and then the server can accurately judge whether the vehicle formation is positive or malicious through signal identification. The asynchronous federal learning and the first strategy based on the signal game can improve the model training quality, and when the server receives the participation training signals sent by the vehicle formation each time, the server judges the nature of the vehicle formation according to the selection strategy, so that malicious attack in the federal learning process is prevented, and the federal learning result is destroyed.
Drawings
In order to more clearly illustrate the embodiments of the present description or the technical solutions in the prior art, the drawings that are required in the embodiments or the description of the prior art will be briefly described below, and it is obvious that the drawings in the following description are only some embodiments described in the embodiments of the present description, and other drawings may be obtained according to these drawings for a person having ordinary skill in the art.
FIG. 1 is a flow chart of a method for asynchronous federal learning of vehicle formation based on signal gaming provided in an embodiment of the present invention;
FIG. 2 is a flowchart for determining whether a vehicle formation is in a positive state in a signal game-based vehicle formation asynchronous federal learning method according to an embodiment of the present invention;
FIG. 3 is a flow chart of determining a vehicle formation status according to a first strategy in a signal game-based vehicle formation asynchronous federal learning method according to an embodiment of the present invention.
Detailed Description
For the purpose of making the objects, technical solutions and advantages of the embodiments of the present invention more apparent, the technical solutions of the embodiments of the present invention will be clearly and completely described below with reference to the accompanying drawings in the embodiments of the present invention, and it is apparent that the described embodiments are some embodiments of the present invention, but not all embodiments of the present invention. All other embodiments, which can be made by those skilled in the art based on the embodiments of the invention without making any inventive effort, are intended to be within the scope of the invention.
For the purpose of facilitating an understanding of the embodiments of the present invention, reference will now be made to the following description of specific embodiments, taken in conjunction with the accompanying drawings, which are not intended to limit the scope of the invention.
Example 1
In this embodiment, the local model refers to a model trained on each participant, and is trained using data local to that participant, including characteristics and patterns of the participant's local data. The global model is a global model obtained by aggregating local models of participants during federal learning. To ensure privacy protection and data security, federal learning typically employs security aggregation algorithms, such as cryptographic aggregation, differential privacy, etc., to aggregate the local models of participants in an encrypted or privacy-preserving manner. By aggregation, the local information of the individual models of the participants is integrated into a global model, which contains the collective knowledge and characteristics of all participants in federal learning for subsequent reasoning or prediction.
The embodiment provides a vehicle formation asynchronous federal learning method based on signal game, as shown in fig. 1, comprising the following steps:
s1: and grouping the vehicles according to the position information to obtain a vehicle formation.
Assuming that there are several vehicles within the system, the set of numbers may be noted as v= {1, …,n},nrepresenting the number of vehicle numbers, which, depending on the geographical distance, respectively constitute a number of vehicle platoons p= {1, …,pthe number of vehicles in each formation is denoted as M 0 ={m 1 ,…,m i ,…m p Simultaneously, any one of the vehicles is formed into a teampThe number set of the vehicle is recorded as. Each vehicle in the formation has its own data set, and the data in the data sets are often different from each other, for example, using a high-definition camera of the vehicle to capture license plate number information. Can belong to a vehiclevIs expressed as a dataset of (a)For two different vehiclesvAnduusually +.>. The vehicle data within a formation is securely shared by Dedicated Short-range communications, such as the vehicle-mounted wireless communication system (DSRC), and thus a data set of the formationD p Is the sum of the vehicle data sets within the formation: />
S2: the server sends a learning task to the Internet of vehicles equipment and receives participation signals sent by vehicle formation.
The server issues machine learning tasks to the internet of vehicles platform, waiting for the vehicle formation to send participation requests. The vehicle formation is used as a participant to send signals to the server, and the server receives the signals sent by the participant.
S3: the server identifies the engagement signals based on a first policy of signal gaming to determine whether the vehicle formation is in a positive state.
As shown in fig. 2, in this step, it includes:
s30: and constructing a game model by taking the server and the vehicle formation as persons in the office.
Building a signal gaming model between a server and a vehicle formationThe method comprises the steps of carrying out a first treatment on the surface of the Wherein,Nrepresenting the collection of persons in the office->,/>,/>,/>,/>,/>Representing a set of types for both parties of the game->Representing a set of types of servers, +.>A set of types representing a formation of vehicles,θ h indicating the type of server that is to be used,θ p indicating the positive type of vehicle formation,θ s representing malicious types of vehicle platoon, Pr{θ=θ p -representing natural probabilities of formation belonging to the positive type, Pr{θ=θ s Natural probability of formation belonging to malicious type, 1-γProbability values representing that the formation belongs to a malicious type, m= {m h ,m l M represents a signal set, n= {π a ,π r And, pi represents the set of policies,π a indicating that the server receives a request for vehicle formation,π r representing that the server refuses the request for vehicle formation; u represents a reward set, U= { U r ,U v },U r Representing server profit, U v Representing vehicle formation benefits.
The details are as follows:
the office set of people is denoted as n= { SER, PL }. The present embodiment defines the server SER and the vehicle formation PL as two persons in the game process, wherein the server is the receiver of the signal and the vehicle formation is the sender of the signal, the server decides its decision by receiving the signal of the vehicle formation.
The type set is recorded asWherein->,/>. The type set describes all possible types of both parties to the game. There is only one type of server, and there are both active and malicious types of vehicle formation. The natural probability of which type of vehicle formation belongs to can be expressed as: />,/>
The signal set is denoted as m= {m h ,m l }. The signal is the information that the vehicle formation uses to send their personal information and their expected rewards to the server, the signal set consists of two offers,m h representing the received first consideration proposal signal,m l representing a received second reward proposal signal, the first reward proposal signal being higher than the second reward proposal signal.
The policy set is noted as pi= {π a ,π r }. The strategy is a decision made by the server according to its belief upon receipt of a signal for vehicle formation. The server may decide to accept or reject the request sent by the vehicle fleet.
The reward set is noted as u= { U r ,U v }. The notation C represents the training overhead, which is a quantity related to the type of vehicle formation, namely C #θ p ) And C%θ S ) Wherein C is%θ p ) Representing training overhead of active vehicle formation C #θ S ) Training for representing formation of malicious vehiclesOverhead is exercised. Assuming that the malicious vehicle formation has smaller training overhead than positive, C #θ p )>C(θ S ). The malicious formation requires some masquerading to deceive the server when sending the signal, the cost required for masquerading is defined as the embodimentF. If a team participates in training, it will inevitably have an impact on the whole model. The present embodiment quantifies the impact of a trained local model receiving a malicious vehicle formation asE s Quantifying the impact of a trained local model that has been actively queued asE p . If a local model of a malicious vehicle formation is accepted for global aggregation, which may lead to a decrease in the accuracy of the global model, it is costly for the server to revert to the previous global aggregation model, which is defined in this embodiment asD
The remuneration set of server and vehicle platoons under various decision possibilities includes: when a=θ p ,B=m h ,C=π a When U r =E p -m h ,U v =m h -C(θ p ) The method comprises the steps of carrying out a first treatment on the surface of the When a=θ p ,B=m l ,C=π a When U r =E p -m l ,U v =m l -C(θ p ) The method comprises the steps of carrying out a first treatment on the surface of the When a=θ p ,B=m h ||m l ,C=π r When U r =0,U v =0; when a=θ s ,B=m h ,C=π a When U r =E s -m h -D,U v =m h -C(θ p )-FThe method comprises the steps of carrying out a first treatment on the surface of the When a=θ s ,B=m h ,C=π r When U r =0,U v =-FThe method comprises the steps of carrying out a first treatment on the surface of the When a=θ s ,B=m l ,C=π a When U r =E s -m l -D,U v =m l -C(θ S ) The method comprises the steps of carrying out a first treatment on the surface of the When a=θ s ,B=m l ,C=π r When U r =0,U v =0, a is the type of vehicle formation, B is the consideration proposal of the vehicle formation received by the server, C is the decision of the server, and is specifically shown in table 1:
the second row of the table shows, for example, when the vehicle is queued as an aggressive typeθ p The server receives a first reward proposal signal sent by the vehicle formationm h While the server decision is to accept requests sent by a vehicle fleetπ a When the server receives the benefit ofE p -m h The benefits of vehicle formation arem h -C(θ p ). The meaning of other rows of the table can be obtained in the same way.
S31: and solving the game model to obtain a first strategy of perfect Bayesian equilibrium.
Defining boundary values for high rewardsBoundary value of low reward +.>The a priori belief of the server may be written asμ={μ h ,μ l -wherein the a priori beliefsμIncluding at the time of receivingm h A priori belief that the time server believes that the vehicle formation is of the positive typeμ h And on receipt ofm l A priori belief that the time server believes that the vehicle formation is of the positive typeμ l The expression is->,/>。/>The present embodiment distributes all possible prior beliefs in four solution spaces, denoted as follows:
and the third stage is that the vehicle formation performs perfect Bayesian equilibrium on the sub-games under different priori belief distribution solution spaces, and finally, separation and mixed equilibrium on four solution spaces are obtained according to comprehensive analysis of an equilibrium result. After solving, the game equalization on the four solution spaces can be expressed as follows:
when the server is a prioriWhen in use; balancing of servers>
Under the first limitation condition,/>,/>,/>The next step is to obtain a mixed equalization +.>The method comprises the steps of carrying out a first treatment on the surface of the Wherein the equilibrium solution of the vehicle formation ∈ ->The method comprises the steps of carrying out a first treatment on the surface of the At this time, the belief of the server is updated asλ={λ h ,λ l }, whereinλ h =γ,/>
Under the second limitation condition,/>,/>Next, a separation balance is obtainedThe method comprises the steps of carrying out a first treatment on the surface of the Wherein the equilibrium solution of the vehicle formation ∈ ->The method comprises the steps of carrying out a first treatment on the surface of the At this time, the belief of the server is updated asλ={λ h ,λ l }, whereinλ h =1,λ l =0。
When the server is a prioriWhen in use; balancing of servers>
In the third limitation condition,/>,/>,/>The next step is to obtain a mixed equalization +.>The method comprises the steps of carrying out a first treatment on the surface of the Wherein the equilibrium solution of the vehicle formation ∈ ->The method comprises the steps of carrying out a first treatment on the surface of the Meanwhile, the belief of the server is updated at the momentλ={λ h ,λ l }, whereinλ h =μ h ,/>
In the fourth limitation condition,/>,/>The next result is a separation equalization->The method comprises the steps of carrying out a first treatment on the surface of the Wherein the equilibrium solution of the vehicle formation ∈ ->The method comprises the steps of carrying out a first treatment on the surface of the At this time, the belief of the server is updated asλ={λ h ,λ l }, whereinλ h =1,λ l =0。
When the server is a prioriWhen in use; balancing of servers>
In the fifth limitation condition,/>,/>,/>The next step is to obtain a mixed equalization +.>The method comprises the steps of carrying out a first treatment on the surface of the Wherein the equilibrium solution of the vehicle formation ∈ ->The method comprises the steps of carrying out a first treatment on the surface of the At this time, the belief of the server is updated asλ={λ h ,λ l }, whereinλ h =μ h ,/>
When the server is a prioriWhen in use; balancing of servers>
In the sixth limiting condition,/>,/>,/>The next step is to obtain a mixed equalization +.>The method comprises the steps of carrying out a first treatment on the surface of the Wherein the equilibrium solution of the vehicle formation ∈ ->The method comprises the steps of carrying out a first treatment on the surface of the At this time, the belief of the server is updated asλ={λ h ,λ l }, whereinλ l =μ l ,/>
In the seventh limitation,/>The next result is a separation equalization->The method comprises the steps of carrying out a first treatment on the surface of the Wherein the equilibrium solution of the vehicle formation ∈ ->The method comprises the steps of carrying out a first treatment on the surface of the At this time, the belief of the server is updated asλ={λ h ,λ l }, whereinλ h =1,λ l =0。
S32: and determining the state of the vehicle formation according to the first strategy.
As shown in fig. 3, in this step, it includes:
s320: according toAnd->Determining the boundary value of the high reward +.>Boundary value of low reward +.>
S321: determining a priori beliefsμAnd according toAnd->The solution space to which the a priori beliefs belong is determined.
In this embodiment, a priori beliefsμThe game can be obtained by analyzing relevant information and experience such as cognition of the server on vehicle requirements, market information and the like before the game starts. During the game, the participants will update according to the observed information, i.e. the a priori beliefs will update, affecting the policy choices and decisions of the participants in the game.
S322: by comparison in the corresponding solution spacem h -m l And (3) withFm h And (3) withE s -Dm l And (3) withE s -DAnd (3) withγm l And C%θ p )-Fm l And C%θ p )、m h And C%θ p )+Fm l And (3) withE p m h And (3) withE p γAnd (3) withμ h γAnd (3) withμ l And determining an equilibrium solution for the vehicle formation.
If it isThe server is balanced +.>The method comprises the steps of carrying out a first treatment on the surface of the When the first constraint is met, a balanced solution of the vehicle platoon is determined>The method comprises the steps of carrying out a first treatment on the surface of the Determining an equalization solution for the vehicle formation when the second constraint is satisfied
If it isThe server is balanced +.>The method comprises the steps of carrying out a first treatment on the surface of the When the third constraint is met, determining an equalization solution for the vehicle platoon>The method comprises the steps of carrying out a first treatment on the surface of the Determining an equalization solution for the vehicle formation when the fourth constraint is satisfied
If it isThe server is balanced +.>The method comprises the steps of carrying out a first treatment on the surface of the When the fifth constraint is met, determining an equalization solution for the vehicle platoon>
If it isThe server is balanced +.>The method comprises the steps of carrying out a first treatment on the surface of the When the sixth constraint is met, determining an equalization solution for the vehicle platoon>The method comprises the steps of carrying out a first treatment on the surface of the Determining an equalization solution for the vehicle formation when the seventh constraint is satisfied
Further, correcting the prior beliefs based on gaming equalization, comprising:
if it isWhen the first constraint is satisfied, the server updates the prior beliefs toλ={λ h ,λ l }, whereinλ h =μ h ,/>The method comprises the steps of carrying out a first treatment on the surface of the When the second constraint is satisfied, the server updates the prior beliefs toλ={λ h ,λ l }, whereinλ h =1,λ l =0。
If it isWhen the third constraint is satisfied, the server updates the prior beliefs toλ={λ h ,λ l }, whereinλ h =μ h ,/>The method comprises the steps of carrying out a first treatment on the surface of the When the fourth constraint is met, the server updates the a priori beliefs toλ={λ h ,λ l }, whereinλ h =1,λ l =0。
If it isWhen the fifth constraint is met, the server updates the a priori beliefs toλ={λ h ,λ l }, whereinλ h =μ h ,/>
If it isWhen the sixth constraint is met, the server updates the a priori beliefs toλ={λ h ,λ l }, whereinλ l =μ l The method comprises the steps of carrying out a first treatment on the surface of the When the seventh constraint is met, the server updates the a priori beliefs toλ={λ h ,λ l }, whereinλ h =1,λ l =0。
The equalization solution refers to an optimal solution for both the server and the vehicle fleet under different constraints. The present embodiment assumes that vehicle platooning does not make a deceptive decision if the revenue received by the deception server may be less than the revenue not deception. Solving equalization is to obtain decisions which are most favorable for both parties under various conditions and different benefits. Illustratively, whenWhen the sixth constraint is satisfied, the server belief is updated toλ l =μ l ,/>. Wherein,λ l =μ l indicating that the server considers that when receiving the signalm l At the time, there areμ l Is considered to be positive.
S4: and if the server judges that the vehicle formation is in a positive state, the global model is sent to the vehicle formation.
The server judges whether the state of the vehicle formation is positive or malicious through perfect Bayesian equilibrium on four solution spaces; if the vehicle formation is active, the server sends the global model of the present wheel to the vehicle formation; if malicious, the server disregards the vehicle formation and continues to wait for participation requests by other vehicle formations.
S5: the server receives a local model of the vehicle formation trained using the local data set, and aggregates the global model in an asynchronous manner based on the local model.
Random gradient descent (Stochastic Gradient Descent, SGD) methods can be used for training in federal learning, with Mini-batch used to control the number of training samples, neither resulting in overfitting nor divergence. Is provided withAs a function of the loss,frepresenting learning rate, the target global model iswVehicle formationpCan be expressed as:
wherein,D p representing the sum of the vehicle data sets within the formation.
First, vehicle formation is trained using a local data set; assume that the received global model isw r-1 And (2) andis a strong convex function->Is local toThe loss function is a function of the loss,ρbeing a hyper-parameter, the training can be expressed as:
wherein,w r-1 a global model representing a server received by a vehicle fleet,is a strong convex function->As a function of the local loss,ρis a super parameter.
Then, the vehicle formation training is completed to obtain a local modelUploading the local model to a server.
In the federal learning process, the server performs global model aggregation in an asynchronous mode, and the local models sent by any vehicle formation are collected to immediately aggregate the global models. Assume that the aggregation round isrThe global model on the server at this time isw r The global model of the last round of server isw r-1 Receiving a local model sent from a vehicle formationAggregate weight isαThe global aggregation formula is:
the training data quantity is set asWhereinkTotal number of training data tags representing the machine learning model trained with federal learning, and +.>Representative modelwIs labeled among training data of (a)kIs a data count of (a). For two modelsw a Andw b Defining model similarity function SIM #w a ,w b ) The method comprises the following steps:
wherein,representative modelw a Is labeled among training data of (a)nData total of>Representative modelw b Is labeled among training data of (a)nIs a data count of (a).
Recording deviceIs a super parameter, the complete global aggregation formula of the server is:
wherein,representation modelw r-1 Model->Is a similarity function of (1).
And taking the updated global model as a new global model, and repeatedly executing S1-S5 until the global model converges or the global model reaches the target accuracy.
Compared with the prior art, the embodiment of the invention provides a vehicle formation asynchronous federal learning method based on signal game. By building an asynchronous federal learning framework based on vehicle formation, the vehicle formation is made a client for federal learning that provides diversified and sufficient data through secure data sharing within the formation while aggregating computing resources to reduce training overhead. Then, asynchronous optimization is performed, and the server immediately aggregates the global model after uploading the trained model for each formation. Further, the embodiment of the invention establishes a first strategy based on the signal game to judge the participants, in the strategy, the server is a receiver of the signal, the vehicle formation is a sender of the signal, the participants can be selected under the condition of incomplete information, the server can obtain the optimal signals of the vehicle formation under different conditions through perfect Bayesian equilibrium, the privacy information of the user is ensured, and then the server can accurately judge whether the vehicle formation is positive or malicious through signal identification. The asynchronous federal learning and the first strategy based on the signal game can improve the model training quality, and when the server receives the participation training signals sent by the vehicle formation each time, the server judges the nature of the vehicle formation according to the selection strategy, so that malicious attack in the federal learning process is prevented, and the federal learning result is destroyed.
Example 2
A computer readable storage medium having stored thereon a computer program having stored thereon a signal game based vehicle formation asynchronous federal learning program which when executed by a processor implements the steps of a signal game based vehicle formation asynchronous federal learning method of embodiment 1.
The foregoing description of the embodiments has been provided for the purpose of illustrating the general principles of the invention, and is not meant to limit the scope of the invention, but to limit the invention to the particular embodiments, and any modifications, equivalents, improvements, etc. that fall within the spirit and principles of the invention are intended to be included within the scope of the invention.

Claims (10)

1. A signal game based asynchronous federal learning method for vehicle formation, comprising:
grouping vehicles according to the position information to obtain a vehicle formation;
the method comprises the steps that a server sends a learning task to the Internet of vehicles equipment and receives participation signals sent by vehicle formation;
the server identifies the participation signals based on a first strategy of signal gaming to determine whether the vehicle formation is in a positive state, including: a server and vehicle formation are taken as persons in a bureau to construct a game model; solving the game model to obtain a first strategy of perfect Bayesian equilibrium, wherein the first strategy comprises game equilibrium on four solution spaces; determining a state of a vehicle formation according to the first strategy, including: according toAnd->Determining the boundary value of the high reward +.>Boundary value of low reward +.>Wherein, the method comprises the steps of, wherein,m h representing the received first consideration proposal signal,m l representing a received second reward proposal signal, the first reward proposal signal being higher than the second reward proposal signal,Drepresenting the cost of recovering to the previous global aggregate model after receiving the formation of the malicious vehicle,E s representing the impact of receiving a trained local model of a malicious vehicle formation,E p representing the effect of the server receiving a trained local model of the active vehicle fleet; determining a priori beliefsμAnd according to the boundary value of the high reward +.>And low rewardsBoundary value->Determining a priori beliefs pertaining to a solution space, the priori beliefsμIncluding at the time of receivingm h A priori belief that the time server believes that the vehicle formation is of the positive typeμ h And on receipt ofm l A priori belief that the time server believes that the vehicle formation is of the positive typeμ l The method comprises the steps of carrying out a first treatment on the surface of the By comparison in the corresponding solution spacem h -m l And (3) withFm h And (3) withE s -Dm l And (3) withE s -DAnd (3) withγm l And C%θ p )-Fm l And C%θ p )、m h And C%θ p )+Fm l And (3) withE p m h And (3) withE p γAnd (3) withμ h γAnd (3) withμ l To determine an equilibrium solution for the vehicle formation, wherein,Frepresents the camouflage cost of a plague vehicle formation,γprobability value representing that the vehicle formation belongs to the positive type C #θ p ) Training overhead representing active vehicle formation;
if the server judges that the vehicle formation is in a positive state, the global model is sent to the vehicle formation;
the server receives a local model of the vehicle formation trained using the local data set, and aggregates the global model in an asynchronous manner based on the local model.
2. The method for asynchronous federal learning of vehicle formation based on signal gaming according to claim 1, wherein the step of constructing a gaming model for persons in the office with the server and the vehicle formation comprises:
building a signal gaming model between a server and a vehicle formationThe method comprises the steps of carrying out a first treatment on the surface of the Wherein N represents the collection of persons in the office, +.>,/>,/>,/>,/>,/>Representing a set of types for both parties of the game->Representing a set of types of servers, +.>A set of types representing a formation of vehicles,θ h indicating the type of server that is to be used,θ p indicating the positive type of vehicle formation,θ s representing malicious types of vehicle platoon, Pr{θ=θ p -representing natural probabilities of formation belonging to the positive type, Pr{θ=θ s Natural probability of formation belonging to malicious type, 1-γProbability values representing that the formation belongs to a malicious type, m= {m h ,m l M represents a signal set, n= {π a ,π r And, pi represents the set of policies,π a indicating that the server receives a request for vehicle formation,π r representing that the server refuses the request for vehicle formation; u represents a reward set, U= { U r ,U v },U r Representing server profit, U v Representing vehicle formation benefits.
3. A method of asynchronous federal learning of vehicle formation based on signal gaming according to claim 2, wherein the set of rewards comprises: when a=θ p ,B=m h ,C=π a When U r =E p -m h ,U v =m h -C(θ p ) The method comprises the steps of carrying out a first treatment on the surface of the When a=θ p ,B=m l ,C=π a When U r =E p -m l ,U v =m l -C(θ p ) The method comprises the steps of carrying out a first treatment on the surface of the When a=θ p ,B=m h ||m l ,C=π r When U r =0,U v =0; when a=θ s ,B=m h ,C=π a When U r =E s -m h -D,U v =m h -C(θ p )-FThe method comprises the steps of carrying out a first treatment on the surface of the When a=θ s ,B=m h ,C=π r When U r =0,U v =-FThe method comprises the steps of carrying out a first treatment on the surface of the When a=θ s ,B=m l ,C=π a When U r =E s -m l -D,U v =m l -C(θ S ) The method comprises the steps of carrying out a first treatment on the surface of the When a=θ s ,B=m l ,C=π r When U r =0,U v =0, wherein, C%θ S ) The training overhead of the malicious vehicle formation is represented, A is the type of the vehicle formation, B is a reward proposal of the vehicle formation received by the server, and C is a decision of the server.
4. A method of asynchronous federal learning of vehicle formation based on signal gaming according to claim 3,
according to the boundary value of the high rewardBoundary value of low reward +.>Determining a solution space to which the prior beliefs belong, including, wherein,D 1 a first one of the solution spaces is represented,D 2 a second one of the solution spaces is represented,D 3 a third one of the solutions is represented by a third one of the solutions,D 4 representing a fourth solution space.
5. A method of asynchronous federal learning of signal-game-based vehicle formation according to claim 4, wherein the first strategy comprises four subspace-based game balancing, comprising:
if it isThe server is balanced +.>
Judging whether the first limiting condition is satisfied,/>,/>,/>If so, determine the equalization solution of the vehicle formation +.>
Judging whether the second limiting condition is satisfied,/>,/>If so, determine the equalization solution of the vehicle formation +.>
If it isThe server is balanced +.>
Judging whether the third limiting condition is satisfied,/>,/>,/>If so, determine the equalization solution of the vehicle formation +.>
Judging whether the fourth limiting condition is satisfied,/>,/>If so, determine the equalization solution of the vehicle formation +.>
If it isThe server is balanced +.>
Judging whether the fifth limiting condition is satisfied,/>,/>,/>If so, determine the equalization solution of the vehicle formation +.>
If it isThe server is balanced +.>
Judging whether the sixth limiting condition is satisfied,/>,/>,/>If so, determine the equalization solution of the vehicle formation +.>
Judging whether the seventh limiting condition is satisfied,/>If so, determining an equalization solution for the vehicle formation
6. The signal game-based vehicle formation asynchronous federal learning method of claim 5, wherein the prior beliefs are modified based on game balancingμComprising:
if it isWhen the first constraint is satisfied, the server updates the prior beliefs toλ={λ h ,λ l }, whereinλ h =μ h The method comprises the steps of carrying out a first treatment on the surface of the When the second constraint is satisfied, the server updates the prior beliefs toλ={λ h ,λ l }, whereinλ h =1,λ l =0;
If it isWhen the third constraint is satisfied, the server updates the prior beliefs toλ={λ h ,λ l }, whereinλ h =μ h The method comprises the steps of carrying out a first treatment on the surface of the When the fourth constraint is met, the server updates the a priori beliefs toλ={λ h ,λ l }, whereinλ h =1,λ l =0;
If it isWhen the fifth constraint is met, the server updates the a priori beliefs toλ={λ h ,λ l }, whereinλ h =μ h
If it isWhen the sixth constraint is met, the server updates the a priori beliefs toλ={λ h ,λ l }, whereinλ l =μ l The method comprises the steps of carrying out a first treatment on the surface of the When the seventh constraint is met, the server updates the a priori beliefs toλ={λ h ,λ l }, whereinλ h =1,λ l =0。
7. The method for asynchronous federal learning of vehicle formation based on signal gaming of claim 4, wherein the local modelBased on training formula->Obtaining; wherein,w r-1 global model representing the round on the server that the vehicle formation received,/for>Is a strong convex function->As a function of the local loss,ρis the parameter of the ultrasonic wave to be used as the ultrasonic wave,D p representing the sum of the vehicle data sets within the formation.
8. The method of asynchronous federal learning of vehicle formation based on signal gaming of claim 7, wherein the server receives a local model of vehicle formation training completion using a local data set, aggregates global models based on the local model in an asynchronous manner, and wherein the formula comprises:
wherein,rthe polymerization run is represented by the number of polymerization runs,w r a global model representing the current turn server,w r-1 a global model representing a round on a server, < +.>Representation modelw r-1 And model->Is used as a similarity function of the (c) for the (c),zrepresenting the super parameter.
9. The method of asynchronous federal learning of vehicle formation based on signal gaming of claim 8, wherein vehicle formation trains the global model with a local data set to obtain a local model, comprising: vehicle formation utilizes a random gradient descent method to find a locally optimal solution, based on a Mini-batch gradient descent method to control the number of training samples.
10. The method of claim 1, wherein if the server determines that the vehicle formation is not in a positive state, the server disregards the vehicle formation and waits for participation signals of other vehicle formations.
CN202311619800.7A 2023-11-30 2023-11-30 Signal game-based vehicle formation asynchronous federal learning method Active CN117313902B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202311619800.7A CN117313902B (en) 2023-11-30 2023-11-30 Signal game-based vehicle formation asynchronous federal learning method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202311619800.7A CN117313902B (en) 2023-11-30 2023-11-30 Signal game-based vehicle formation asynchronous federal learning method

Publications (2)

Publication Number Publication Date
CN117313902A true CN117313902A (en) 2023-12-29
CN117313902B CN117313902B (en) 2024-02-06

Family

ID=89281621

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202311619800.7A Active CN117313902B (en) 2023-11-30 2023-11-30 Signal game-based vehicle formation asynchronous federal learning method

Country Status (1)

Country Link
CN (1) CN117313902B (en)

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110214178A1 (en) * 2009-08-31 2011-09-01 Telcordia Technologies, Inc. System and Method for Detecting and Evicting Malicious Vehicles in a Vehicle Communications Network
CN107181793A (en) * 2017-04-27 2017-09-19 长安大学 The transportation service information forwarding mechanism discussed based on dynamic game
US20180018605A1 (en) * 2016-07-15 2018-01-18 Cummins Inc. Method and apparatus for platooning of vehicles
CN111798002A (en) * 2020-05-31 2020-10-20 北京科技大学 Local model proportion controllable federated learning global model aggregation method
CN112742028A (en) * 2021-01-22 2021-05-04 中国人民解放军国防科技大学 Formation decision method, system, medium and equipment for fighting game
CN115099417A (en) * 2022-06-28 2022-09-23 贵州大学 Multi-factor federal learning incentive mechanism based on Starkeberg game

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110214178A1 (en) * 2009-08-31 2011-09-01 Telcordia Technologies, Inc. System and Method for Detecting and Evicting Malicious Vehicles in a Vehicle Communications Network
US20180018605A1 (en) * 2016-07-15 2018-01-18 Cummins Inc. Method and apparatus for platooning of vehicles
CN107181793A (en) * 2017-04-27 2017-09-19 长安大学 The transportation service information forwarding mechanism discussed based on dynamic game
CN111798002A (en) * 2020-05-31 2020-10-20 北京科技大学 Local model proportion controllable federated learning global model aggregation method
CN112742028A (en) * 2021-01-22 2021-05-04 中国人民解放军国防科技大学 Formation decision method, system, medium and equipment for fighting game
CN115099417A (en) * 2022-06-28 2022-09-23 贵州大学 Multi-factor federal learning incentive mechanism based on Starkeberg game

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
刘润坤,于海洋: "间歇性通信失效下网联自动驾驶队列控制策略", 交通运输系统工程与信息 *
董文远;朱研;王永红;张光华;: "车联网环境下基于重复博弈的恶意车辆节点检测机制", 计算机应用研究, no. 05 *

Also Published As

Publication number Publication date
CN117313902B (en) 2024-02-06

Similar Documents

Publication Publication Date Title
CN111460443A (en) Security defense method for data manipulation attack in federated learning
CN112101404B (en) Image classification method and system based on generation countermeasure network and electronic equipment
CN113408209A (en) Cross-sample federal classification modeling method and device, storage medium and electronic equipment
CN114418109A (en) Node selection and aggregation optimization system and method for federal learning under micro-service architecture
CN112101403B (en) Classification method and system based on federal few-sample network model and electronic equipment
CN116471286A (en) Internet of things data sharing method based on block chain and federal learning
Wang et al. Blockchain-based federated learning in mobile edge networks with application in internet of vehicles
Cai et al. 2cp: Decentralized protocols to transparently evaluate contributivity in blockchain federated learning environments
CN117999562A (en) Method and system for quantifying client contribution in federal learning
Hu et al. Solving the crowdsourcing dilemma using the zero-determinant strategies
CN112085051B (en) Image classification method and system based on weighted voting and electronic equipment
CN117313902B (en) Signal game-based vehicle formation asynchronous federal learning method
You et al. Accuracy degrading: Towards participation-fair federated learning
CN114866272B (en) Multi-round data delivery system of true value discovery algorithm in crowd-sourced sensing environment
CN116187469A (en) Client member reasoning attack method based on federal distillation learning framework
Fu et al. ATWR-SMR: an area-constrained truthful worker recruitment based sensing map recovery scheme for sparse MCS in extreme-environment internet-of-things
CN115481215A (en) Partner prediction method and prediction system based on temporal partner knowledge graph
Gao et al. Multi-source feedback based light-weight trust mechanism for edge computing
CN116541831B (en) Dual defense method based on blockchain and federal learning
Wang et al. Secured and Cooperative Publish/Subscribe Scheme in Autonomous Vehicular Networks
CN114124784B (en) Intelligent routing decision protection method and system based on vertical federation
CN116306986B (en) Federal learning method based on dynamic affinity aggregation and related equipment
CN117557870B (en) Classification model training method and system based on federal learning client selection
Janjua et al. A Blockchain-Enabled Decentralized Gossip Federated Learning Framework
CN116957067B (en) Reinforced federal learning method and device for public safety event prediction model

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant