CN113923128A - Intelligent coding caching method based on federal reinforcement learning in fog wireless access network - Google Patents
Intelligent coding caching method based on federal reinforcement learning in fog wireless access network Download PDFInfo
- Publication number
- CN113923128A CN113923128A CN202111258088.3A CN202111258088A CN113923128A CN 113923128 A CN113923128 A CN 113923128A CN 202111258088 A CN202111258088 A CN 202111258088A CN 113923128 A CN113923128 A CN 113923128A
- Authority
- CN
- China
- Prior art keywords
- fog
- access point
- local
- global
- fog access
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 33
- 230000002787 reinforcement Effects 0.000 title claims abstract description 20
- 230000005540 biological transmission Effects 0.000 claims abstract description 18
- 238000012549 training Methods 0.000 claims abstract description 12
- 239000013598 vector Substances 0.000 claims description 18
- 238000013528 artificial neural network Methods 0.000 claims description 6
- 230000010354 integration Effects 0.000 claims description 4
- 230000003139 buffering effect Effects 0.000 claims description 3
- 239000003795 chemical substances by application Substances 0.000 claims description 3
- 239000003595 mist Substances 0.000 claims description 3
- 238000005070 sampling Methods 0.000 claims description 3
- 239000000126 substance Substances 0.000 claims description 3
- 238000004891 communication Methods 0.000 description 2
- 238000010295 mobile communication Methods 0.000 description 2
- 238000004088 simulation Methods 0.000 description 2
- 238000013480 data collection Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 239000000463 material Substances 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 239000002699 waste material Substances 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L41/00—Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
- H04L41/14—Network analysis or design
- H04L41/145—Network analysis or design involving simulating, designing, planning or modelling of a network
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L41/00—Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
- H04L41/14—Network analysis or design
- H04L41/147—Network analysis or design for predicting network behaviour
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L67/00—Network arrangements or protocols for supporting network services or applications
- H04L67/01—Protocols
- H04L67/10—Protocols in which an application is distributed across nodes in the network
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Computer Networks & Wireless Communication (AREA)
- Signal Processing (AREA)
- Data Mining & Analysis (AREA)
- Computing Systems (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Evolutionary Computation (AREA)
- General Health & Medical Sciences (AREA)
- Molecular Biology (AREA)
- Artificial Intelligence (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Mobile Radio Communication Systems (AREA)
- Information Transfer Between Computers (AREA)
Abstract
The invention discloses an intelligent coding caching method based on federal reinforcement learning in a fog wireless access network, which comprises the following steps: 1. the cloud center server builds a global prediction model and initializes the global prediction model, and a plurality of distributed learning agents are arranged at the fog access point and initializes the local prediction model; 2. the system observes the global state; 3. each fog access point observes local conditions; 4. the system uses a multicast coding transmission mode to meet the request received by each network edge node according to the implemented content placement decision; 5. each fog access point calculates theoretical time delay and virtual feedback according to local request conditions and stores local learning experience; 6. each fog access point independently performs local model training by using local learning and remembering experience; 7. and each fog access point uploads a local model to the cloud center server every a plurality of time slots to update the global model. The invention can reduce the forward load of the system, reduce the transmission time delay and protect the privacy of users.
Description
Technical Field
The invention belongs to the field of intelligent caching of edge networks in mobile communication systems, and particularly relates to an intelligent coding caching method based on federal reinforcement learning in a fog wireless access network.
Background
With the wide popularization of intelligent devices and the rapid development of mobile applications, mobile communication networks will bear increasingly large traffic loads. In order to handle these increased data traffic, a Fog wireless Access network (Fog Access Point, F-AP) is proposed as a new network architecture, where the Fog wireless Access Point (Fog Access Point, F-AP) is disposed near the network edge of the user and uses its edge storage capability and edge computing capability to reduce the traffic pressure of the cloud center server, thereby improving the user experience. Meanwhile, the coding cache is used as a new cache mode, the local cache and the multicast transmission are effectively combined, and the limited storage space of the edge equipment can be more fully utilized.
However, most of the existing researches on the inconsistent popularity coding caching method are focused on the upper and lower bound analysis of the theoretical performance of the fixed content placement strategy, so that the potential of the coding caching cannot be fully utilized by the non-optimal content placement strategy to reduce the forward load. In addition, the current non-uniform popularity coding and caching method assumes an ideal condition that the popularity of the content is fixed, and ignores the fact that the popularity of the content changes along with time, so that the performance of forwarding load, time delay and the like cannot reach an ideal state in practice. Therefore, there is a need for a coding caching method under the condition of time-varying popularity, which reduces the content transmission delay and improves the stability at the same time, and provides a higher-quality and more reliable communication service for mobile users.
Disclosure of Invention
The invention aims to provide an intelligent coding caching method based on federal reinforcement learning in a fog wireless access network, and aims to solve the technical problems of reducing forward load, reducing request time delay and protecting user privacy.
In order to solve the technical problems, the specific technical scheme of the invention is as follows:
a code caching method based on federal reinforcement learning in a fog wireless access network comprises the following steps:
Step 2, at the end of the time slot t, the system observes the global state s (t), and each fog access point observes the local state s thereofk(t);
Step 3, the system uses the global model to predict and implement the actual code cache content placement decision a (t) according to the observed global state s (t), and each fog access point simultaneously uses the observed local state sk(t) predicting and recording virtual content placement decisions ak(t);
Step 4, in the time slot t +1, the system places a decision a (t) according to the implemented content and uses a multicast coding transmission mode to meet the request received by each fog access point;
step 5, at the end of time slot t +1, the local state is changed into sk(t +1), each fog access point counts the content request condition of the service user in the time slot t +1 and makes a decision according to the virtual placement thereofk(t) calculating theoretical time delay and virtual feedback rk(t) storing local learning experience [ s ]k(t),ak(t),rk(t),sk(t+1)]T;
Step 6, randomly sampling local learning experience of each fog access point for training to realize local model updating;
step 7, every TAIn each time slot, all the fog access points upload the local models to the cloud center server for integration, and download the integrated global model for replacing the local models;
and 8, repeatedly executing the step 2 to the step 7 until the predicted performance fluctuation of the global model in the continuous time slot is less than 5 percent.
Further, the step 1 specifically includes the following steps:
step 1.1, the cloud center server constructs a neural network Q (s, a; theta)G) Where s is the current global state vector, a is the global content placement decision vector, θGAre global network model parameters and are randomly initialized;
step 1.2, forFor the fog access point index set, K is the number of fog access points, and the fog access points K construct a neural network Q(s)k,ak;θk) Wherein s iskFor the current local state vector, akPlacement of decision vectors, θ, for local contentkAre local network model parameters and are randomly initialized.
Further, the step 2 specifically includes the following steps:
step 2.1, at the end of the time slot t, the system observes the global stateWhere a (t-1) is the global buffering decision made by the t-1 slot system,requesting frequency vectors for statistics of all N files in the t time slot system;
step 2.2, at the same time, forFog access point k observes its local stateWherein a isk(t-1) local caching decisions for t-1 time slot records,the requested frequency vectors for all N files in the slot tmouch access point k.
Further, the step 3 specifically includes the following steps:
step 3.1, at the end of the time slot t, the system uses the global model to predict and implement the actual coding cache content placement decision according to the observed global state s (t)Wherein N isc(t) represents the number of files of the encoding buffer, cn(t) < 1 > indicates that file n is selected, cnWhen (t) is 0, it means that the file n is not writtenSelecting;
step 3.2, forFog access point k based on observed local state sk(t) predicting and recording virtual code cache content placement decisions using its local modelWhereinIndicating the number of files of the encoding buffer,the presentation file n is selected to indicate that,it indicates that file n is not selected.
Further, the step 3.1 specifically includes the following steps:
step 3.1.1, at the end of the time slot t, the system predicts the actual coding cache content placement decision by using a global model according to the observed global state s (t):
step 3.1.2, implementing a content placement strategy: by usingRepresents a subset of fog access points, whereinIs the number of elements is LtOf fog access points, LtIs a variable and Lt=KM/Nc(t), M is the buffer size of the fog access point, | □ | represents the number of the collection elements; splitting a file n intoSubfiles of the same sizeWherein Representing a selected set of file indexes, shaped asA combination number representing the number of B elements taken out of A different elements; for theThe placement content of the fog access point k is as follows:
further, the step 3.2 specifically includes the following steps:
step 3.2.1, forFog access point k based on observed local state sk(t) according to a greedy action selection strategy, randomly selecting a virtual content placement decision with a probability of epsilon, and predicting the virtual content placement decision with a local model thereof with a probability of 1-epsilon:
step 3.2.2, record content Placement strategy ak(t) but not practiced;
further, the step 4 specifically includes the following steps:
step 4.1, each fog access point receives a user request;
step 4.2, for the cached request, useRepresenting its index set; by usingRepresenting a fog access point set cached with the request file; letRepresents a set of mist access point subsets, whereinFor the fog access point subset, the cloud center server is connected toThe content of the multicast transmission is:
step 4.3, for the request which is not cached, the cloud center server sends the request to the serverThe request content is unicast transmitted.
Further, the step 5 specifically includes:
step 5.2, making each request collected by fog access pointIs received in fog, where k1,k2,...,kK-1Is a virtually existing virtual fog access point;
step 5.3, collecting station for fog access pointsThe number of buffered requests of the ith request of all K fog access points is expressed asWhereinThe ith request representing a virtual fog access point k' is file n,the ith request, representing a virtual fog access point k', is not file n and thus it is loaded from the forward link byWhereinIs a variable andthe form min (A, B) represents that the smaller value is selected from A, B;
step 5.4, in the time slot t, the theoretical time delay of the fog access point k is expressed asWherein d isfTime delay for complete transmission of one file to the cloud center serveraA time delay for complete transmission of a file from the fog access point to the user;
and 5.5, the virtual feedback comprises the following steps:
Step 5.6, learning experience [ s ]k(t),ak(t),rk(t),sk(t+1)]TStored in a local experience playback pool.
Further, the step 6 specifically includes the following steps:
step 6.1, forFog access point k randomly draws experience s from within local experience playback poolk(j),ak(j),rk(j),sk(j+1)]T;
Step 6.2, training by using gradient descent and carrying out theta on the local modelkAnd (6) updating.
The intelligent coding caching method based on federal reinforcement learning in the fog wireless access network has the following advantages:
1. aiming at the more challenging scene of the time-varying content popularity in the fog wireless access network, the invention tracks the time-varying popularity by using federal reinforcement learning and adaptively makes a content placement decision, can effectively reduce the forward load, reduce the transmission delay and keep the stable performance, and is closer to the reality in the applicable scene.
2. The invention uses a virtual coding cache method, calculates the theoretical time delay and virtual feedback of a virtual content placement strategy by assuming a virtual fog access point, thereby solving the problem of local training data collection caused by the fact that a single fog access point cannot execute coding cache and realizing distributed local training.
3. The method uses federated learning, fully utilizes the edge computing capability of the fog access node, obtains the global prediction model of the coding cache content placement strategy through distributed training and model integration, and avoids repeatedly uploading training data to the cloud center server, thereby reducing the bandwidth resource waste of a fronthaul link and reducing the privacy disclosure risk of a user.
Drawings
Fig. 1 is a schematic flow chart of a code caching method based on federal reinforcement learning in a fog wireless access network according to the present invention;
fig. 2 is a simulation result diagram of average delay performance of the code caching method based on federal reinforcement learning in the fog wireless access network.
Detailed Description
In order to better understand the purpose, structure and function of the present invention, the following describes an intelligent code caching method based on federal reinforcement learning in a fog radio access network in detail with reference to the accompanying drawings.
The embodiment provides a fog radio access network asynchronous request code caching method, as shown in fig. 1, which includes the following steps:
The step 1 specifically comprises the following steps:
step 1.1, the cloud center server constructs a neural network Q (s, a; theta)G) Where s is the current global state vector, a is the global content placement decision vector, θGAre global network model parameters and are randomly initialized;
step 1.2, forFor the fog access point index set, K is the number of fog access points, and the fog access points K construct a neural network Q(s)k,ak;θk) Wherein s iskFor the current local state vector, akPlacement of decision vectors, θ, for local contentkIs a local network model parameter andis randomly initialized.
Step 2, at the end of the time slot t, the system observes the global state s (t), and each fog access point observes the local state s thereofk(t)。
The step 2 specifically comprises the following steps:
step 2.1, at the end of the time slot t, the system observes the global stateWhere a (t-1) is the global buffering decision made by the t-1 slot system,requesting frequency vectors for statistics of all N files in the t time slot system;
step 2.2, at the same time, forFog access point k observes its local stateWherein a isk(t-1) local caching decisions for t-1 time slot records,the requested frequency vectors for all N files in the slot tmouch access point k.
Step 2.3, the system uses the global model to predict and implement the actual code cache content placement decision a (t) according to the observed global state s (t), and each fog access point simultaneously uses the observed local state sk(t) predicting and recording virtual content placement decisions ak(t)。
The step 3 specifically comprises the following steps:
step 3.1, at the end of the time slot t, the system uses a global model to predict the actual code cache content placement decision according to the observed global state s (t)
Wherein N isc(t) represents the number of files of the encoding buffer, cn(t) < 1 > indicates that file n is selected, cnIf (t) ═ 0, then file n is not selected. The content placement policy is then enforced: by usingRepresents a subset of fog access points, whereinIs the number of elements is LtOf fog access points, LtIs a variable and Lt=KM/Nc(t), M is the buffer size of the fog access point, and the form | □ | represents the number of the collection elements; splitting a file n intoSubfiles of the same sizeWherein Representing a selected set of file indexes, shaped asThe number of combinations of B elements taken out of a different elements is indicated. For theThe placement content of the fog access point k is as follows:
step 3.2, at the same time, forFog access point k based on observed local state sk(t) selecting a strategy according to a greedy action, randomly selecting a virtual content placement decision with a probability of epsilon, and predicting the virtual content placement decision with a local model thereof with a probability of 1-epsilon
Record the content placement policy ak(t) but not implemented.
And 4, in the time slot t +1, the system uses a multicast coding transmission mode to meet the request received by each fog access point according to the content placement decision a (t).
The step 4 specifically comprises the following steps:
step 4.1, each fog access point receives a user request;
step 4.2, for the cached request, useRepresenting its index set. By usingRepresenting a set of fog access points that have cached the request file. LetRepresents a set of mist access point subsets, whereinFor the fog access point subset, the cloud center server is connected toThe content of the multicast transmission is:
step 4.3, for the request which is not cached, the cloud center server sends the request to the serverThe request content is unicast transmitted.
Step 5, at the end of time slot t +1, the local state is changed into sk(t +1), each fog access point counts the content request condition of the service user in the time slot t +1 and makes a decision according to the virtual placement thereofk(t) calculating theoretical time delay and virtual feedback rk(t) storing local learning experience [ s ]k(t),ak(t),rk(t),sk(t+1)]T。
The step 5 specifically comprises the following steps:
step 5.2, assume that each request is assembled by a fog access pointIs received in fog, where k1,k2,...,kK-1Is a virtually existing virtual fog access point;
step 5.3, for fog Access Point setThe ith request of all K fog access points in the system, a cached request number tableShown asWhereinThe ith request representing a virtual fog access point k' is file n,the ith request, representing a virtual fog access point k', is not file n and thus it is loaded from the forward link byWhereinIs a variable andthe form min (A, B) represents that the smaller value is selected from A, B;
step 5.4, in the time slot t, the theoretical time delay of the fog access point k is expressed asWherein d isfTime delay for complete transmission of one file to the cloud center serveraA time delay for complete transmission of a file from the fog access point to the user;
and 5.5, the virtual feedback comprises the following steps:
Step 5.6, learning experience [ s ]k(t),ak(t),rk(t),sk(t+1)]TStoring in a local experience playback pool;
and 6, randomly sampling local learning experience of each fog access point for training to realize local model updating.
The step 6 specifically comprises the following steps:
step 6.1, forFog access point k randomly draws experience s from within local experience playback poolk(j),ak(j),rk(j),sk(j+1)]T;
Step 6.2, training by using gradient descent and carrying out theta on the local modelkAnd (6) updating.
Step 7, every TAAnd in each time slot, all the fog access points upload the local models to the cloud center server for integration, and download the integrated global model for replacing the local models.
The step 7 specifically comprises the following steps:
step 7.2, the cloud center server integrates:
wherein DkFor fog access point k at TAThe total number of the extracted learning experiences in the training of the time slots;
And 8, repeatedly executing the step 2 to the step 7 until the predicted performance fluctuation of the global model in the continuous time slot is less than 5 percent.
In the simulation results shown in FIG. 2, LFU (Least frequency used) is a conventional Non-coding Caching method, NUCC (Abdollah Ghaffari Sheshjavani et al, authored in 2020 Wireless Communications and network Conference), Coded Cable Under Non-Uniform Content delivery distribution with Multiple Requests, and APCC (Coded Cable architecture Power distribution, authored in 2018, IEEE Transactions on Information Theory) are two coding methods designed for Non-Uniform Popularity, CentralizedCentralizedmethod is a Centralized extension method of the present invention method, and Propoad is the present patent method. Compared with LFU, NUCC, APCC and other methods, the method realizes more stable and lower-delay content transmission, and has no obvious performance loss compared with a centralized extension method.
It is to be understood that the present invention has been described with reference to certain embodiments, and that various changes in the features and embodiments, or equivalent substitutions may be made therein by those skilled in the art without departing from the spirit and scope of the invention. In addition, many modifications may be made to adapt a particular situation or material to the teachings of the invention without departing from the essential scope thereof. Therefore, it is intended that the invention not be limited to the particular embodiment disclosed, but that the invention will include all embodiments falling within the scope of the appended claims.
Claims (9)
1. A code caching method based on federal reinforcement learning in a fog wireless access network is characterized by comprising the following steps:
step 1, constructing a global model theta on a cloud center serverGAnd initializing, arranging a learning agent at each fog access point and initializing its local model θk;
Step 2, at the end of the time slot t, the system observes the global state s (t), and each fog access point observes the local state thereofsk(t);
Step 3, the system uses the global model to predict and implement the actual code cache content placement decision a (t) according to the observed global state s (t), and each fog access point simultaneously uses the observed local state sk(t) predicting and recording virtual content placement decisions ak(t);
Step 4, in the time slot t +1, the system places a decision a (t) according to the implemented content and uses a multicast coding transmission mode to meet the request received by each fog access point;
step 5, at the end of time slot t +1, the local state is changed into sk(t +1), each fog access point counts the content request condition of the service user in the time slot t +1 and makes a decision according to the virtual placement thereofk(t) calculating theoretical time delay and virtual feedback rk(t) storing local learning experience [ s ]k(t),ak(t),rk(t),sk(t+1)]T;
Step 6, randomly sampling local learning experience of each fog access point for training to realize local model updating;
step 7, every TAIn each time slot, all the fog access points upload the local models to the cloud center server for integration, and download the integrated global model for replacing the local models;
and 8, repeatedly executing the step 2 to the step 7 until the predicted performance fluctuation of the global model in the continuous time slot is less than 5 percent.
2. The code caching method based on federal reinforcement learning in a fog radio access network according to claim 1, wherein the step 1 specifically comprises the following steps:
step 1.1, the cloud center server constructs a neural network Q (s, a; theta)G) Where s is the current global state vector, a is the global content placement decision vector, θGAre global network model parameters and are randomly initialized;
step 1.2, for For the fog access point index set, K is the number of fog access points, and the fog access points K construct a neural network Q(s)k,ak;θk) Wherein s iskFor the current local state vector, akPlacement of decision vectors, θ, for local contentkAre local network model parameters and are randomly initialized.
3. The code caching method based on federal reinforcement learning in a fog radio access network according to claim 1, wherein the step 2 specifically comprises the following steps:
step 2.1, at the end of the time slot t, the system observes the global stateWhere a (t-1) is the global buffering decision made by the t-1 slot system,requesting frequency vectors for statistics of all N files in the t time slot system;
4. The code caching method based on federal reinforcement learning in a fog radio access network according to claim 1, wherein the step 3 specifically comprises the following steps:
step 3.1, at the end of the time slot t, the system uses the global model to predict and implement the actual coding cache content placement decision according to the observed global state s (t)Wherein N isc(t) represents the number of files of the encoding buffer, cn(t) < 1 > indicates that file n is selected, cn(t) ═ 0 then indicates that file n has not been selected;
step 3.2, forFog access point k based on observed local state sk(t) predicting and recording virtual code cache content placement decisions using its local modelWhereinIndicating the number of files of the encoding buffer,the presentation file n is selected to indicate that,it indicates that file n is not selected.
5. The code caching method based on federal reinforcement learning in a fog wireless access network as claimed in claim 4, wherein the step 3.1 specifically comprises the following steps:
step 3.1.1, at the end of the time slot t, the system predicts the actual coding cache content placement decision by using a global model according to the observed global state s (t):
step 3.1.2, implementing a content placement strategy: by usingRepresents a subset of fog access points, whereinIs the number of elements is LtOf fog access points, LtIs a variable and Lt=KM/Nc(t), M is the buffer size of the fog access point, | □ | represents the number of the collection elements; splitting a file n intoSubfiles of the same sizeWherein Representing a selected set of file indexes, shaped asA combination number representing the number of B elements taken out of A different elements; for theThe placement content of the fog access point k is as follows:
6. the code caching method based on federal reinforcement learning in a fog radio access network as claimed in claim 4, wherein said step 3.2 specifically comprises the following steps:
step 3.2.1, forFog access point k based on observed local state sk(t) according to a greedy action selection strategy, randomly selecting a virtual content placement decision with a probability of epsilon, and predicting the virtual content placement decision with a local model thereof with a probability of 1-epsilon:
step 3.2.2, record content Placement strategy ak(t) but not implemented.
7. The code caching method based on federal reinforcement learning in a fog radio access network according to claim 1, wherein the step 4 specifically comprises the following steps:
step 4.1, each fog access point receives a user request;
step 4.2, for the cached request, useRepresenting its index set; by usingRepresenting a fog access point set cached with the request file; letRepresents a set of mist access point subsets, whereinAs fog access pointSubset, cloud-centric Server toThe content of the multicast transmission is:
8. The code caching method based on federal reinforcement learning in a fog radio access network according to claim 1, wherein the step 5 specifically comprises:
step 5.2, making each request collected by fog access pointIs received in fog, where k1,k2,...,kK-1Is a virtually existing virtual fog access point;
step 5.3, for fog Access Point setThe number of buffered requests of the ith request of all K fog access points is expressed asWhereinThe ith request representing a virtual fog access point k' is file n,the ith request, representing a virtual fog access point k', is not file n and thus it is loaded from the forward link byWhereinIs a variable andthe form min (A, B) represents that the smaller value is selected from A, B;
step 5.4, in the time slot t, the theoretical time delay of the fog access point k is expressed asWherein d isfTime delay for complete transmission of one file to the cloud center serveraA time delay for complete transmission of a file from the fog access point to the user;
and 5.5, the virtual feedback comprises the following steps:
Step 5.6, learning experience [ s ]k(t),ak(t),rk(t),sk(t+1)]TStored in a local experience playback pool.
9. The code caching method based on federal reinforcement learning in a fog radio access network according to claim 1, wherein the step 6 specifically comprises the following steps:
step 6.1, forFog access point k randomly draws experience s from within local experience playback poolk(j),ak(j),rk(j),sk(j+1)]T;
Step 6.2, training by using gradient descent and carrying out theta on the local modelkAnd (6) updating.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202111258088.3A CN113923128B (en) | 2021-10-27 | 2021-10-27 | Intelligent coding caching method based on federal reinforcement learning in fog wireless access network |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202111258088.3A CN113923128B (en) | 2021-10-27 | 2021-10-27 | Intelligent coding caching method based on federal reinforcement learning in fog wireless access network |
Publications (2)
Publication Number | Publication Date |
---|---|
CN113923128A true CN113923128A (en) | 2022-01-11 |
CN113923128B CN113923128B (en) | 2024-02-13 |
Family
ID=79243126
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202111258088.3A Active CN113923128B (en) | 2021-10-27 | 2021-10-27 | Intelligent coding caching method based on federal reinforcement learning in fog wireless access network |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN113923128B (en) |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108184252A (en) * | 2017-11-28 | 2018-06-19 | 东南大学 | A kind of asynchronous request code cache method of mist wireless access network |
CN111314862A (en) * | 2020-02-19 | 2020-06-19 | 东南大学 | Caching method with recommendation under deep reinforcement learning in fog wireless access network |
CN111340277A (en) * | 2020-02-19 | 2020-06-26 | 东南大学 | Popularity prediction model and method based on federal learning in fog wireless access network |
CN111935784A (en) * | 2020-08-12 | 2020-11-13 | 重庆邮电大学 | Content caching method based on federal learning in fog computing network |
CN112579544A (en) * | 2020-12-18 | 2021-03-30 | 北京邮电大学 | File caching method and device, electronic equipment and storage medium |
CN113255004A (en) * | 2021-06-16 | 2021-08-13 | 大连理工大学 | Safe and efficient federal learning content caching method |
CN113382059A (en) * | 2021-06-08 | 2021-09-10 | 东南大学 | Collaborative caching method based on federal reinforcement learning in fog wireless access network |
-
2021
- 2021-10-27 CN CN202111258088.3A patent/CN113923128B/en active Active
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108184252A (en) * | 2017-11-28 | 2018-06-19 | 东南大学 | A kind of asynchronous request code cache method of mist wireless access network |
CN111314862A (en) * | 2020-02-19 | 2020-06-19 | 东南大学 | Caching method with recommendation under deep reinforcement learning in fog wireless access network |
CN111340277A (en) * | 2020-02-19 | 2020-06-26 | 东南大学 | Popularity prediction model and method based on federal learning in fog wireless access network |
CN111935784A (en) * | 2020-08-12 | 2020-11-13 | 重庆邮电大学 | Content caching method based on federal learning in fog computing network |
CN112579544A (en) * | 2020-12-18 | 2021-03-30 | 北京邮电大学 | File caching method and device, electronic equipment and storage medium |
CN113382059A (en) * | 2021-06-08 | 2021-09-10 | 东南大学 | Collaborative caching method based on federal reinforcement learning in fog wireless access network |
CN113255004A (en) * | 2021-06-16 | 2021-08-13 | 大连理工大学 | Safe and efficient federal learning content caching method |
Also Published As
Publication number | Publication date |
---|---|
CN113923128B (en) | 2024-02-13 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110312231B (en) | Content caching decision and resource allocation optimization method based on MEC in Internet of vehicles | |
CN111031102B (en) | Multi-user, multi-task mobile edge computing system cacheable task migration method | |
Baştuğ et al. | Proactive small cell networks | |
CN109218747B (en) | Video service classification caching method based on user mobility in super-dense heterogeneous network | |
CN111935784B (en) | Content caching method based on federal learning in fog computing network | |
CN105979274B (en) | The distributed caching laying method of dynamic self-adapting video stream media | |
CN110580199B (en) | Service migration method based on particle swarm in edge computing environment | |
CN112954385A (en) | Self-adaptive shunt decision method based on control theory and data driving | |
CN113255004A (en) | Safe and efficient federal learning content caching method | |
CN114863683B (en) | Heterogeneous Internet of vehicles edge computing unloading scheduling method based on multi-objective optimization | |
CN113012013A (en) | Cooperative edge caching method based on deep reinforcement learning in Internet of vehicles | |
Huang et al. | A more refined mobile edge cache replacement scheme for adaptive video streaming with mutual cooperation in multi-mec servers | |
CN115720237A (en) | Caching and resource scheduling method for edge network self-adaptive bit rate video | |
CN109672626B (en) | Service aggregation method based on queuing delay utilization | |
Wang et al. | An efficient content popularity prediction of privacy preserving based on federated learning and wasserstein gan | |
Liu et al. | Mobility-aware video prefetch caching and replacement strategies in mobile-edge computing networks | |
CN113993168A (en) | Multi-agent reinforcement learning-based cooperative caching method in fog wireless access network | |
CN113923128A (en) | Intelligent coding caching method based on federal reinforcement learning in fog wireless access network | |
Kim et al. | Impacts of device caching of content fractions on expected content quality | |
CN115756873B (en) | Mobile edge computing and unloading method and platform based on federation reinforcement learning | |
CN115587266A (en) | Air-space-ground integrated internet intelligent edge caching method | |
KR101966588B1 (en) | Method and apparatus for receiving video contents | |
CN115484314A (en) | Edge cache optimization method for recommending performance under mobile edge computing network | |
CN115883601A (en) | Method, system, equipment and medium for allocating cooperative resources of Internet of vehicles | |
CN111124298B (en) | Mist computing network content cache replacement method based on value function |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |