CN106358308A - Resource allocation method for reinforcement learning in ultra-dense network - Google Patents
Resource allocation method for reinforcement learning in ultra-dense network Download PDFInfo
- Publication number
- CN106358308A CN106358308A CN201510409462.3A CN201510409462A CN106358308A CN 106358308 A CN106358308 A CN 106358308A CN 201510409462 A CN201510409462 A CN 201510409462A CN 106358308 A CN106358308 A CN 106358308A
- Authority
- CN
- China
- Prior art keywords
- strategy
- resource allocation
- formula
- guess
- home enodeb
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04W—WIRELESS COMMUNICATION NETWORKS
- H04W72/00—Local resource management
- H04W72/50—Allocation or scheduling criteria for wireless resources
- H04W72/54—Allocation or scheduling criteria for wireless resources based on quality criteria
- H04W72/542—Allocation or scheduling criteria for wireless resources based on quality criteria using measured or perceived quality
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04W—WIRELESS COMMUNICATION NETWORKS
- H04W52/00—Power management, e.g. TPC [Transmission Power Control], power saving or power classes
- H04W52/04—TPC
- H04W52/18—TPC being performed according to specific parameters
- H04W52/24—TPC being performed according to specific parameters using SIR [Signal to Interference Ratio] or other wireless path parameters
- H04W52/241—TPC being performed according to specific parameters using SIR [Signal to Interference Ratio] or other wireless path parameters taking into account channel quality metrics, e.g. SIR, SNR, CIR, Eb/lo
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04W—WIRELESS COMMUNICATION NETWORKS
- H04W52/00—Power management, e.g. TPC [Transmission Power Control], power saving or power classes
- H04W52/04—TPC
- H04W52/18—TPC being performed according to specific parameters
- H04W52/24—TPC being performed according to specific parameters using SIR [Signal to Interference Ratio] or other wireless path parameters
- H04W52/243—TPC being performed according to specific parameters using SIR [Signal to Interference Ratio] or other wireless path parameters taking into account interferences
- H04W52/244—Interferences in heterogeneous networks, e.g. among macro and femto or pico cells or other sector / system interference [OSI]
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04W—WIRELESS COMMUNICATION NETWORKS
- H04W84/00—Network topologies
- H04W84/18—Self-organising networks, e.g. ad-hoc networks or sensor networks
Landscapes
- Engineering & Computer Science (AREA)
- Computer Networks & Wireless Communication (AREA)
- Signal Processing (AREA)
- Quality & Reliability (AREA)
- Mobile Radio Communication Systems (AREA)
Abstract
A resource allocation method for reinforcement learning in an ultra-dense network is provided. The invention relates to the field of ultra-dense networks in 5G (fifth generation) mobile communications and provides a method for allocating resources between a home node B and a macro node B, between a home node B and another home node B and between a home node B and a mobile user in a dense deployment network; the method is implemented through power control, each femotcell is considered as an intelligent body to jointly adjust transmitting powers of home node Bs, the densely deployed home node Bs are avoided causing severe jamming to a macro node B and an adjacent B when transmitting at maximum powder, and system throughput is maximized; user delay QoS is considered, and traditional 'Shannon capacity' is replaced with 'available capacity' that may ensure user delay; a supermodular game model is utilized such that whole network power distribution gains Nash equilibrium; the reinforcement learning method Q-learning is utilized such that the home node B has learning function, and optimal power distribution can be achieved; by using the resource allocation method, it is possible to effectively improve the system capacity of an ultra-dense network at the premise of satisfying user delay.
Description
Technical field
Present document relates to moving communicating field, especially, the present invention is the method for a kind of resource allocation of super-intensive heterogeneous network (ultra dense network, udn) in the 5th generation (5th-generation) GSM.
Background technology
Nowadays mobile network has been enter into the quick stage of popularization, and meanwhile 5g technology is all energetically being studied in countries in the world, and 5g standard also begins to show up prominently.Take the lead in, using cognitive radio power technology, being automatically determined the frequency range of offer by network, realize the distinguishing feature that multi-internet integration is 5g.The 5g of China works also achieved with preliminary achievement.The main target of 5g network is Consumer's Experience, network is carried out redesigning in terms of capacity, speed, time delay three, optimize.5g network will accommodate substantial amounts of terminal device simultaneously, therefore will jointly meet the demand of mobile service flow growth by higher spectrum reuse efficiency, more frequency spectrum resource and more dense cell deployment.This makes the new challenge of 5g network faces.
Because, in traditional GSM, network design, operation maintenance depend on greatly manually, need to expend a large amount of human and material resources.Therefore, just have self-organizing network (self-organization network,
Son concept), realizes deployment, maintenance and the optimization of network by the self organization ability of communication network.And have the access of many low power nodes in 5g system so that network structure is more complicated, form super-intensive heterogeneous network.Due to the rare of Radio Resource so that many operators wish to make full use of frequency spectrum resource in high band, use double-layer network in following 5g system, there is two kinds of user is grand user and domestic consumer respectively simultaneously.In order to make full use of frequency spectrum resource, double-layer network shares identical frequency range, but have also been introduced co-channel interference simultaneously, and existing technology can not solve cross-layer interference problem present in 5g network well, based on this, the self-optimization techniques in self-organization network are applied among super-intensive network distribute to realize the self-organizing of resource by primary study of the present invention.
In order to realize resource self-organizing distribution, classical intensified learning q-learning algorithm is applied in super-intensive network.Q-learning algorithm is based on discrete state, and discrete variable refers to the size of power in the present invention.The housing choice behavior of q-learning and behavior value function select different strategies respectively, by measuring channel parameter in real time, dynamically carry out resource self-organizing distribution, and then reach the purpose of suppression interference.
Content of the invention
Present invention is generally directed to energy saving resources assignment problem is it is proposed that a kind of resource allocation methods based on q-learning algorithm and equilibria of Supermodular Games in super-intensive self-organizing network.The method optimizes the energy utilization efficiency in network, ensures qos, improves power system capacity.
In order to solve the above problems, the invention provides a kind of Resource Allocation Formula of efficiency optimization:
Step 1: initialization Studying factors, for each state s and each action a, initialize evaluation function, transmission strategy, guess, positive scalar.
Step 2: initialization Home eNodeB state s, transmission power p, signal interference ratio etc..
Step 3: according to transmission strategy, the action of selected active procedure.
Step 4: according to the current signal interference ratio of the feedback information detection of target receiver, by identifying whether current transmission power rank and relatively current signal interference ratio determine subsequent time state more than thresholding.
Step 5: if the current Signal to Interference plus Noise Ratio of domestic consumer is more than threshold value, return (return, i.e. Efficiency Function) is calculated by Reward Program, otherwise, zero setting will be returned.
Step 6: the return value obtaining in step 4 is asked with expectation, using the q more new formula based on guess, obtains new q value, update evaluation function.
Step 7: according to the q value obtaining in step 5, according to greedy strategy, update the strategy of user.
Step 8: using guess more new formula, obtain the guess of other Home eNodeB subsequent time behaviors, make user enter next state.Proceed to step 2.
Step 9: terminate this learning process, the radio resource allocation of each Home eNodeB completes, prepare scheduling of resource next time.
In step 1, each Home eNodeB only allows a user to access, and the setting of q value needs estimation in advance.
In step 3, strategy, i.e. the probability of selection action, the maximum action of select probability.
In step 4, the computing formula of signal interference ratio
In formula,Represent Home eNodeB to the channel gain of internal user, the channel gain of Home eNodeB to other base station user.Represent the interference to Home eNodeB for the grand user., represent the transmission power of this base station and interference base station respectively, refer to Gaussian noise.
In steps of 5, the Efficiency Function of each base station refers to
In formula, refer to the available capacity of Home eNodeB k, under the requirement ensureing a fixed response time qos, wireless channel can reach the transfer rate of maximum.Represent qos delay parameter, bigger, delay requirement is higher.Represent the transmission power to user for the Home eNodeB.Represent the power being consumed in communication network.
Above formula shows that subchannel is linear, and the transmission power obtaining on every sub-channels is linear with general power
(in formula, refer to a frame data transmission duration in conjunction with Shannon public affairs capacity formula, b represents communication bandwidth.Obtain the expression of utility function
In step 6, the more new formula of q value,
In formula, it is Studying factors, there is attenuation characteristic, it is discount factor.Refer to the state in next moment.
In step 7, the more new formula of strategy
Value is the positive parameter of temperature, and value is bigger, and the probability convergence of action is identical.
In step 8, guess more new formula
In formula, refer to the guess of previous moment.Refer to subsequent time, in current state and strategy when taking current action, refer to the strategy of previous moment.
From technical method, this method with maximize each Home eNodeB efficiency as target, consider the same layer interference in super-intensive self-organizing network, cross-layer is disturbed and qos time delay, constantly adjust the transmission power of Home eNodeB using q-learning algorithm, it is finally reached Nash Equilibrium, so that the efficiency of each Home eNodeB is optimized.Both ensure the qos time delay of user, improve the efficiency of Home eNodeB it is achieved that the RRM of home base station network again.
Below by the drawings and specific embodiments, technical scheme is further elaborated.
Brief description
For the elaboration embodiments of the invention that become apparent from and existing technical scheme, below technical scheme is illustrated that the explanation accompanying drawing used in accompanying drawing and description of the prior art does simple introduction, obviously, on the premise of not paying creative work, those of ordinary skill in the art can obtain other accompanying drawings by this accompanying drawing.
Fig. 1 show the system architecture diagram comprising Home eNodeB and macro base station in the embodiment of the present invention;
The flow chart that Fig. 2 show Home eNodeB power distribution in the embodiment of the present invention.
Specific embodiment
Main idea is that, by being simulated to communication environment, set up model, initialize Studying factors, guess, transmission strategy and evaluation function q, the state of detection present channel, state instruction parameter has signal interference ratio, transmission power, status etc..According to transmission policy selection current action, the signal interference ratio detecting is compared with given threshold value, if being more than threshold value, obtain a return, if being less than threshold value, the return obtaining is set to zero, obtain new q value using based on the q more new formula of guess, and according to q value by the strategy of greedy strategy acquisition subsequent time and guess, renewal subsequent time state, and enter next communications status, repeat above-mentioned learning process.Power allocation scheme is assessed as Performance evaluation criterion using q value, finds and make the maximum power allocation scheme of Total Return function in super-intensive subzone network.
Fig. 1 show the system architecture diagram simultaneously including that Home eNodeB is disposed with frequency with macro base station, and it comprises a macro base station, multiple femto base station and its user, grand user.
Step 101: setting learning process initial time t=0.
Step 102: initialization Studying factors, for each state s and each action a, initialize evaluation function, transmission strategy, guess, positive scalar.
Step 103: initialization Home eNodeB state s, transmission power p, signal interference ratio etc..
Step 104: according to transmission strategy, select the action of active procedure.
Step 105: according to the current signal interference ratio of the feedback information detection of target receiver, by identifying whether current transmission power rank and relatively current signal interference ratio determine subsequent time state more than thresholding.
Step 106: if current signal interference ratio is more than thresholding, will pass through Reward Program and calculate a return, otherwise, return is set to zero.
Step 107: in step 4, all return value obtaining are asked expectation, using the q more new formula based on guess, obtain new q value, update evaluation function.
Step 108: by the q value obtaining according to the new transmission strategy of the acquisition of greedy strategy.
Step 109: using guess more new formula, obtain the guess to other Home eNodeB subsequent time behaviors.
Step 110: enter next state determined by step 105, enter next moment t=t+1, and proceed to step 102.
Step 111: terminate this learning process, the radio resource allocation of each Home eNodeB completes, prepare scheduling of resource next time.
Claims (5)
1. a kind of resource allocation methods of the intensified learning in super-intensive network are it is characterised in that comprise the following steps:
Step 1: initialization Studying factors, for each state s and each action a, initialize evaluation function, transmission strategy, guess, positive scalar;
Step 2: initialization Home eNodeB state s, transmission power p, signal interference ratio etc.;
Step 3: according to transmission strategy, the action of selected active procedure;
Step 4: according to the current signal interference ratio of the feedback information detection of target receiver, by identifying whether current transmission power rank and relatively current signal interference ratio determine subsequent time state more than thresholding;
Step 5: if the current Signal to Interference plus Noise Ratio of domestic consumer is more than threshold value, return (return, i.e. Efficiency Function) is calculated by Reward Program, otherwise, zero setting will be returned;
Step 6: the return value obtaining in step 4 is asked with expectation, using the more new formula based on guess q, obtains new q value, update evaluation function;
Step 7: according to the q value obtaining in step 5, according to greedy strategy, update the strategy of user;
Step 8: using guess more new formula, obtain the guess of other Home eNodeB subsequent time behaviors, make user enter next state, proceed to step 2;
Step 9: terminate this learning process, the radio resource allocation of each Home eNodeB completes, prepare scheduling of resource next time.
2. super-intensive network according to claim 1 resource allocation methods it is characterised in that:
In described step 1, using equilibria of Supermodular Games model, the intelligent body that each Home eNodeB is viewed as having fair competition relation tries to achieve optimal objective so as to reach Nash Equilibrium.
3. super-intensive network according to claim 1 resource allocation methods it is characterised in that:
In described step 5, using Efficiency Function as q-learning algorithm evaluation function, Efficiency Function formula is
, the available capacity of its middle finger Home eNodeB k, under the requirement ensureing a fixed response time qos, wireless channel can reach the transfer rate of maximum;Represent qos delay parameter, bigger, delay requirement is higher, represent the transmission power to user for the Home eNodeB, represent the power being consumed in communication network,
Above formula shows that subchannel is linear, and the transmission power obtaining on every sub-channels is linear with general power,
(in formula, refer to a frame data transmission duration in conjunction with Shannon public affairs capacity formula, b represents communication bandwidth, obtains the expression of utility function.
4. super-intensive network according to claim 1 resource allocation methods it is characterised in that:
In described step 8, target household base station requires no knowledge about the strategy to other base station strategies, but the strategy of other base stations is guessed, adjusts the strategy of oneself so as to convergence Nash Equilibrium, its guess more new formula is in formula, refers to the guess of previous moment,Refer to subsequent time, in current state and strategy when taking current action, refer to the strategy of previous moment.
5. super-intensive network according to claim 1 resource allocation methods it is characterised in that:
In described step 6, based on the q more new formula of guess it is:
In formula, it is Studying factors, there is attenuation characteristic, it is discount factor, refers to the state in next moment, go to assess the quality of return value in this iteration with q value.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201510409462.3A CN106358308A (en) | 2015-07-14 | 2015-07-14 | Resource allocation method for reinforcement learning in ultra-dense network |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201510409462.3A CN106358308A (en) | 2015-07-14 | 2015-07-14 | Resource allocation method for reinforcement learning in ultra-dense network |
Publications (1)
Publication Number | Publication Date |
---|---|
CN106358308A true CN106358308A (en) | 2017-01-25 |
Family
ID=57842125
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201510409462.3A Pending CN106358308A (en) | 2015-07-14 | 2015-07-14 | Resource allocation method for reinforcement learning in ultra-dense network |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN106358308A (en) |
Cited By (26)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106954234A (en) * | 2017-04-24 | 2017-07-14 | 东南大学 | User's connection and virtual resource allocation method in a kind of super-intensive heterogeneous network |
CN107333333A (en) * | 2017-08-10 | 2017-11-07 | 北京邮电大学 | A kind of resource allocation methods based on user traffic flow |
CN107734700A (en) * | 2017-08-23 | 2018-02-23 | 南京邮电大学 | The network resource allocation method of user's QoS demand is ensured in super-intensive network |
CN108112082A (en) * | 2017-12-18 | 2018-06-01 | 北京工业大学 | A kind of wireless network distributed freedom resource allocation methods based on statelessly Q study |
CN108124306A (en) * | 2017-12-29 | 2018-06-05 | 河海大学 | A kind of ultra dense set network spectrum efficiency optimization method of cross operator |
CN108307510A (en) * | 2018-02-28 | 2018-07-20 | 北京科技大学 | A kind of power distribution method in isomery subzone network |
CN108307412A (en) * | 2018-02-08 | 2018-07-20 | 北京邮电大学 | The super-intensive network interferences management method based on grouping game of user-center |
CN108401254A (en) * | 2018-02-27 | 2018-08-14 | 苏州经贸职业技术学院 | A kind of wireless network resource distribution method based on intensified learning |
CN108430103A (en) * | 2018-02-27 | 2018-08-21 | 北京科技大学 | A kind of software definition mist car networking resource allocation methods and its system |
CN108718463A (en) * | 2018-05-09 | 2018-10-30 | 南京邮电大学 | A kind of resource allocation methods based on Multiple Time Scales collaboration optimization under H-CRAN |
CN108834158A (en) * | 2018-05-02 | 2018-11-16 | 北京交通大学 | A kind of interference management method for super-intensive networking |
CN109362113A (en) * | 2018-11-06 | 2019-02-19 | 哈尔滨工程大学 | A kind of water sound sensor network cooperation exploration intensified learning method for routing |
CN109474980A (en) * | 2018-12-14 | 2019-03-15 | 北京科技大学 | A kind of wireless network resource distribution method based on depth enhancing study |
CN109496305A (en) * | 2018-08-01 | 2019-03-19 | 东莞理工学院 | Nash equilibrium strategy on continuous action space and social network public opinion evolution model |
CN109617991A (en) * | 2018-12-29 | 2019-04-12 | 东南大学 | Based on value function approximate super-intensive heterogeneous network small station coding cooperative caching method |
CN109743778A (en) * | 2019-01-14 | 2019-05-10 | 长沙学院 | A kind of resource allocation optimization method and system based on intensified learning |
CN109982434A (en) * | 2019-03-08 | 2019-07-05 | 西安电子科技大学 | Wireless resource scheduling integrated intelligent control system and method, wireless communication system |
CN110049565A (en) * | 2019-04-18 | 2019-07-23 | 天津大学 | A kind of 5G network power distribution method based on available capacity |
CN110392377A (en) * | 2019-07-19 | 2019-10-29 | 北京信息科技大学 | A kind of 5G super-intensive networking resources distribution method and device |
CN110472764A (en) * | 2018-05-09 | 2019-11-19 | 沃尔沃汽车公司 | Coordinate the method and system serviced in many ways using half cooperation nash banlance based on intensified learning |
CN110770761A (en) * | 2017-07-06 | 2020-02-07 | 华为技术有限公司 | Deep learning system and method and wireless network optimization using deep learning |
CN110972309A (en) * | 2019-11-08 | 2020-04-07 | 厦门大学 | Ultra-dense wireless network power distribution method combining graph signals and reinforcement learning |
CN111988834A (en) * | 2020-08-27 | 2020-11-24 | 长安大学 | Heterogeneous Internet of vehicles network selection method, system and device |
CN112351433A (en) * | 2021-01-05 | 2021-02-09 | 南京邮电大学 | Heterogeneous network resource allocation method based on reinforcement learning |
CN113038583A (en) * | 2021-03-11 | 2021-06-25 | 南京南瑞信息通信科技有限公司 | Inter-cell downlink interference control method, device and system suitable for ultra-dense network |
CN113490219A (en) * | 2021-07-06 | 2021-10-08 | 香港中文大学(深圳) | Dynamic resource allocation method for ultra-dense networking |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102711123A (en) * | 2012-06-08 | 2012-10-03 | 北京邮电大学 | Method for inhibiting interference in intensively-deployed household base station network |
CN102811443A (en) * | 2012-07-27 | 2012-12-05 | 南京邮电大学 | Interference management method based on spectrum allocation and power control in family base station system |
CN104023384A (en) * | 2014-05-28 | 2014-09-03 | 北京邮电大学 | Double-layer network resource allocation method in consideration of time delay limitation in dense home base station deployment scene |
CN104113903A (en) * | 2014-07-31 | 2014-10-22 | 厦门大学 | Interactive cognitive learning based downlink power adjusting method and device |
WO2015054264A1 (en) * | 2013-10-08 | 2015-04-16 | Google Inc. | Methods and apparatus for reinforcement learning |
-
2015
- 2015-07-14 CN CN201510409462.3A patent/CN106358308A/en active Pending
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102711123A (en) * | 2012-06-08 | 2012-10-03 | 北京邮电大学 | Method for inhibiting interference in intensively-deployed household base station network |
CN102811443A (en) * | 2012-07-27 | 2012-12-05 | 南京邮电大学 | Interference management method based on spectrum allocation and power control in family base station system |
WO2015054264A1 (en) * | 2013-10-08 | 2015-04-16 | Google Inc. | Methods and apparatus for reinforcement learning |
CN104023384A (en) * | 2014-05-28 | 2014-09-03 | 北京邮电大学 | Double-layer network resource allocation method in consideration of time delay limitation in dense home base station deployment scene |
CN104113903A (en) * | 2014-07-31 | 2014-10-22 | 厦门大学 | Interactive cognitive learning based downlink power adjusting method and device |
Non-Patent Citations (2)
Title |
---|
WENPENG JING,ZHAOMING LU,ZHICAI ZHANG,HAIJUN ZHANG,ETC: "Energy-efficient power allocation with QoS provisioning in OFDMA femtocell networks", 《2014 IEEE WIRELESS COMMUNICATIONS AND NETWORKING CONFERENCE (WCNC)》 * |
XIANFU CHEN, ZHIFENG ZHAO, HONGGANG ZHANG: "Stochastic Power Adaptation with Multiagent Reinforcement Learning for Cognitive Wireless Mesh Networks", 《IEEE TRANSACTIONS ON MOBILE COMPUTING》 * |
Cited By (48)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106954234B (en) * | 2017-04-24 | 2020-08-14 | 东南大学 | User connection and virtual resource allocation method in ultra-dense heterogeneous network |
CN106954234A (en) * | 2017-04-24 | 2017-07-14 | 东南大学 | User's connection and virtual resource allocation method in a kind of super-intensive heterogeneous network |
CN110770761B (en) * | 2017-07-06 | 2022-07-22 | 华为技术有限公司 | Deep learning system and method and wireless network optimization using deep learning |
CN110770761A (en) * | 2017-07-06 | 2020-02-07 | 华为技术有限公司 | Deep learning system and method and wireless network optimization using deep learning |
CN107333333A (en) * | 2017-08-10 | 2017-11-07 | 北京邮电大学 | A kind of resource allocation methods based on user traffic flow |
CN107333333B (en) * | 2017-08-10 | 2019-10-08 | 北京邮电大学 | A kind of resource allocation methods based on user traffic flow |
CN107734700A (en) * | 2017-08-23 | 2018-02-23 | 南京邮电大学 | The network resource allocation method of user's QoS demand is ensured in super-intensive network |
CN107734700B (en) * | 2017-08-23 | 2021-09-14 | 南京邮电大学 | Network resource allocation method for guaranteeing QoS (quality of service) requirements of users in ultra-dense network |
CN108112082A (en) * | 2017-12-18 | 2018-06-01 | 北京工业大学 | A kind of wireless network distributed freedom resource allocation methods based on statelessly Q study |
CN108112082B (en) * | 2017-12-18 | 2021-05-25 | 北京工业大学 | Wireless network distributed autonomous resource allocation method based on stateless Q learning |
CN108124306A (en) * | 2017-12-29 | 2018-06-05 | 河海大学 | A kind of ultra dense set network spectrum efficiency optimization method of cross operator |
CN108307412A (en) * | 2018-02-08 | 2018-07-20 | 北京邮电大学 | The super-intensive network interferences management method based on grouping game of user-center |
CN108307412B (en) * | 2018-02-08 | 2020-08-07 | 北京邮电大学 | User-centered ultra-dense network interference management method based on grouping game |
CN108430103A (en) * | 2018-02-27 | 2018-08-21 | 北京科技大学 | A kind of software definition mist car networking resource allocation methods and its system |
CN108401254A (en) * | 2018-02-27 | 2018-08-14 | 苏州经贸职业技术学院 | A kind of wireless network resource distribution method based on intensified learning |
CN108430103B (en) * | 2018-02-27 | 2020-02-07 | 北京科技大学 | Software-defined fog internet-of-vehicles resource allocation method and system |
CN108307510A (en) * | 2018-02-28 | 2018-07-20 | 北京科技大学 | A kind of power distribution method in isomery subzone network |
CN108834158A (en) * | 2018-05-02 | 2018-11-16 | 北京交通大学 | A kind of interference management method for super-intensive networking |
CN108834158B (en) * | 2018-05-02 | 2020-12-25 | 北京交通大学 | Interference management method for ultra-dense networking |
CN110472764B (en) * | 2018-05-09 | 2023-08-11 | 沃尔沃汽车公司 | Method and system for coordinating multiparty services using semi-collaborative Nash balancing based on reinforcement learning |
CN108718463A (en) * | 2018-05-09 | 2018-10-30 | 南京邮电大学 | A kind of resource allocation methods based on Multiple Time Scales collaboration optimization under H-CRAN |
CN110472764A (en) * | 2018-05-09 | 2019-11-19 | 沃尔沃汽车公司 | Coordinate the method and system serviced in many ways using half cooperation nash banlance based on intensified learning |
CN108718463B (en) * | 2018-05-09 | 2022-07-05 | 南京邮电大学 | Resource allocation method based on multi-time scale collaborative optimization under H-CRAN |
CN109496305B (en) * | 2018-08-01 | 2022-05-13 | 东莞理工学院 | Social network public opinion evolution method |
CN109496305A (en) * | 2018-08-01 | 2019-03-19 | 东莞理工学院 | Nash equilibrium strategy on continuous action space and social network public opinion evolution model |
CN109362113B (en) * | 2018-11-06 | 2022-03-18 | 哈尔滨工程大学 | Underwater acoustic sensor network cooperation exploration reinforcement learning routing method |
CN109362113A (en) * | 2018-11-06 | 2019-02-19 | 哈尔滨工程大学 | A kind of water sound sensor network cooperation exploration intensified learning method for routing |
CN109474980B (en) * | 2018-12-14 | 2020-04-28 | 北京科技大学 | Wireless network resource allocation method based on deep reinforcement learning |
CN109474980A (en) * | 2018-12-14 | 2019-03-15 | 北京科技大学 | A kind of wireless network resource distribution method based on depth enhancing study |
CN109617991B (en) * | 2018-12-29 | 2021-03-30 | 东南大学 | Value function approximation-based cooperative caching method for codes of small stations of ultra-dense heterogeneous network |
CN109617991A (en) * | 2018-12-29 | 2019-04-12 | 东南大学 | Based on value function approximate super-intensive heterogeneous network small station coding cooperative caching method |
CN109743778A (en) * | 2019-01-14 | 2019-05-10 | 长沙学院 | A kind of resource allocation optimization method and system based on intensified learning |
CN109743778B (en) * | 2019-01-14 | 2022-05-10 | 长沙学院 | Resource allocation optimization method and system based on reinforcement learning |
CN109982434A (en) * | 2019-03-08 | 2019-07-05 | 西安电子科技大学 | Wireless resource scheduling integrated intelligent control system and method, wireless communication system |
CN109982434B (en) * | 2019-03-08 | 2022-04-01 | 西安电子科技大学 | Wireless resource scheduling integrated intelligent control system and method and wireless communication system |
CN110049565A (en) * | 2019-04-18 | 2019-07-23 | 天津大学 | A kind of 5G network power distribution method based on available capacity |
CN110049565B (en) * | 2019-04-18 | 2021-11-02 | 天津大学 | 5G network power distribution method based on effective capacity |
CN110392377B (en) * | 2019-07-19 | 2022-07-12 | 北京信息科技大学 | 5G ultra-dense networking resource allocation method and device |
CN110392377A (en) * | 2019-07-19 | 2019-10-29 | 北京信息科技大学 | A kind of 5G super-intensive networking resources distribution method and device |
CN110972309A (en) * | 2019-11-08 | 2020-04-07 | 厦门大学 | Ultra-dense wireless network power distribution method combining graph signals and reinforcement learning |
CN110972309B (en) * | 2019-11-08 | 2022-07-19 | 厦门大学 | Ultra-dense wireless network power distribution method combining graph signals and reinforcement learning |
CN111988834A (en) * | 2020-08-27 | 2020-11-24 | 长安大学 | Heterogeneous Internet of vehicles network selection method, system and device |
CN111988834B (en) * | 2020-08-27 | 2022-11-04 | 长安大学 | Heterogeneous Internet of vehicles network selection method, system and device |
CN112351433B (en) * | 2021-01-05 | 2021-05-25 | 南京邮电大学 | Heterogeneous network resource allocation method based on reinforcement learning |
CN112351433A (en) * | 2021-01-05 | 2021-02-09 | 南京邮电大学 | Heterogeneous network resource allocation method based on reinforcement learning |
CN113038583A (en) * | 2021-03-11 | 2021-06-25 | 南京南瑞信息通信科技有限公司 | Inter-cell downlink interference control method, device and system suitable for ultra-dense network |
CN113490219B (en) * | 2021-07-06 | 2022-02-25 | 香港中文大学(深圳) | Dynamic resource allocation method for ultra-dense networking |
CN113490219A (en) * | 2021-07-06 | 2021-10-08 | 香港中文大学(深圳) | Dynamic resource allocation method for ultra-dense networking |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN106358308A (en) | Resource allocation method for reinforcement learning in ultra-dense network | |
CN109474980B (en) | Wireless network resource allocation method based on deep reinforcement learning | |
CN109729528A (en) | A kind of D2D resource allocation methods based on the study of multiple agent deeply | |
CN107071914B (en) | Dynamic mode selection and energy distributing method in a kind of energy capture D2D network | |
CN107426773B (en) | Energy efficiency-oriented distributed resource allocation method and device in wireless heterogeneous network | |
Hasan et al. | A novel HGBBDSA-CTI approach for subcarrier allocation in heterogeneous network | |
CN108112082A (en) | A kind of wireless network distributed freedom resource allocation methods based on statelessly Q study | |
CN104717755A (en) | Downlink frequency spectrum resource distribution method with D2D technology introduced in cellular network | |
CN107613555A (en) | Non-orthogonal multiple accesses honeycomb and terminal direct connection dense network resource management-control method | |
CN104918257B (en) | Relaying cooperation isomery cellular network D2D communication resource allocation methods | |
CN110191489B (en) | Resource allocation method and device based on reinforcement learning in ultra-dense network | |
CN107708157A (en) | Intensive small cell network resource allocation methods based on efficiency | |
CN104378772B (en) | Towards the small base station deployment method of the amorphous covering of cell in a kind of cellular network | |
CN108064077B (en) | The power distribution method of full duplex D2D in cellular network | |
Ahmad et al. | Quality-of-service aware game theory-based uplink power control for 5G heterogeneous networks | |
CN105490794B (en) | The packet-based resource allocation methods of the Femto cell OFDMA double-layer network | |
Ghosh et al. | E 2 M 3: energy-efficient massive MIMO–MISO 5G HetNet using Stackelberg game | |
CN105764068B (en) | A kind of small Base Station Compatibility and coverage optimization method based on TABU search | |
CN110139281A (en) | A method of the full duplex D2D sub-clustering resource allocation based on K-means | |
CN103139800A (en) | Node adjustment method, device and system of relay cellular network | |
Zhang et al. | Energy efficient resource allocation in millimeter-wave-based fog radio access networks | |
CN110225494A (en) | A kind of machine type communication resource allocation methods based on external effect and matching algorithm | |
Naparstek et al. | Distributed energy efficient channel allocation | |
CN108322271A (en) | The dynamic clustering method of customer-centric based on load | |
CN114423070A (en) | D2D-based heterogeneous wireless network power distribution method and system |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
WD01 | Invention patent application deemed withdrawn after publication |
Application publication date: 20170125 |
|
WD01 | Invention patent application deemed withdrawn after publication |