CN106358308A - Resource allocation method for reinforcement learning in ultra-dense network - Google Patents

Resource allocation method for reinforcement learning in ultra-dense network Download PDF

Info

Publication number
CN106358308A
CN106358308A CN201510409462.3A CN201510409462A CN106358308A CN 106358308 A CN106358308 A CN 106358308A CN 201510409462 A CN201510409462 A CN 201510409462A CN 106358308 A CN106358308 A CN 106358308A
Authority
CN
China
Prior art keywords
strategy
resource allocation
formula
guess
home enodeb
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201510409462.3A
Other languages
Chinese (zh)
Inventor
张海君
王文韬
孙梦颖
郝匀琴
周平
阳欣豪
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing University of Chemical Technology
Original Assignee
Beijing University of Chemical Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing University of Chemical Technology filed Critical Beijing University of Chemical Technology
Priority to CN201510409462.3A priority Critical patent/CN106358308A/en
Publication of CN106358308A publication Critical patent/CN106358308A/en
Pending legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W72/00Local resource management
    • H04W72/50Allocation or scheduling criteria for wireless resources
    • H04W72/54Allocation or scheduling criteria for wireless resources based on quality criteria
    • H04W72/542Allocation or scheduling criteria for wireless resources based on quality criteria using measured or perceived quality
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W52/00Power management, e.g. TPC [Transmission Power Control], power saving or power classes
    • H04W52/04TPC
    • H04W52/18TPC being performed according to specific parameters
    • H04W52/24TPC being performed according to specific parameters using SIR [Signal to Interference Ratio] or other wireless path parameters
    • H04W52/241TPC being performed according to specific parameters using SIR [Signal to Interference Ratio] or other wireless path parameters taking into account channel quality metrics, e.g. SIR, SNR, CIR, Eb/lo
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W52/00Power management, e.g. TPC [Transmission Power Control], power saving or power classes
    • H04W52/04TPC
    • H04W52/18TPC being performed according to specific parameters
    • H04W52/24TPC being performed according to specific parameters using SIR [Signal to Interference Ratio] or other wireless path parameters
    • H04W52/243TPC being performed according to specific parameters using SIR [Signal to Interference Ratio] or other wireless path parameters taking into account interferences
    • H04W52/244Interferences in heterogeneous networks, e.g. among macro and femto or pico cells or other sector / system interference [OSI]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W84/00Network topologies
    • H04W84/18Self-organising networks, e.g. ad-hoc networks or sensor networks

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • Quality & Reliability (AREA)
  • Mobile Radio Communication Systems (AREA)

Abstract

A resource allocation method for reinforcement learning in an ultra-dense network is provided. The invention relates to the field of ultra-dense networks in 5G (fifth generation) mobile communications and provides a method for allocating resources between a home node B and a macro node B, between a home node B and another home node B and between a home node B and a mobile user in a dense deployment network; the method is implemented through power control, each femotcell is considered as an intelligent body to jointly adjust transmitting powers of home node Bs, the densely deployed home node Bs are avoided causing severe jamming to a macro node B and an adjacent B when transmitting at maximum powder, and system throughput is maximized; user delay QoS is considered, and traditional 'Shannon capacity' is replaced with 'available capacity' that may ensure user delay; a supermodular game model is utilized such that whole network power distribution gains Nash equilibrium; the reinforcement learning method Q-learning is utilized such that the home node B has learning function, and optimal power distribution can be achieved; by using the resource allocation method, it is possible to effectively improve the system capacity of an ultra-dense network at the premise of satisfying user delay.

Description

A kind of resource allocation methods of the intensified learning in super-intensive network
Technical field
Present document relates to moving communicating field, especially, the present invention is the method for a kind of resource allocation of super-intensive heterogeneous network (ultra dense network, udn) in the 5th generation (5th-generation) GSM.
Background technology
Nowadays mobile network has been enter into the quick stage of popularization, and meanwhile 5g technology is all energetically being studied in countries in the world, and 5g standard also begins to show up prominently.Take the lead in, using cognitive radio power technology, being automatically determined the frequency range of offer by network, realize the distinguishing feature that multi-internet integration is 5g.The 5g of China works also achieved with preliminary achievement.The main target of 5g network is Consumer's Experience, network is carried out redesigning in terms of capacity, speed, time delay three, optimize.5g network will accommodate substantial amounts of terminal device simultaneously, therefore will jointly meet the demand of mobile service flow growth by higher spectrum reuse efficiency, more frequency spectrum resource and more dense cell deployment.This makes the new challenge of 5g network faces.
Because, in traditional GSM, network design, operation maintenance depend on greatly manually, need to expend a large amount of human and material resources.Therefore, just have self-organizing network (self-organization network, Son concept), realizes deployment, maintenance and the optimization of network by the self organization ability of communication network.And have the access of many low power nodes in 5g system so that network structure is more complicated, form super-intensive heterogeneous network.Due to the rare of Radio Resource so that many operators wish to make full use of frequency spectrum resource in high band, use double-layer network in following 5g system, there is two kinds of user is grand user and domestic consumer respectively simultaneously.In order to make full use of frequency spectrum resource, double-layer network shares identical frequency range, but have also been introduced co-channel interference simultaneously, and existing technology can not solve cross-layer interference problem present in 5g network well, based on this, the self-optimization techniques in self-organization network are applied among super-intensive network distribute to realize the self-organizing of resource by primary study of the present invention.
In order to realize resource self-organizing distribution, classical intensified learning q-learning algorithm is applied in super-intensive network.Q-learning algorithm is based on discrete state, and discrete variable refers to the size of power in the present invention.The housing choice behavior of q-learning and behavior value function select different strategies respectively, by measuring channel parameter in real time, dynamically carry out resource self-organizing distribution, and then reach the purpose of suppression interference.
Content of the invention
Present invention is generally directed to energy saving resources assignment problem is it is proposed that a kind of resource allocation methods based on q-learning algorithm and equilibria of Supermodular Games in super-intensive self-organizing network.The method optimizes the energy utilization efficiency in network, ensures qos, improves power system capacity.
In order to solve the above problems, the invention provides a kind of Resource Allocation Formula of efficiency optimization:
Step 1: initialization Studying factors, for each state s and each action a, initialize evaluation function, transmission strategy, guess, positive scalar.
Step 2: initialization Home eNodeB state s, transmission power p, signal interference ratio etc..
Step 3: according to transmission strategy, the action of selected active procedure.
Step 4: according to the current signal interference ratio of the feedback information detection of target receiver, by identifying whether current transmission power rank and relatively current signal interference ratio determine subsequent time state more than thresholding.
Step 5: if the current Signal to Interference plus Noise Ratio of domestic consumer is more than threshold value, return (return, i.e. Efficiency Function) is calculated by Reward Program, otherwise, zero setting will be returned.
Step 6: the return value obtaining in step 4 is asked with expectation, using the q more new formula based on guess, obtains new q value, update evaluation function.
Step 7: according to the q value obtaining in step 5, according to greedy strategy, update the strategy of user.
Step 8: using guess more new formula, obtain the guess of other Home eNodeB subsequent time behaviors, make user enter next state.Proceed to step 2.
Step 9: terminate this learning process, the radio resource allocation of each Home eNodeB completes, prepare scheduling of resource next time.
In step 1, each Home eNodeB only allows a user to access, and the setting of q value needs estimation in advance.
In step 3, strategy, i.e. the probability of selection action, the maximum action of select probability.
In step 4, the computing formula of signal interference ratio
In formula,Represent Home eNodeB to the channel gain of internal user, the channel gain of Home eNodeB to other base station user.Represent the interference to Home eNodeB for the grand user., represent the transmission power of this base station and interference base station respectively, refer to Gaussian noise.
In steps of 5, the Efficiency Function of each base station refers to
In formula, refer to the available capacity of Home eNodeB k, under the requirement ensureing a fixed response time qos, wireless channel can reach the transfer rate of maximum.Represent qos delay parameter, bigger, delay requirement is higher.Represent the transmission power to user for the Home eNodeB.Represent the power being consumed in communication network.
Above formula shows that subchannel is linear, and the transmission power obtaining on every sub-channels is linear with general power
(in formula, refer to a frame data transmission duration in conjunction with Shannon public affairs capacity formula, b represents communication bandwidth.Obtain the expression of utility function
In step 6, the more new formula of q value,
In formula, it is Studying factors, there is attenuation characteristic, it is discount factor.Refer to the state in next moment.
In step 7, the more new formula of strategy
Value is the positive parameter of temperature, and value is bigger, and the probability convergence of action is identical.
In step 8, guess more new formula
In formula, refer to the guess of previous moment.Refer to subsequent time, in current state and strategy when taking current action, refer to the strategy of previous moment.
From technical method, this method with maximize each Home eNodeB efficiency as target, consider the same layer interference in super-intensive self-organizing network, cross-layer is disturbed and qos time delay, constantly adjust the transmission power of Home eNodeB using q-learning algorithm, it is finally reached Nash Equilibrium, so that the efficiency of each Home eNodeB is optimized.Both ensure the qos time delay of user, improve the efficiency of Home eNodeB it is achieved that the RRM of home base station network again.
Below by the drawings and specific embodiments, technical scheme is further elaborated.
Brief description
For the elaboration embodiments of the invention that become apparent from and existing technical scheme, below technical scheme is illustrated that the explanation accompanying drawing used in accompanying drawing and description of the prior art does simple introduction, obviously, on the premise of not paying creative work, those of ordinary skill in the art can obtain other accompanying drawings by this accompanying drawing.
Fig. 1 show the system architecture diagram comprising Home eNodeB and macro base station in the embodiment of the present invention;
The flow chart that Fig. 2 show Home eNodeB power distribution in the embodiment of the present invention.
Specific embodiment
Main idea is that, by being simulated to communication environment, set up model, initialize Studying factors, guess, transmission strategy and evaluation function q, the state of detection present channel, state instruction parameter has signal interference ratio, transmission power, status etc..According to transmission policy selection current action, the signal interference ratio detecting is compared with given threshold value, if being more than threshold value, obtain a return, if being less than threshold value, the return obtaining is set to zero, obtain new q value using based on the q more new formula of guess, and according to q value by the strategy of greedy strategy acquisition subsequent time and guess, renewal subsequent time state, and enter next communications status, repeat above-mentioned learning process.Power allocation scheme is assessed as Performance evaluation criterion using q value, finds and make the maximum power allocation scheme of Total Return function in super-intensive subzone network.
Fig. 1 show the system architecture diagram simultaneously including that Home eNodeB is disposed with frequency with macro base station, and it comprises a macro base station, multiple femto base station and its user, grand user.
Step 101: setting learning process initial time t=0.
Step 102: initialization Studying factors, for each state s and each action a, initialize evaluation function, transmission strategy, guess, positive scalar.
Step 103: initialization Home eNodeB state s, transmission power p, signal interference ratio etc..
Step 104: according to transmission strategy, select the action of active procedure.
Step 105: according to the current signal interference ratio of the feedback information detection of target receiver, by identifying whether current transmission power rank and relatively current signal interference ratio determine subsequent time state more than thresholding.
Step 106: if current signal interference ratio is more than thresholding, will pass through Reward Program and calculate a return, otherwise, return is set to zero.
Step 107: in step 4, all return value obtaining are asked expectation, using the q more new formula based on guess, obtain new q value, update evaluation function.
Step 108: by the q value obtaining according to the new transmission strategy of the acquisition of greedy strategy.
Step 109: using guess more new formula, obtain the guess to other Home eNodeB subsequent time behaviors.
Step 110: enter next state determined by step 105, enter next moment t=t+1, and proceed to step 102.
Step 111: terminate this learning process, the radio resource allocation of each Home eNodeB completes, prepare scheduling of resource next time.

Claims (5)

1. a kind of resource allocation methods of the intensified learning in super-intensive network are it is characterised in that comprise the following steps:
Step 1: initialization Studying factors, for each state s and each action a, initialize evaluation function, transmission strategy, guess, positive scalar;
Step 2: initialization Home eNodeB state s, transmission power p, signal interference ratio etc.;
Step 3: according to transmission strategy, the action of selected active procedure;
Step 4: according to the current signal interference ratio of the feedback information detection of target receiver, by identifying whether current transmission power rank and relatively current signal interference ratio determine subsequent time state more than thresholding;
Step 5: if the current Signal to Interference plus Noise Ratio of domestic consumer is more than threshold value, return (return, i.e. Efficiency Function) is calculated by Reward Program, otherwise, zero setting will be returned;
Step 6: the return value obtaining in step 4 is asked with expectation, using the more new formula based on guess q, obtains new q value, update evaluation function;
Step 7: according to the q value obtaining in step 5, according to greedy strategy, update the strategy of user;
Step 8: using guess more new formula, obtain the guess of other Home eNodeB subsequent time behaviors, make user enter next state, proceed to step 2;
Step 9: terminate this learning process, the radio resource allocation of each Home eNodeB completes, prepare scheduling of resource next time.
2. super-intensive network according to claim 1 resource allocation methods it is characterised in that:
In described step 1, using equilibria of Supermodular Games model, the intelligent body that each Home eNodeB is viewed as having fair competition relation tries to achieve optimal objective so as to reach Nash Equilibrium.
3. super-intensive network according to claim 1 resource allocation methods it is characterised in that:
In described step 5, using Efficiency Function as q-learning algorithm evaluation function, Efficiency Function formula is
, the available capacity of its middle finger Home eNodeB k, under the requirement ensureing a fixed response time qos, wireless channel can reach the transfer rate of maximum;Represent qos delay parameter, bigger, delay requirement is higher, represent the transmission power to user for the Home eNodeB, represent the power being consumed in communication network,
Above formula shows that subchannel is linear, and the transmission power obtaining on every sub-channels is linear with general power, (in formula, refer to a frame data transmission duration in conjunction with Shannon public affairs capacity formula, b represents communication bandwidth, obtains the expression of utility function.
4. super-intensive network according to claim 1 resource allocation methods it is characterised in that:
In described step 8, target household base station requires no knowledge about the strategy to other base station strategies, but the strategy of other base stations is guessed, adjusts the strategy of oneself so as to convergence Nash Equilibrium, its guess more new formula is in formula, refers to the guess of previous moment,Refer to subsequent time, in current state and strategy when taking current action, refer to the strategy of previous moment.
5. super-intensive network according to claim 1 resource allocation methods it is characterised in that:
In described step 6, based on the q more new formula of guess it is:
In formula, it is Studying factors, there is attenuation characteristic, it is discount factor, refers to the state in next moment, go to assess the quality of return value in this iteration with q value.
CN201510409462.3A 2015-07-14 2015-07-14 Resource allocation method for reinforcement learning in ultra-dense network Pending CN106358308A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510409462.3A CN106358308A (en) 2015-07-14 2015-07-14 Resource allocation method for reinforcement learning in ultra-dense network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510409462.3A CN106358308A (en) 2015-07-14 2015-07-14 Resource allocation method for reinforcement learning in ultra-dense network

Publications (1)

Publication Number Publication Date
CN106358308A true CN106358308A (en) 2017-01-25

Family

ID=57842125

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510409462.3A Pending CN106358308A (en) 2015-07-14 2015-07-14 Resource allocation method for reinforcement learning in ultra-dense network

Country Status (1)

Country Link
CN (1) CN106358308A (en)

Cited By (26)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106954234A (en) * 2017-04-24 2017-07-14 东南大学 User's connection and virtual resource allocation method in a kind of super-intensive heterogeneous network
CN107333333A (en) * 2017-08-10 2017-11-07 北京邮电大学 A kind of resource allocation methods based on user traffic flow
CN107734700A (en) * 2017-08-23 2018-02-23 南京邮电大学 The network resource allocation method of user's QoS demand is ensured in super-intensive network
CN108112082A (en) * 2017-12-18 2018-06-01 北京工业大学 A kind of wireless network distributed freedom resource allocation methods based on statelessly Q study
CN108124306A (en) * 2017-12-29 2018-06-05 河海大学 A kind of ultra dense set network spectrum efficiency optimization method of cross operator
CN108307510A (en) * 2018-02-28 2018-07-20 北京科技大学 A kind of power distribution method in isomery subzone network
CN108307412A (en) * 2018-02-08 2018-07-20 北京邮电大学 The super-intensive network interferences management method based on grouping game of user-center
CN108401254A (en) * 2018-02-27 2018-08-14 苏州经贸职业技术学院 A kind of wireless network resource distribution method based on intensified learning
CN108430103A (en) * 2018-02-27 2018-08-21 北京科技大学 A kind of software definition mist car networking resource allocation methods and its system
CN108718463A (en) * 2018-05-09 2018-10-30 南京邮电大学 A kind of resource allocation methods based on Multiple Time Scales collaboration optimization under H-CRAN
CN108834158A (en) * 2018-05-02 2018-11-16 北京交通大学 A kind of interference management method for super-intensive networking
CN109362113A (en) * 2018-11-06 2019-02-19 哈尔滨工程大学 A kind of water sound sensor network cooperation exploration intensified learning method for routing
CN109474980A (en) * 2018-12-14 2019-03-15 北京科技大学 A kind of wireless network resource distribution method based on depth enhancing study
CN109496305A (en) * 2018-08-01 2019-03-19 东莞理工学院 Nash equilibrium strategy on continuous action space and social network public opinion evolution model
CN109617991A (en) * 2018-12-29 2019-04-12 东南大学 Based on value function approximate super-intensive heterogeneous network small station coding cooperative caching method
CN109743778A (en) * 2019-01-14 2019-05-10 长沙学院 A kind of resource allocation optimization method and system based on intensified learning
CN109982434A (en) * 2019-03-08 2019-07-05 西安电子科技大学 Wireless resource scheduling integrated intelligent control system and method, wireless communication system
CN110049565A (en) * 2019-04-18 2019-07-23 天津大学 A kind of 5G network power distribution method based on available capacity
CN110392377A (en) * 2019-07-19 2019-10-29 北京信息科技大学 A kind of 5G super-intensive networking resources distribution method and device
CN110472764A (en) * 2018-05-09 2019-11-19 沃尔沃汽车公司 Coordinate the method and system serviced in many ways using half cooperation nash banlance based on intensified learning
CN110770761A (en) * 2017-07-06 2020-02-07 华为技术有限公司 Deep learning system and method and wireless network optimization using deep learning
CN110972309A (en) * 2019-11-08 2020-04-07 厦门大学 Ultra-dense wireless network power distribution method combining graph signals and reinforcement learning
CN111988834A (en) * 2020-08-27 2020-11-24 长安大学 Heterogeneous Internet of vehicles network selection method, system and device
CN112351433A (en) * 2021-01-05 2021-02-09 南京邮电大学 Heterogeneous network resource allocation method based on reinforcement learning
CN113038583A (en) * 2021-03-11 2021-06-25 南京南瑞信息通信科技有限公司 Inter-cell downlink interference control method, device and system suitable for ultra-dense network
CN113490219A (en) * 2021-07-06 2021-10-08 香港中文大学(深圳) Dynamic resource allocation method for ultra-dense networking

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102711123A (en) * 2012-06-08 2012-10-03 北京邮电大学 Method for inhibiting interference in intensively-deployed household base station network
CN102811443A (en) * 2012-07-27 2012-12-05 南京邮电大学 Interference management method based on spectrum allocation and power control in family base station system
CN104023384A (en) * 2014-05-28 2014-09-03 北京邮电大学 Double-layer network resource allocation method in consideration of time delay limitation in dense home base station deployment scene
CN104113903A (en) * 2014-07-31 2014-10-22 厦门大学 Interactive cognitive learning based downlink power adjusting method and device
WO2015054264A1 (en) * 2013-10-08 2015-04-16 Google Inc. Methods and apparatus for reinforcement learning

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102711123A (en) * 2012-06-08 2012-10-03 北京邮电大学 Method for inhibiting interference in intensively-deployed household base station network
CN102811443A (en) * 2012-07-27 2012-12-05 南京邮电大学 Interference management method based on spectrum allocation and power control in family base station system
WO2015054264A1 (en) * 2013-10-08 2015-04-16 Google Inc. Methods and apparatus for reinforcement learning
CN104023384A (en) * 2014-05-28 2014-09-03 北京邮电大学 Double-layer network resource allocation method in consideration of time delay limitation in dense home base station deployment scene
CN104113903A (en) * 2014-07-31 2014-10-22 厦门大学 Interactive cognitive learning based downlink power adjusting method and device

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
WENPENG JING,ZHAOMING LU,ZHICAI ZHANG,HAIJUN ZHANG,ETC: "Energy-efficient power allocation with QoS provisioning in OFDMA femtocell networks", 《2014 IEEE WIRELESS COMMUNICATIONS AND NETWORKING CONFERENCE (WCNC)》 *
XIANFU CHEN, ZHIFENG ZHAO, HONGGANG ZHANG: "Stochastic Power Adaptation with Multiagent Reinforcement Learning for Cognitive Wireless Mesh Networks", 《IEEE TRANSACTIONS ON MOBILE COMPUTING》 *

Cited By (48)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106954234B (en) * 2017-04-24 2020-08-14 东南大学 User connection and virtual resource allocation method in ultra-dense heterogeneous network
CN106954234A (en) * 2017-04-24 2017-07-14 东南大学 User's connection and virtual resource allocation method in a kind of super-intensive heterogeneous network
CN110770761B (en) * 2017-07-06 2022-07-22 华为技术有限公司 Deep learning system and method and wireless network optimization using deep learning
CN110770761A (en) * 2017-07-06 2020-02-07 华为技术有限公司 Deep learning system and method and wireless network optimization using deep learning
CN107333333A (en) * 2017-08-10 2017-11-07 北京邮电大学 A kind of resource allocation methods based on user traffic flow
CN107333333B (en) * 2017-08-10 2019-10-08 北京邮电大学 A kind of resource allocation methods based on user traffic flow
CN107734700A (en) * 2017-08-23 2018-02-23 南京邮电大学 The network resource allocation method of user's QoS demand is ensured in super-intensive network
CN107734700B (en) * 2017-08-23 2021-09-14 南京邮电大学 Network resource allocation method for guaranteeing QoS (quality of service) requirements of users in ultra-dense network
CN108112082A (en) * 2017-12-18 2018-06-01 北京工业大学 A kind of wireless network distributed freedom resource allocation methods based on statelessly Q study
CN108112082B (en) * 2017-12-18 2021-05-25 北京工业大学 Wireless network distributed autonomous resource allocation method based on stateless Q learning
CN108124306A (en) * 2017-12-29 2018-06-05 河海大学 A kind of ultra dense set network spectrum efficiency optimization method of cross operator
CN108307412A (en) * 2018-02-08 2018-07-20 北京邮电大学 The super-intensive network interferences management method based on grouping game of user-center
CN108307412B (en) * 2018-02-08 2020-08-07 北京邮电大学 User-centered ultra-dense network interference management method based on grouping game
CN108430103A (en) * 2018-02-27 2018-08-21 北京科技大学 A kind of software definition mist car networking resource allocation methods and its system
CN108401254A (en) * 2018-02-27 2018-08-14 苏州经贸职业技术学院 A kind of wireless network resource distribution method based on intensified learning
CN108430103B (en) * 2018-02-27 2020-02-07 北京科技大学 Software-defined fog internet-of-vehicles resource allocation method and system
CN108307510A (en) * 2018-02-28 2018-07-20 北京科技大学 A kind of power distribution method in isomery subzone network
CN108834158A (en) * 2018-05-02 2018-11-16 北京交通大学 A kind of interference management method for super-intensive networking
CN108834158B (en) * 2018-05-02 2020-12-25 北京交通大学 Interference management method for ultra-dense networking
CN110472764B (en) * 2018-05-09 2023-08-11 沃尔沃汽车公司 Method and system for coordinating multiparty services using semi-collaborative Nash balancing based on reinforcement learning
CN108718463A (en) * 2018-05-09 2018-10-30 南京邮电大学 A kind of resource allocation methods based on Multiple Time Scales collaboration optimization under H-CRAN
CN110472764A (en) * 2018-05-09 2019-11-19 沃尔沃汽车公司 Coordinate the method and system serviced in many ways using half cooperation nash banlance based on intensified learning
CN108718463B (en) * 2018-05-09 2022-07-05 南京邮电大学 Resource allocation method based on multi-time scale collaborative optimization under H-CRAN
CN109496305B (en) * 2018-08-01 2022-05-13 东莞理工学院 Social network public opinion evolution method
CN109496305A (en) * 2018-08-01 2019-03-19 东莞理工学院 Nash equilibrium strategy on continuous action space and social network public opinion evolution model
CN109362113B (en) * 2018-11-06 2022-03-18 哈尔滨工程大学 Underwater acoustic sensor network cooperation exploration reinforcement learning routing method
CN109362113A (en) * 2018-11-06 2019-02-19 哈尔滨工程大学 A kind of water sound sensor network cooperation exploration intensified learning method for routing
CN109474980B (en) * 2018-12-14 2020-04-28 北京科技大学 Wireless network resource allocation method based on deep reinforcement learning
CN109474980A (en) * 2018-12-14 2019-03-15 北京科技大学 A kind of wireless network resource distribution method based on depth enhancing study
CN109617991B (en) * 2018-12-29 2021-03-30 东南大学 Value function approximation-based cooperative caching method for codes of small stations of ultra-dense heterogeneous network
CN109617991A (en) * 2018-12-29 2019-04-12 东南大学 Based on value function approximate super-intensive heterogeneous network small station coding cooperative caching method
CN109743778A (en) * 2019-01-14 2019-05-10 长沙学院 A kind of resource allocation optimization method and system based on intensified learning
CN109743778B (en) * 2019-01-14 2022-05-10 长沙学院 Resource allocation optimization method and system based on reinforcement learning
CN109982434A (en) * 2019-03-08 2019-07-05 西安电子科技大学 Wireless resource scheduling integrated intelligent control system and method, wireless communication system
CN109982434B (en) * 2019-03-08 2022-04-01 西安电子科技大学 Wireless resource scheduling integrated intelligent control system and method and wireless communication system
CN110049565A (en) * 2019-04-18 2019-07-23 天津大学 A kind of 5G network power distribution method based on available capacity
CN110049565B (en) * 2019-04-18 2021-11-02 天津大学 5G network power distribution method based on effective capacity
CN110392377B (en) * 2019-07-19 2022-07-12 北京信息科技大学 5G ultra-dense networking resource allocation method and device
CN110392377A (en) * 2019-07-19 2019-10-29 北京信息科技大学 A kind of 5G super-intensive networking resources distribution method and device
CN110972309A (en) * 2019-11-08 2020-04-07 厦门大学 Ultra-dense wireless network power distribution method combining graph signals and reinforcement learning
CN110972309B (en) * 2019-11-08 2022-07-19 厦门大学 Ultra-dense wireless network power distribution method combining graph signals and reinforcement learning
CN111988834A (en) * 2020-08-27 2020-11-24 长安大学 Heterogeneous Internet of vehicles network selection method, system and device
CN111988834B (en) * 2020-08-27 2022-11-04 长安大学 Heterogeneous Internet of vehicles network selection method, system and device
CN112351433B (en) * 2021-01-05 2021-05-25 南京邮电大学 Heterogeneous network resource allocation method based on reinforcement learning
CN112351433A (en) * 2021-01-05 2021-02-09 南京邮电大学 Heterogeneous network resource allocation method based on reinforcement learning
CN113038583A (en) * 2021-03-11 2021-06-25 南京南瑞信息通信科技有限公司 Inter-cell downlink interference control method, device and system suitable for ultra-dense network
CN113490219B (en) * 2021-07-06 2022-02-25 香港中文大学(深圳) Dynamic resource allocation method for ultra-dense networking
CN113490219A (en) * 2021-07-06 2021-10-08 香港中文大学(深圳) Dynamic resource allocation method for ultra-dense networking

Similar Documents

Publication Publication Date Title
CN106358308A (en) Resource allocation method for reinforcement learning in ultra-dense network
CN109474980B (en) Wireless network resource allocation method based on deep reinforcement learning
CN109729528A (en) A kind of D2D resource allocation methods based on the study of multiple agent deeply
CN107071914B (en) Dynamic mode selection and energy distributing method in a kind of energy capture D2D network
CN107426773B (en) Energy efficiency-oriented distributed resource allocation method and device in wireless heterogeneous network
Hasan et al. A novel HGBBDSA-CTI approach for subcarrier allocation in heterogeneous network
CN108112082A (en) A kind of wireless network distributed freedom resource allocation methods based on statelessly Q study
CN104717755A (en) Downlink frequency spectrum resource distribution method with D2D technology introduced in cellular network
CN107613555A (en) Non-orthogonal multiple accesses honeycomb and terminal direct connection dense network resource management-control method
CN104918257B (en) Relaying cooperation isomery cellular network D2D communication resource allocation methods
CN110191489B (en) Resource allocation method and device based on reinforcement learning in ultra-dense network
CN107708157A (en) Intensive small cell network resource allocation methods based on efficiency
CN104378772B (en) Towards the small base station deployment method of the amorphous covering of cell in a kind of cellular network
CN108064077B (en) The power distribution method of full duplex D2D in cellular network
Ahmad et al. Quality-of-service aware game theory-based uplink power control for 5G heterogeneous networks
CN105490794B (en) The packet-based resource allocation methods of the Femto cell OFDMA double-layer network
Ghosh et al. E 2 M 3: energy-efficient massive MIMO–MISO 5G HetNet using Stackelberg game
CN105764068B (en) A kind of small Base Station Compatibility and coverage optimization method based on TABU search
CN110139281A (en) A method of the full duplex D2D sub-clustering resource allocation based on K-means
CN103139800A (en) Node adjustment method, device and system of relay cellular network
Zhang et al. Energy efficient resource allocation in millimeter-wave-based fog radio access networks
CN110225494A (en) A kind of machine type communication resource allocation methods based on external effect and matching algorithm
Naparstek et al. Distributed energy efficient channel allocation
CN108322271A (en) The dynamic clustering method of customer-centric based on load
CN114423070A (en) D2D-based heterogeneous wireless network power distribution method and system

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20170125

WD01 Invention patent application deemed withdrawn after publication