CN110084414B - Empty pipe anti-collision method based on K-time control deep reinforcement learning - Google Patents

Empty pipe anti-collision method based on K-time control deep reinforcement learning Download PDF

Info

Publication number
CN110084414B
CN110084414B CN201910311467.0A CN201910311467A CN110084414B CN 110084414 B CN110084414 B CN 110084414B CN 201910311467 A CN201910311467 A CN 201910311467A CN 110084414 B CN110084414 B CN 110084414B
Authority
CN
China
Prior art keywords
control
airplane
reinforcement learning
neural network
sector
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910311467.0A
Other languages
Chinese (zh)
Other versions
CN110084414A (en
Inventor
李辉
王壮
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
CHENGDU RONGAO TECHNOLOGY Co Ltd
Original Assignee
CHENGDU RONGAO TECHNOLOGY Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by CHENGDU RONGAO TECHNOLOGY Co Ltd filed Critical CHENGDU RONGAO TECHNOLOGY Co Ltd
Priority to CN201910311467.0A priority Critical patent/CN110084414B/en
Publication of CN110084414A publication Critical patent/CN110084414A/en
Application granted granted Critical
Publication of CN110084414B publication Critical patent/CN110084414B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/04Forecasting or optimisation specially adapted for administrative or management purposes, e.g. linear programming or "cutting stock problem"
    • G06Q10/047Optimisation of routes or paths, e.g. travelling salesman problem
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q50/00Information and communication technology [ICT] specially adapted for implementation of business processes of specific business sectors, e.g. utilities or tourism
    • G06Q50/40Business processes related to the transportation industry

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Business, Economics & Management (AREA)
  • Theoretical Computer Science (AREA)
  • Human Resources & Organizations (AREA)
  • General Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Strategic Management (AREA)
  • Health & Medical Sciences (AREA)
  • Economics (AREA)
  • Biomedical Technology (AREA)
  • Computational Linguistics (AREA)
  • Software Systems (AREA)
  • General Business, Economics & Management (AREA)
  • Mathematical Physics (AREA)
  • General Engineering & Computer Science (AREA)
  • Marketing (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Computing Systems (AREA)
  • Biophysics (AREA)
  • Tourism & Hospitality (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Molecular Biology (AREA)
  • Development Economics (AREA)
  • Entrepreneurship & Innovation (AREA)
  • Operations Research (AREA)
  • Game Theory and Decision Science (AREA)
  • Quality & Reliability (AREA)
  • Primary Health Care (AREA)
  • Traffic Control Systems (AREA)

Abstract

The invention discloses an empty pipe anti-collision method based on K times of control depth reinforcement learning, which comprises the following steps: firstly, setting the number of airplanes in a sector in a use scene, and setting the control times K in an anti-collision process; then, performing K times of control in a training mode, determining the next position point through the action of the neural network in the former K-1 times of control according to a two-dimensional normal distribution method, updating the neural network parameters according to a reinforcement learning method, taking the destination as the next position point in the Kth time of control, and repeating the steps to finish the training of the neural network; finally, in the application mode, the shortest path without conflict can be obtained by using the trained neural network. The method can be applied to the existing air traffic management system, obtains the shortest path to the destination on the premise of not colliding with other airplanes in the sector, and has practical significance for planning the air pipeline path.

Description

Empty pipe anti-collision method based on K-time control deep reinforcement learning
Technical Field
The invention relates to the field of air traffic management, in particular to an air traffic control anti-collision method based on K-time control deep reinforcement learning.
Background
In recent years, civil aviation is rapidly developed, and the continuous development brings serious air traffic jam and great pressure to air managers. When an airplane flies from one sector to another sector, the flight path of the airplane needs to be planned, and correct guidance is given so as to avoid conflict with the existing airplane in the sector. The existing algorithm can generate an optimal or suboptimal flight path and conduct aircraft guidance, but the calculation efficiency is low, the real-time requirement in real air traffic control cannot be met, and further research is still needed. The deep reinforcement learning execution efficiency is high, the use is flexible, and the improved deep reinforcement learning execution system can be applied to an air traffic control system to quickly give a guide track.
Disclosure of Invention
The invention aims to overcome the defects and shortcomings of the prior art, provides an air traffic control anti-collision method based on K times of control depth reinforcement learning, realizes that an airplane enters a sector and arrives at a destination on the premise of not colliding with the existing airplane in the sector, and can quickly form a plurality of schemes for an air traffic controller to select.
In order to realize the purpose, the invention adopts the following technical scheme:
an empty pipe anti-collision method based on K times of control depth reinforcement learning comprises the following steps:
(1) numbering existing airplanes in the sector, and generating a coordinate matrix P from the current moment to the moment when the airplanes fly out of the sector according to the existing flight plan of the existing airplanes and the time step;
(2) training a deep neural network by using a K-time control deep reinforcement learning method, and generating a path of a control airplane according to the current position of the control airplane and a coordinate matrix P of the airplane in a sector;
the calculation process of the K times of control depth reinforcement learning algorithm is as follows: setting a control frequency K; constructing a deep neural network, inputting a coordinate matrix P for controlling the current position of the airplane and the existing airplane, and outputting a polar coordinate of the next position point for controlling the airplane
Figure GDA0002300374960000011
If the control is not the Kth control, obtaining the polar coordinates of the next position point according to a two-dimensional normal distribution method, and updating the deep neural network parameters according to a reinforcement learning method according to the guide result; if the control is the Kth time, taking the destination of the airplane as the next position point, finishing the training and entering the next training;
(3) after massive training, the deep neural network has guiding capability, and can quickly generate a shortest path which does not conflict with other airplanes and reaches a destination for the control airplane according to the input position of the control airplane and the coordinate matrix of the existing airplane;
(4) in practical use, a plurality of deep neural networks with different K values can be trained, and a guide path can be quickly generated for an empty manager according to specific problems.
As a preferred technical solution, in the step (1), the coordinate matrix P of the existing airplane in the sector not only includes the coordinates of the current airplane, but also includes future coordinates according to the flight plan.
As a preferable technical scheme, in the step (2), the control times are adjusted through a parameter K, and the control times can be flexibly set in the empty pipe guide; and selecting the polar angle and the polar diameter of the next position point through two-dimensional normal distribution, wherein the point selection formula is as follows:
Figure GDA0002300374960000021
wherein, muρ,σρRepresents the mean and standard deviation of a normal distribution of the pole diameters,
Figure GDA0002300374960000022
the normal distribution mean value and the standard deviation of the representative polar angle meet the exploratory requirement in the reinforcement learning training process by the point selection method;
by adopting the operator-critic double-nerve network structure, the updating formula of the critic nerve network is as follows:
Figure GDA0002300374960000023
the update formula of the actor neural network is as follows:
Figure GDA0002300374960000024
wherein, αw,αθFor the learning rate of neural networks, δ ═ Rt+γV(St+1,W)-V(St,W),RtFor reinforcement learning of the return function, V (S)tW) is a function of the state value at time t, and γ is a discount factor.
As a preferred technical solution, in the step (3), according to the characteristic of neural network feature identification, non-conflicting K position points can be quickly generated according to state input, and the aircraft is controlled to sequentially fly through the K control points, so as to form a non-conflicting shortest path.
As a preferred technical solution, in the step (4), a plurality of alternatives can be generated at the same time for the air manager to flexibly select according to the air situation.
Compared with the prior art, the invention has the following advantages and effects:
(1) compared with the traditional method, the method has higher calculation efficiency and can generate the optimal path within 200 ms.
(2) The invention improves the deep reinforcement learning, the control times can be selected, and the reasonable control times can be selected according to the actual air condition.
(3) The invention applies the empty pipe anti-collision method based on K times of control depth reinforcement learning to the air traffic management system, realizes that the airplane enters the sector and reaches the destination on the premise of not colliding with the existing airplane in the sector, can quickly form a plurality of schemes for an empty pipe manager to select, and has practical significance for planning the empty pipe path.
Drawings
Fig. 1 is a flowchart of an empty pipe anti-collision method based on K times of control depth reinforcement learning according to this embodiment;
fig. 2 is an inside-sector air management schematic diagram of an air management anti-collision method based on K times of control depth reinforcement learning according to this embodiment;
fig. 3 is a schematic control diagram of K times of the empty pipe anti-collision method based on K times of control depth reinforcement learning according to this embodiment;
FIG. 4 is a diagram of an actor neural network structure of the empty pipe anti-collision method based on K times of control depth reinforcement learning according to the present embodiment;
fig. 5 is a flight path diagram between two points of an empty pipe anti-collision method based on K times of control depth reinforcement learning according to the embodiment.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, the present invention will be described in further detail below with reference to the accompanying drawings and embodiments. It should be understood that the specific embodiments described herein are merely illustrative of the invention and are not limiting of the invention.
An empty pipe anti-collision method based on K times of control depth reinforcement learning is shown in fig. 1 and comprises the following steps:
(1) numbering existing airplanes in the sector, and generating a coordinate matrix P from the current moment to the moment when the airplanes fly out of the sector according to the existing flight plan of the existing airplanes and the time step;
(2) training a deep neural network by using a K-time control deep reinforcement learning method, and generating a path of a control airplane according to the current position of the control airplane and a coordinate matrix P of the airplane in a sector;
the calculation process of the K times of control depth reinforcement learning algorithm is as follows: setting a control frequency K; constructing a deep neural network, inputting a coordinate matrix P for controlling the current position of the airplane and the existing airplane, and outputting a polar coordinate of the next position point for controlling the airplane
Figure GDA0002300374960000031
If the control is not the Kth control, obtaining the polar coordinates of the next position point according to a two-dimensional normal distribution method, and updating the deep neural network parameters according to a reinforcement learning method according to the guide result; if the control is the Kth time, taking the destination of the airplane as the next position point, finishing the training and entering the next training;
(3) after massive training, the deep neural network has guiding capability, and can quickly generate a shortest path which does not conflict with other airplanes and reaches a destination for the control airplane according to the input position of the control airplane and the coordinate matrix of the existing airplane;
(4) in practical use, a plurality of deep neural networks with different K values can be trained, and a guide path can be quickly generated for an empty manager according to specific problems.
In the embodiment, an airplane flying according to a set flight path exists in a sector, and the airplane is controlled to fly into the sector, and the air traffic control anti-collision method based on the K-time control depth reinforcement learning realizes that the airplane enters the sector and arrives at a destination on the premise of not colliding with the existing airplane in the sector;
as shown in fig. 2, existing airplanes in a sector are numbered, and a coordinate matrix P from the current moment to the moment when the airplane flies out of the sector is generated according to the existing flight plan of the existing airplanes and the time step;
as shown in fig. 3, there are four airplanes in the sector, and the coordinate matrix P of the airplanes in the sector not only contains the coordinates of the current airplane, but also includes the future coordinates according to the flight plan.
In this embodiment, the number of times of control is adjusted by a parameter K, and the number of times of control can be flexibly set in the empty pipe guidance, as shown in fig. 3, the value of K is 3;
in this embodiment, as shown in fig. 4, the actor neural network is composed of three layers of fully connected networks, and the output is the normal distribution mean and standard deviation of the polar angle and the polar diameter of the next position point.
In this embodiment, according to the characteristic of neural network feature identification, non-conflicting K position points can be quickly generated according to state input, the aircraft is controlled to sequentially fly through the K control points, and the flight trajectory between the two control points is as shown in fig. 5, so as to form a non-conflicting shortest path.
In this embodiment, a collision avoidance solution can be generated within 200ms using the present method. Five different solutions can be generated within one second for an air manager to select, and the efficiency is obviously superior to that of the existing method for generating one solution for several seconds or even tens of seconds.

Claims (5)

1. An empty pipe anti-collision method based on K times of control depth reinforcement learning is characterized by comprising the following steps:
(1) numbering existing airplanes in the sector, and generating a coordinate matrix P from the current moment to the moment when the airplanes fly out of the sector according to the existing flight plan of the existing airplanes and the time step;
(2) training a deep neural network by using a K-time control deep reinforcement learning method, and generating a path of a control airplane according to the current position of the control airplane and a coordinate matrix P of the airplane in a sector;
the calculation process of the K times of control depth reinforcement learning algorithm is as follows: setting a control frequency K; constructing a deep neural network, inputting a coordinate matrix P for controlling the current position of the airplane and the existing airplane, and outputting a polar coordinate of the next position point for controlling the airplane
Figure FDA0002300374950000011
If the control is not the Kth control, obtaining the polar coordinates of the next position point according to a two-dimensional normal distribution method, and updating the deep neural network parameters according to a reinforcement learning method according to the guide result; if the control is the Kth control, the destination of the airplane is taken as the next position point, and the control is finishedThe training enters the next training;
(3) after mass training, according to the input position of the control plane and the coordinate matrix of the existing plane, a shortest path which does not conflict with other planes and reaches a destination is quickly generated for the control plane;
(4) in actual use, a plurality of deep neural networks with different K values are trained according to actual air conditions, and a guide path is quickly generated for an air traffic manager aiming at specific problems.
2. The air traffic control anti-collision method based on K times of control depth reinforcement learning of claim 1, characterized in that in step (1), the coordinate matrix P of the existing airplane in the sector not only contains the current airplane coordinates, but also includes the future coordinates according to the flight plan.
3. The empty pipe anti-collision method based on K times of control depth reinforcement learning of claim 1, characterized in that in step (2), the control times are adjusted by a parameter K, and the control times are flexibly set in empty pipe guidance; and selecting the polar angle and the polar diameter of the next position point through two-dimensional normal distribution, wherein the point selection formula is as follows:
Figure FDA0002300374950000012
wherein, muρ,σρRepresents the mean and standard deviation of a normal distribution of the pole diameters,
Figure FDA0002300374950000013
normal distribution mean and standard deviation representing polar angle;
by adopting the operator-critic double-nerve network structure, the updating formula of the critic nerve network is as follows:
Figure FDA0002300374950000014
the update formula of the actor neural network is as follows:
Figure FDA0002300374950000021
wherein, αw,αθFor the learning rate of neural networks, δ ═ Rt+γV(St+1,w)-V(St,w),RtFor reinforcement learning of the return function, V (S)tW) is a function of the state value at time t, and γ is a discount factor.
4. The empty pipe anti-collision method based on K times of control depth reinforcement learning of claim 1, characterized in that in step (3), according to the characteristic of neural network feature identification, K non-collision position points are rapidly generated according to state input, and the airplane is controlled to sequentially fly through K control points to form a shortest non-collision path.
5. The empty pipe anti-collision method based on K times of control deep reinforcement learning of claim 1, characterized in that in step (4), a plurality of alternatives are generated simultaneously for the empty pipe administrator to flexibly select according to the empty situation.
CN201910311467.0A 2019-04-18 2019-04-18 Empty pipe anti-collision method based on K-time control deep reinforcement learning Active CN110084414B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910311467.0A CN110084414B (en) 2019-04-18 2019-04-18 Empty pipe anti-collision method based on K-time control deep reinforcement learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910311467.0A CN110084414B (en) 2019-04-18 2019-04-18 Empty pipe anti-collision method based on K-time control deep reinforcement learning

Publications (2)

Publication Number Publication Date
CN110084414A CN110084414A (en) 2019-08-02
CN110084414B true CN110084414B (en) 2020-03-06

Family

ID=67415491

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910311467.0A Active CN110084414B (en) 2019-04-18 2019-04-18 Empty pipe anti-collision method based on K-time control deep reinforcement learning

Country Status (1)

Country Link
CN (1) CN110084414B (en)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111882047B (en) * 2020-09-28 2021-01-15 四川大学 Rapid empty pipe anti-collision method based on reinforcement learning and linear programming
CN113393495B (en) * 2021-06-21 2022-02-01 暨南大学 High-altitude parabolic track identification method based on reinforcement learning
CN114141062B (en) * 2021-11-30 2022-11-01 中国电子科技集团公司第二十八研究所 Aircraft interval management decision method based on deep reinforcement learning
CN113962031B (en) * 2021-12-20 2022-03-29 北京航空航天大学 Heterogeneous platform conflict resolution method based on graph neural network reinforcement learning

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103530704B (en) * 2013-10-16 2016-06-29 南京航空航天大学 A kind of air dynamic traffic volume in terminal airspace prognoses system and method thereof
KR101747393B1 (en) * 2015-06-29 2017-06-15 인하공업전문대학산학협력단 Collision avoidance control method for unmanned air vehicle
CN106601033B (en) * 2017-02-28 2018-05-08 中国人民解放军空军装备研究院雷达与电子对抗研究所 A kind of detection method and device of air traffic control mid-term conflict
CN106814744A (en) * 2017-03-14 2017-06-09 吉林化工学院 A kind of UAV Flight Control System and method
CN109597839B (en) * 2018-12-04 2022-11-04 中国航空无线电电子研究所 Data mining method based on avionic combat situation

Also Published As

Publication number Publication date
CN110084414A (en) 2019-08-02

Similar Documents

Publication Publication Date Title
CN110084414B (en) Empty pipe anti-collision method based on K-time control deep reinforcement learning
CN109933086B (en) Unmanned aerial vehicle environment perception and autonomous obstacle avoidance method based on deep Q learning
CN105353766B (en) A kind of distributed fault-tolerance management method of multiple no-manned plane formation structure
CN110673637A (en) Unmanned aerial vehicle pseudo path planning method based on deep reinforcement learning
CN105953800B (en) A kind of unmanned vehicle trajectory planning grid space division methods
CN111882047B (en) Rapid empty pipe anti-collision method based on reinforcement learning and linear programming
JP2020201958A (en) Three-dimensional aircraft autonomous navigation under restriction
CN111026157B (en) Intelligent aircraft guiding method based on reward remodeling reinforcement learning
CN111121784B (en) Unmanned reconnaissance aircraft route planning method
CN114330115B (en) Neural network air combat maneuver decision-making method based on particle swarm search
CN109240335B (en) Aerospace vehicle approach landing guidance method
CN105185163A (en) Flight path selection method, flight path selection device, aircraft and air traffic management system
CN104850009A (en) Coordination control method for multi-unmanned aerial vehicle team based on predation escape pigeon optimization
CN103578299B (en) A kind of method simulating aircraft process
CN113296537A (en) Electric power unmanned aerial vehicle inspection method and system based on electric power tower model matching
CN114879716B (en) Law enforcement unmanned aerial vehicle path planning method for countering low-altitude airspace aircraft
CN115454115A (en) Rotor unmanned aerial vehicle path planning method based on hybrid wolf-particle swarm algorithm
CN115903888A (en) Rotor unmanned aerial vehicle autonomous path planning method based on longicorn swarm algorithm
Wang et al. Design of agent training environment for aircraft landing guidance based on deep reinforcement learning
CN105759630B (en) Aircraft 4D track Simulations system and emulation mode based on fuzzy-adaptation PID control
CN113283727B (en) Airport taxiway scheduling method based on quantum heuristic algorithm
CN114679729A (en) Radar communication integrated unmanned aerial vehicle cooperative multi-target detection method
CN114020009A (en) Terrain penetration planning method for small-sized fixed-wing unmanned aerial vehicle
Zhang et al. UAV path planning based on receding horizon control with adaptive strategy
Lee et al. Predictive control for soaring of unpowered autonomous UAVs

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant