CN115574826A - National park unmanned aerial vehicle patrol path optimization method based on reinforcement learning - Google Patents

National park unmanned aerial vehicle patrol path optimization method based on reinforcement learning Download PDF

Info

Publication number
CN115574826A
CN115574826A CN202211572414.2A CN202211572414A CN115574826A CN 115574826 A CN115574826 A CN 115574826A CN 202211572414 A CN202211572414 A CN 202211572414A CN 115574826 A CN115574826 A CN 115574826A
Authority
CN
China
Prior art keywords
path
unmanned aerial
aerial vehicle
energy consumption
task
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202211572414.2A
Other languages
Chinese (zh)
Other versions
CN115574826B (en
Inventor
郭强辉
殷虹娇
张鹏
王永峰
宋尚源
刘兆泽
高琳
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Deepiot Technology Co ltd
Nankai University
Original Assignee
Beijing Deepiot Technology Co ltd
Nankai University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Deepiot Technology Co ltd, Nankai University filed Critical Beijing Deepiot Technology Co ltd
Priority to CN202211572414.2A priority Critical patent/CN115574826B/en
Publication of CN115574826A publication Critical patent/CN115574826A/en
Application granted granted Critical
Publication of CN115574826B publication Critical patent/CN115574826B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C21/00Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
    • G01C21/20Instruments for performing navigational calculations
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/04Forecasting or optimisation specially adapted for administrative or management purposes, e.g. linear programming or "cutting stock problem"
    • G06Q10/047Optimisation of routes or paths, e.g. travelling salesman problem
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Business, Economics & Management (AREA)
  • Radar, Positioning & Navigation (AREA)
  • Remote Sensing (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Human Resources & Organizations (AREA)
  • Strategic Management (AREA)
  • Economics (AREA)
  • Tourism & Hospitality (AREA)
  • Biophysics (AREA)
  • Entrepreneurship & Innovation (AREA)
  • Marketing (AREA)
  • Operations Research (AREA)
  • Quality & Reliability (AREA)
  • Development Economics (AREA)
  • General Business, Economics & Management (AREA)
  • Automation & Control Theory (AREA)
  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Biomedical Technology (AREA)
  • Game Theory and Decision Science (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)
  • Traffic Control Systems (AREA)

Abstract

The invention discloses a national park Unmanned Aerial Vehicle (UAV) patrol path optimization method based on reinforcement learning, which comprises the steps of taking an unmanned aerial vehicle flight path as an optimization target, adding constraint conditions of unmanned aerial vehicle traversal path points, unmanned aerial vehicle electric quantity limitation and path point task execution energy consumption, and establishing an UAV path planning model with a self-service charging function; then respectively corresponding the unmanned aerial vehicle, the path points, the charging base station, the energy, the battery capacity, the flight path energy consumption and the path point task energy consumption in the unmanned aerial vehicle path planning model to a CVRP problem model; the unmanned aerial vehicle patrol route planning problem which needs to consider side energy consumption constraint and point energy consumption constraint originally is reduced into a CVRP problem which takes the route length as an optimization target and takes the customer demand and the vehicle load as constraints by using a feedforward weighting method; and finally, solving the reduced CVRP problem by using a multi-decoder attention model.

Description

National park unmanned aerial vehicle patrol path optimization method based on reinforcement learning
Technical Field
The invention belongs to the technical field of computer intelligent calculation and unmanned aerial vehicle flight control, and particularly relates to a national park unmanned aerial vehicle patrol route optimization method based on reinforcement learning.
Background
The field patrol monitoring is the most important ecological monitoring and daily supervision means in national parks and natural conservation places, and a patrol guard collects data in the aspects of wild species population, habitat, phenology and the like through patrol monitoring, can timely discover ecological environment problems, inhibit illegal activities and the like, realizes effective protection on the national parks and the natural conservation places, and provides decision basis for natural resource supervision. However, national parks and natural protection lands have large areas, wide ranges and complex terrains, people and vehicles in most regions are difficult to reach, and the traditional manual patrol mode has low efficiency, wastes time and labor. Therefore, in recent years, unmanned aerial vehicles are increasingly used for patrol monitoring work of various natural protection places.
The unmanned aerial vehicle technology is an unmanned aerial vehicle remote sensing technology which is realized by fusing an aircraft technology, a communication technology, a GPS (global positioning system), a differential positioning technology and an image technology, and automatic acquisition and transmission of monitoring data are realized by carrying sensing equipment such as a high-definition camera and an intelligent sensor and combining a wireless communication network. The existing unmanned aerial vehicle used for patrol monitoring of national parks and natural conservation places has the challenges of short endurance, high requirement on flight control personnel, difficult storage and transportation of airplanes, high application integration difficulty and the like, and is difficult to meet the application requirements of normalized monitoring.
The automatic airport of unmanned aerial vehicle is the ground automation facility of assisting unmanned aerial vehicle full flow operation, for unmanned aerial vehicle provides all-weather protection, through automatic opening and shutting, go up and down, get and unload structural design, let unmanned aerial vehicle take off, descend, deposit and battery management all can accomplish automatically, need not artificial intervention. The unmanned aerial vehicle is stored in the automatic airport, and when flight demands exist, the unmanned aerial vehicle takes off from the airport autonomously, and automatically lands in the automatic airport after the task is finished, so that charging is carried out in the automatic airport, preparation is made for the next task, and full-automatic operation is realized.
For realizing the normalized development of unmanned aerial vehicle in national park and the ecological monitoring work of nature protected area, satisfy the field and patrol and protect the monitoring management demand, this patent carries out path planning, electric quantity state control, commander's dispatch to unmanned aerial vehicle based on the automatic airport of unmanned aerial vehicle, and very big degree promotes unmanned aerial vehicle and patrols and protects monitoring efficiency.
Disclosure of Invention
The invention aims to overcome the defects of the prior art and provides a national park unmanned aerial vehicle patrol route optimization method based on reinforcement learning.
The invention is realized by the following technical scheme:
a national park unmanned aerial vehicle patrol path optimization method based on reinforcement learning comprises the following steps:
step 1: inputting three-dimensional terrain data to generate a bounded three-dimensional region
Figure 654531DEST_PATH_IMAGE001
According to the performance and patrol requirement of the airborne camera of the unmanned aerial vehicle, a path point set is set above the area in the air
Figure 681393DEST_PATH_IMAGE002
The unmanned aerial vehicle is required to complete the visual coverage task after traversing all path points;
and 2, step: taking the flight path of the unmanned aerial vehicle as an optimization target, adding constraint conditions of traversal path points of the unmanned aerial vehicle, electric quantity limitation of the unmanned aerial vehicle and task execution energy consumption of the path points, and establishing an unmanned aerial vehicle path planning model with a self-service charging function;
and step 3: respectively corresponding unmanned aerial vehicle, path points, charging base stations, energy, battery capacity, flight path energy consumption and path point task energy consumption in the established unmanned aerial vehicle path planning model with the self-service charging function to vehicles, customers, warehouses, goods, the maximum cargo capacity of the vehicles, the path length and customer requirements in the CVRP problem model; defining new path point task energy consumption by using a feedforward weighting method, so that the new path point task energy consumption comprises the task energy consumption of a path point and the average edge energy consumption reaching the path point; corresponding the obtained new path point task energy consumption to the client requirement of the CVRP problem model, and further reducing the unmanned aerial vehicle patrol path planning problem into a CVRP problem which takes the path length as an optimization target and takes the client requirement and the vehicle cargo load as constraints;
and 4, step 4: the CVRP problem reduced in step 3 is solved using a multi-decoder attention model.
In the above technical solution, in step 2, an unmanned aerial vehicle path planning model with a self-service charging function is established, and the specific steps are as follows:
step 2.1: defining flight path decision variables for dronesx ij
x ij =1, representing unmanned aerial vehicle from a waypointiFly to the waypointj
x ij =0, meaning that the drone is not following a waypointiFly to the waypointj
Defining an objective function:
Figure 459862DEST_PATH_IMAGE003
(1)
wherein the content of the first and second substances,
Figure 888569DEST_PATH_IMAGE004
is flight path energy consumption and represents the path point of the unmanned planeiAnd a waypointjEnergy consumption is needed;
the flight path decision variables are to form a complete and feasible one-time traversal path, and the constraints are as follows:
Figure 336868DEST_PATH_IMAGE005
(2)
Figure 116605DEST_PATH_IMAGE006
(3)
step 2.2: aiming at the self-service charging function of the unmanned aerial vehicle, the route planning with the charging base station is adjusted, the energy consumption of the unmanned aerial vehicle is measured according to the flight path, and the maximum endurance of the unmanned aerial vehicle is recorded asQDefining the energy loss variable
Figure 183918DEST_PATH_IMAGE007
The charging base station is the starting point of the unmanned aerial vehicle and is recorded as
Figure 467132DEST_PATH_IMAGE008
Remaining range of the drone during performance of the mission not exceeding maximum range
Figure 351911DEST_PATH_IMAGE009
Is expressed as follows:
Figure 353365DEST_PATH_IMAGE010
(4)
Figure 441014DEST_PATH_IMAGE011
(5)
wherein the content of the first and second substances,
Figure 906630DEST_PATH_IMAGE012
is a path point
Figure 899994DEST_PATH_IMAGE013
Task energy consumption, representing the point of unmanned aerial vehicle completing path
Figure 60848DEST_PATH_IMAGE013
The required energy consumption of the patrol task is reduced,
Figure 797860DEST_PATH_IMAGE014
representing points of a path
Figure 117983DEST_PATH_IMAGE015
Points of other paths
Figure 16669DEST_PATH_IMAGE016
To the path point
Figure 179666DEST_PATH_IMAGE015
The decision variables of the edges of (a) are,
Figure 454789DEST_PATH_IMAGE017
indicating unmanned aerial vehicle slave waypoints
Figure 629419DEST_PATH_IMAGE016
Performing a mission to fly to a waypoint
Figure 636689DEST_PATH_IMAGE015
The residual energy after the reaction;
when the unmanned aerial vehicle leaves the charging base station, the electric quantity is full, and the formula is as follows:
Figure 834452DEST_PATH_IMAGE018
(6)
Figure 975583DEST_PATH_IMAGE019
indicating that the unmanned aerial vehicle leaves the charging base station to reach the waypoint
Figure 207982DEST_PATH_IMAGE020
The residual energy of the waste water is the energy,
Figure 136886DEST_PATH_IMAGE021
indicating that the unmanned aerial vehicle flies to a waypoint from a charging base station
Figure 87524DEST_PATH_IMAGE015
The decision-making variables of (a) are,
Figure 766767DEST_PATH_IMAGE022
is a path point
Figure 525776DEST_PATH_IMAGE015
Task energy consumption, representing the point of unmanned aerial vehicle completing path
Figure 937165DEST_PATH_IMAGE015
Energy consumption required by the patrol task.
In the above technical scheme, in step 3, firstly, under the condition that the edge energy consumption constraint between the path point and the path point is not considered, a deep reinforcement learning method is used to independently solve the CVRP problem corresponding to the unmanned aerial vehicle patrol path for multiple times, and the number of the solution times is recorded as
Figure 171838DEST_PATH_IMAGE023
And training the neural network in the deep reinforcement learning model again every time of solving, and using the neural network trained every time for predicting the CVRP problem corresponding to the original unmanned aerial vehicle patrol problem
Figure 592455DEST_PATH_IMAGE023
The secondary solution is obtained
Figure 720817DEST_PATH_IMAGE023
Grouping different solutions to form a solution set
Figure 568687DEST_PATH_IMAGE024
Solution set
Figure 290655DEST_PATH_IMAGE024
Therein comprises
Figure 249384DEST_PATH_IMAGE023
Planting a patrol path scheme;
redefining new task point energy consumption variables on the basis of the known solution set
Figure 248564DEST_PATH_IMAGE025
Figure 267336DEST_PATH_IMAGE026
(7)
Wherein the content of the first and second substances,
Figure 211021DEST_PATH_IMAGE027
representing points of a path
Figure 973440DEST_PATH_IMAGE028
To the path point
Figure 574930DEST_PATH_IMAGE029
Is in the solution set
Figure 764603DEST_PATH_IMAGE024
The number of occurrences in (1) is equivalent to the weighted average of the path energy consumption required for reaching any path point, and the weight is
Figure 461163DEST_PATH_IMAGE030
Then the solution set is obtained by optimizing the path length of the reference total patrol task
Figure 761694DEST_PATH_IMAGE024
In the above technical solution, the solving process of step 4 includes the following steps:
step 4.1: firstly, according to the scale of input information, several groups of data sets with identical path point quantity are produced, and said data sets are equipped with
Figure 469887DEST_PATH_IMAGE031
Group data set, first
Figure 830462DEST_PATH_IMAGE032
The information in the group dataset comprises a randomly generated starting point
Figure 748739DEST_PATH_IMAGE033
And the position of the path point
Figure 852961DEST_PATH_IMAGE034
And randomly generated waypoint task energy consumption
Figure 930508DEST_PATH_IMAGE035
Wherein
Figure 461983DEST_PATH_IMAGE036
Step 4.2: using generated
Figure 867557DEST_PATH_IMAGE031
Training the multi-decoder attention model in a block data set, where the parameters of the encoder and decoder are
Figure 509891DEST_PATH_IMAGE037
The model is trained by a strategy gradient algorithm with baseline, and parameters of the optimized model are continuously updated circularly to obtain a trained attention model of the multi-decoder;
step 4.3: after the training of the model parameters is finished, inputting the data of the task planning problem of the original unmanned aerial vehicle as a reduced CVRP problem example into the trained model, and taking the output sequence of the model at the moment as a path point access scheme of the unmanned aerial vehicle patrol problem.
In the above technical solution, in step 4.3, the data of the original unmanned aerial vehicle mission planning problem includes a starting point
Figure 192676DEST_PATH_IMAGE038
Figure 895053DEST_PATH_IMAGE039
A path point
Figure 787922DEST_PATH_IMAGE040
And information of energy consumption of each path point task, wherein the energy consumption of the path point task refers to the energy consumption of the new path point task defined in the step 2.
The invention has the advantages and beneficial effects that:
the base station is introduced to provide real-time charging service for the working unmanned aerial vehicle, and the unmanned aerial vehicle can access the base station to perform charging for multiple times when executing tasks. Under the system, a constraint formula is constructed by taking the optimized unmanned aerial vehicle task path length as a target, a multi-unmanned aerial vehicle path planning model is established, and the problem is converted into a combined optimization problem. A known combined optimization solver is utilized, a feedforward weighting method is designed to calculate the path energy consumption constraint, and the problem is further converted into a vehicle path problem (CVRP) with capacity limitation. In addition, the deep reinforcement learning method based on the multi-decoder attention model can stably output a high-quality solution of a visual coverage problem for a specific scene, has generalization capability for solving the reduced unmanned aerial vehicle path planning problem, has strong adaptability to a training data set, and can guarantee an efficient training network for path planning under different scenes to obtain the high-quality solution. Based on a trained learning model, the result can be quickly obtained by only calling neural network prediction after the unmanned aerial vehicle path problem example is reduced, the solving speed is higher than the efficiency of the traditional search algorithm, and the decision requirement of the unmanned aerial vehicle quick scheduling planning can be met.
Drawings
FIG. 1 is a flow chart of the national park unmanned aerial vehicle patrol route optimization method based on reinforcement learning.
FIG. 2 is a flow chart of a solution of a multi-decoder attention model to an example problem.
For a person skilled in the art, other relevant figures can be obtained from the above figures without inventive effort.
Detailed Description
In order to make the technical solution of the present invention better understood, the technical solution of the present invention is further described below with reference to specific examples.
A national park unmanned aerial vehicle patrol path optimization method based on reinforcement learning is disclosed, referring to the attached figure 1, and comprises the following steps:
step 1: inputting three-dimensional terrain data to generate a bounded three-dimensional region
Figure 233947DEST_PATH_IMAGE041
According to the unmanned plane on-board shootingHead-like performance and patrol requirements set a set of waypoints in the air above the area
Figure 787550DEST_PATH_IMAGE042
Obtaining initial data
Figure 660828DEST_PATH_IMAGE043
And the unmanned aerial vehicle is required to complete the visual coverage task after traversing all path points.
Step 2: and establishing a constraint formula, taking the flight path of the unmanned aerial vehicle as an optimization target, adding constraint conditions of traversal path points of the unmanned aerial vehicle, electric quantity limitation of the unmanned aerial vehicle and energy consumption of task execution of the path points, and establishing an unmanned aerial vehicle path planning model with a self-service charging function without considering uncontrollable factors such as wind power, visibility and unmanned aerial vehicle faults. The method comprises the following specific steps.
Step 2.1: defining flight path decision variables for an unmanned aerial vehiclex ij
x ij =1, representing unmanned aerial vehicle from a waypointiFly to the waypointj
x ij =0, meaning that the drone is not following a waypointiFly to the waypointj
Defining an objective function:
Figure 775415DEST_PATH_IMAGE044
(1)
wherein the content of the first and second substances,
Figure 25131DEST_PATH_IMAGE045
is flight path energy consumption and represents the path point of the unmanned planeiAnd a waypointjThe energy consumption generated between the unmanned aerial vehicle and the unmanned aerial vehicle is in direct proportion to the distance between the path points, and the aim of the task is to optimize the flight path of the unmanned aerial vehicle and minimize the flight path on the premise of completing the task aim. Meanwhile, the flight path decision variables need to ensure that a complete and feasible one-time traversal path can be formed, and the specific constraints are as follows:
Figure 682508DEST_PATH_IMAGE046
(2)
Figure 726688DEST_PATH_IMAGE047
(3)
Step 2.2: aiming at the self-service charging function of the unmanned aerial vehicle, the route planning with the charging base station is adjusted, the energy consumption of the unmanned aerial vehicle is measured according to the flight path, and the maximum endurance of the unmanned aerial vehicle is recorded asQDefining the energy loss variable
Figure 328570DEST_PATH_IMAGE048
Charging base station is the departure point of the unmanned aerial vehicle and is recorded
Figure 116398DEST_PATH_IMAGE049
First, the drone consumes energy as it moves between waypoints and the remaining range of the drone during the mission should not exceed the maximum range
Figure 877549DEST_PATH_IMAGE050
Is given by the following equation:
Figure 92630DEST_PATH_IMAGE051
(4)
Figure 447388DEST_PATH_IMAGE011
(5)
wherein, the first and the second end of the pipe are connected with each other,
Figure 773327DEST_PATH_IMAGE052
is a path point
Figure 405297DEST_PATH_IMAGE053
Task energy consumption, representing the point of unmanned aerial vehicle completing path
Figure 588016DEST_PATH_IMAGE053
The energy consumption required by the patrol task is reduced,
Figure 367753DEST_PATH_IMAGE054
representing points of a path
Figure 497383DEST_PATH_IMAGE055
Points of other routes
Figure 466083DEST_PATH_IMAGE056
To the path point
Figure 85283DEST_PATH_IMAGE055
The decision variables of the edges of (a) are,
Figure 758841DEST_PATH_IMAGE057
representing unmanned aerial vehicle slave waypoints
Figure 692162DEST_PATH_IMAGE056
Performing a mission to fly to a waypoint
Figure 157778DEST_PATH_IMAGE055
The remaining energy (i.e., electricity).
Secondly, when unmanned aerial vehicle leaves charging base station, the electric quantity is full, and the formula is expressed as follows:
Figure 885563DEST_PATH_IMAGE058
(6)
Figure 561264DEST_PATH_IMAGE059
indicating that the unmanned aerial vehicle leaves the charging base station to reach the waypoint
Figure 298276DEST_PATH_IMAGE055
The residual energy of the waste water is the energy,
Figure 618399DEST_PATH_IMAGE060
indicating that the drone is flying from the charging base stationTo the path point
Figure 517084DEST_PATH_IMAGE055
The decision variable(s) of (a),
Figure 430814DEST_PATH_IMAGE061
is a path point
Figure 705937DEST_PATH_IMAGE055
Task energy consumption, representing the point of unmanned aerial vehicle completing path
Figure 880567DEST_PATH_IMAGE055
Energy consumption required by the patrol task.
In conclusion, an unmanned aerial vehicle path planning model with a self-service charging function is established, and the model comprises an objective function (1) and constraint formulas (2), (3), (4), (5) and (6). The solution of this model is a combinatorial optimization problem promptly, that is to say, the unmanned aerial vehicle patrols the route planning problem and transforms for a combinatorial optimization problem.
And step 3: referring to table 1, the unmanned aerial vehicle, the waypoints, the charging base station, the energy (i.e., the electric quantity), the battery capacity, the flight path energy consumption, and the waypoint task energy consumption in the unmanned aerial vehicle path planning model with the self-service charging function, which are established as above, are respectively corresponding to the maximum cargo capacity, the path length, and the customer demand of the vehicle, the customer, the warehouse, the goods, and the vehicle in the CVRP problem (the capacity-limited vehicle path solving problem) model, and then the unmanned aerial vehicle path planning model is converted into the capacity-limited vehicle path solving problem (CVRP).
Table 1: correspondence between unmanned aerial vehicle path planning and CVRP problem model
Figure 215733DEST_PATH_IMAGE062
The energy consumption of the unmanned aerial vehicle comprises the side energy consumption from the path point to the path point and the point energy consumption required by the path point to complete the patrol task, but in the CVRP problem model, the side energy consumption is only used as an optimization target for planning the vehicle path, and only the point energy consumption is used as a constraint condition of the vehicle path. Therefore, the invention uses a feedforward weighting method to enable point energy consumption to replace 'point plus edge energy consumption', and then add edge energy consumption into the constraint condition, so that the problem of unmanned aerial vehicle patrol route planning which originally needs to consider edge energy consumption constraint and point energy consumption constraint is reduced to a CVRP problem which takes the route length as an optimization target and takes customer requirements and vehicle cargo as constraints. The specific treatment method is as follows.
Firstly, under the condition of not considering the limit energy consumption constraint, a deep reinforcement learning method is used for independently solving the CVRP problem corresponding to the unmanned aerial vehicle patrol path for multiple times, and the solving times are recorded as
Figure 101912DEST_PATH_IMAGE063
And (2) training the neural network in the deep reinforcement learning model again (or independently) every time of solving, using the neural network trained every time for predicting the CVRP problem corresponding to the original unmanned aerial vehicle patrol problem, wherein the generation and extraction of the training set are random, so that the method has the advantages of high reliability, high accuracy and low cost
Figure 180726DEST_PATH_IMAGE063
Total of sub-training
Figure 209862DEST_PATH_IMAGE063
The neural networks are different, and the prediction results of the neural networks are different, so that the neural networks can obtain
Figure 715930DEST_PATH_IMAGE063
Grouping different solutions to form a solution set
Figure 338672DEST_PATH_IMAGE064
Solution set
Figure 955598DEST_PATH_IMAGE064
Therein comprises
Figure 104820DEST_PATH_IMAGE063
And a patrol path scheme is adopted.
Redefining new path point task energy consumption based on known solution set
Figure 781789DEST_PATH_IMAGE065
(i.e., waypoints)
Figure 141095DEST_PATH_IMAGE066
Energy consumption required for completing the patrol task):
Figure 561712DEST_PATH_IMAGE067
(7)
wherein the content of the first and second substances,
Figure 565440DEST_PATH_IMAGE068
representing points of a path
Figure 413310DEST_PATH_IMAGE069
To the path point
Figure 10645DEST_PATH_IMAGE070
Is in the solution set
Figure 969374DEST_PATH_IMAGE071
The number of occurrences in (1) is equivalent to weighted average of the path energy consumption required for reaching any path point, and the weight is
Figure 93188DEST_PATH_IMAGE072
Then refer to the solution set optimized by the total patrol task path length
Figure 846380DEST_PATH_IMAGE071
The obtained new path point task energy consumption
Figure 678813DEST_PATH_IMAGE065
Customer requirements for the CVRP problem model so that new waypoint tasks consume energy
Figure 441233DEST_PATH_IMAGE065
The task energy consumption of a path point and the average side energy consumption for reaching the path point are included, and the patrol path problem which originally needs to consider side energy consumption constraint and point energy consumption constraint is reduced to a CVRP problem which takes path length as an optimization target and takes customer demand and vehicle cargo as constraints.
And 4, step 4: the CVRP problem reduced in step 3 is solved using a multi-decoder attention model.
The data of the unmanned aerial vehicle path planning problem comprises a starting point
Figure 419553DEST_PATH_IMAGE073
Information and
Figure 609226DEST_PATH_IMAGE074
a path point
Figure 915574DEST_PATH_IMAGE075
And information of task energy consumption of each path point (the path point task energy consumption refers to new path point task energy consumption defined in the step 2), and the information is reduced to information of warehouse, client demand and the like in the CVRP problem according to the step 3 and is used as input information of the model. The encoder structure of the model is based on a transformer model, a plurality of decoders with the same structure and independent parameters are used in a decoder part, the difference degree of construction solutions between the decoders is measured by Kullback-Leibler divergence (abbreviated as 'KL divergence') between probability distributions calculated by different decoders, and in addition, each decoder increases the masking of nodes when calculating attention weights and is used for realizing task path constraint in the CVRP problem. The model is trained by a policy gradient algorithm with baseline and a plurality of data sets which are randomly generated and have the same scale with the problem to be solved. Referring to fig. 2, the specific solving process is as follows.
Step 4.1: firstly, groups with the same path point number (namely, the same path point number) are generated according to the scale of input information
Figure 216105DEST_PATH_IMAGE076
) Assuming common data sets of
Figure 314511DEST_PATH_IMAGE077
Group data set, in order
Figure 675085DEST_PATH_IMAGE078
For the example of a group dataset, the information therein includes a randomly generated starting point
Figure 717996DEST_PATH_IMAGE079
And the position of the path point
Figure 822219DEST_PATH_IMAGE080
And randomly generated waypoint task energy consumption
Figure 775131DEST_PATH_IMAGE081
Wherein
Figure 978711DEST_PATH_IMAGE082
Step 4.2: using generated
Figure 321967DEST_PATH_IMAGE083
Training the multi-decoder attention model in a block data set, where the parameters of the encoder and decoder are
Figure 229880DEST_PATH_IMAGE084
The model is trained by a policy gradient algorithm with baseline, model parameters are continuously updated and optimized in a circulating mode, the training target is the model parameters for optimizing the shortest path length of a client access scheme and KL divergence of decoder parameters, and the model parameters are recorded
Figure 37299DEST_PATH_IMAGE085
The total length of the task path is obtained for the solution under the model parameters, and is recorded
Figure 162512DEST_PATH_IMAGE086
And (4) carrying out parameter training for the KL divergence of the decoder parameters under the model parameters according to the following algorithm to obtain the trained attention model of the multi-decoder.
The reinforcement learning algorithm with baseline is as follows:
1, inputting
Figure 258644DEST_PATH_IMAGE087
Group dataset, significance level
Figure 704669DEST_PATH_IMAGE088
Training period
Figure 366595DEST_PATH_IMAGE089
2, initializing model parameters
Figure 239873DEST_PATH_IMAGE084
3, recording baseline parameters
Figure 495405DEST_PATH_IMAGE090
4, current number of training times
Figure 541858DEST_PATH_IMAGE091
5, combining the optimization objectives according to the current
Figure 261553DEST_PATH_IMAGE087
Group dataset and parameters
Figure 305732DEST_PATH_IMAGE084
Calculating the task path length and KL divergence of the output result of the model
Figure 297828DEST_PATH_IMAGE084
Optimizing direction
Figure 85655DEST_PATH_IMAGE092
6 according to the optimization direction
Figure 722173DEST_PATH_IMAGE092
Updating parameters using Adam function
Figure 937253DEST_PATH_IMAGE084
7, using t test parameters
Figure 167378DEST_PATH_IMAGE084
And Baseline parameters
Figure 493317DEST_PATH_IMAGE090
If the significance is less than
Figure 984341DEST_PATH_IMAGE088
Update, update
Figure 635902DEST_PATH_IMAGE084
8, if
Figure 835546DEST_PATH_IMAGE093
Figure 965176DEST_PATH_IMAGE094
Returning to the step 5; otherwise, turning to the next step;
9, training is finished, and the obtained parameters are
Figure 576286DEST_PATH_IMAGE084
The multi-decoder attention model of (1).
Step 4.3: after the training of the model parameters is finished, the data (including the starting point) of the original unmanned aerial vehicle mission planning problem is processed
Figure 133169DEST_PATH_IMAGE095
Figure 72306DEST_PATH_IMAGE096
A path point
Figure 5627DEST_PATH_IMAGE097
And information of task energy consumption of each path point) as a reduced CVRP problem instance, inputting the trained model, and taking an output sequence of the model at the moment as a path of the unmanned aerial vehicle patrol problemA point access scheme.
The invention being thus described by way of example, it should be understood that any simple alterations, modifications or other equivalent alterations as would be within the skill of the art without the exercise of inventive faculty, are within the scope of the invention.

Claims (5)

1. A national park Unmanned Aerial Vehicle (UAV) patrol path optimization method based on reinforcement learning is characterized by comprising the following steps of:
step 1: inputting three-dimensional terrain data to generate a bounded three-dimensional region
Figure 538132DEST_PATH_IMAGE001
According to the performance and patrol requirement of the airborne camera of the unmanned aerial vehicle, a path point set is set above the area in the air
Figure 675852DEST_PATH_IMAGE002
The unmanned aerial vehicle is required to complete the visual coverage task after traversing all path points;
step 2: taking the flight path of the unmanned aerial vehicle as an optimization target, adding constraint conditions of traversal path points of the unmanned aerial vehicle, electric quantity limitation of the unmanned aerial vehicle and energy consumption of task execution of the path points, and establishing an unmanned aerial vehicle path planning model with a self-service charging function;
and 3, step 3: respectively corresponding unmanned aerial vehicle, path points, charging base stations, energy, battery capacity, flight path energy consumption and path point task energy consumption in the established unmanned aerial vehicle path planning model with the self-service charging function to vehicles, customers, warehouses, goods, the maximum cargo capacity of the vehicles, the path length and customer requirements in the CVRP problem model; defining new path point task energy consumption by using a feedforward weighting method, so that the new path point task energy consumption comprises the task energy consumption of a path point and the average edge energy consumption reaching the path point; corresponding the obtained new path point task energy consumption to the client requirement of the CVRP problem model, and further reducing the unmanned aerial vehicle patrol path planning problem into a CVRP problem which takes the path length as an optimization target and takes the client requirement and the vehicle cargo load as constraints;
and 4, step 4: the CVRP problem reduced in step 3 is solved using a multi-decoder attention model.
2. The reinforcement learning-based national park unmanned aerial vehicle patrol route optimization method according to claim 1, wherein: in step 2, an unmanned aerial vehicle path planning model with a self-service charging function is established, and the method specifically comprises the following steps:
step 2.1: defining flight path decision variables for an unmanned aerial vehiclex ij
x ij =1, representing unmanned aerial vehicle from a waypointiFly to waypointj
x ij =0, meaning that the drone is not following a waypointiFly to waypointj
Defining an objective function:
Figure 606899DEST_PATH_IMAGE003
(1)
wherein the content of the first and second substances,
Figure 95650DEST_PATH_IMAGE004
is flight path energy consumption and represents the path point of the unmanned planeiAnd a waypointjEnergy consumption is needed;
the flight path decision variables need to form a complete and feasible one-time traversal path, and the constraints are as follows:
Figure 629399DEST_PATH_IMAGE005
(2)
Figure 887205DEST_PATH_IMAGE006
(3)
step 2.2: for no oneThe self-service charging function of the unmanned aerial vehicle is adjusted to adjust the route planning with the charging base station, the energy consumption of the unmanned aerial vehicle is measured according to the flight path, and the maximum endurance of the unmanned aerial vehicle is recorded asQDefining the energy loss variable
Figure 739886DEST_PATH_IMAGE007
The charging base station is the starting point of the unmanned aerial vehicle and is recorded as
Figure 450353DEST_PATH_IMAGE008
The remaining range of the drone during execution of the mission does not exceed the maximum range
Figure 522214DEST_PATH_IMAGE009
Is expressed as follows:
Figure 837789DEST_PATH_IMAGE010
(4)
Figure 907376DEST_PATH_IMAGE011
(5)
wherein the content of the first and second substances,
Figure 433035DEST_PATH_IMAGE012
is a path point
Figure 246270DEST_PATH_IMAGE013
Task energy consumption, representing the point of path completed by the unmanned aerial vehicle
Figure 665619DEST_PATH_IMAGE013
The energy consumption required by the patrol task is reduced,
Figure 906108DEST_PATH_IMAGE014
representing points of a path
Figure 919063DEST_PATH_IMAGE015
Points of other routes
Figure 535989DEST_PATH_IMAGE016
To the path point
Figure 560577DEST_PATH_IMAGE015
The decision variables of the edges of (a) are,
Figure 34284DEST_PATH_IMAGE017
indicating unmanned aerial vehicle slave waypoints
Figure 206639DEST_PATH_IMAGE016
Performing a mission to fly to a waypoint
Figure 47163DEST_PATH_IMAGE015
The residual energy after the reaction;
when unmanned aerial vehicle leaves charging base station, the electric quantity is full, and the formula is as follows:
Figure 254153DEST_PATH_IMAGE018
(6)
Figure 898761DEST_PATH_IMAGE019
indicating that the unmanned aerial vehicle leaves the charging base station to reach the waypoint
Figure 558413DEST_PATH_IMAGE015
The residual energy of the waste water is the energy,
Figure 454824DEST_PATH_IMAGE020
indicating that the unmanned aerial vehicle flies to a waypoint from a charging base station
Figure 516321DEST_PATH_IMAGE015
The decision-making variables of (a) are,
Figure 597410DEST_PATH_IMAGE021
is a path point
Figure 744357DEST_PATH_IMAGE015
Task energy consumption, representing the point of path completed by the unmanned aerial vehicle
Figure 428149DEST_PATH_IMAGE015
Energy consumption required by the patrol task.
3. The reinforcement learning-based national park unmanned aerial vehicle patrol route optimization method according to claim 2, wherein: in step 3, firstly, under the condition of not considering the edge energy consumption constraint between the path points, a deep reinforcement learning method is used for independently solving the CVRP problem corresponding to the unmanned aerial vehicle patrol path for multiple times, and the solving times are recorded as
Figure 140890DEST_PATH_IMAGE022
And (3) retraining the neural network in the deep reinforcement learning model every time of solving, and using the neural network trained every time to predict the CVRP problem corresponding to the original unmanned aerial vehicle patrol problem
Figure 330562DEST_PATH_IMAGE022
The secondary solution is obtained
Figure 902489DEST_PATH_IMAGE022
Grouping different solutions to form solution sets
Figure 203021DEST_PATH_IMAGE023
Solution set
Figure 35847DEST_PATH_IMAGE023
Therein comprises
Figure 396422DEST_PATH_IMAGE022
A patrol path scheme is planted;
in a known solution setOn the basis, redefining new task point energy consumption variables
Figure 206377DEST_PATH_IMAGE024
Figure 45020DEST_PATH_IMAGE025
(7)
Wherein, the first and the second end of the pipe are connected with each other,
Figure 997932DEST_PATH_IMAGE026
representing points of a path
Figure 201512DEST_PATH_IMAGE015
To the path point
Figure 544768DEST_PATH_IMAGE013
Is in the solution set
Figure 514998DEST_PATH_IMAGE027
The number of occurrences in (1) is equivalent to the weighted average of the path energy consumption required for reaching any path point, and the weight is
Figure 260101DEST_PATH_IMAGE026
Then the solution set is obtained by optimizing the path length of the reference total patrol task
Figure 149428DEST_PATH_IMAGE027
4. The reinforcement learning-based national park unmanned aerial vehicle patrol route optimization method according to claim 1, wherein: the solving process of the step 4 comprises the following steps:
step 4.1: firstly, according to the scale of input information, several groups of data sets with same path point quantity are generated, and said method is characterized by that
Figure 979981DEST_PATH_IMAGE028
Group data set, first
Figure 222743DEST_PATH_IMAGE015
The information in the group dataset comprises a randomly generated starting point
Figure 87931DEST_PATH_IMAGE029
And the position of the path point
Figure 633313DEST_PATH_IMAGE030
And randomly generated waypoint task energy consumption
Figure 13479DEST_PATH_IMAGE031
Wherein
Figure 997615DEST_PATH_IMAGE032
Step 4.2: using generated
Figure 402795DEST_PATH_IMAGE016
Training the multi-decoder attention model in a block data set, where the parameters of the encoder and decoder are
Figure 446975DEST_PATH_IMAGE033
The model is trained by a strategy gradient algorithm with baseline, and parameters of the optimized model are continuously updated circularly to obtain a trained attention model of the multi-decoder;
step 4.3: after the training of the model parameters is finished, inputting the data of the task planning problem of the original unmanned aerial vehicle as a reduced CVRP problem example into the trained model, and taking the output sequence of the model at the moment as a path point access scheme of the unmanned aerial vehicle patrol problem.
5. The reinforcement learning-based national park unmanned aerial vehicle patrol route optimization method according to claim 4, wherein: in step 4.3, the data of the original unmanned aerial vehicle mission planning problem comprises a starting point
Figure 314437DEST_PATH_IMAGE029
Figure 102264DEST_PATH_IMAGE034
A path point
Figure 348569DEST_PATH_IMAGE035
And information of energy consumption of each path point task, wherein the energy consumption of the path point task refers to the energy consumption of the new path point task defined in the step 2.
CN202211572414.2A 2022-12-08 2022-12-08 National park unmanned aerial vehicle patrol path optimization method based on reinforcement learning Active CN115574826B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202211572414.2A CN115574826B (en) 2022-12-08 2022-12-08 National park unmanned aerial vehicle patrol path optimization method based on reinforcement learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202211572414.2A CN115574826B (en) 2022-12-08 2022-12-08 National park unmanned aerial vehicle patrol path optimization method based on reinforcement learning

Publications (2)

Publication Number Publication Date
CN115574826A true CN115574826A (en) 2023-01-06
CN115574826B CN115574826B (en) 2023-04-07

Family

ID=84590469

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202211572414.2A Active CN115574826B (en) 2022-12-08 2022-12-08 National park unmanned aerial vehicle patrol path optimization method based on reinforcement learning

Country Status (1)

Country Link
CN (1) CN115574826B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116894519A (en) * 2023-07-21 2023-10-17 江苏舟行时空智能科技股份有限公司 Position point optimization determination method meeting user service coverage requirement

Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110263983A (en) * 2019-05-31 2019-09-20 中国人民解放军国防科技大学 Double-layer path planning method and system for logistics distribution of vehicles and unmanned aerial vehicles
CN110428111A (en) * 2019-08-08 2019-11-08 西安工业大学 Multi-Tasking method for planning track when UAV/UGV collaboration is long
CN110470301A (en) * 2019-08-13 2019-11-19 上海交通大学 Unmanned plane paths planning method under more dynamic task target points
CN111429052A (en) * 2020-03-16 2020-07-17 北京航空航天大学 Initial solution structure for vehicle path problem distributed by cooperating unmanned aerial vehicle
CN111536979A (en) * 2020-07-08 2020-08-14 浙江浙能天然气运行有限公司 Unmanned aerial vehicle routing inspection path planning method based on random optimization
CN112132312A (en) * 2020-08-14 2020-12-25 蓝海(福建)信息科技有限公司 Path planning method based on evolution multi-objective multi-task optimization
US20210020051A1 (en) * 2017-07-27 2021-01-21 Beihang University Airplane flight path planning method and device based on the pigeon-inspired optimization
US20210325195A1 (en) * 2020-04-20 2021-10-21 Insurance Services Office, Inc. Systems and Methods for Automated Vehicle Routing Using Relaxed Dual Optimal Inequalities for Relaxed Columns
CN114422363A (en) * 2022-01-11 2022-04-29 北京科技大学 Unmanned aerial vehicle loaded RIS auxiliary communication system capacity optimization method and device
CN115065939A (en) * 2022-06-08 2022-09-16 电子科技大学长三角研究院(衢州) Auxiliary communication unmanned aerial vehicle trajectory planning and power control method capable of charging in flight
CN115185303A (en) * 2022-09-14 2022-10-14 南开大学 Unmanned aerial vehicle patrol path planning method for national parks and natural protected areas
CN115280103A (en) * 2020-03-23 2022-11-01 支付宝实验室(新加坡)有限公司 System and method for determining a path by learning selective optimization

Patent Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20210020051A1 (en) * 2017-07-27 2021-01-21 Beihang University Airplane flight path planning method and device based on the pigeon-inspired optimization
CN110263983A (en) * 2019-05-31 2019-09-20 中国人民解放军国防科技大学 Double-layer path planning method and system for logistics distribution of vehicles and unmanned aerial vehicles
CN110428111A (en) * 2019-08-08 2019-11-08 西安工业大学 Multi-Tasking method for planning track when UAV/UGV collaboration is long
CN110470301A (en) * 2019-08-13 2019-11-19 上海交通大学 Unmanned plane paths planning method under more dynamic task target points
CN111429052A (en) * 2020-03-16 2020-07-17 北京航空航天大学 Initial solution structure for vehicle path problem distributed by cooperating unmanned aerial vehicle
CN115280103A (en) * 2020-03-23 2022-11-01 支付宝实验室(新加坡)有限公司 System and method for determining a path by learning selective optimization
US20210325195A1 (en) * 2020-04-20 2021-10-21 Insurance Services Office, Inc. Systems and Methods for Automated Vehicle Routing Using Relaxed Dual Optimal Inequalities for Relaxed Columns
CN111536979A (en) * 2020-07-08 2020-08-14 浙江浙能天然气运行有限公司 Unmanned aerial vehicle routing inspection path planning method based on random optimization
CN112132312A (en) * 2020-08-14 2020-12-25 蓝海(福建)信息科技有限公司 Path planning method based on evolution multi-objective multi-task optimization
CN114422363A (en) * 2022-01-11 2022-04-29 北京科技大学 Unmanned aerial vehicle loaded RIS auxiliary communication system capacity optimization method and device
CN115065939A (en) * 2022-06-08 2022-09-16 电子科技大学长三角研究院(衢州) Auxiliary communication unmanned aerial vehicle trajectory planning and power control method capable of charging in flight
CN115185303A (en) * 2022-09-14 2022-10-14 南开大学 Unmanned aerial vehicle patrol path planning method for national parks and natural protected areas

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
闵桂龙等: "军事后勤中的多目标无人机任务规划", 《计算机仿真》 *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116894519A (en) * 2023-07-21 2023-10-17 江苏舟行时空智能科技股份有限公司 Position point optimization determination method meeting user service coverage requirement

Also Published As

Publication number Publication date
CN115574826B (en) 2023-04-07

Similar Documents

Publication Publication Date Title
CN113342046B (en) Power transmission line unmanned aerial vehicle routing inspection path optimization method based on ant colony algorithm
CN105045274B (en) A kind of intelligent shaft tower connected graph construction method for unmanned plane inspection trajectory planning
CN106959700B (en) A kind of unmanned aerial vehicle group collaboration patrol tracing path planing method based on upper limit confidence interval algorithm
CN114169066B (en) Space target characteristic measuring and reconnaissance method based on micro-nano constellation approaching reconnaissance
CN110597286B (en) Method for realizing unmanned aerial vehicle autonomous inspection of power transmission line by using smart hangar
Liu et al. Application of unmanned aerial vehicle hangar in transmission tower inspection considering the risk probabilities of steel towers
CN115185303B (en) Unmanned aerial vehicle patrol path planning method for national parks and natural protected areas
CN113268081B (en) Small unmanned aerial vehicle prevention and control command decision method and system based on reinforcement learning
Liang et al. Drone fleet deployment strategy for large scale agriculture and forestry surveying
CN114638155A (en) Unmanned aerial vehicle task allocation and path planning method based on intelligent airport
Zheng et al. Robustness of the planning algorithm for ocean observation tasks
Zheng et al. The collaborative power inspection task allocation method of “unmanned aerial vehicle and operating vehicle”
CN115574826B (en) National park unmanned aerial vehicle patrol path optimization method based on reinforcement learning
CN116578120A (en) Unmanned aerial vehicle scheduling method and device, unmanned aerial vehicle system and computer equipment
Qiu et al. Improved F‐RRT∗ Algorithm for Flight‐Path Optimization in Hazardous Weather
Gaowei et al. Using multi-layer coding genetic algorithm to solve time-critical task assignment of heterogeneous UAV teaming
CN113283827B (en) Two-stage unmanned aerial vehicle logistics path planning method based on deep reinforcement learning
Hehtke et al. An Autonomous Mission Management System to Assist Decision Making of a HALE Operator
Li et al. Intelligent Early Warning Method Based on Drone Inspection
Sehrawat et al. A power prediction approach for a solar-powered aerial vehicle enhanced by stacked machine learning technique
Zheng Multimachine Collaborative Path Planning Method Based on A* Mechanism Connection Depth Neural Network Model
Dai et al. A genetic algorithm-based research on drone trajectory planning strategy of cooperative inspection of transmission lines, substations and distribution lines
Yin et al. Multi UAV cooperative task allocation method for intensive corridors of transmission lines inspection
CN117472083B (en) Multi-unmanned aerial vehicle collaborative marine search path planning method
Liu et al. Research on Unmanned Aerial Vehicle Trajectory Planning Based on Agent Reinforcement Learning in Alpine Forest Environment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant