WO2022045982A1 - Unmanned aerial vehicle and localization method for unmanned aerial vehicle - Google Patents

Unmanned aerial vehicle and localization method for unmanned aerial vehicle Download PDF

Info

Publication number
WO2022045982A1
WO2022045982A1 PCT/SG2021/050512 SG2021050512W WO2022045982A1 WO 2022045982 A1 WO2022045982 A1 WO 2022045982A1 SG 2021050512 W SG2021050512 W SG 2021050512W WO 2022045982 A1 WO2022045982 A1 WO 2022045982A1
Authority
WO
WIPO (PCT)
Prior art keywords
aerial vehicle
unmanned aerial
localization
ultra
measurements
Prior art date
Application number
PCT/SG2021/050512
Other languages
French (fr)
Inventor
Lihua Xie
Pham Nhat Thien Minh NGUYEN
Muqing Cao
Shenghai YUAN
Original Assignee
Nanyang Technological University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanyang Technological University filed Critical Nanyang Technological University
Priority to US18/040,148 priority Critical patent/US20230314548A1/en
Publication of WO2022045982A1 publication Critical patent/WO2022045982A1/en

Links

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C21/00Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
    • G01C21/20Instruments for performing navigational calculations
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S5/00Position-fixing by co-ordinating two or more direction or position line determinations; Position-fixing by co-ordinating two or more distance determinations
    • G01S5/02Position-fixing by co-ordinating two or more direction or position line determinations; Position-fixing by co-ordinating two or more distance determinations using radio waves
    • G01S5/0257Hybrid positioning
    • G01S5/0263Hybrid positioning by combining or switching between positions derived from two or more separate positioning systems
    • G01S5/0264Hybrid positioning by combining or switching between positions derived from two or more separate positioning systems at least one of the systems being a non-radio wave positioning system
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B64AIRCRAFT; AVIATION; COSMONAUTICS
    • B64UUNMANNED AERIAL VEHICLES [UAV]; EQUIPMENT THEREFOR
    • B64U10/00Type of UAV
    • B64U10/10Rotorcrafts
    • B64U10/13Flying platforms
    • B64U10/14Flying platforms with four distinct rotor axes, e.g. quadcopters
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B64AIRCRAFT; AVIATION; COSMONAUTICS
    • B64UUNMANNED AERIAL VEHICLES [UAV]; EQUIPMENT THEREFOR
    • B64U20/00Constructional aspects of UAVs
    • B64U20/80Arrangement of on-board electronics, e.g. avionics systems or wiring
    • B64U20/87Mounting of imaging devices, e.g. mounting of gimbals
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S13/00Systems using the reflection or reradiation of radio waves, e.g. radar systems; Analogous systems using reflection or reradiation of waves whose nature or wavelength is irrelevant or unspecified
    • G01S13/74Systems using reradiation of radio waves, e.g. secondary radar systems; Analogous systems
    • G01S13/76Systems using reradiation of radio waves, e.g. secondary radar systems; Analogous systems wherein pulse-type signals are transmitted
    • G01S13/765Systems using reradiation of radio waves, e.g. secondary radar systems; Analogous systems wherein pulse-type signals are transmitted with exchange of information between interrogator and responder
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S13/00Systems using the reflection or reradiation of radio waves, e.g. radar systems; Analogous systems using reflection or reradiation of waves whose nature or wavelength is irrelevant or unspecified
    • G01S13/87Combinations of radar systems, e.g. primary radar and secondary radar
    • G01S13/874Combination of several systems for attitude determination
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S17/00Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
    • G01S17/02Systems using the reflection of electromagnetic waves other than radio waves
    • G01S17/06Systems determining position data of a target
    • G01S17/08Systems determining position data of a target for measuring distance only
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S5/00Position-fixing by co-ordinating two or more direction or position line determinations; Position-fixing by co-ordinating two or more distance determinations
    • G01S5/02Position-fixing by co-ordinating two or more direction or position line determinations; Position-fixing by co-ordinating two or more distance determinations using radio waves
    • G01S5/14Determining absolute distances from a plurality of spaced points of known location
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S7/00Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
    • G01S7/003Transmission of data between radar, sonar or lidar systems and remote stations
    • G01S7/006Transmission of data between radar, sonar or lidar systems and remote stations using shared front-end circuitry, e.g. antennas
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04BTRANSMISSION
    • H04B7/00Radio transmission systems, i.e. using radiation field
    • H04B7/14Relay systems
    • H04B7/15Active relay systems
    • H04B7/185Space-based or airborne stations; Stations for satellite systems
    • H04B7/18502Airborne stations
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B64AIRCRAFT; AVIATION; COSMONAUTICS
    • B64UUNMANNED AERIAL VEHICLES [UAV]; EQUIPMENT THEREFOR
    • B64U2101/00UAVs specially adapted for particular uses or applications
    • B64U2101/30UAVs specially adapted for particular uses or applications for imaging, photography or videography
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B64AIRCRAFT; AVIATION; COSMONAUTICS
    • B64UUNMANNED AERIAL VEHICLES [UAV]; EQUIPMENT THEREFOR
    • B64U2201/00UAVs characterised by their flight controls
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B64AIRCRAFT; AVIATION; COSMONAUTICS
    • B64UUNMANNED AERIAL VEHICLES [UAV]; EQUIPMENT THEREFOR
    • B64U30/00Means for producing lift; Empennages; Arrangements thereof
    • B64U30/20Rotors; Rotor supports
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C21/00Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
    • G01C21/10Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 by using measurements of speed or acceleration
    • G01C21/12Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 by using measurements of speed or acceleration executed aboard the object being navigated; Dead reckoning
    • G01C21/16Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 by using measurements of speed or acceleration executed aboard the object being navigated; Dead reckoning by integrating acceleration or speed, i.e. inertial navigation
    • G01C21/165Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 by using measurements of speed or acceleration executed aboard the object being navigated; Dead reckoning by integrating acceleration or speed, i.e. inertial navigation combined with non-inertial navigation instruments
    • G01C21/1652Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 by using measurements of speed or acceleration executed aboard the object being navigated; Dead reckoning by integrating acceleration or speed, i.e. inertial navigation combined with non-inertial navigation instruments with ranging devices, e.g. LIDAR or RADAR
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S13/00Systems using the reflection or reradiation of radio waves, e.g. radar systems; Analogous systems using reflection or reradiation of waves whose nature or wavelength is irrelevant or unspecified
    • G01S13/86Combinations of radar systems with non-radar systems, e.g. sonar, direction finder
    • G01S13/865Combination of radar systems with lidar systems
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S13/00Systems using the reflection or reradiation of radio waves, e.g. radar systems; Analogous systems using reflection or reradiation of waves whose nature or wavelength is irrelevant or unspecified
    • G01S13/86Combinations of radar systems with non-radar systems, e.g. sonar, direction finder
    • G01S13/867Combination of radar systems with cameras
    • GPHYSICS
    • G08SIGNALLING
    • G08GTRAFFIC CONTROL SYSTEMS
    • G08G5/00Traffic control systems for aircraft, e.g. air-traffic control [ATC]
    • G08G5/0004Transmission of traffic-related information to or from an aircraft
    • G08G5/0013Transmission of traffic-related information to or from an aircraft with a ground station
    • GPHYSICS
    • G08SIGNALLING
    • G08GTRAFFIC CONTROL SYSTEMS
    • G08G5/00Traffic control systems for aircraft, e.g. air-traffic control [ATC]
    • G08G5/0017Arrangements for implementing traffic-related aircraft activities, e.g. arrangements for generating, displaying, acquiring or managing traffic information
    • G08G5/0026Arrangements for implementing traffic-related aircraft activities, e.g. arrangements for generating, displaying, acquiring or managing traffic information located on the ground
    • GPHYSICS
    • G08SIGNALLING
    • G08GTRAFFIC CONTROL SYSTEMS
    • G08G5/00Traffic control systems for aircraft, e.g. air-traffic control [ATC]
    • G08G5/0047Navigation or guidance aids for a single aircraft
    • G08G5/0052Navigation or guidance aids for a single aircraft for cruising
    • GPHYSICS
    • G08SIGNALLING
    • G08GTRAFFIC CONTROL SYSTEMS
    • G08G5/00Traffic control systems for aircraft, e.g. air-traffic control [ATC]
    • G08G5/0047Navigation or guidance aids for a single aircraft
    • G08G5/0069Navigation or guidance aids for a single aircraft specially adapted for an unmanned aircraft

Definitions

  • the present disclosure relates to unmanned aerial vehicles and in particular to localization of unmanned aerial vehicles.
  • UAVs unmanned aerial vehicles
  • the first approach is the so-called onboard self-localization (OSL) approach, in which a variety of methods based on visual-inertial odometry (VIO) or Lidar Odometry and Mapping (LOAM) techniques are employed. These techniques have become ubiquitous now due to their flexibility and effectiveness. However, the main characteristics of these methods is that while they have high accuracy in the short term, they can suffer from accumulated error in the long term.
  • OSL onboard self-localization
  • the second approach is the so-called Infrastructure-based localization (ISL) approach in which the UAV can rely on fixed anchors installed in the environment to localize itself with respect to a fixed coordinate system.
  • ISL Infrastructure-based localization
  • This approach will ensure that the positioning information will not drift in long term as in OSL methods, which is desired in a great number of applications as a user often seeks to have the UAV perform a predefined trajectory to cover an area, inspect an object, or coordinated large-scale performance.
  • the installation of the anchors also incurs additional costs, deployment effort, and maintenance responsibility.
  • an unmanned aerial vehicle comprises: a first ultra-wide band node; a second ultra-wide band node; and a localization processor configured to use signals received at the first ultra-wide band node and the second ultra-wide band node from a plurality of anchor nodes located within an environment to estimate a pose state of the unmanned aerial vehicle, wherein the first ultra-wide band node and the second ultra-wide band node are arranged on the unmanned aerial vehicle at positions offset from one another.
  • the unmanned aerial vehicle has a body-offset configuration of ultra-wide band (UWB) nodes.
  • UWB ultra-wide band
  • This configuration is employed to achieve direct coupling of the UAV’s orientation in the range measurements.
  • This configuration also allows a simple and efficient scheme to be developed for deployment of a network of UWB nodes in the field with minimal time and effort.
  • the use of a few UWBs that can be quickly and effortlessly deployed allows user to define a global coordinates system that aligns well with the object or structure to be inspected on site, thus avoiding misalignment between the pre-defined trajectory optimized in the path planning simulation with the real object in the actual site.
  • the network can automatically calculate the position of the ground nodes by using ranging and communication capabilities of UWB radio modules.
  • the use of the body-offset ranging ensures that the orientation of the UAV will not drift as with OSL methods and conventional UWB- based localization system.
  • the UAV further comprises an inertial measurement unit configured to detect linear acceleration and angular velocity of the unmanned aerial vehicle and wherein the localization processor is further configured to use measurements from the inertial measurement unit to predict the pose state of the unmanned aerial vehicle.
  • the localization processor is further configured to pre-integrate the measurements from the inertial measurement unit to obtain a set of pre-integrated inertial measurement unit measurements and to use the set of pre-integrated inertial movement unit measurements to estimate the pose state of the unmanned aerial vehicle.
  • Mathematical models have been developed for ranging and pre-integration observations that couple the UWB and IMU measurements with the state estimates. The use of IMU pre-integration and VIO/LOAM observations can increase the efficiency and robustness of the method.
  • the localization processor is further configured to compare a predicted distance calculated from the pose state of the unmanned aerial vehicle determined from the inertial measurement measurements with distance predictions determined from measurements by the first ultra-wideband node and the second ultra- wideband node and thereby identify outliners in the measurements by the first ultra- wideband node and the second ultra-wideband node.
  • the approach is robust to outliers.
  • the outliers can be effectively- detected and rejected by its contribution in the cost function, on the other hand, the effect of a wrongly-admitted outlier can be mitigated by the presence of the overwhelming number of inliers, and the estimation can quickly return to the true value.
  • filter-based methods there is no straightforward way to do this as all information we have is the previous estimate, and the wrong admission of an outlier can lead to catastrophic effect, causing the system to diverge without any mechanism to reset.
  • the UAV further comprises an on-board self-localization module configured to generate on-board self-localization data.
  • the on-board self-localization module may comprise a lidar sensor and the on-board self-localization data comprises laser scan data and / or a camera and the on-board self-localization data comprises image data.
  • VIO/LOAM IMU, UWB and OSL
  • the sensor fusion framework can be easily adjusted to integrate any number of VIO or LOAM sources. Thus, increasing the level of redundancy and reliability of the localization method employed by the UAV can be achieved.
  • the localization processor is further configured to estimate the pose state of the unmanned aerial vehicle using measurement data comprising measurements by the first ultra-wideband node and the second ultra-wideband node, the measurements from the inertial measurement unit and the on-board selflocalization data by minimizing a cost function between the measurement data and a model predicted pose state.
  • the on-board self-localization module comprises a lidar sensor and the on-board self-localization data comprises laser scan data and the localization processor is further configured to use the on-board self-localization data to extract lidar features from the laser scan data and match the lidar features with a local map to obtain a set of feature-map-matching coefficients, wherein the cost function depends on the set of feature-map-matching coefficients.
  • the on-board self-localization module comprises a camera and the on-board self-localization data comprises image data and localization processor is further configured to use the on-board self-localization data to extract a set of visual features from image data, wherein the cost function depends on the set of visual features.
  • the localization processor is further configured to estimate the pose state of the unmanned aerial vehicle using measurement data corresponding to a sliding time window. Mathematical models are developed of OSL observations, i.e. lidar and visual features, from VIO and LOAM systems, which couple multiple poses of the state vectors in the sliding window.
  • the ranging, pre-integration, OSL observations over the whole sliding window are combined together to construct a single cost function, which is then minimized via a search-and-optimize strategy to obtain the optimal values for the state estimates.
  • embodiments of the present disclosure provide a way to effectively combine the advantages of the aforementioned two approaches. Specifically, the short-term accuracy and reliability of OSL techniques is preserved, while also ensuring the that the localization is continuous, does not drift and is aligned well with the chosen frame of reference in the long-term as with infrastructure-based methods. Moreover, while the system can be assisted by the installation of the beacons, is does not incur too much cost in installation and starting up in fieid operations since fewer beacons may be required due to the provision of more than one UWB node on the UAV.
  • a localization method for an unmanned aerial vehicle comprises: a first ultra-wide band node; and a second ultra-wide band node, wherein the first ultra-wide band node and the second ultra-wide band node are arranged on the unmanned aerial vehicle at positions offset from one another.
  • the method comprises: using signals received at the first ultra-wideband node and the second ultra-wideband node from a plurality of anchor nodes located within an environment to estimate a pose state of the unmanned aerial vehicle.
  • the unmanned aerial vehicle further comprises: an inertial measurement unit configured to detect linear acceleration and angular velocity of the unmanned aerial vehicle, and the method further comprises using measurements from the inertial measurement unit to predict the pose state of the unmanned aerial vehicle.
  • the method further comprises pre-integrating the measurements from the inertial measurement unit to obtain a set of pre-integrated inertial measurement unit measurements and using the set of pre-integrated inertial measurement unit measurements to estimate the pose state of the unmanned aerial vehicle.
  • the method further comprises comparing the predicted distance from the pose prediction of the unmanned aerial vehicle determined from measurements from the inertial measurement unit with the distance measurements by the first ultra-wideband node and the second ultra-wideband node and thereby identify outliners in the measurements by the first ultra-wideband node and the second ultra- wideband node.
  • the method further comprises comprising receiving on-board seiflocalization data generated by on-board self-localization module of the unmanned aerial vehicle.
  • the on-board self-localization module comprises a lidar sensor and the on-board self-localization data comprises laser scan data.
  • the on-board self-localization module comprises a camera and the on-board self-localization data comprises image data.
  • the method further comprises estimating the pose state of the unmanned aerial vehicle using measurement data comprising measurements by the first ultra-wideband node and the second ultra-wideband node, the measurements from the inertial measurement unit and the on-board self-localization data by minimizing a cost function between the measurement data and a model predicted pose state.
  • the on-board self-localization data comprises laser scan data
  • the method further comprises extracting lidar features from the laser scan data and matching the lidar features with a local map and thereby obtaining a set of feature- map-matching coefficients, wherein the cost function depends on the set of feature- map-matching coefficients.
  • the on-board self-localization data comprises image data
  • the method further comprises extracting visual features from the image data and thereby obtaining a set of visual features, wherein the cost function depends on the set of visual features.
  • the method further comprises the estimating the pose state of the unmanned aerial vehicle using measurement data corresponding to a sliding time window.
  • a computer readable medium storing processor executable instructions which when executed on a processor cause the processor to carry out a method as set out above is provided.
  • FIG.1 shows an unmanned aerial vehicle according to an embodiment of the present invention
  • FIG.2 shows the deployment of anchor nodes for use with an unmanned aerial vehicle according to an embodiment of the present invention
  • FIG.3 is a table showing communication slots for anchor nodes used in embodiments of the present invention.
  • FIG.4 is a block diagram showing an unmanned aerial vehicle according to an embodiment of the present invention.
  • FIG.5 is a flowchart showing a localization method for an unmanned aerial vehicle according to an embodiment of the present invention
  • FIG.6 is a flow chart showing the workflow in a localization method for an unmanned aerial vehicle according to an embodiment of the present invention.
  • FIG.7 is a timing diagram showing a sliding window synchronization scheme used in embodiments of the present invention.
  • FIG.8 shows multiple poses of an unmanned aerial vehicle and the coupling of these poses with available observations
  • FIG.9 is a table showing a comparison absolute trajectory error of localization methods according to embodiments of the present invention with other localization methods
  • FIG.10 shows a set of ultra-wide band measurements for a test flight of an unmanned aerial vehicle according to an embodiment of the present invention
  • FIG.11 is a table showing position estimates of the anchors in the flight tests
  • FIG.12 is a table showing a comparison of absolute trajectory error over building inspection datasets.
  • FIG.13A to FIG.13D show the results of a test flight experiment using an unmanned aerial vehicle according to an embodiment of the present invention.
  • FIG.1 shows an unmanned aerial vehicle according to an embodiment of the present invention.
  • the unmanned aerial vehicle 100 comprises a body 101 from which six arms 102 extend.
  • a rotor 104 is mounted on each of the arms 102.
  • a landing gear 106 extends downward from the body 101.
  • Two cameras are mounted on the body: a left camera 108 and a right camera 109 which both face forward.
  • a vertical lidar sensor 110 and a horizontal lidar sensor 111 are mounted on the body 101.
  • Each of the vertical lidar sensor 110 and the horizontal lidar sensor 111 comprise a laser and a receiver and operate to determine the range of objects by measuring the return time for reflected light to return to the receiver.
  • the vertical lidar sensor 110 is arranged to carry out ranging in vertical directions and the horizontal lidar sensor 111 is arranged to carryout ranging in horizontal directions.
  • An inertial measurement unit (IMU) 112 is arranged on the body 101.
  • the inertial measurement unit 112 comprises gyroscopes and accelerometers and is configured to detect linear acceleration and angular velocity of the unmanned aerial vehicle.
  • a first ultra-wide band antenna node 120 is arrange on an arm extending to the front left of the body 101
  • a second ultra-wide band antenna node 121 is arranged on an arm extending to the rear left of the body 101
  • a third ultra-wide band antenna node 122 is arrange on an arm extending to the rear right of the body 101
  • a fourth ultra-wide band antenna node 123 is arranged on an arm extending to the front right of the body 101 .
  • the UWB signal has a frequency band from 3.1 GHz to 4.8GHz. Ranging measurement is carried out by a two-way time-of-flight technique.
  • the network is synchronized via a Time-Division Multiple Access technique where ranging measurements are scheduled on a slot map.
  • a crystal prism 130 which acts as a ground truth tracking point is mounted on top of the body 101.
  • the crystal prism 130 is used to verify the localization scheme in experiments such as those described below with reference to FIG.13A to FIG.13D and is not used as part of the localization schemes. Thus, the localization schemes described herein would operate without the crystal prism 130.
  • the ultra-wide band antenna nodes are configured to send and receive signals from anchor nodes located at fixed locations in the environment in which the unmanned aerial vehicle operates. By sending and receiving signals from these anchor nodes, the location and orientation of the unmanned aerial vehicle can be estimated.
  • FIG.2 shows the deployment of anchor nodes in an environment for use with an unmanned aerial vehicle according to an embodiment of the present invention.
  • three ultra-wide band (UWB) ground nodes or anchor nodes are arranged in an environment 200 around a structure 210 under inspection.
  • a first anchor node 222, a second anchor node 224 and a third anchor node 226 are arranged in the environment 200.
  • the UAV 230 having four UWB ranging nodes 232 flies in the environment 200.
  • a world frame 240 is defined having an x-axis 242 in the direction from the first anchor node 222 towards the second anchor node 224 and a y-axis in the direction towards the structure 210 under inspection.
  • a z-axis is defined to point upwards by the right-hand rule.
  • the third anchor node 226 is positioned on the -y half plane. In this scheme by having the +x direction point from one node to another, together with the preferred choice of z pointing upwards, we can easily define a fixed reference frame that can be easily aligned with the environment we wish to conduct the UAV operation in. Additionally, the three anchor nodes are arranged substantially in a horizonal plane and the four UWB ranging nodes 232 on the UAV 230 are also arranged in a horizontal plane.
  • FIG.3 is a table showing communication slots for anchor nodes used in embodiments of the present invention.
  • FIG.3 shows the slot map designed for one anchor node, for example, the first anchor node 222 having ID101. Notice that grayed slots where ranging between the anchors are scheduled (i.e. slots 0, 7, 14) are the same for all anchors.
  • all nodes in the network should be able to obtain the distances among the three nodes ID100, ID101 , ID102 after 2 cycles and calculate the corresponding coordinates x1 , x2, y2 of anchor ID 101 and ID102, assuming that the plane made by the three anchors are parallel with the x-y plane of the coordinates system.
  • ranging between the anchors are the same for all slot maps.
  • slot 0, 7 and 14 are used to conduct ranging between the anchor nodes.
  • all other nodes can receive a message containing the distance between anchors. These distances can then be used to calculate the coordinates of the ground node.
  • the important issue here is to design such a set of slot maps with minimal number of nodes needed for ranging between the anchors (thus allowing more time for UAV to range to the anchors), yet still ensure enough range measurements to efficiently calculate the coordinates of the anchors.
  • FIG.4 is a block diagram showing an unmanned aerial vehicle according to an embodiment of the present invention.
  • the unmanned aerial vehicle 400 comprises a flight controller 402 which is controls rotors 404 of the unmanned aerial vehicle 400.
  • the unmanned aerial vehicle 400 further comprises a memory 406, a localization processor 408, a first ultra-wide band node 412, a second ultra-wide band node 414, an on-board self-localization module 420 and an inertial measurement unit 430.
  • the memory 406 may store flight plans which are executed by the flight controller 402 in order to carry out an inspection or other task.
  • the memory may also store images, video or other data captured by sensors of the unmanned aerial vehicle during a task.
  • the localization processor 408 executes methods according to embodiments of the present invention to determine the location of the unmanned aerial vehicle using information captured by sensors on the unmanned aerial vehicle.
  • the first ultra-wide band node 412, and the second ultra-wide band node 414 are ultra-wide band antennas which determine the range to anchor ultra-wide band nodes arranged in the environment.
  • the on-board self-localization module 420 comprises a lidar module 422 and a camera 424.
  • the lidar module 422 comprises a laser and a receiver and operate to determine the range of objects by measuring the return time for reflected light to return to the receiver.
  • the inertial measurement unit 430 comprises gyroscopes and accelerometers and is configured to detect linear acceleration and angular velocity of the unmanned aerial vehicle 400.
  • the flight controller 402, the memory 406 and the localization processor 408 may be implemented as a small form factor computer having a processor and a memory storing computer executable instructions which are executed on the processor to carry out the methods described in this disclosure.
  • the memory 406 may be referred to in some contexts as computer readable storage media and/or non-transitory computer readable media.
  • a software implementation of the computer program modules is described herein, these may alternatively be implemented as one or more hardware modules (such as field-programmable gate array(s) or applicationspecific integrated circuit(s)) comprising circuitry which implements equivalent functionality to that implemented in software.
  • FIG.5 is a flowchart showing a localization method for an unmanned aerial vehicle according to an embodiment of the present invention.
  • the method 500 shown in FIG.5 is carried out by the unmanned aerial vehicle 400 shown in FIG.4.
  • the first ultra-wide band node 412 and the second ultra-wide band node 414 of the unmanned aerial vehicle 400 receive signals from anchor nodes. From these signal, the localization processor 408 determines the range to each of the anchor nodes. In step 504, the localization processor 408 of the unmanned aerial vehicle 400 estimates the pose state of the unmanned aerial vehicle.
  • the pose state comprises the location and orientation of the unmanned aerial vehicle relative to the anchor nodes in the environment.
  • FIG.6 is a flow chart showing the workflow in a localization method for an unmanned aerial vehicle according to an embodiment of the present invention.
  • the method 600 shown in FIG.6 is carried out by the unmanned aerial vehicle 400 shown in FIG.4.
  • the inputs to the method are ultra-wide band (UWB) data 610, inertial measurement unit (IMU) data 620, pointclouds 630 which are generated by the lidar module 422 of the unmanned aerial vehicle 400, and images 650 which are generated by the camera 424 of the unmanned aerial vehicle 400.
  • the method 600 also uses UWB measurements and pose predictions x m from previous stages in key frame management 640.
  • the UWB ranging and communication 602 network is employed to gather the distance between the anchors, after they have been deployed in the field. These distances will be then used to calculate the position of the anchors by using an anchor self-localization 604 algorithm. Once the anchors' positions 605 have been calculated, they will be associated with the UWB measurements 603.
  • a UWB measurement which makes up the UWB data 610 consists of three pieces of information, first is the position of the UWB node in the UAV's body frame, second is the UWB anchor position in the chosen world frame, and third is the distance between these two UWB nodes.
  • UWB, IMU and OSL data will go through multiple stages before they are fused together in the Optimization-based Sensor Fusion (OSF) algorithm.
  • OSF Optimization-based Sensor Fusion
  • the system is synchronized by the lidar messages, whose time stamps are used to determine the time steps from time t w to t k . This synchronization process is described in more detail below with reference to FIG.7.
  • the UWB data 610 is subjected to some preliminary checks 612 to eliminate bad measurements. These checks include comparing the measurement's signal over noise ratio (SNR), leading-edge-detection quality, and rate-of-change with some user- defined thresholds, etc. Only those that meet these thresholds will be passed to the buffer in a bundling 614 stage for subsequent stages. UWB measurements that arrive in the interval are grouped into a set Finally, this set will be put into a final outlier check 616 to see if any measurement in it is an outlier. This is done by comparing the measurement with the IMU-predicted range value. As shown in FIG.6, the IMU propagated state 629 determined in by the pose prediction 628 algorithm is used in the outlier check 616.
  • SNR signal over noise ratio
  • leading-edge-detection quality leading-edge-detection quality
  • rate-of-change rate-of-change
  • FIG.7 is a timing diagram showing a sliding window synchronization scheme used in embodiments of the present invention.
  • the system is synchronized by the lidar messages 710, whose time stamps are used to determine the time steps from time t w to t k .
  • the IMU data 720, the UWB data 730 and the image data 740 is then allocated into the sliding windows according to the timing relative to the lidar messages 710.
  • IMU data from a subsequent frame 722 is pre-integrated and used in the processing of earlier frames.
  • IMU data from one or more frames 724 is propagated and used to compensate for motion distortion.
  • a group of UWB data points 732 in a frame may be defined having a timestamp .
  • An image 742 may be defined as having a visual time delay t d with respect to the timing of the corresponding frame defined by the lidar messages 710.
  • the IMU data 620 is subjected to bundling 622 to arrange the data into frames as described above with reference to FIG.7, and pre-integration before being input into a local optimization algorithm 626.
  • the local optimization algorithm 626 generates a pose prediction 628.
  • the plane and edge features are then extracted using the common smoothness criteria of LOAM systems in a feature extraction 632 step.
  • the motion distortion in these features is compensated by the IMU-propagated states, and then transform to the inertial frame in the transform and deskew step 634.
  • these features are then matched with the local map 642 built from marginalized point clouds in an association 644 step to obtain the set of feature-map-matching (FMM) coefficients which are used to create the lidar factors in the final cost function calculated in the local optimization algorithm 626.
  • FMM feature-map-matching
  • the feature-map-matching coefficients are calculated using the following algorithm:
  • the binary robust independent elementary features (BRIEF features) are extracted in a feature extraction sync step 652 and then use the latest pose estimates to triangulate these features for depth initialization in a triangulation step 645.
  • BRIEF features binary robust independent elementary features
  • a Z i , b Z i are the projected coordinates of the feature in camera frames a and b
  • a f i , b f i are the feature’s 3D coordinates w.r.t camera frames a and b
  • ⁇ 1 is the inverse depth of visual feature i, which is only associated with the camera frame that first observes the feature.
  • the cost function will be then optimized to provide the estimate of the robot states in the sliding windows.
  • the latest state estimate in the sliding windows will be combined with the latest IMU measurements in the buffer, which are leftovers from the last extraction plus the newly arrived during the optimization process, to calculate the high- rate pose estimate.
  • This high-rate pose estimate is also used to perform the outlier check on the UWB measurements that was mentioned earlier.
  • the sliding window will be slid forward, the oldest state estimate and its associated measurements will be abandoned, and a new time step will be registered when the conditions mentioned earlier are met.
  • FIG.8 shows multiple poses of an unmanned aerial vehicle and the coupling of these poses with available observations.
  • the unmanned aerial vehicle 802 having UWB nodes 804 moves through a series of positions which are labelled k, k+1 , k+2 and k+3.
  • Three anchor UWB nodes 810 are arranged in the environment.
  • the pose of the unmanned aerial vehicle 802 is expressed relative to axes 820.
  • the pose of the UAV 802 at position k is X k
  • the poses at positions k+1 , k+2 and k+3 are X k+1 , X k+2 and X k+3 respectively.
  • the IMU measurements between the positions are Z k ,Z k+1 and Z k+2 .
  • the UWB measurements between the UWB nodes 802 of the UAV and the anchor nodes are expressed as d k+3 where the superscript number indicates a label of the anchor node.
  • a cost function based on the observations from UWB, IMU and OSL measurements.
  • the cost function can be constructed at each time step t k+M as follows: where are the residuals constructed from IMU, UWB and OSL measurements are the covariance matrices of the measurement errors, and are the number of UWB measurement in the intervals.
  • the cost function above summarizes the coupling of all the states and the observations.
  • each factor in the cost function. Essentially, for each measurement, we will find an observation model that links the actual measurements with the states. Then we will calculate the residual (or loosely- speaking, the difference) between the measurement and the model-predicted value. Then, we need to derive the covariance matrix of the measurement errors, to properly apply the weightage among the factor. Finally, to facilitate real-time optimization, the Jacobians of each factor have to be analytically derived.
  • Ranging factor At each time step t k we have the following observation from a UAV ranging node: where is the distance measurement, and is the position of the UWB anchor w.r.t. the world frame, is the coordinate of the UWB ranging node in the body frame, are the time stamps defined in FIG.7.
  • the residual of a range measurement is calculated by taking the difference between the distance estimate d m and the distance measurement .
  • the estimated distance d m .
  • IMU pre-integration factors For visual feature factors, the residual and covariance from each feature’s observation is defined as iii. IMU pre-integration factors:
  • the IMU measurements are defined as where g is the gravitational acceleration, are respectively the biases of the gyroscope and the accelerometer, and are some zero- mean Gaussian noises, whose standard deviations are respectively.
  • the above integration can be implemented by zero-order-hold (ZOH) or higher order methods (Runge-Kutta methods).
  • ZOH zero-order-hold
  • Runge-Kutta methods higher order methods
  • the residual of the IMU preintegration factor is defined as follows ,
  • the covariance of the IMU pre-integration errors are calculated based on a propagation scheme.
  • FIG.9 is a table showing a comparison absolute trajectory error of localization methods according to embodiments of the present invention with other localization methods.
  • VINS-Mono, VINS-Fusion, A-LOAM, LIO-SAM, M-LOAM are prior art methods and the methods labelled VIRAL (horiz. lidar) and VIRAL (2 lidars) correspond to localization methods according to embodiments of the using a single horizontal lidar and two lidars respectively. All algorithms are run on an NUC 10 computer with core i7 processor. Each method is slightly modified and configured for their best performance with the dataset. Since several lidar-based methods are not designed to incorporate multiple lidar inputs, we also include experiments of VIRAL using only the horizontal lidar for a fairer comparison. The Absolute Trajectory Error (ATE) result is summarized in FIG.9. The best odometry result is highlighted in bold, and the second best is underlined. All values are in m. From this table we can clearly see that VIRAL consistently achieves better performance compared to existing methods, even when only one lidar is used.
  • VIRAL horiz. lidar
  • VIRAL (2 lidars) correspond to localization methods
  • the data rate of the stereo image, IMU, LIDAR 1 , LIDAR 2, and UWB topics are 10 Hz, 400 Hz, 10 Hz, 10 Hz, and 69 Hz respectively.
  • a Leica MS60 station is used to provide ground truth for the experiment. All software is run on a small form factor computer with Intel Core i7-8650U processor.
  • FIG.10 shows a set of ultra-wide band measurements for a test flight of an unmanned aerial vehicle according to an embodiment of the present invention.
  • the UWB nodes on the UAV are labelled 200. A, 200. B, 201 .A and 201 .B.
  • the anchor UWB nodes are labelled 100, 101 and 102.
  • Each of the plots shown in FIG.10 corresponds to measurements from one of the UWB nodes on the UAV to one of the anchor nodes. From the zoomed-in part of the 200.A ⁇ 101 plot, the variation of the distance is about 5cm. We also notice some outliers in the 200.A ⁇ 102 and 200.B ⁇ 102 ranging measurements.
  • FIG.11 is a table showing position estimates of the anchors in the flight tests. The position estimates were determined as described above with reference to FIG.2 and FIG.3.
  • FIG.12 is a table showing a comparison of absolute trajectory error over building inspection datasets. A total of four flight tests were conducted. As having introduced earlier, since our work is meant to deliver a solution for UAV-based inspection application, the UAV’s trajectory is different from those in common datasets, which are often extensive in x and y directions, but have limited movement in the z direction. In our case, for inspection purposes, the UAV only takes off and moves in a vertical plane parallel to the object. Hence the movements along the x and z directions are extensive, while movement along the y direction is minimal.
  • FIG.13A to FIG.13D show the results of a test flight experiment. The results correspond to the flight labelled bid_01 in FIG.11 and FIG.12.
  • FIG.13A shows a 3D view of the UAV trajectory as estimated by individual OSL systems and the sensor fusion result.
  • FIG.13B shows Orientation estimate from the system over time.
  • FIG.13C shows velocity estimate from the system over time.
  • FIG.13D shows position estimates from the OSL systems and sensor fusion results over time.
  • the ground truth position was obtained using a prism mounted on the UAV as shown in FIG.1 .
  • UAVs implementing the methods described above will find applications in aerial surveillance by UAVs; automated inspection operations by UAVs; and graphical modelling and structural health of 3D objects.

Landscapes

  • Engineering & Computer Science (AREA)
  • Remote Sensing (AREA)
  • Radar, Positioning & Navigation (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Aviation & Aerospace Engineering (AREA)
  • Electromagnetism (AREA)
  • Mechanical Engineering (AREA)
  • Astronomy & Astrophysics (AREA)
  • Automation & Control Theory (AREA)
  • Signal Processing (AREA)
  • Microelectronics & Electronic Packaging (AREA)
  • Navigation (AREA)

Abstract

An unmanned aerial vehicle aerial vehicle and a localization method for an unmanned aerial vehicle aerial vehicle are described. In an embodiment, an unmanned aerial vehicle comprises: a first ultra-wide band node; a second ultra-wide band node; and a localization processor configured to use signals received at the first ultra-wide band node and the second ultra-wide band node from a plurality of anchor nodes located within an environment to estimate a pose state of the unmanned aerial vehicle, wherein the first ultra-wide band node and the second ultra-wide band node are arranged on the unmanned aerial vehicle at positions offset from one another.

Description

UNMANNED AERIAL VEHICLE AND LOCALIZATION METHOD FOR UNMANNED
AERIAL VEHICLE
TECHNICAL FIELD
The present disclosure relates to unmanned aerial vehicles and in particular to localization of unmanned aerial vehicles.
BACKGROUND
Robust and reliable state estimation is undoubtedly one of the most crucial tasks for development of unmanned aerial vehicles (UAVs). There are two main approaches to localization of UAVs that are usually employed, depending on the scenario in which the UAV is applied.
The first approach is the so-called onboard self-localization (OSL) approach, in which a variety of methods based on visual-inertial odometry (VIO) or Lidar Odometry and Mapping (LOAM) techniques are employed. These techniques have become ubiquitous now due to their flexibility and effectiveness. However, the main characteristics of these methods is that while they have high accuracy in the short term, they can suffer from accumulated error in the long term.
The second approach is the so-called Infrastructure-based localization (ISL) approach in which the UAV can rely on fixed anchors installed in the environment to localize itself with respect to a fixed coordinate system. This approach will ensure that the positioning information will not drift in long term as in OSL methods, which is desired in a great number of applications as a user often seeks to have the UAV perform a predefined trajectory to cover an area, inspect an object, or coordinated large-scale performance. However, the installation of the anchors also incurs additional costs, deployment effort, and maintenance responsibility. SUMMARY
According to a first aspect of the present disclosure, an unmanned aerial vehicle is provided. The unmanned aerial vehicle comprises: a first ultra-wide band node; a second ultra-wide band node; and a localization processor configured to use signals received at the first ultra-wide band node and the second ultra-wide band node from a plurality of anchor nodes located within an environment to estimate a pose state of the unmanned aerial vehicle, wherein the first ultra-wide band node and the second ultra-wide band node are arranged on the unmanned aerial vehicle at positions offset from one another.
The unmanned aerial vehicle has a body-offset configuration of ultra-wide band (UWB) nodes. This configuration is employed to achieve direct coupling of the UAV’s orientation in the range measurements. This configuration also allows a simple and efficient scheme to be developed for deployment of a network of UWB nodes in the field with minimal time and effort. The use of a few UWBs that can be quickly and effortlessly deployed allows user to define a global coordinates system that aligns well with the object or structure to be inspected on site, thus avoiding misalignment between the pre-defined trajectory optimized in the path planning simulation with the real object in the actual site. After the deployment, the network can automatically calculate the position of the ground nodes by using ranging and communication capabilities of UWB radio modules. The use of the body-offset ranging ensures that the orientation of the UAV will not drift as with OSL methods and conventional UWB- based localization system.
In an embodiment, the UAV further comprises an inertial measurement unit configured to detect linear acceleration and angular velocity of the unmanned aerial vehicle and wherein the localization processor is further configured to use measurements from the inertial measurement unit to predict the pose state of the unmanned aerial vehicle.
In an embodiment, the localization processor is further configured to pre-integrate the measurements from the inertial measurement unit to obtain a set of pre-integrated inertial measurement unit measurements and to use the set of pre-integrated inertial movement unit measurements to estimate the pose state of the unmanned aerial vehicle. Mathematical models have been developed for ranging and pre-integration observations that couple the UWB and IMU measurements with the state estimates. The use of IMU pre-integration and VIO/LOAM observations can increase the efficiency and robustness of the method. in an embodiment, the localization processor is further configured to compare a predicted distance calculated from the pose state of the unmanned aerial vehicle determined from the inertial measurement measurements with distance predictions determined from measurements by the first ultra-wideband node and the second ultra- wideband node and thereby identify outliners in the measurements by the first ultra- wideband node and the second ultra-wideband node.
Thus, the approach is robust to outliers. On one hand, the outliers can be effectively- detected and rejected by its contribution in the cost function, on the other hand, the effect of a wrongly-admitted outlier can be mitigated by the presence of the overwhelming number of inliers, and the estimation can quickly return to the true value. In contrast, in filter-based methods, there is no straightforward way to do this as all information we have is the previous estimate, and the wrong admission of an outlier can lead to catastrophic effect, causing the system to diverge without any mechanism to reset.
In an embodiment, the UAV further comprises an on-board self-localization module configured to generate on-board self-localization data. The on-board self-localization module may comprise a lidar sensor and the on-board self-localization data comprises laser scan data and / or a camera and the on-board self-localization data comprises image data. Such embodiments provide for comprehensive localization approach that can fuse IMU, UWB and OSL (VIO/LOAM) information to achieve accurate and non- drifting estimation of a UAV’s extended state. The sensor fusion framework can be easily adjusted to integrate any number of VIO or LOAM sources. Thus, increasing the level of redundancy and reliability of the localization method employed by the UAV can be achieved.
In an embodiment, the localization processor is further configured to estimate the pose state of the unmanned aerial vehicle using measurement data comprising measurements by the first ultra-wideband node and the second ultra-wideband node, the measurements from the inertial measurement unit and the on-board selflocalization data by minimizing a cost function between the measurement data and a model predicted pose state. in an embodiment, the on-board self-localization module comprises a lidar sensor and the on-board self-localization data comprises laser scan data and the localization processor is further configured to use the on-board self-localization data to extract lidar features from the laser scan data and match the lidar features with a local map to obtain a set of feature-map-matching coefficients, wherein the cost function depends on the set of feature-map-matching coefficients.
In an embodiment, the on-board self-localization module comprises a camera and the on-board self-localization data comprises image data and localization processor is further configured to use the on-board self-localization data to extract a set of visual features from image data, wherein the cost function depends on the set of visual features. in an embodiment, the localization processor is further configured to estimate the pose state of the unmanned aerial vehicle using measurement data corresponding to a sliding time window. Mathematical models are developed of OSL observations, i.e. lidar and visual features, from VIO and LOAM systems, which couple multiple poses of the state vectors in the sliding window.
Further, the ranging, pre-integration, OSL observations over the whole sliding window are combined together to construct a single cost function, which is then minimized via a search-and-optimize strategy to obtain the optimal values for the state estimates.
Thus, embodiments of the present disclosure provide a way to effectively combine the advantages of the aforementioned two approaches. Specifically, the short-term accuracy and reliability of OSL techniques is preserved, while also ensuring the that the localization is continuous, does not drift and is aligned well with the chosen frame of reference in the long-term as with infrastructure-based methods. Moreover, while the system can be assisted by the installation of the beacons, is does not incur too much cost in installation and starting up in fieid operations since fewer beacons may be required due to the provision of more than one UWB node on the UAV.
It is envisaged that such a system will find many applications in inspection of various structures (building, cranes, warehouse...), area coverage, or multi-UAV performance.
According to a second aspect of the present disclosure a localization method for an unmanned aerial vehicle is provided. The unmanned aerial vehicle comprises: a first ultra-wide band node; and a second ultra-wide band node, wherein the first ultra-wide band node and the second ultra-wide band node are arranged on the unmanned aerial vehicle at positions offset from one another. The method comprises: using signals received at the first ultra-wideband node and the second ultra-wideband node from a plurality of anchor nodes located within an environment to estimate a pose state of the unmanned aerial vehicle.
In an embodiment, the unmanned aerial vehicle further comprises: an inertial measurement unit configured to detect linear acceleration and angular velocity of the unmanned aerial vehicle, and the method further comprises using measurements from the inertial measurement unit to predict the pose state of the unmanned aerial vehicle.
In an embodiment, the method further comprises pre-integrating the measurements from the inertial measurement unit to obtain a set of pre-integrated inertial measurement unit measurements and using the set of pre-integrated inertial measurement unit measurements to estimate the pose state of the unmanned aerial vehicle.
In an embodiment, the method further comprises comparing the predicted distance from the pose prediction of the unmanned aerial vehicle determined from measurements from the inertial measurement unit with the distance measurements by the first ultra-wideband node and the second ultra-wideband node and thereby identify outliners in the measurements by the first ultra-wideband node and the second ultra- wideband node. In an embodiment, the method further comprises comprising receiving on-board seiflocalization data generated by on-board self-localization module of the unmanned aerial vehicle.
In an embodiment, the on-board self-localization module comprises a lidar sensor and the on-board self-localization data comprises laser scan data.
In an embodiment, the on-board self-localization module comprises a camera and the on-board self-localization data comprises image data.
In an embodiment, the method further comprises estimating the pose state of the unmanned aerial vehicle using measurement data comprising measurements by the first ultra-wideband node and the second ultra-wideband node, the measurements from the inertial measurement unit and the on-board self-localization data by minimizing a cost function between the measurement data and a model predicted pose state.
In an embodiment, the on-board self-localization data comprises laser scan data, and the method further comprises extracting lidar features from the laser scan data and matching the lidar features with a local map and thereby obtaining a set of feature- map-matching coefficients, wherein the cost function depends on the set of feature- map-matching coefficients.
In an embodiment, the on-board self-localization data comprises image data, and the method further comprises extracting visual features from the image data and thereby obtaining a set of visual features, wherein the cost function depends on the set of visual features.
In an embodiment, the method further comprises the estimating the pose state of the unmanned aerial vehicle using measurement data corresponding to a sliding time window. According to a third aspect of the present disclosure, a computer readable medium storing processor executable instructions which when executed on a processor cause the processor to carry out a method as set out above is provided.
BRIEF DESCRIPTION OF THE DRAWINGS
In the following, embodiments of the present invention will be described as non-limiting examples with reference to the accompanying drawings in which:
FIG.1 shows an unmanned aerial vehicle according to an embodiment of the present invention;
FIG.2 shows the deployment of anchor nodes for use with an unmanned aerial vehicle according to an embodiment of the present invention;
FIG.3 is a table showing communication slots for anchor nodes used in embodiments of the present invention;
FIG.4 is a block diagram showing an unmanned aerial vehicle according to an embodiment of the present invention;
FIG.5 is a flowchart showing a localization method for an unmanned aerial vehicle according to an embodiment of the present invention;
FIG.6 is a flow chart showing the workflow in a localization method for an unmanned aerial vehicle according to an embodiment of the present invention;
FIG.7 is a timing diagram showing a sliding window synchronization scheme used in embodiments of the present invention;
FIG.8 shows multiple poses of an unmanned aerial vehicle and the coupling of these poses with available observations; FIG.9 is a table showing a comparison absolute trajectory error of localization methods according to embodiments of the present invention with other localization methods;
FIG.10 shows a set of ultra-wide band measurements for a test flight of an unmanned aerial vehicle according to an embodiment of the present invention;
FIG.11 is a table showing position estimates of the anchors in the flight tests;
FIG.12 is a table showing a comparison of absolute trajectory error over building inspection datasets; and
FIG.13A to FIG.13D show the results of a test flight experiment using an unmanned aerial vehicle according to an embodiment of the present invention.
DETAILED DESCRIPTION
FIG.1 shows an unmanned aerial vehicle according to an embodiment of the present invention. The unmanned aerial vehicle 100 comprises a body 101 from which six arms 102 extend. A rotor 104 is mounted on each of the arms 102. A landing gear 106 extends downward from the body 101. Two cameras are mounted on the body: a left camera 108 and a right camera 109 which both face forward. A vertical lidar sensor 110 and a horizontal lidar sensor 111 are mounted on the body 101. Each of the vertical lidar sensor 110 and the horizontal lidar sensor 111 comprise a laser and a receiver and operate to determine the range of objects by measuring the return time for reflected light to return to the receiver. The vertical lidar sensor 110 is arranged to carry out ranging in vertical directions and the horizontal lidar sensor 111 is arranged to carryout ranging in horizontal directions. An inertial measurement unit (IMU) 112 is arranged on the body 101. The inertial measurement unit 112 comprises gyroscopes and accelerometers and is configured to detect linear acceleration and angular velocity of the unmanned aerial vehicle.
Four ultra-wide band antenna nodes are arranged on arms extending from the body 101 of the unmanned aerial vehicle. As shown in FIG.1 , a first ultra-wide band antenna node 120 is arrange on an arm extending to the front left of the body 101 , a second ultra-wide band antenna node 121 is arranged on an arm extending to the rear left of the body 101 , a third ultra-wide band antenna node 122 is arrange on an arm extending to the rear right of the body 101 and a fourth ultra-wide band antenna node 123 is arranged on an arm extending to the front right of the body 101 . The UWB signal has a frequency band from 3.1 GHz to 4.8GHz. Ranging measurement is carried out by a two-way time-of-flight technique. The network is synchronized via a Time-Division Multiple Access technique where ranging measurements are scheduled on a slot map.
In the unmanned aerial vehicle shown in FIG.1 , a crystal prism 130 which acts as a ground truth tracking point is mounted on top of the body 101. The crystal prism 130 is used to verify the localization scheme in experiments such as those described below with reference to FIG.13A to FIG.13D and is not used as part of the localization schemes. Thus, the localization schemes described herein would operate without the crystal prism 130.
The ultra-wide band antenna nodes are configured to send and receive signals from anchor nodes located at fixed locations in the environment in which the unmanned aerial vehicle operates. By sending and receiving signals from these anchor nodes, the location and orientation of the unmanned aerial vehicle can be estimated.
FIG.2 shows the deployment of anchor nodes in an environment for use with an unmanned aerial vehicle according to an embodiment of the present invention. As shown in FIG.2 three ultra-wide band (UWB) ground nodes or anchor nodes are arranged in an environment 200 around a structure 210 under inspection. A first anchor node 222, a second anchor node 224 and a third anchor node 226 are arranged in the environment 200. The UAV 230 having four UWB ranging nodes 232 flies in the environment 200. A world frame 240 is defined having an x-axis 242 in the direction from the first anchor node 222 towards the second anchor node 224 and a y-axis in the direction towards the structure 210 under inspection. A z-axis is defined to point upwards by the right-hand rule. The third anchor node 226 is positioned on the -y half plane. In this scheme by having the +x direction point from one node to another, together with the preferred choice of z pointing upwards, we can easily define a fixed reference frame that can be easily aligned with the environment we wish to conduct the UAV operation in. Additionally, the three anchor nodes are arranged substantially in a horizonal plane and the four UWB ranging nodes 232 on the UAV 230 are also arranged in a horizontal plane.
It is noted that the scheme shown in FIG.2 may be extended to more than three anchor nodes, however, three are shown for simplicity as this is the minimum number required for localization of the UAV.
FIG.3 is a table showing communication slots for anchor nodes used in embodiments of the present invention. FIG.3 shows the slot map designed for one anchor node, for example, the first anchor node 222 having ID101. Notice that grayed slots where ranging between the anchors are scheduled (i.e. slots 0, 7, 14) are the same for all anchors. By configuring the UWB to echo the last range, all nodes in the network should be able to obtain the distances among the three nodes ID100, ID101 , ID102 after 2 cycles and calculate the corresponding coordinates x1 , x2, y2 of anchor ID 101 and ID102, assuming that the plane made by the three anchors are parallel with the x-y plane of the coordinates system. In this scheme, while ranging between the UAV- mounted nodes and the anchor nodes are different across the slot maps, ranging between the anchors are the same for all slot maps. In this case, slot 0, 7 and 14 are used to conduct ranging between the anchor nodes. By configuring the anchor nodes to echo their last range, all other nodes can receive a message containing the distance between anchors. These distances can then be used to calculate the coordinates of the ground node. The important issue here is to design such a set of slot maps with minimal number of nodes needed for ranging between the anchors (thus allowing more time for UAV to range to the anchors), yet still ensure enough range measurements to efficiently calculate the coordinates of the anchors.
FIG.4 is a block diagram showing an unmanned aerial vehicle according to an embodiment of the present invention. The unmanned aerial vehicle 400 comprises a flight controller 402 which is controls rotors 404 of the unmanned aerial vehicle 400. The unmanned aerial vehicle 400 further comprises a memory 406, a localization processor 408, a first ultra-wide band node 412, a second ultra-wide band node 414, an on-board self-localization module 420 and an inertial measurement unit 430. The memory 406 may store flight plans which are executed by the flight controller 402 in order to carry out an inspection or other task. The memory may also store images, video or other data captured by sensors of the unmanned aerial vehicle during a task. The localization processor 408 executes methods according to embodiments of the present invention to determine the location of the unmanned aerial vehicle using information captured by sensors on the unmanned aerial vehicle. The first ultra-wide band node 412, and the second ultra-wide band node 414 are ultra-wide band antennas which determine the range to anchor ultra-wide band nodes arranged in the environment. The on-board self-localization module 420 comprises a lidar module 422 and a camera 424. The lidar module 422 comprises a laser and a receiver and operate to determine the range of objects by measuring the return time for reflected light to return to the receiver. The inertial measurement unit 430 comprises gyroscopes and accelerometers and is configured to detect linear acceleration and angular velocity of the unmanned aerial vehicle 400.
The flight controller 402, the memory 406 and the localization processor 408 may be implemented as a small form factor computer having a processor and a memory storing computer executable instructions which are executed on the processor to carry out the methods described in this disclosure. The memory 406 may be referred to in some contexts as computer readable storage media and/or non-transitory computer readable media. Alternatively, while a software implementation of the computer program modules is described herein, these may alternatively be implemented as one or more hardware modules (such as field-programmable gate array(s) or applicationspecific integrated circuit(s)) comprising circuitry which implements equivalent functionality to that implemented in software.
FIG.5 is a flowchart showing a localization method for an unmanned aerial vehicle according to an embodiment of the present invention. The method 500 shown in FIG.5 is carried out by the unmanned aerial vehicle 400 shown in FIG.4.
In step 502, the first ultra-wide band node 412 and the second ultra-wide band node 414 of the unmanned aerial vehicle 400 receive signals from anchor nodes. From these signal, the localization processor 408 determines the range to each of the anchor nodes. In step 504, the localization processor 408 of the unmanned aerial vehicle 400 estimates the pose state of the unmanned aerial vehicle. The pose state comprises the location and orientation of the unmanned aerial vehicle relative to the anchor nodes in the environment.
FIG.6 is a flow chart showing the workflow in a localization method for an unmanned aerial vehicle according to an embodiment of the present invention. The method 600 shown in FIG.6 is carried out by the unmanned aerial vehicle 400 shown in FIG.4.
As shown in FIG.6, the inputs to the method are ultra-wide band (UWB) data 610, inertial measurement unit (IMU) data 620, pointclouds 630 which are generated by the lidar module 422 of the unmanned aerial vehicle 400, and images 650 which are generated by the camera 424 of the unmanned aerial vehicle 400. The method 600 also uses UWB measurements
Figure imgf000014_0001
and pose predictions xm from previous stages in key frame management 640.
As described above with reference to FIG.2, the UWB ranging and communication 602 network is employed to gather the distance between the anchors, after they have been deployed in the field. These distances will be then used to calculate the position of the anchors by using an anchor self-localization 604 algorithm. Once the anchors' positions 605 have been calculated, they will be associated with the UWB measurements 603. Thus, a UWB measurement which makes up the UWB data 610 consists of three pieces of information, first is the position of the UWB node in the UAV's body frame, second is the UWB anchor position in the chosen world frame, and third is the distance between these two UWB nodes.
After the self-localization stage, UWB, IMU and OSL data will go through multiple stages before they are fused together in the Optimization-based Sensor Fusion (OSF) algorithm. The system is synchronized by the lidar messages, whose time stamps are used to determine the time steps from time tw to tk. This synchronization process is described in more detail below with reference to FIG.7.
The UWB data 610 is subjected to some preliminary checks 612 to eliminate bad measurements. These checks include comparing the measurement's signal over noise ratio (SNR), leading-edge-detection quality, and rate-of-change with some user- defined thresholds, etc. Only those that meet these thresholds will be passed to the buffer in a bundling 614 stage for subsequent stages. UWB measurements that arrive in the interval
Figure imgf000015_0001
are grouped into a set
Figure imgf000015_0002
Finally, this set will be put into a final outlier check 616 to see if any measurement in it is an outlier. This is done by comparing the measurement with the IMU-predicted range value. As shown in FIG.6, the IMU propagated state 629 determined in by the pose prediction 628 algorithm is used in the outlier check 616.
FIG.7 is a timing diagram showing a sliding window synchronization scheme used in embodiments of the present invention. As shown in FIG.7, The system is synchronized by the lidar messages 710, whose time stamps are used to determine the time steps from time tw to tk. The IMU data 720, the UWB data 730 and the image data 740 is then allocated into the sliding windows according to the timing relative to the lidar messages 710. IMU data from a subsequent frame 722 is pre-integrated and used in the processing of earlier frames. IMU data from one or more frames 724 is propagated and used to compensate for motion distortion. A group of UWB data points 732 in a frame may be defined having a timestamp . An image 742 may be defined as having
Figure imgf000015_0003
a visual time delay td with respect to the timing of the corresponding frame defined by the lidar messages 710.
Returning now to FIG.6, the IMU data 620 is subjected to bundling 622 to arrange the data into frames as described above with reference to FIG.7, and pre-integration before being input into a local optimization algorithm 626. The local optimization algorithm 626 generates a pose prediction 628.
For the lidar point cloud data 630, the plane and edge features are then extracted using the common smoothness criteria of LOAM systems in a feature extraction 632 step. The motion distortion in these features is compensated by the IMU-propagated states, and then transform to the inertial frame in the transform and deskew step 634. Hence, these features are then matched with the local map 642 built from marginalized point clouds in an association 644 step to obtain the set of feature-map-matching (FMM) coefficients which are used to create the lidar factors in the final cost function calculated in the local optimization algorithm 626.
The feature-map-matching coefficients are calculated using the following algorithm:
Figure imgf000016_0001
For the camera images 650, the binary robust independent elementary features (BRIEF features) are extracted in a feature extraction sync step 652 and then use the latest pose estimates to triangulate these features for depth initialization in a triangulation step 645. Hence, we can obtain a set of coefficients
Figure imgf000016_0002
which can be used to construct the visual factors in the final cost functions used in the local optimization algorithm 626.
Specifically, is defined as:
Figure imgf000017_0001
where aZi, bZi are the projected coordinates of the feature in camera frames a and b, afi, bfi are the feature’s 3D coordinates w.r.t camera frames a and b, λ1 is the inverse depth of visual feature i, which is only associated with the camera frame that first observes the feature.
The cost function will be then optimized to provide the estimate of the robot states in the sliding windows. The latest state estimate in the sliding windows will be combined with the latest IMU measurements in the buffer, which are leftovers from the last extraction plus the newly arrived during the optimization process, to calculate the high- rate pose estimate. This high-rate pose estimate is also used to perform the outlier check on the UWB measurements that was mentioned earlier.
After the optimization has been completed, the sliding window will be slid forward, the oldest state estimate and its associated measurements will be abandoned, and a new time step will be registered when the conditions mentioned earlier are met.
FIG.8 shows multiple poses of an unmanned aerial vehicle and the coupling of these poses with available observations. As shown in FIG.8, the unmanned aerial vehicle 802 having UWB nodes 804 moves through a series of positions which are labelled k, k+1 , k+2 and k+3. Three anchor UWB nodes 810 are arranged in the environment. The pose of the unmanned aerial vehicle 802 is expressed relative to axes 820. As shown in FIG.8, the pose of the UAV 802 at position k is Xk, the poses at positions k+1 , k+2 and k+3 are Xk+1, Xk+2 and Xk+3 respectively. The IMU measurements between the positions are Zk,Zk+1 and Zk+2. The UWB measurements between the UWB nodes 802 of the UAV and the anchor nodes are expressed as dk+3 where the superscript number indicates a label of the anchor node.
We denote all of the state estimates on the sliding window at time tk as Xk as follows:
Figure imgf000018_0003
where is the orientation quaternion estimate, and
Figure imgf000018_0001
are the robot’s position
Figure imgf000018_0004
and velocity estimates w.r.t the world frame, and are the estimates of the
Figure imgf000018_0005
Figure imgf000018_0006
gyroscope and accelerometer biases of the IMU, are the estimates of the
Figure imgf000018_0007
inverse depth of Nk visual features that are tracked in the sliding window.
Hence, we can construct a cost function based on the observations from UWB, IMU and OSL measurements. The cost function can be constructed at each time step tk+M as follows:
Figure imgf000018_0002
where are the residuals constructed from IMU, UWB and
Figure imgf000018_0008
OSL measurements are the covariance matrices of the
Figure imgf000018_0009
measurement errors, and are the number of UWB measurement in the
Figure imgf000018_0010
intervals. The cost function above summarizes the coupling of all the states and the observations. Once the cost function has been constructed, we can use open-source non-linear solvers such as g2o or ceres to execute the optimization.
Below, we will explain in detail the process to construct each factor in the cost function. Essentially, for each measurement, we will find an observation model that links the actual measurements with the states. Then we will calculate the residual (or loosely- speaking, the difference) between the measurement and the model-predicted value. Then, we need to derive the covariance matrix of the measurement errors, to properly apply the weightage among the factor. Finally, to facilitate real-time optimization, the Jacobians of each factor have to be analytically derived. i. Ranging factor: At each time step tk we have the following observation from a UAV ranging node:
Figure imgf000019_0001
where is the distance measurement, and is the position of the UWB anchor w.r.t. the world frame,
Figure imgf000019_0002
is the coordinate of the UWB ranging node in the body frame, are the time stamps defined in FIG.7.
Figure imgf000019_0007
The residual of a range measurement
Figure imgf000019_0008
is calculated by taking the difference between the distance estimate dm and the distance measurement
Figure imgf000019_0009
.
Figure imgf000019_0003
where represents the estimated distance:
Figure imgf000019_0010
And the covariance of the measurement error is chosen as
Figure imgf000019_0004
ii. OSL factors:
For lidar factors, the residual and covariance constructed from each feature’s FMM coefficients is defined as
Figure imgf000019_0011
Figure imgf000019_0012
For visual feature factors, the residual and covariance from each feature’s observation is defined as
Figure imgf000019_0005
Figure imgf000019_0006
iii. IMU pre-integration factors:
The IMU measurements are defined as
Figure imgf000020_0003
where g is the gravitational acceleration,
Figure imgf000020_0004
are respectively the
Figure imgf000020_0006
biases of the gyroscope and the accelerometer, and are some zero-
Figure imgf000020_0005
mean Gaussian noises, whose standard deviations are
Figure imgf000020_0007
respectively.
Given the noisy IMU measurements and some nominal bias values ,
Figure imgf000020_0008
Figure imgf000020_0009
we can calculate the following tuple of IMU pre-integrations:
Figure imgf000020_0001
where
Figure imgf000020_0002
In practice, the above integration can be implemented by zero-order-hold (ZOH) or higher order methods (Runge-Kutta methods). During the period from , we
Figure imgf000020_0010
have a sequence of m IMU samples
Figure imgf000020_0011
. Thus, if
ZOH is chosen, we use the following recursive rule to update the pre-integrated measurements:
Figure imgf000020_0012
for n = 0,1, ... m - 1, where τn is the time stamp of the n-th sample, ,
Figure imgf000021_0001
and .
Figure imgf000021_0002
The residual of the IMU preintegration factor is defined as follows
Figure imgf000021_0003
Figure imgf000021_0004
,
C are computed via an iterative scheme as follows:
Figure imgf000021_0005
The covariance of the IMU pre-integration errors are calculated based on a propagation scheme.
FIG.9 is a table showing a comparison absolute trajectory error of localization methods according to embodiments of the present invention with other localization methods.
The methods VINS-Mono, VINS-Fusion, A-LOAM, LIO-SAM, M-LOAM are prior art methods and the methods labelled VIRAL (horiz. lidar) and VIRAL (2 lidars) correspond to localization methods according to embodiments of the using a single horizontal lidar and two lidars respectively. All algorithms are run on an NUC 10 computer with core i7 processor. Each method is slightly modified and configured for their best performance with the dataset. Since several lidar-based methods are not designed to incorporate multiple lidar inputs, we also include experiments of VIRAL using only the horizontal lidar for a fairer comparison. The Absolute Trajectory Error (ATE) result is summarized in FIG.9. The best odometry result is highlighted in bold, and the second best is underlined. All values are in m. From this table we can clearly see that VIRAL consistently achieves better performance compared to existing methods, even when only one lidar is used.
To further verify the efficacy of the localization scheme, we develop a platform consisting of a camera system, one IMU, two LIDAR sensors, and two UWB modules. We improve the utility of the UWB sensor by using the two antennas mounted at different points on the platform, Thus, effectively we have a total of four UAV ranging nodes. The data rate of the stereo image, IMU, LIDAR 1 , LIDAR 2, and UWB topics are 10 Hz, 400 Hz, 10 Hz, 10 Hz, and 69 Hz respectively. A Leica MS60 station is used to provide ground truth for the experiment. All software is run on a small form factor computer with Intel Core i7-8650U processor.
FIG.10 shows a set of ultra-wide band measurements for a test flight of an unmanned aerial vehicle according to an embodiment of the present invention. The UWB nodes on the UAV are labelled 200. A, 200. B, 201 .A and 201 .B. The anchor UWB nodes are labelled 100, 101 and 102. Each of the plots shown in FIG.10 corresponds to measurements from one of the UWB nodes on the UAV to one of the anchor nodes. From the zoomed-in part of the 200.A→101 plot, the variation of the distance is about 5cm. We also notice some outliers in the 200.A→102 and 200.B→102 ranging measurements.
FIG.11 is a table showing position estimates of the anchors in the flight tests. The position estimates were determined as described above with reference to FIG.2 and FIG.3.
FIG.12 is a table showing a comparison of absolute trajectory error over building inspection datasets. A total of four flight tests were conducted. As having introduced earlier, since our work is meant to deliver a solution for UAV-based inspection application, the UAV’s trajectory is different from those in common datasets, which are often extensive in x and y directions, but have limited movement in the z direction. In our case, for inspection purposes, the UAV only takes off and moves in a vertical plane parallel to the object. Hence the movements along the x and z directions are extensive, while movement along the y direction is minimal.
FIG.13A to FIG.13D show the results of a test flight experiment. The results correspond to the flight labelled bid_01 in FIG.11 and FIG.12. FIG.13A shows a 3D view of the UAV trajectory as estimated by individual OSL systems and the sensor fusion result. FIG.13B shows Orientation estimate from the system over time. FIG.13C shows velocity estimate from the system over time. FIG.13D shows position estimates from the OSL systems and sensor fusion results over time. The ground truth position was obtained using a prism mounted on the UAV as shown in FIG.1 .
From the results in FIG.13A to FIG13D, we can reaffirm the benefit of UWB-IMU-OSL fusion over the state-of-the-art methods. Moreover, as can be seen on FIG.13A that while all of the OSL estimates exhibit significant drift over the operation range, our sensor fusion’s estimation error remains consistent, which is the most important quality we are looking for. Besides the position error, though we don’t have a ground truth for the orientation, during the flight test, the drone’s heading was kept relatively unchanged, always pointing towards the object. Hence the yaw value should be around 90 degrees. Indeed, we can see that the heading estimate stays around this value in FIG.13B.
It is anticipated that UAVs implementing the methods described above will find applications in aerial surveillance by UAVs; automated inspection operations by UAVs; and graphical modelling and structural health of 3D objects.
Whilst the foregoing description has described exemplary embodiments, it will be understood by those skilled in the art that many variations of the embodiments can be made within the scope and spirit of the present invention.

Claims

1 . An unmanned aerial vehicle comprising: a first ultra-wide band node; a second ultra-wide band node; and a localization processor configured to use signals received at the first ultra-wide band node and the second ultra-wide band node from a plurality of anchor nodes located within an environment to estimate a pose state of the unmanned aerial vehicle, wherein the first ultra-wide band node and the second ultra-wide band node are arranged on the unmanned aerial vehicle at positions offset from one another.
2. An unmanned aerial vehicle according to claim 1 , further comprising an inertial measurement unit configured to detect linear acceleration and angular velocity of the unmanned aerial vehicle and wherein the localization processor is further configured to use measurements from the inertial measurement unit to predict the pose state of the unmanned aerial vehicle.
3. An unmanned aerial vehicle according to claim 2, wherein the localization processor is further configured to pre-integrate the measurements from the inertial measurement unit to obtain a set of pre-integrated inertial measurement unit measurements and to use the set of pre-integrated inertial movement unit measurements to estimate the pose state of the unmanned aerial vehicle.
4. An unmanned aerial vehicle according to claim 2 or claim 3, wherein the localization processor is further configured to compare a predicted distance calculated from the pose state of the unmanned aerial vehicle determined from the inertial measurement measurements with distance predictions determined from measurements by the first ultra-wideband node and the second ultra-wideband node and thereby identify outliners in the measurements by the first ultra-wideband node and the second ultra-wideband node.
5. An unmanned aerial vehicle according to any one of claims 2 to 4, further comprising an on-board self-localization module configured to generate on-board selflocalization data.
6. An unmanned aerial vehicle according to claim 5, wherein the on-board selflocalization module comprises a lidar sensor and the on-board self-localization data comprises laser scan data.
7. An unmanned aerial vehicle according to claim 5, wherein the on-board selflocalization module comprises a camera and the on-board self-localization data comprises image data.
8. An unmanned aerial vehicle according to any one of claims 5 to 7, wherein the localization processor is further configured to estimate the pose state of the unmanned aerial vehicle using measurement data comprising measurements by the first ultra- wideband node and the second ultra-wideband node, the measurements from the inertial measurement unit and the on-board self-localization data by minimizing a cost function between the measurement data and a model predicted pose state.
9. An unmanned aerial vehicle according to claim 8, wherein the on-board selflocalization module comprises a lidar sensor and the on-board self-localization data comprises laser scan data and the localization processor is further configured to use the on-board self-localization data to extract lidar features from the laser scan data and match the lidar features with a local map to obtain a set of feature-map-matching coefficients, wherein the cost function depends on the set of feature-map-matching coefficients.
10. An unmanned aerial vehicle according to claim 8, wherein the on-board selflocalization module comprises a camera and the on-board self-localization data comprises image data and localization processor is further configured to use the onboard self-localization data to extract a set of visual features from image data, wherein the cost function depends on the set of visual features.
11. An unmanned aerial vehicle according to any one of claims 8 to 10, wherein the localization processor is further configured to estimate the pose state of the unmanned aerial vehicle using measurement data corresponding to a sliding time window.
12. A localization method for an unmanned aerial vehicle, the unmanned aerial vehicle comprising: a first ultra-wide band node; and a second ultra-wide band node, wherein the first ultra-wide band node and the second ultra-wide band node are arranged on the unmanned aerial vehicle at positions offset from one another, the method comprising using signals received at the first ultra-wideband node and the second ultra- wideband node from a plurality of anchor nodes located within an environment to estimate a pose state of the unmanned aerial vehicle.
13. A localization method according to claim 12, wherein the unmanned aerial vehicle further comprises: an inertial measurement unit configured to detect linear acceleration and angular velocity of the unmanned aerial vehicle, and the method further comprises using measurements from the inertial measurement unit to predict the pose state of the unmanned aerial vehicle.
14. A localization method according to claim 13, further comprising pre-integrating the measurements from the inertial measurement unit to obtain a set of pre-integrated inertial measurement unit measurements and using the set of pre-integrated inertial measurement unit measurements to estimate the pose state of the unmanned aerial vehicle.
15. A localization method according to claim 13 or claim 14, further comprising comparing the predicted distance from the pose prediction of the unmanned aerial vehicle determined from measurements from the inertial measurement unit with the distance measurements by the first ultra-wideband node and the second ultra- wideband node and thereby identify outliners in the measurements by the first ultra- wideband node and the second ultra-wideband node.
16. A localization method according to any one of claims 13 to 15, further comprising receiving on-board self-localization data generated by on-board selflocalization module of the unmanned aerial vehicle.
17. A localization method according to claim 18, wherein the on-board self- localization module comprises a lidar sensor and the on-board seif-localization data comprises laser scan data.
18. A localization method according to claim 16 or claim 17, wherein the on-board seif-localization module comprises a camera and the on-board self-localization data comprises image data.
19. A localization method according to any one of claims 16 to 18, further comprising estimating the pose state of the unmanned aerial vehicle using measurement data comprising measurements by the first ultra-wideband node and the second ultra-wideband node, the measurements from the inertial measurement unit and the on-board self-localization data by minimizing a cost function between the measurement data and a model predicted pose state.
20. A localization method according to claim 19, wherein the on-board self- localization data comprises laser scan data, and the method further comprises extracting lidar features from the laser scan data and matching the lidar features with a local map and thereby obtaining a set of feature-map-matching coefficients, wherein the cost function depends on the set of feature-map-matching coefficients.
21. A localization method according to claim 19, wherein the on-board self- localization data comprises image data, and the method further comprises extracting visual features from the image data and thereby obtaining a set of visual features, wherein the cost function depends on the set of visual features.
22. A localization method according to any one of claim 19 to 21 , further comprising the estimating the pose state of the unmanned aerial vehicle using measurement data corresponding to a sliding time window.
23. A computer readable medium storing processor executable instructions which when executed on a processor cause the processor to carry out a method according to any one of claims 12 to 22.
PCT/SG2021/050512 2020-08-31 2021-08-26 Unmanned aerial vehicle and localization method for unmanned aerial vehicle WO2022045982A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US18/040,148 US20230314548A1 (en) 2020-08-31 2021-08-26 Unmanned aerial vehicle and localization method for unmanned aerial vehicle

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
SG10202008405P 2020-08-31
SG10202008405P 2020-08-31

Publications (1)

Publication Number Publication Date
WO2022045982A1 true WO2022045982A1 (en) 2022-03-03

Family

ID=80353009

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/SG2021/050512 WO2022045982A1 (en) 2020-08-31 2021-08-26 Unmanned aerial vehicle and localization method for unmanned aerial vehicle

Country Status (2)

Country Link
US (1) US20230314548A1 (en)
WO (1) WO2022045982A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117784120A (en) * 2024-02-23 2024-03-29 南京新航线无人机科技有限公司 Unmanned aerial vehicle flight state monitoring method and system
CN117784120B (en) * 2024-02-23 2024-05-28 南京新航线无人机科技有限公司 Unmanned aerial vehicle flight state monitoring method and system

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN206202676U (en) * 2016-12-02 2017-05-31 国网四川省电力公司电力应急中心 A kind of four rotor unmanned aircraft obstacle avoidance apparatus of use UWB
CN108303995A (en) * 2018-03-02 2018-07-20 三峡大学 A kind of substation inspection unmanned plane during flying security system
US20180244385A1 (en) * 2015-08-20 2018-08-30 Lg Electronics Inc. Unmanned aircraft and method of controlling the same

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20180244385A1 (en) * 2015-08-20 2018-08-30 Lg Electronics Inc. Unmanned aircraft and method of controlling the same
CN206202676U (en) * 2016-12-02 2017-05-31 国网四川省电力公司电力应急中心 A kind of four rotor unmanned aircraft obstacle avoidance apparatus of use UWB
CN108303995A (en) * 2018-03-02 2018-07-20 三峡大学 A kind of substation inspection unmanned plane during flying security system

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
BALDERAS ET AL.: "Low-Profile Conformal UWB Antenna for UAV Applications", IEEE ACCESS, vol. 7, 4 September 2019 (2019-09-04), pages 127486 - 127494, XP011746453, DOI: 10.1109/ACCESS.2019.2939511 *
NGUYEN ET AL.: "Integrated UWB-Vision Approach for Autonomous Docking of UAVs in GPS-denied Environments", 2019 INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA, 24 May 2019 (2019-05-24), pages 9603 - 9609, XP033593848, DOI: 10.1109/ICRA.2019.8793851 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117784120A (en) * 2024-02-23 2024-03-29 南京新航线无人机科技有限公司 Unmanned aerial vehicle flight state monitoring method and system
CN117784120B (en) * 2024-02-23 2024-05-28 南京新航线无人机科技有限公司 Unmanned aerial vehicle flight state monitoring method and system

Also Published As

Publication number Publication date
US20230314548A1 (en) 2023-10-05

Similar Documents

Publication Publication Date Title
CN112347840B (en) Vision sensor laser radar integrated unmanned aerial vehicle positioning and image building device and method
JP7141403B2 (en) Laser scanner with real-time online self-motion estimation
US20230130320A1 (en) Laser scanner with real-time, online ego-motion estimation
US10962370B2 (en) Laser scanner with real-time, online ego-motion estimation
US9709404B2 (en) Iterative Kalman Smoother for robust 3D localization for vision-aided inertial navigation
CN109887057B (en) Method and device for generating high-precision map
Bijjahalli et al. Advances in intelligent and autonomous navigation systems for small UAS
US20190346271A1 (en) Laser scanner with real-time, online ego-motion estimation
US11466990B2 (en) Square-root multi-state constraint Kalman filter for vision-aided inertial navigation system
Eustice et al. Exactly sparse delayed-state filters
US10254118B2 (en) Extrinsic parameter calibration of a vision-aided inertial navigation system
Sazdovski et al. Inertial navigation aided by vision-based simultaneous localization and mapping
WO2018071416A1 (en) Laser scanner with real-time, online ego-motion estimation
CN113625774B (en) Local map matching and end-to-end ranging multi-unmanned aerial vehicle co-location system and method
Steiner et al. A vision-aided inertial navigation system for agile high-speed flight in unmapped environments: Distribution statement a: Approved for public release, distribution unlimited
Zachariah et al. Camera-aided inertial navigation using epipolar points
Ivancsits et al. Visual navigation system for small unmanned aerial vehicles
US20210319568A1 (en) Cooperative Aircraft Navigation
US20230314548A1 (en) Unmanned aerial vehicle and localization method for unmanned aerial vehicle
Panahandeh et al. IMU-camera data fusion: Horizontal plane observation with explicit outlier rejection
Sakamaki et al. Tracking multiple ground objects using a team of unmanned air vehicles
Özbek et al. Research on the Availability of VINS-Mono and ORB-SLAM3 Algorithms for Aviation
Fan et al. An incremental LiDAR/POS online calibration method
Li et al. Dynamic obstacle tracking based on high-definition map in urban scene
Hidaka et al. Terrain relative navigation enhanced with sar for moon’s shadowed regions

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21862220

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 21862220

Country of ref document: EP

Kind code of ref document: A1