WO2021248636A1 - System and method for detecting and positioning autonomous driving object - Google Patents

System and method for detecting and positioning autonomous driving object Download PDF

Info

Publication number
WO2021248636A1
WO2021248636A1 PCT/CN2020/102996 CN2020102996W WO2021248636A1 WO 2021248636 A1 WO2021248636 A1 WO 2021248636A1 CN 2020102996 W CN2020102996 W CN 2020102996W WO 2021248636 A1 WO2021248636 A1 WO 2021248636A1
Authority
WO
WIPO (PCT)
Prior art keywords
module
measurement data
inertial navigation
stereo vision
satellite
Prior art date
Application number
PCT/CN2020/102996
Other languages
French (fr)
Chinese (zh)
Inventor
王峰
潘观潮
刘进辉
王宏武
王晓洒
Original Assignee
东莞市普灵思智能电子有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 东莞市普灵思智能电子有限公司 filed Critical 东莞市普灵思智能电子有限公司
Publication of WO2021248636A1 publication Critical patent/WO2021248636A1/en

Links

Images

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S19/00Satellite radio beacon positioning systems; Determining position, velocity or attitude using signals transmitted by such systems
    • G01S19/38Determining a navigation solution using signals transmitted by a satellite radio beacon positioning system
    • G01S19/39Determining a navigation solution using signals transmitted by a satellite radio beacon positioning system the satellite radio beacon positioning system transmitting time-stamped messages, e.g. GPS [Global Positioning System], GLONASS [Global Orbiting Navigation Satellite System] or GALILEO
    • G01S19/42Determining position
    • G01S19/45Determining position by combining measurements of signals from the satellite radio beacon positioning system with a supplementary measurement
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C21/00Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
    • G01C21/26Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 specially adapted for navigation in a road network
    • G01C21/28Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 specially adapted for navigation in a road network with correlation of data from several navigational instruments
    • G01C21/30Map- or contour-matching
    • G01C21/32Structuring or formatting of map data
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C21/00Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
    • G01C21/26Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 specially adapted for navigation in a road network
    • G01C21/34Route searching; Route guidance
    • G01C21/3407Route searching; Route guidance specially adapted for specific applications
    • G01C21/343Calculating itineraries, i.e. routes leading from a starting point to a series of categorical destinations using a global route restraint, round trips, touristic trips
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C21/00Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
    • G01C21/26Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 specially adapted for navigation in a road network
    • G01C21/34Route searching; Route guidance
    • G01C21/3446Details of route searching algorithms, e.g. Dijkstra, A*, arc-flags, using precalculated routes
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C25/00Manufacturing, calibrating, cleaning, or repairing instruments or devices referred to in the other groups of this subclass
    • G01C25/005Manufacturing, calibrating, cleaning, or repairing instruments or devices referred to in the other groups of this subclass initial alignment, calibration or starting-up of inertial devices
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S19/00Satellite radio beacon positioning systems; Determining position, velocity or attitude using signals transmitted by such systems
    • G01S19/38Determining a navigation solution using signals transmitted by a satellite radio beacon positioning system
    • G01S19/39Determining a navigation solution using signals transmitted by a satellite radio beacon positioning system the satellite radio beacon positioning system transmitting time-stamped messages, e.g. GPS [Global Positioning System], GLONASS [Global Orbiting Navigation Satellite System] or GALILEO
    • G01S19/42Determining position
    • G01S19/45Determining position by combining measurements of signals from the satellite radio beacon positioning system with a supplementary measurement
    • G01S19/47Determining position by combining measurements of signals from the satellite radio beacon positioning system with a supplementary measurement the supplementary measurement being an inertial measurement, e.g. tightly coupled inertial

Definitions

  • the present invention relates to the field of computer application technology, in particular to an automatic driving object detection and positioning system and method.
  • Autonomous driving means that the car perceives the road environment through the on-board sensor system, and controls the steering and speed of the vehicle based on the road, vehicle position and obstacle information obtained by the perception, and then automatically plans the driving route and controls the vehicle to reach the predetermined target Technology.
  • the existing technology has a combination system of binocular direct method vision system and inertial navigation module, but the error generated by the vision system and inertial navigation module in the combined system cannot be Effective restriction, the error of the combined system will increase indefinitely when there is no image gradient for a long time, resulting in the failure of the combined system's perception.
  • the prior art also has monocular feature point method vision systems, inertial navigation modules, and satellite navigation close-coupled automatic driving perception systems, but monocular cameras cannot detect uncharacteristic obstacles, such as isolated guardrails on highways, bicycles or animals, etc. .
  • the existing vision system also uses a binocular stereo vision system for coupling, but still uses the feature point method, which requires a large amount of calculation and high hardware performance requirements.
  • the most advanced binocular stereo vision environment detection method only uses the visual difference information in binocular vision, and does not use the images collected by the camera at different times and positions to realize the three-dimensional modeling of the environment.
  • the purpose of the present invention is to provide an automatic driving object detection and positioning system and method, which has the advantages of maximizing the positioning accuracy of the automatic driving perception system, and improving the calculation efficiency and reliability, so as to solve the problems raised in the background art.
  • An automatic driving object detection and positioning system including a stereo vision image processing module, a satellite navigation module, an inertial navigation module and a system tight coupling module, in which:
  • Stereo vision image processing module using or multi-eye camera to obtain the image data of the stereo vision module
  • the satellite navigation module is used to obtain the original measurement data of satellite navigation through the receiver
  • the inertial navigation module is used to obtain the measurement data of the inertial navigation module by using inertial sensors;
  • the system tight coupling module is used to tightly couple the measurement data of the inertial navigation module, the image data of the stereo vision module and the original measurement data of the satellite navigation, and establish three-dimensional environment information, and finally use the deep neural network module to detect Targets and obstacles in the environment;
  • the deep neural network is composed of a three-dimensional sparse convolutional neural network, a dot network neural network, and a combination of the two;
  • the stereo vision image processing module, the satellite navigation module and the inertial navigation module are all connected to the system tight coupling module.
  • the stereo vision image processing module includes a binocular or multi-eye camera.
  • a method for automatic driving object detection and positioning including the following steps:
  • S1 Obtain the image data of the stereo vision module, the measurement data of the inertial navigation module and the original measurement data of the satellite navigation;
  • S2 tightly couple the image data of the stereo vision module, the original measurement data of the satellite navigation and the measurement data of the inertial navigation module to correct the drift error of the inertial navigation module;
  • S3 Tightly couple the image data of the stereo vision module, the original measurement data of the satellite navigation and the measurement data of the inertial navigation module to establish three-dimensional environment information, and then use the deep neural network module to judge and classify the targets and obstacles in the environment;
  • the deep neural network uses a three-dimensional sparse convolutional neural network, a dot network neural network, or a combination of the two.
  • the stereo vision module adopts multi-camera, including binocular camera direct method for processing, and the image data includes the visual difference between multi-camera or binocular stereo vision cameras at the same time and each camera at different time,
  • the image information captured by the location is used to establish the three-dimensional information of the environment.
  • the tight coupling consists of a weighted reprojection error of stereo vision, satellite navigation error, and state error from inertial navigation to form a cost function.
  • the S1 specifically includes the following steps:
  • S11 Use multi-eye or binocular cameras to obtain image data of the stereo vision module.
  • the image data includes the visual difference between the multi-eye or binocular stereo vision cameras at the same time and the images captured by each camera at different times and positions information;
  • S12 Use inertial sensors to obtain measurement data of the inertial navigation module
  • S14 Perform tight coupling processing on the measurement data of the inertial navigation module, the image data, and the original measurement data of the satellite navigation.
  • the S11 specifically includes the following steps:
  • S111 Use a multi-eye or binocular camera to collect environmental image signals
  • S112 Combine the visual difference captured by multiple different cameras at the same time and the images captured by each camera at different times and positions to form a direct method of stereo vision observation;
  • S113 Combine the measurement data of inertial navigation and satellite navigation to establish three-dimensional environment information, and finally use the deep neural network module to classify and judge the targets and obstacles in the environment.
  • the S12 specifically includes the following steps:
  • S121 Use inertial sensors to measure the 3-axis acceleration and 3-axis angular velocity of the autonomous vehicle in a fixed coordinate system
  • S123 Combine image data, inertial navigation data and satellite navigation data to establish three-dimensional environment information, and then use deep neural network modules to classify and judge objects and obstacles in the environment.
  • the S14 specifically includes: correcting the drift error of the inertial navigation module by combining the original measurement data of the satellite navigation with the image data of the stereo vision module.
  • the beneficial effect of the present invention is: the present invention tightly couples the measurement data of the inertial navigation module, the image data of the stereo vision module, and the original measurement data of the satellite navigation, and measures the error of the measurement data of the inertial navigation module. Correction, thereby improving the positioning accuracy, and then using the three-dimensional environment modeling data and deep neural network realized by the direct method of stereo vision to identify and judge objects in the environment.
  • the system uses high-precision three-dimensional sparse convolutional neural networks, dot-net neural networks, or their combination to improve the accuracy of obstacle judgment, and no longer resorts to expensive laser scanning radar, thereby reducing the cost of autonomous vehicles.
  • FIG. 1 is a flowchart of an automatic driving positioning and object detection method according to an embodiment of the present invention
  • FIG. 2 is a structural diagram of a three-dimensional sparse convolutional neural network of an automatic driving object detection and positioning system according to an embodiment of the present invention
  • FIG. 3 is a structural diagram of a point network neural network of an automatic driving object detection and positioning system according to an embodiment of the present invention
  • Fig. 4 is a schematic structural diagram of an automatic driving object detection and positioning system according to an embodiment of the present invention.
  • Fig. 5 is an architecture diagram of the automatic driving object detection and positioning system of the present invention.
  • An automatic driving object detection and positioning system and method based on the tight coupling of a stereo vision module, an inertial navigation module and a satellite navigation module can continuously provide information with high accuracy in a short time, but positioning errors will accumulate over time; satellite navigation will be long-term Stability is good, but it is susceptible to interference, and the data update frequency is low; the usual stereo vision uses the feature point method, first select multiple feature points in the image, and then use the visual difference of these feature points captured by the left and right eyes to match. Finally, the triangulation method is used to determine the distance information of these feature points to the camera, so as to firstly detect and locate the three-dimensional environment.
  • the stereo vision method based on feature points needs to select many feature points in each frame of image, which consumes a lot of computing resources, and in many cases, there are no suitable feature points in the image. For example, when the image intensity has only one fixed gradient, the feature point method will fail completely.
  • the stereo vision method based on feature points does not use the images captured by the camera at different positions to achieve distance measurement, which wastes valuable measurement data and leads to environmental modeling accuracy errors and positioning errors.
  • a better algorithm is to use the direct method of stereo vision, it does not need to select feature points, saving computing resources, and as long as there are gradient images, environment modeling and positioning can be achieved.
  • it not only uses the image data captured by different cameras at the same time, but also uses the image information captured by the cameras at different times and positions for three-dimensional modeling and positioning.
  • the stereo vision calculates the obstacles to the obstacles according to the parallax of different cameras. The distance of the camera, but the visual system cannot effectively locate and detect the distance of obstacles in an environment where the image lacks gradient.
  • the stereo vision direct method, inertial navigation and satellite navigation form a combined navigation system, which can assist each other in perceiving the state of the vehicle and the environment, complement each other in different environments, improve the reliability and navigation accuracy, can accurately extract the distance, and can replace the current situation.
  • Some laser scanning radars reduce costs.
  • Fig. 1 is a flowchart of a tightly coupled automatic driving perception method according to an embodiment of the present invention. As shown in Fig. 1, the tightly coupled automatic driving perception method specifically includes:
  • the image data of the stereo vision module is obtained by the stereo vision image processing module of the binocular direct method.
  • the detailed process of using the binocular camera to obtain the image data of the stereo vision module is described below:
  • the direct method is based on the assumption that the gray level is invariable.
  • the pixel gray level of the same spatial point is fixed in each image.
  • the direct method does not need to extract features. At the same time, it lacks corners and edges or the light changes are not obvious.
  • the environment also has better results; and the direct method requires less data to be processed, which can achieve high computational efficiency.
  • the binocular camera uses the binocular camera to obtain the grayscale image of two or more images of the measured object from different positions (the color image can use the intensity of red, green, and blue, respectively use this method), and record the binocular camera at t
  • the images obtained at time and t+1 are I i and I j .
  • R ji is a 3*3 rotation matrix
  • t is the translation vector
  • the luminosity error energy function is:
  • ⁇ p is the weight value, in order to reduce the weight corresponding to the pixels with large gradients in the image, so that the final energy function can reflect most of the pixels
  • the luminosity error is not the error of individual pixels with large gradients
  • I i [p] and I j [p'] are respectively the gray value of Pi at the corresponding point of the image.
  • the camera position and attitude angle are obtained by optimization by minimizing the photometric error energy.
  • the photometric error energy is at the minimum, the image data of the camera pose and environmental information of the stereo vision module can be obtained.
  • the objective function is:
  • the above formulas (1-3, 1-4) are the photometric error energy function of the monocular camera.
  • the direct method is a stereo vision module.
  • the combined image data includes the visual static difference data collected by the binocular cameras at the same time and the dynamic time series data collected by each camera at different times and positions.
  • the binocular direct method balances the relative weight of each image frame of the camera and the static stereo vision by introducing a coupling factor ⁇ , and its error energy function is:
  • obs(p) represents all image frames where p points can be observed
  • i belongs to different image frames
  • p belongs to multiple points in the same frame
  • j is all frames where p points can be observed.
  • the objective function of the new binocular camera can be obtained.
  • R ji and t appearing in the conversion between the above two image frames are the pose of the camera.
  • the pose of the camera and the three-dimensional space point Pi are obtained. coordinate.
  • the camera pose is used to correct the errors in the measurement data of the inertial navigation module of the autonomous vehicle in the inertial navigation.
  • the camera pose acquired by the binocular direct method does not require feature extraction, and the ability to respond to road and environmental changes is significantly improved.
  • the collection of the three-dimensional coordinates of many spatial points Pi becomes the three-dimensional point cloud of the environment.
  • the measurement data of the inertial navigation module is obtained through the inertial measurement unit.
  • the measurement data of the inertial navigation module includes the acceleration vector and the rotation angular rate vector of the autonomous vehicle. The detailed process of obtaining the measurement data of the inertial navigation module is described below:
  • inertial sensors are used to measure the 3-axis acceleration and 3-axis angular velocity of the autonomous vehicle in a fixed coordinate system.
  • the inertial measurement unit installed on the vehicle is an accelerometer and a gyroscope.
  • the accelerometer is used to measure the acceleration of the vehicle
  • the gyroscope is used to measure the angular velocity of the vehicle.
  • the measurement sensor has measurement errors such as zero drift. These measurement errors cause the positioning error to increase in the square of time and the attitude angle error to increase in proportion to time. If it is not restricted, the navigation system will quickly lose its ability. Then, error compensation is carried out through modeling. Can reduce deterministic error and random drift error.
  • the navigation system selected for the update of the pose value and the error compensation is the n system, which is usually the northeast sky.
  • the elements are each uncorrelated zero-mean Gaussian white noise process.
  • Is the accelerometer measurement g W is the earth’s gravitational acceleration vector, Is the three-dimensional position of the autonomous vehicle, q ws is the 4 elements representing the attitude angle of the vehicle, Is the speed vector of the vehicle, b a and b g are the acceleration and angular velocity deviation vectors respectively.
  • the time constant ⁇ >0 is used to model the accelerometer bias as a bounded random walk.
  • the matrix ⁇ is estimated by the measurement accompanied by the gyroscope Angular rate composition:
  • the middle three items on the right are the 1, 2, and 3 components of the quaternion. It is the velocity vector of the inertial navigation at time k, the deviation vector of the acceleration and angular velocity of the inertial measurement unit.
  • the usual satellite navigation module outputs the three-dimensional position, speed and local time of the satellite receiving antenna installed on the vehicle.
  • the tight coupling method of the present invention requires the original measurement data of satellite navigation, including pseudorange measurement of visible satellites, Doppler frequency drift measurement and carrier phase measurement, and uses the satellite navigation data stream to assist the inertial measurement and stereo vision system of the present invention, Comprehensively determine the position, attitude angle and environmental three-dimensional information of the autonomous vehicle. The detailed process of obtaining the original measurement data of satellite navigation is described below:
  • the Global Navigation Satellite System uses wireless broadcast signals to provide navigation positioning and time synchronization services for the public. Get the satellite antenna receiver on the autonomous vehicle to receive the signal of the navigation satellite; analyze the ephemeris information of each satellite according to the received satellite signal, and calculate the satellite position and satellite of each satellite based on the ephemeris information speed. At the same time, the satellite shuttle plane uses the radio signal sent by the satellite to calculate the pseudo-range, Doppler frequency drift and carrier phase of the satellite to the local receiver.
  • the satellite navigation receiver uses a single-point positioning method to determine the pseudo-distance.
  • the pseudo-distance is the flight time between a certain satellite and the user's antenna, multiplied by the speed of light.
  • the pseudorange ⁇ (n) (t) is calculated by using the time difference between the time when the n satellite signal is received t u (t) and the transmission time t s (n) (t- ⁇ ) and the vacuum
  • the speed c of the radio wave is multiplied, and the expression is as follows:
  • the symbol ⁇ represents the actual time interval from when the GNSS signal is transmitted to when it is received by the user.
  • the clocks of GNSS satellites and user receivers are usually not synchronized with GNSS time t, the satellite time is ahead of GNSS time.
  • the receiver time is ahead of the GNSS time by ⁇ t u (t), namely:
  • Is the satellite clock error, iono ionospheric delay and d trop tropospheric delay are all known quantities
  • ⁇ (n) represents the unknown pseudorange measurement noise
  • r (n) is the receiver in physical space (x, y, z)
  • the geometric distance to the nth satellite (x (n) , y (n) , z (n) ), the coordinates of each satellite position (x (n) , y (n) , z (n) ) can be It is calculated from the ephemeris broadcast by each satellite.
  • is the carrier wavelength
  • N is the ambiguity of the whole cycle
  • t s is the time when the satellite transmits the signal
  • t r is the time when the receiver receives the signal
  • dt s and dt r are the clock difference between the satellite and the receiver, respectively
  • c is the speed of light
  • d iono ionospheric delay is the time when the satellite transmits the signal
  • d SA is the SA influence
  • is the carrier measurement noise
  • ⁇ (t s, t r ) is t s time satellite and The geometric distance between the receiver antennas at the time t r , which includes the station coordinates, satellite orbit and earth rotation parameters, etc.
  • v (i) is the moving speed of satellite i
  • Is the moving speed of user u Is the unit vector of user u pointing to satellite i
  • c is the speed of light in vacuum
  • ⁇ f ⁇ is the clock drift of user u
  • ⁇ f (i) is the clock drift of satellite i
  • It is the Doppler frequency measurement noise of user u relative to satellite i. It can be seen from equations 3-4, 3-5 and 3-7 that each navigation satellite can provide 2 independent measurement results (pseudo-range and carrier phase). These measurement equations can be used to calibrate the positioning deviation of inertial navigation and stereo vision. .
  • the data input of the three parts of the stereo vision image processing module, the inertial navigation module and the satellite navigation module are tightly coupled.
  • the data includes the measurement data of the inertial navigation module, the image data of the stereo vision module and the original measurement data of the satellite navigation.
  • the original measurement data is combined with the image data of the stereo vision module to correct the drift error of the inertial navigation module, which can assist the inertial navigation module to limit the drift error.
  • Figure 2 is a specific flow chart of the tightly coupled module of the system.
  • the inertial sensor is used to obtain the 3-axis acceleration and 3-axis angular velocity of the vehicle in a fixed coordinate system and input to the inertial navigation module;
  • the binocular camera collects pictures for direct method input;
  • the GPS receiver is used to obtain satellites Raw measurement data.
  • tight coupling When performing tight coupling, tight coupling is expressed by the cost function of weighted reprojection error of stereo vision, satellite navigation error, and time error from inertial navigation.
  • the expression is as follows:
  • e r is the weighted reprojection error of stereo vision
  • e g is the satellite navigation error
  • e s is the time error term of the inertial navigation
  • i is the camera index
  • k is the camera frame index
  • j is the landmark index.
  • the landmark labels visible in the k-th frame and the i-th camera are written as the set J(i,k).
  • e r is the weighted re-projection error of stereo vision
  • the manifestation of the re-projection error of stereo vision is:
  • h i denotes a projection camera model
  • z i, j, k represents the coordinates of the image features.
  • e g is the satellite navigation error
  • the error of each navigation satellite s at each time t contains three parts of error
  • the expression is as follows:
  • e p represents a pseudorange error
  • ed denotes the Doppler error
  • e c represents the carrier phase error
  • the value of the cost function J(x) is minimized by linear or non-linear optimization methods, thereby completing the tight coupling between the stereo vision image processing module, the inertial navigation module and the satellite navigation module, and combining the original measurement data of the satellite navigation
  • the image data of the stereo vision module corrects the drift error of the inertial navigation module.
  • the receiver cannot be positioned.
  • the system error of the inertial navigation module is corrected through the original measurement data of satellite navigation combined with the image data of the stereo vision module, thereby improving the accuracy of navigation and greatly enhancing the navigation Robustness of the system.
  • the environment perception system judges the targets and obstacles in the environment through the deep learning neural network.
  • the central problem of object detection is to segment different objects from the environmental three-dimensional point cloud information, and to determine the nature of the object. This question is handed to semantic segmentation.
  • the three-dimensional environmental information generated by this system is a sparse point cloud composed of points with varying intensity in the image. And under the color camera, each point also carries the intensity information of yellow, green and blue.
  • the deep neural network module that can process three-dimensional point clouds, realize semantic segmentation, and target detection is usually divided into three-dimensional sparse convolutional neural network (3D CNN), point network neural network (PointNet) or a combination of the two.
  • 3D CNN three-dimensional sparse convolutional neural network
  • PointNet point network neural network
  • the three-dimensional sparse convolutional neural network can efficiently process the three-dimensional point cloud under the grid point approximation.
  • it is engineered by multiple convolutional layers and pooling layers.
  • Figure 2 is the convolutional layer of the three-dimensional convolutional neural network, which consists of a 3x3x3 convolution kernel with a step size of 1 to perform convolution operations on the three-dimensional data.
  • Figure 2 shows that for a 3x3x3 three-dimensional convolution kernel, when its input is 9x9x9, it outputs a new 7x7x7 three-dimensional data. In more cases, filling the input data edge with 0 can make the input and output have the same scale.
  • a complete three-dimensional convolutional neural network is composed of multiple three-dimensional convolutional layers and multiple three-dimensional pooling layers ( Figure 3). It has multiple convolution kernels, which produce different features. Then the maximum pooling layer takes the maximum value of the pooling unit as the output of the unit. If 2x2x2 is used as the size of the pooling unit and the step size is 2, the three-dimensional data after pooling will be reduced by half in each direction. After the multi-level convolutional layer and the pooling layer, the data passes through the multi-layer fully connected layer, and finally the result of the classification judgment is output from the output layer, usually a normalized exponential function (Softmax).
  • Softmax normalized exponential function
  • the three-dimensional convolutional neural network can efficiently process the three-dimensional point cloud data after gridding, the gridding loses part of the subtle position information, causing errors.
  • the PointNet method does not require gridded 3D data. It can directly use 3D point clouds as input to achieve semantic segmentation of 3D environmental point clouds.
  • the basic unit of the dot network is called Set Abstraction. It is composed of sampling combined layer and dot network layer.
  • the sampling combination layer uses the iterative farthest point sampling method to divide the points in the three-dimensional space into a combination of points near the centroid region, and then use the multi-layer perceptron network to extract the point cloud features.
  • the dot network is generally composed of multiple layers of combined abstract units to form a multi-level combined abstract function, and finally a fully connected layer and an output layer are used.
  • the output layer uses the normalized index (Softmax) fully connected to output the classification results ( Figure 4).
  • the tightly coupled autonomous driving perception system can help provide high-precision measurements of the environment and build maps without resorting to expensive laser scanning radar, reducing the cost of autonomous vehicles.
  • FIG. 3 is a schematic structural diagram of the automatic driving object detection and positioning system 100 according to an embodiment of the present invention, as shown in FIG. 5.
  • the automatic driving object detection and positioning system 100 includes the stereo vision image processing module 10 of the binocular direct method, the satellite navigation module 20, the inertial navigation module 30, the system tight coupling module 40 and the deep neural network module 50, which can judge the environment in the environment. Goals and obstacles.
  • the inertial navigation module 30 is used to obtain the measurement data of the inertial navigation module by using an inertial sensor; the details are as above.
  • the stereo vision image processing module 10 adopts the binocular direct method to obtain the image data of the stereo vision module; the details are as above.
  • the satellite navigation module 20 is used to obtain the original satellite navigation measurement data of the navigation satellite through the receiver; the details are as above.
  • the system tight coupling module 40 is used to perform tight coupling processing on the measurement data of the inertial navigation module, the image data of the stereo vision module, and the original measurement data of the satellite navigation; the details are as above.
  • the connection relationship between the various modules is: the stereo vision image processing module 10, the satellite navigation module 20, and the inertial navigation module 30 are all connected to the system tight coupling module 40.
  • the stereo vision image processing module 10 includes a binocular camera, and the binocular camera is installed on an autonomous vehicle; the details are as described above.
  • the inertial navigation module 30 includes inertial sensors, and the inertial sensors are fixed on the autonomous vehicle; the details are as described above.
  • the deep neural network module uses a three-dimensional convolutional neural network or a three-dimensional dot network to realize object detection and semantic segmentation. The details are as described above.
  • the invention tightly couples the measurement data of the inertial navigation module, the image data of the stereo vision module, and the original measurement data of the satellite navigation to correct the error of the measurement data of the inertial navigation module, thereby improving the positioning accuracy without resorting to expensive Laser scanning radar, thereby reducing the cost of self-driving cars.

Abstract

A system and method for detecting and positioning an autonomous driving object. The method comprises the following steps: obtaining the measurement data of an inertial navigation module of an autonomous driving vehicle, the image data of a stereoscopic vision module, satellite navigation original measurement data, and a deep neural network module detection and determination software module; tightly coupling the image data of the stereoscopic vision module, the satellite navigation original measurement data, and the measurement data of the inertial navigation module, so as to limit the increase of a drift error of the inertial navigation module, and ensure the precision of positioning; and establishing three-dimensional environment information by using a satellite-inertial navigation assisted stereoscopic vision method, and then using a deep neural network module to determine and classify targets and obstacles in an environment. The method has the advantages of maximally improving the positioning precision of an autonomous driving sensing system, and improving the calculation efficiency and reliability.

Description

一种自动驾驶对象探测和定位系统及方法Automatic driving object detection and positioning system and method 技术领域Technical field
本发明涉及计算机应用技术领域,特别涉及一种自动驾驶对象探测和定位系统及方法。The present invention relates to the field of computer application technology, in particular to an automatic driving object detection and positioning system and method.
背景技术Background technique
近年来,随着人们对汽车安全意识的增强以尽信息技术的发展,自动驾驶领域越来越受关注,世界上许多公司和科研机构都开始投入研发自动驾驶相关产品,预计2021年自动驾驶车辆将进入市场,给汽车行业带来巨大的变革。相关研究表明自动驾驶技术的发展将会在多种领域带来颠覆性的发展,例如其发展可以增强公路的交通安全、缓解交通拥堵状况和减少环境污染等方面,同时自动驾驶技术也是衡量一个国家科研实力和工业水平的一个重要标志,在国防和国民经济领域具有广阔的应用前景。In recent years, as people’s awareness of car safety has increased and the development of information technology has made the field of autonomous driving more and more concerned, many companies and scientific research institutions in the world have begun to invest in research and development of autonomous driving related products. It is expected that autonomous vehicles will be driven in 2021. Will enter the market and bring huge changes to the automotive industry. Relevant research shows that the development of autonomous driving technology will bring disruptive development in many fields. For example, its development can enhance road traffic safety, alleviate traffic congestion and reduce environmental pollution. At the same time, autonomous driving technology is also a measure of a country. An important symbol of scientific research strength and industrial level, it has broad application prospects in the fields of national defense and national economy.
自动驾驶是指汽车通过车载传感系统来对道路环境进行感知,并根据感知所获得的道路、车辆位置和障碍物信息等控制车辆的转向和速度,进而自动规划行车路线并控制车辆到达预定目标的技术。Autonomous driving means that the car perceives the road environment through the on-board sensor system, and controls the steering and speed of the vehicle based on the road, vehicle position and obstacle information obtained by the perception, and then automatically plans the driving route and controls the vehicle to reach the predetermined target Technology.
如今在自动驾驶方面,各大公司都有自己的技术方向,现有技术已有双目直接方法的视觉系统和惯性导航模块的组合系统,但是组合系统中视觉系统和惯性导航模块产生的误差不能有效限制,组合系统在长时间无图像梯度情况下误差会无限制增长,导致组合系统感知失败。Nowadays, in terms of autonomous driving, major companies have their own technical direction. The existing technology has a combination system of binocular direct method vision system and inertial navigation module, but the error generated by the vision system and inertial navigation module in the combined system cannot be Effective restriction, the error of the combined system will increase indefinitely when there is no image gradient for a long time, resulting in the failure of the combined system's perception.
现有技术也有单目特征点方法的视觉系统、惯性导航模块及卫星导航的紧耦合自动驾驶感知系统,但是单目摄像头无法探测到无 特征障碍物,比如高速路的隔离护栏、自行车或动物等。而现有视觉系统也有采用双目立体视觉系统进行耦合,但是仍然采用特征点法,计算量大,且对硬件性能要求高。目前最先进的双目立体视觉环境检测方法,也只使用了双目视觉中的视觉差信息,并没有使用摄像头在不同时间、位置采集到的图像来实现环境的三维建模。The prior art also has monocular feature point method vision systems, inertial navigation modules, and satellite navigation close-coupled automatic driving perception systems, but monocular cameras cannot detect uncharacteristic obstacles, such as isolated guardrails on highways, bicycles or animals, etc. . The existing vision system also uses a binocular stereo vision system for coupling, but still uses the feature point method, which requires a large amount of calculation and high hardware performance requirements. At present, the most advanced binocular stereo vision environment detection method only uses the visual difference information in binocular vision, and does not use the images collected by the camera at different times and positions to realize the three-dimensional modeling of the environment.
发明内容Summary of the invention
本发明的目的在于提供一种自动驾驶对象探测和定位系统及方法,具有最大限度地提高自动驾驶感知系统的定位精度、提高计算效率和可靠性的优点,以解决上述背景技术中提出的问题。The purpose of the present invention is to provide an automatic driving object detection and positioning system and method, which has the advantages of maximizing the positioning accuracy of the automatic driving perception system, and improving the calculation efficiency and reliability, so as to solve the problems raised in the background art.
为实现上述目的,本发明提供如下技术方案:In order to achieve the above objectives, the present invention provides the following technical solutions:
一种自动驾驶对象探测和定位系统,包括立体视觉图像处理模块、卫星导航模块、惯性导航模块及系统紧耦合模块,其中:An automatic driving object detection and positioning system, including a stereo vision image processing module, a satellite navigation module, an inertial navigation module and a system tight coupling module, in which:
立体视觉图像处理模块,采用或者多目摄像头获取立体视觉模块的图像数据;Stereo vision image processing module, using or multi-eye camera to obtain the image data of the stereo vision module;
卫星导航模块,用于通过接收机获取卫星导航原始测量数据;The satellite navigation module is used to obtain the original measurement data of satellite navigation through the receiver;
惯性导航模块,用于采用惯性传感器获取惯性导航模块的测量数据;The inertial navigation module is used to obtain the measurement data of the inertial navigation module by using inertial sensors;
系统紧耦合模块,用于将所述惯性导航模块测量数据、所述立体视觉模块的图像数据及所述卫星导航原始测量数据进行紧耦合处理,并建立三维环境信息,最后利用深度神经网络模块检测环境中的目标和障碍物;The system tight coupling module is used to tightly couple the measurement data of the inertial navigation module, the image data of the stereo vision module and the original measurement data of the satellite navigation, and establish three-dimensional environment information, and finally use the deep neural network module to detect Targets and obstacles in the environment;
所述深度神经网络由三维稀疏卷积神经网络、点网神经网络和二者的组合构成;The deep neural network is composed of a three-dimensional sparse convolutional neural network, a dot network neural network, and a combination of the two;
所述立体视觉图像处理模块、所述卫星导航模块及所述惯性导航模块均与所述系统紧耦合模块连接。The stereo vision image processing module, the satellite navigation module and the inertial navigation module are all connected to the system tight coupling module.
进一步地,所述立体视觉图像处理模块包含双目或者多目摄像头。Further, the stereo vision image processing module includes a binocular or multi-eye camera.
本发明要解决的另一种技术方案:一种自动驾驶对象探测和定位的方法,包括以下步骤:Another technical solution to be solved by the present invention: a method for automatic driving object detection and positioning, including the following steps:
S1:获取立体视觉模块的图像数据、惯性导航模块的测量数据及卫星导航原始测量数据;S1: Obtain the image data of the stereo vision module, the measurement data of the inertial navigation module and the original measurement data of the satellite navigation;
S2:将立体视觉模块的图像数据、卫星导航原始测量数据及所述惯性导航模块的测量数据进行紧耦合,对惯性导航模块的漂移误差进行修正;S2: tightly couple the image data of the stereo vision module, the original measurement data of the satellite navigation and the measurement data of the inertial navigation module to correct the drift error of the inertial navigation module;
S3:将立体视觉模块的图像数据、卫星导航原始测量数据及所述惯性导航模块的测量数据进行紧耦合,建立三维环境信息,再使用深度神经网络模块判断和分类环境中的目标和障碍物;S3: Tightly couple the image data of the stereo vision module, the original measurement data of the satellite navigation and the measurement data of the inertial navigation module to establish three-dimensional environment information, and then use the deep neural network module to judge and classify the targets and obstacles in the environment;
S4:深度神经网络使用三维稀疏卷积神经网络、点网神经网络或者是二者的组合。S4: The deep neural network uses a three-dimensional sparse convolutional neural network, a dot network neural network, or a combination of the two.
进一步地,所述立体视觉模块采用多目,包括双目摄像头直接方法进行处理,所述图像数据包括在同一时刻多目或者双目立体视觉摄像头之间的视觉差和每一个摄像头在不同时间、位置拍摄到的图像信息来建立环境的三维信息。Further, the stereo vision module adopts multi-camera, including binocular camera direct method for processing, and the image data includes the visual difference between multi-camera or binocular stereo vision cameras at the same time and each camera at different time, The image information captured by the location is used to establish the three-dimensional information of the environment.
进一步地,所述紧耦合由立体视觉的加权重投影误差、卫星导航误差和来自惯性导航的状态误差构成成本函数。Further, the tight coupling consists of a weighted reprojection error of stereo vision, satellite navigation error, and state error from inertial navigation to form a cost function.
进一步地,所述S1具体包括如下步骤:Further, the S1 specifically includes the following steps:
S11:采用多目或者是双目摄像头获取立体视觉模块的图像数据、图像数据包括在同一时刻多目或者双目立体视觉摄像头之间的视觉差和每一个摄像头在不同时间、位置拍摄到的图像信息;S11: Use multi-eye or binocular cameras to obtain image data of the stereo vision module. The image data includes the visual difference between the multi-eye or binocular stereo vision cameras at the same time and the images captured by each camera at different times and positions information;
S12:采用惯性传感器获取惯性导航模块的测量数据;S12: Use inertial sensors to obtain measurement data of the inertial navigation module;
S13:通过接收机获取卫星导航原始测量数据;S13: Obtain the original measurement data of satellite navigation through the receiver;
S14:将所述惯性导航模块测量数据、所述图像数据及所述卫星导航原始测量数据进行紧耦合处理。S14: Perform tight coupling processing on the measurement data of the inertial navigation module, the image data, and the original measurement data of the satellite navigation.
进一步地,所述S11具体包括如下步骤:Further, the S11 specifically includes the following steps:
S111:用多目或者双目摄像头采集环境图像信号;S111: Use a multi-eye or binocular camera to collect environmental image signals;
S112:组合多个不同摄像头在同一个时刻拍摄到的视觉差和每一个摄像头在不同时间、位置拍摄到的图像,形成立体视觉观测的直接方法;S112: Combine the visual difference captured by multiple different cameras at the same time and the images captured by each camera at different times and positions to form a direct method of stereo vision observation;
S113:并组合惯性导航、卫星导航的测量数据,建立三维环境信息,最后使用深度神经网络模块分类和判断环境中的目标和障碍物。S113: Combine the measurement data of inertial navigation and satellite navigation to establish three-dimensional environment information, and finally use the deep neural network module to classify and judge the targets and obstacles in the environment.
进一步地,所述S12具体包括如下步骤:Further, the S12 specifically includes the following steps:
S121:利用惯性传感器测量自动驾驶车辆在固定坐标系下的3轴加速度和3轴角速度;S121: Use inertial sensors to measure the 3-axis acceleration and 3-axis angular velocity of the autonomous vehicle in a fixed coordinate system;
S122:将所述加速度和所述角速度转动到导航坐标系下,求解惯性导航机械编排方程并计算出自动驾驶车辆的位置和姿态角;S122: Rotate the acceleration and the angular velocity to the navigation coordinate system, solve the inertial navigation mechanical arrangement equation, and calculate the position and attitude angle of the autonomous vehicle;
S123:图像数据、惯性导航数据和卫星导航数据组合,建立三维环境信息,再利用深度神经网络模块,分类和判断环境中的目标和障碍物。S123: Combine image data, inertial navigation data and satellite navigation data to establish three-dimensional environment information, and then use deep neural network modules to classify and judge objects and obstacles in the environment.
进一步地,所述S14具体包括:通过所述卫星导航原始测量数据结合所述立体视觉模块的图像数据对所述惯性导航模块的漂移误差进行校正。Further, the S14 specifically includes: correcting the drift error of the inertial navigation module by combining the original measurement data of the satellite navigation with the image data of the stereo vision module.
与现有技术相比,本发明的有益效果是:本发明将惯性导航模块测量数据、立体视觉模块的图像数据及卫星导航原始测量数据三者进行紧耦合,对惯性导航模块测量数据的误差进行修正,从而提高了定 位精度,再利用立体视觉直接方法实现的三维环境建模数据和深度神经网络,识别和判断环境中的物体。被系统借助高精度的三维稀疏卷积神经网络、点网神经网络或者是它们的组合,提高障碍物判断的准确度,不再借助于昂贵的激光扫描雷达,从而降低自动驾驶汽车的成本。Compared with the prior art, the beneficial effect of the present invention is: the present invention tightly couples the measurement data of the inertial navigation module, the image data of the stereo vision module, and the original measurement data of the satellite navigation, and measures the error of the measurement data of the inertial navigation module. Correction, thereby improving the positioning accuracy, and then using the three-dimensional environment modeling data and deep neural network realized by the direct method of stereo vision to identify and judge objects in the environment. The system uses high-precision three-dimensional sparse convolutional neural networks, dot-net neural networks, or their combination to improve the accuracy of obstacle judgment, and no longer resorts to expensive laser scanning radar, thereby reducing the cost of autonomous vehicles.
附图说明Description of the drawings
图1为本发明实施例自动驾驶定位和对象探测方法的流程图;FIG. 1 is a flowchart of an automatic driving positioning and object detection method according to an embodiment of the present invention;
图2为本发明实施例自动驾驶对象探测和定位系统的三维稀疏卷积神经网络的结构图;2 is a structural diagram of a three-dimensional sparse convolutional neural network of an automatic driving object detection and positioning system according to an embodiment of the present invention;
图3为本发明实施例自动驾驶对象探测和定位系统的点网神经网络的结构图;FIG. 3 is a structural diagram of a point network neural network of an automatic driving object detection and positioning system according to an embodiment of the present invention;
[根据细则91更正 27.10.2020] 
图4为本发明实施例自动驾驶对象探测和定位系统的结构示意图。
图5为本发明自动驾驶对象探测和定位系统的架构图。
[Corrected according to Rule 91 27.10.2020]
Fig. 4 is a schematic structural diagram of an automatic driving object detection and positioning system according to an embodiment of the present invention.
Fig. 5 is an architecture diagram of the automatic driving object detection and positioning system of the present invention.
具体实施方式detailed description
下面将结合本发明实施例中的附图,对本发明实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例仅仅是本发明一部分实施例,而不是全部的实施例。基于本发明中的实施例,本领域普通技术人员在没有做出创造性劳动前提下所获得的所有其他实施例,都属于本发明保护的范围。The technical solutions in the embodiments of the present invention will be clearly and completely described below in conjunction with the accompanying drawings in the embodiments of the present invention. Obviously, the described embodiments are only a part of the embodiments of the present invention, rather than all the embodiments. Based on the embodiments of the present invention, all other embodiments obtained by those of ordinary skill in the art without creative work shall fall within the protection scope of the present invention.
一种基于立体视觉模块、惯性导航模块及卫星导航模块紧耦合的自动驾驶对象探测和定位系统及方法,惯性导航可连续提供信息,短时间精度高,但是定位误差会随时间积累;卫星导航长期稳定性好,但易受到干扰,数据更新频率低;通常的立体视觉使用特征点方法,首先在图像中挑选多个特征点,再利用左右眼拍到的这些特征点的视觉差,进行匹配。最后使用三角方法确定这些特征点的到照相机的距离信息,从而首先三维环境探测和定位。基于特征点的立体视 觉方法需要在每一帧图像中挑选很多特征点,花费大量的计算资源,而且在很多情况下,图像中并没有合适的特征点。比如图像强度只有一个固定梯度的情况下,特征点方法就会完全失败。基于特征点的立体视觉方法,也没有利用摄像头在不同位置拍摄到的图像来实现距离测量,浪费了宝贵的测量数据,导致环境建模精度误差和定位误差。An automatic driving object detection and positioning system and method based on the tight coupling of a stereo vision module, an inertial navigation module and a satellite navigation module. Inertial navigation can continuously provide information with high accuracy in a short time, but positioning errors will accumulate over time; satellite navigation will be long-term Stability is good, but it is susceptible to interference, and the data update frequency is low; the usual stereo vision uses the feature point method, first select multiple feature points in the image, and then use the visual difference of these feature points captured by the left and right eyes to match. Finally, the triangulation method is used to determine the distance information of these feature points to the camera, so as to firstly detect and locate the three-dimensional environment. The stereo vision method based on feature points needs to select many feature points in each frame of image, which consumes a lot of computing resources, and in many cases, there are no suitable feature points in the image. For example, when the image intensity has only one fixed gradient, the feature point method will fail completely. The stereo vision method based on feature points does not use the images captured by the camera at different positions to achieve distance measurement, which wastes valuable measurement data and leads to environmental modeling accuracy errors and positioning errors.
更好的算法是使用立体视觉直接方法,它不需要挑选特征点,节约计算资源,而且只要有梯度的图像就可以实现环境建模和定位。再加上它不仅利用不同摄像头在同一时刻拍摄到的图像数据,它还利用摄像头在不同时间、位置拍摄到的图像信息进行三维建模和定位,立体视觉根据不同照相机的视差,计算障碍物到照相机的距离,但视觉系统在图像缺乏梯度的环境下,不能有效定位和探测障碍物距离。将立体视觉直接方法、惯性导航及卫星导航构成组合导航系统,从而可相互协助感知车辆的状态和环境状态,在不同环境下互补,提高了可靠性和导航精度,能精确提取距离,可以替代现有的激光扫描雷达,降低成本。A better algorithm is to use the direct method of stereo vision, it does not need to select feature points, saving computing resources, and as long as there are gradient images, environment modeling and positioning can be achieved. In addition, it not only uses the image data captured by different cameras at the same time, but also uses the image information captured by the cameras at different times and positions for three-dimensional modeling and positioning. The stereo vision calculates the obstacles to the obstacles according to the parallax of different cameras. The distance of the camera, but the visual system cannot effectively locate and detect the distance of obstacles in an environment where the image lacks gradient. The stereo vision direct method, inertial navigation and satellite navigation form a combined navigation system, which can assist each other in perceiving the state of the vehicle and the environment, complement each other in different environments, improve the reliability and navigation accuracy, can accurately extract the distance, and can replace the current situation. Some laser scanning radars reduce costs.
图1为本发明实施例紧耦合的自动驾驶感知方法的流程图,如图1所示,该紧耦合的自动驾驶感知方法具体包括:Fig. 1 is a flowchart of a tightly coupled automatic driving perception method according to an embodiment of the present invention. As shown in Fig. 1, the tightly coupled automatic driving perception method specifically includes:
S1、获取自动驾驶车辆的惯性导航模块测量数据、立体视觉模块的图像数据及卫星导航原始测量数据;S2、在紧耦合过程中,将立体视觉模块的图像数据及卫星导航原始测量数据与惯性导航模块测量数据进行紧耦合,对惯性导航模块测量数据的误差进行修正,从而感知车辆的状态和环境的状态。下面对具体的自动驾驶感知方法的过程进行详细描述:S1. Obtain the measurement data of the inertial navigation module of the autonomous vehicle, the image data of the stereo vision module and the original measurement data of the satellite navigation; S2, in the tight coupling process, combine the image data of the stereo vision module and the original measurement data of the satellite navigation with the inertial navigation The module measurement data is tightly coupled to correct the error of the inertial navigation module measurement data, so as to perceive the state of the vehicle and the state of the environment. The process of the specific autonomous driving perception method is described in detail below:
(1)立体视觉模块的图像数据通过双目直接方法的立体视觉图像处理模块获取,下面对采用双目摄像头获取立体视觉模块的图像数据的详细过程进行描述:(1) The image data of the stereo vision module is obtained by the stereo vision image processing module of the binocular direct method. The detailed process of using the binocular camera to obtain the image data of the stereo vision module is described below:
直接方法是基于灰度不变假设的,同一个空间点的像素灰度,在各个图像中是固定不变的,直接方法不需要提取特征,同时在缺乏角点和边缘或光线变化不明显的环境,也有较好效果;且直接方法需处理的数据较少,可实现高计算效率。The direct method is based on the assumption that the gray level is invariable. The pixel gray level of the same spatial point is fixed in each image. The direct method does not need to extract features. At the same time, it lacks corners and edges or the light changes are not obvious. The environment also has better results; and the direct method requires less data to be processed, which can achieve high computational efficiency.
采用双目直接方法处理摄像头立体视觉拍摄到的图像具体过程如下:The specific process of using the binocular direct method to process the images captured by the camera stereo vision is as follows:
1.利用双目摄像头从不同位置获取被测物体的两幅或多幅图像的灰度图像(彩色的图可以使用红、绿、兰的强度,分别使用本方法),记录双目摄像头在t和t+1时刻获得的图像为I i和I j1. Use the binocular camera to obtain the grayscale image of two or more images of the measured object from different positions (the color image can use the intensity of red, green, and blue, respectively use this method), and record the binocular camera at t The images obtained at time and t+1 are I i and I j .
2.通过matlab或者opencv进行相机标定,获得相机的内参。2. Carry out camera calibration through matlab or opencv to obtain the internal parameters of the camera.
3.根据相机内参对获得的图像进行畸变矫正处理。3. Perform distortion correction processing on the acquired image according to the camera's internal parameters.
4.将上述的图像数据输入到系统紧耦合模块中,根据拍摄图像的计算出光度测量能量,再计算误差能量函数,通过最小化误差函数,实现照相机本体的定位和环境三维感知。具体的,图像Ii中的空间点Pi出现在另一帧图像Ij中,其中Ii和Ij为同一个照相机在不同位置和姿态下拍摄到的图像,选取图像Ii和Ij中出现的一个空间点Pi,根据同一空间三维点在各个视角下测到的灰度值不变假设原理,图像Ii中的点p在图像Ij的投影p’由下式计算:4. Input the above-mentioned image data into the tightly coupled module of the system, calculate the photometric measurement energy according to the captured image, and then calculate the error energy function. By minimizing the error function, the positioning of the camera body and the three-dimensional perception of the environment are realized. Specifically, the spatial point Pi in the image Ii appears in another frame of image Ij, where Ii and Ij are images captured by the same camera at different positions and attitudes, and a spatial point Pi appearing in the images Ii and Ij is selected. , According to the principle of the assumption that the gray values of the three-dimensional points in the same space are not changed under various viewing angles, the projection p'of the point p in the image Ii on the image Ij is calculated by the following formula:
Figure PCTCN2020102996-appb-000001
Figure PCTCN2020102996-appb-000001
其中,Π k
Figure PCTCN2020102996-appb-000002
是相机图像帧点的投影和反投影函数,d p是p点的逆深度,T ji是图像帧之间的转换关系:
Where Π k and
Figure PCTCN2020102996-appb-000002
Is the projection and back projection function of the camera image frame point, d p is the inverse depth of p point, and T ji is the conversion relationship between image frames:
Figure PCTCN2020102996-appb-000003
Figure PCTCN2020102996-appb-000003
其中,R ji是一个3*3的旋转矩阵,t是平移向量。 Among them, R ji is a 3*3 rotation matrix, and t is the translation vector.
其光度误差能量函数为:The luminosity error energy function is:
Figure PCTCN2020102996-appb-000004
Figure PCTCN2020102996-appb-000004
其中,
Figure PCTCN2020102996-appb-000005
是Huber范数,是为了防止误差能量函数随光度误差增长过快;ω p是权值,为了将图像中梯度大的像素点对应的权值缩小,这样最终的能量函数能够反映大多数像素点的光度误差,而不是个别梯度大的像素点的误差;I i[p]和I j[p']分别为 Pi在图像对应点的灰度值。
in,
Figure PCTCN2020102996-appb-000005
It is the Huber norm to prevent the error energy function from increasing too fast with the luminosity error; ω p is the weight value, in order to reduce the weight corresponding to the pixels with large gradients in the image, so that the final energy function can reflect most of the pixels The luminosity error is not the error of individual pixels with large gradients; I i [p] and I j [p'] are respectively the gray value of Pi at the corresponding point of the image.
通过最小化光度误差能量进行优化从而得到相机位置和姿态角,当光度误差能量为最小值,求解得到立体视觉模块相机位姿和环境信息的图像数据。得到目标函数为:The camera position and attitude angle are obtained by optimization by minimizing the photometric error energy. When the photometric error energy is at the minimum, the image data of the camera pose and environmental information of the stereo vision module can be obtained. The objective function is:
Figure PCTCN2020102996-appb-000006
Figure PCTCN2020102996-appb-000006
上述的公式(1-3,1-4)是单目摄像头的光度误差能量函数。当摄像头为双目或者多目摄像头时,直接方法的立体视觉模块,组合图像数据包括双目摄像头同时采集到的视觉静态差数据和每一个摄像头在不同时间、位置采集到的动态时间序列数据。双目直接法通过引入耦合因子λ来平衡运动中相机每个图像帧和静态立体视觉的相对权重,其误差能量函数为:The above formulas (1-3, 1-4) are the photometric error energy function of the monocular camera. When the camera is a binocular or multi-camera, the direct method is a stereo vision module. The combined image data includes the visual static difference data collected by the binocular cameras at the same time and the dynamic time series data collected by each camera at different times and positions. The binocular direct method balances the relative weight of each image frame of the camera and the static stereo vision by introducing a coupling factor λ, and its error energy function is:
Figure PCTCN2020102996-appb-000007
Figure PCTCN2020102996-appb-000007
其中,obs(p)表示所有可以观察到p点的图像帧,
Figure PCTCN2020102996-appb-000008
是静态立体视觉中的误差能量函数,i属于为不同图像帧,p属于同一帧的多个点,j为可以观察到p点的所有帧。则根据公式(1-4)及公式(1-5)可以得出新的双目摄像头的目标函数。上述两个图像帧之间的转换中出现的R ji和t,即是相机的位姿,通过采用梯度下降法或者高斯牛顿法求解目标函数,从而得到相机的位姿以及和空间点Pi的三维坐标。进而通过该相机位姿对惯性导航中自动驾驶车辆的惯性导航模块测量数据的误差进行修正,该双目直接方法获取的相机位姿不需要提取特征,对道路和环境变化反应能力显著提高。很多空间点Pi的三维坐标的集合就成为环境的三维点云。
Among them, obs(p) represents all image frames where p points can be observed,
Figure PCTCN2020102996-appb-000008
Is the error energy function in static stereo vision, i belongs to different image frames, p belongs to multiple points in the same frame, and j is all frames where p points can be observed. According to formula (1-4) and formula (1-5), the objective function of the new binocular camera can be obtained. R ji and t appearing in the conversion between the above two image frames are the pose of the camera. By using the gradient descent method or Gauss Newton method to solve the objective function, the pose of the camera and the three-dimensional space point Pi are obtained. coordinate. Furthermore, the camera pose is used to correct the errors in the measurement data of the inertial navigation module of the autonomous vehicle in the inertial navigation. The camera pose acquired by the binocular direct method does not require feature extraction, and the ability to respond to road and environmental changes is significantly improved. The collection of the three-dimensional coordinates of many spatial points Pi becomes the three-dimensional point cloud of the environment.
(2)惯性导航模块测量数据通过惯性测量单元获取,惯性导航模块测量数据包括自动驾驶车辆的加速度矢量及转动角速率矢量,下面对获取惯性导航模块测量数据的详细过程进行描述:(2) The measurement data of the inertial navigation module is obtained through the inertial measurement unit. The measurement data of the inertial navigation module includes the acceleration vector and the rotation angular rate vector of the autonomous vehicle. The detailed process of obtaining the measurement data of the inertial navigation module is described below:
首先,利用惯性传感器测量自动驾驶车辆在固定坐标系下的3轴加速度和3轴角速度。安装在车辆上的惯性测量单元为为加速度计和陀螺仪,其中,加速度计用于测量车辆加速度,陀螺仪用于测量车辆角速度。测量传感器存在零飘等测量误差,这些测量误差使得定位误差以时间平方增长、姿态角误差以时间成比例增长,如果不加以限制,导航系统很快失去能力,则,通过建模进行误差补偿,可减小确定性误差和随机漂移误差。First, inertial sensors are used to measure the 3-axis acceleration and 3-axis angular velocity of the autonomous vehicle in a fixed coordinate system. The inertial measurement unit installed on the vehicle is an accelerometer and a gyroscope. The accelerometer is used to measure the acceleration of the vehicle, and the gyroscope is used to measure the angular velocity of the vehicle. The measurement sensor has measurement errors such as zero drift. These measurement errors cause the positioning error to increase in the square of time and the attitude angle error to increase in proportion to time. If it is not restricted, the navigation system will quickly lose its ability. Then, error compensation is carried out through modeling. Can reduce deterministic error and random drift error.
将加速度和角速度转动到导航坐标系下,求解惯性导航机械编排方程并计算出自动驾驶车辆的位置和姿态角。具体的,进行位姿数值更新和误差补偿时所选用的导航系均为n系,通常是东北天。Rotate the acceleration and angular velocity to the navigation coordinate system, solve the inertial navigation mechanical arrangement equation and calculate the position and attitude angle of the autonomous vehicle. Specifically, the navigation system selected for the update of the pose value and the error compensation is the n system, which is usually the northeast sky.
在东北天坐标系下,车辆的位置更新公式如下:In the northeast sky coordinate system, the vehicle's position update formula is as follows:
Figure PCTCN2020102996-appb-000009
Figure PCTCN2020102996-appb-000009
Figure PCTCN2020102996-appb-000010
Figure PCTCN2020102996-appb-000010
Figure PCTCN2020102996-appb-000011
Figure PCTCN2020102996-appb-000011
在上述公式中,
Figure PCTCN2020102996-appb-000012
为载体速度在n系中的投影;λ、L和h分别表示载体的经度、纬度和高度;a表示WGS-84坐标系下基本大地参数-椭球体的长半轴长,e为椭球体的偏心率。最终得到关于λ、L、h的微分方程,根据求解可以得到λ、L、h的值,从而可以计算出自动驾驶车辆的位置,R N为卯酉面曲率,R M为子午圈曲率半径。
In the above formula,
Figure PCTCN2020102996-appb-000012
Is the projection of the carrier velocity in the n system; λ, L and h represent the longitude, latitude and height of the carrier respectively; a represents the basic geodetic parameter under the WGS-84 coordinate system-the length of the semi-major axis of the ellipsoid, and e is the length of the ellipsoid Eccentricity. Finally obtained differential equation λ, L, h, and can be solved in accordance with λ, L value, h, which can calculate the position of the automatic driving of the vehicle, R N is prime vertical curvature, R M is the radius of curvature of the meridian.
在东北天坐标系下,车辆的速度更新公式如下:In the northeast sky coordinate system, the vehicle speed update formula is as follows:
Figure PCTCN2020102996-appb-000013
Figure PCTCN2020102996-appb-000013
Figure PCTCN2020102996-appb-000014
Figure PCTCN2020102996-appb-000014
Figure PCTCN2020102996-appb-000015
Figure PCTCN2020102996-appb-000015
在上述公式中,
Figure PCTCN2020102996-appb-000016
为四元数方向余弦矩阵
Figure PCTCN2020102996-appb-000017
的转置;v n为载体的速度在n系中的投影;
Figure PCTCN2020102996-appb-000018
为通过载体的相对速度计算得到的位移角速率;
Figure PCTCN2020102996-appb-000019
表示的是地球的自转角速率在n下的投影;g n为 当地的重力加速度在n上的投影;f b为加速度计的测量输出值。符号“×”代表向量叉乘。最终得到关于v n
Figure PCTCN2020102996-appb-000020
的微分方程,根据求解可以得到v n
Figure PCTCN2020102996-appb-000021
的值,从而可以计算出自动驾驶车辆的速度及位移角速率。
In the above formula,
Figure PCTCN2020102996-appb-000016
Is the quaternion direction cosine matrix
Figure PCTCN2020102996-appb-000017
The transpose of; v n is the projection of the velocity of the carrier in the n system;
Figure PCTCN2020102996-appb-000018
Is the displacement angular rate calculated from the relative velocity of the carrier;
Figure PCTCN2020102996-appb-000019
It represents the projection of the earth's rotation angular rate under n; g n is the projection of the local gravitational acceleration on n; f b is the measured output value of the accelerometer. The symbol "×" represents a vector cross product. Finally get about v n
Figure PCTCN2020102996-appb-000020
The differential equation of, according to the solution, v n,
Figure PCTCN2020102996-appb-000021
Therefore, the speed and angular rate of displacement of the autonomous vehicle can be calculated.
再者,将惯性导航运动学与简单的动态偏差模型结合起来,得到如下方程组:Furthermore, combining the inertial navigation kinematics with a simple dynamic deviation model, the following equations are obtained:
Figure PCTCN2020102996-appb-000022
Figure PCTCN2020102996-appb-000022
Figure PCTCN2020102996-appb-000023
Figure PCTCN2020102996-appb-000023
Figure PCTCN2020102996-appb-000024
Figure PCTCN2020102996-appb-000024
Figure PCTCN2020102996-appb-000025
Figure PCTCN2020102996-appb-000025
Figure PCTCN2020102996-appb-000026
Figure PCTCN2020102996-appb-000026
其中,
Figure PCTCN2020102996-appb-000027
的元素是每个不相关的零均值高斯白噪声过程。
Figure PCTCN2020102996-appb-000028
是加速度计测量,g W是地球的重力加速度矢量,
Figure PCTCN2020102996-appb-000029
是自动驾驶车辆三维位置,q ws是代表车辆姿态角的4元素,
Figure PCTCN2020102996-appb-000030
是车辆的速度矢量,b a,b g分别是加速度和角速度偏差矢量。与作为随机游走建模的陀螺仪偏差相比,使用时间常数τ>0来将加速度计偏差建模为有界随机游走。矩阵Ω由估计的伴随着陀螺仪测量
Figure PCTCN2020102996-appb-000031
的角速率
Figure PCTCN2020102996-appb-000032
组成:
in,
Figure PCTCN2020102996-appb-000027
The elements are each uncorrelated zero-mean Gaussian white noise process.
Figure PCTCN2020102996-appb-000028
Is the accelerometer measurement, g W is the earth’s gravitational acceleration vector,
Figure PCTCN2020102996-appb-000029
Is the three-dimensional position of the autonomous vehicle, q ws is the 4 elements representing the attitude angle of the vehicle,
Figure PCTCN2020102996-appb-000030
Is the speed vector of the vehicle, b a and b g are the acceleration and angular velocity deviation vectors respectively. Compared with the gyroscope bias modeled as a random walk, the time constant τ>0 is used to model the accelerometer bias as a bounded random walk. The matrix Ω is estimated by the measurement accompanied by the gyroscope
Figure PCTCN2020102996-appb-000031
Angular rate
Figure PCTCN2020102996-appb-000032
composition:
Figure PCTCN2020102996-appb-000033
Figure PCTCN2020102996-appb-000033
最后得出来组合导航中惯性导航部分的预测误差e s,e s为惯性导航的状态误差项: Finally, the prediction error e s of the inertial navigation part of the integrated navigation is obtained, and e s is the state error term of the inertial navigation:
Figure PCTCN2020102996-appb-000034
Figure PCTCN2020102996-appb-000034
其中右边中间三项为四元数的1、2、3分量,
Figure PCTCN2020102996-appb-000035
是在时间k的惯性导航的速度矢量,惯性测量单元加速度和角速度的偏差矢量。
Among them, the middle three items on the right are the 1, 2, and 3 components of the quaternion.
Figure PCTCN2020102996-appb-000035
It is the velocity vector of the inertial navigation at time k, the deviation vector of the acceleration and angular velocity of the inertial measurement unit.
(3)通常的卫星导航模块输出安装在车辆上的卫星接收天线的三维位置、速度和当地时间。本发明的紧耦合方法需要卫星导航原始测量数据,包括可见卫星的伪距离测量、多普勒频飘测量及载波相位测量,并使用卫星的导航数据流协助本发明的惯性测量和立体视觉系统,综合决定自动驾驶车辆的位置、姿态角和环境三维信息,下面对获取卫星导航原始测量数据的详细过程进行描述:(3) The usual satellite navigation module outputs the three-dimensional position, speed and local time of the satellite receiving antenna installed on the vehicle. The tight coupling method of the present invention requires the original measurement data of satellite navigation, including pseudorange measurement of visible satellites, Doppler frequency drift measurement and carrier phase measurement, and uses the satellite navigation data stream to assist the inertial measurement and stereo vision system of the present invention, Comprehensively determine the position, attitude angle and environmental three-dimensional information of the autonomous vehicle. The detailed process of obtaining the original measurement data of satellite navigation is described below:
全球导航卫星系统(GNSS)使用无线广播信号为大众提供导航定位和时间同步服务。搞定在自动驾驶车辆上的卫星天线接收机接收导航卫星的信号;根据接收的卫星信号解析出每个卫星的星历信息,根据所述星历信息计算出每个所述卫星的卫星位置和卫星速度。同时,卫星接送机利用卫星发送的无线电信号,计算卫星的到本地接收机的伪距离、多普勒频飘及载波相位。The Global Navigation Satellite System (GNSS) uses wireless broadcast signals to provide navigation positioning and time synchronization services for the public. Get the satellite antenna receiver on the autonomous vehicle to receive the signal of the navigation satellite; analyze the ephemeris information of each satellite according to the received satellite signal, and calculate the satellite position and satellite of each satellite based on the ephemeris information speed. At the same time, the satellite shuttle plane uses the radio signal sent by the satellite to calculate the pseudo-range, Doppler frequency drift and carrier phase of the satellite to the local receiver.
具体的,卫星导航接收机使用单点定位方法来测定伪距离,伪距离即是某颗卫星与用户天线的的飞行时间,乘以光速。卫星导航 中,伪距ρ (n)(t)的计算是用n号卫星信号被接收的时间t u(t)与发射时间t s (n)(t-τ)之间的时间差与真空中无线电波的速度c相乘,表达式如下: Specifically, the satellite navigation receiver uses a single-point positioning method to determine the pseudo-distance. The pseudo-distance is the flight time between a certain satellite and the user's antenna, multiplied by the speed of light. In satellite navigation, the pseudorange ρ (n) (t) is calculated by using the time difference between the time when the n satellite signal is received t u (t) and the transmission time t s (n) (t-τ) and the vacuum The speed c of the radio wave is multiplied, and the expression is as follows:
Figure PCTCN2020102996-appb-000036
Figure PCTCN2020102996-appb-000036
其中,符号τ代表GNSS信号从发射到被用户接收的实际时间间隔。但由于GNSS卫星与用户接收机的时钟通常不会与GNSS时间t同步,将卫星时间超前GNSS时间用
Figure PCTCN2020102996-appb-000037
表示,将接收机时间超前GNSS时间用δt u(t)表示,即:
Among them, the symbol τ represents the actual time interval from when the GNSS signal is transmitted to when it is received by the user. However, since the clocks of GNSS satellites and user receivers are usually not synchronized with GNSS time t, the satellite time is ahead of GNSS time.
Figure PCTCN2020102996-appb-000037
Indicates that the receiver time is ahead of the GNSS time by δt u (t), namely:
Figure PCTCN2020102996-appb-000038
Figure PCTCN2020102996-appb-000038
t u(t)=t+δt u(t)     (3-3) t u (t)=t+δt u (t) (3-3)
电离层、对流层等结构会对电磁波的传播造成一定程度上的延时,因此τ需要减去电离层延时I (n)(t)和对流层延时T (n)(t)才是卫星信号从卫星位置到接收机位置的几何距离r (n)的传播时间,即: Structures such as the ionosphere and troposphere will delay the propagation of electromagnetic waves to a certain extent, so τ needs to be subtracted from the ionospheric delay I (n) (t) and the tropospheric delay T (n) (t) to be the satellite signal The propagation time of the geometric distance r (n) from the position of the satellite to the position of the receiver, namely:
Figure PCTCN2020102996-appb-000039
Figure PCTCN2020102996-appb-000039
Figure PCTCN2020102996-appb-000040
Figure PCTCN2020102996-appb-000040
d trop=cT (n)(t) d trop = cT (n) (t)
d iono=cI (n)(t) d iono = cI (n) (t)
其中,
Figure PCTCN2020102996-appb-000041
为卫星钟差、d iono电离层延迟和d trop对流层延迟均为已知量,ε (n)表示的是未知的伪距测量噪声,r (n)是物理空间中接 收机(x,y,z)到第n颗卫星(x (n),y (n),z (n))的几何距离,各个卫星位置的坐标(x (n),y (n),z (n))均可从各个卫星播发的星历解算得到。
in,
Figure PCTCN2020102996-appb-000041
Is the satellite clock error, iono ionospheric delay and d trop tropospheric delay are all known quantities, ε (n) represents the unknown pseudorange measurement noise, r (n) is the receiver in physical space (x, y, z) The geometric distance to the nth satellite (x (n) , y (n) , z (n) ), the coordinates of each satellite position (x (n) , y (n) , z (n) ) can be It is calculated from the ephemeris broadcast by each satellite.
类似地,卫星导航中的载波相位的观测方程如下:Similarly, the observation equation of carrier phase in satellite navigation is as follows:
φλ=ρ(t s,t r)+c(dt r-dt s)+d trop-d iono+d rel+d SA+d multi+Nλ+ε    (3-5) φλ=ρ(t s ,t r )+c(dt r -dt s )+d trop -d iono +d rel +d SA +d multi +Nλ+ε (3-5)
其中,
Figure PCTCN2020102996-appb-000042
为载波相位观测值,λ为载波波长,N为整周模糊度,t s为卫星发射信号时刻,t r为接收机接受信号时刻,dt s和dt r分别为卫星和接收机的钟差,c为光速,d iono为电离层延迟,d trop为对流层延迟,d SA为SA影响,d multi为多路径效应,ε为载波观测噪声,ρ(t s,t r)为t s时刻卫星和t r时刻接收机天线之间的几何距离,它含有测站坐标、卫星轨道和地球自转参数等。
in,
Figure PCTCN2020102996-appb-000042
Is the observed value of the carrier phase, λ is the carrier wavelength, N is the ambiguity of the whole cycle, t s is the time when the satellite transmits the signal, t r is the time when the receiver receives the signal, dt s and dt r are the clock difference between the satellite and the receiver, respectively, c is the speed of light, d iono ionospheric delay, d trop delay of the troposphere, d SA is the SA influence, d multi multi-path effect, ε is the carrier measurement noise, ρ (t s, t r ) is t s time satellite and The geometric distance between the receiver antennas at the time t r , which includes the station coordinates, satellite orbit and earth rotation parameters, etc.
由于多普勒频移观测量是载波相位率的瞬时观测值,忽略电离层、对流层延迟对时间的变化。对载波相位观测方程进行微分:Since the Doppler frequency shift observation is an instantaneous observation of the carrier phase rate, the changes in the ionospheric and tropospheric delays with respect to time are ignored. Differentiate the carrier phase observation equation:
Figure PCTCN2020102996-appb-000043
Figure PCTCN2020102996-appb-000043
其中,
Figure PCTCN2020102996-appb-000044
则,可得多普勒频移测量方程为:
in,
Figure PCTCN2020102996-appb-000044
Then, the Doppler frequency shift measurement equation is:
Figure PCTCN2020102996-appb-000045
Figure PCTCN2020102996-appb-000045
其中,λ是载波L1(f1=1575.42MHz)对应的波长,
Figure PCTCN2020102996-appb-000046
是用户u相对卫星i的多普勒频移,v (i)是卫星i的移动速度,
Figure PCTCN2020102996-appb-000047
是用户u的移动速度,
Figure PCTCN2020102996-appb-000048
是用户u指向卫星i的单位向量,c是真空中的 光速,δf α是用户u的时钟钟漂,δf (i)是卫星i的时钟钟漂,
Figure PCTCN2020102996-appb-000049
是用户u相对卫星i的多普勒频率测量噪声。由方程3-4,3-5和3-7可见,每一颗导航卫星可以提供2个独立测量结果(伪距离和载波相位)、这些测量方程可以用来校准惯性导航和立体视觉的定位偏差。
Among them, λ is the wavelength corresponding to carrier L1 (f1=1575.42MHz),
Figure PCTCN2020102996-appb-000046
Is the Doppler shift of user u relative to satellite i, v (i) is the moving speed of satellite i,
Figure PCTCN2020102996-appb-000047
Is the moving speed of user u,
Figure PCTCN2020102996-appb-000048
Is the unit vector of user u pointing to satellite i, c is the speed of light in vacuum, δf α is the clock drift of user u, δf (i) is the clock drift of satellite i,
Figure PCTCN2020102996-appb-000049
It is the Doppler frequency measurement noise of user u relative to satellite i. It can be seen from equations 3-4, 3-5 and 3-7 that each navigation satellite can provide 2 independent measurement results (pseudo-range and carrier phase). These measurement equations can be used to calibrate the positioning deviation of inertial navigation and stereo vision. .
(4)将立体视觉图像处理模块、惯性导航模块及卫星导航模块三部分的数据输入进行紧耦合,数据包括惯性导航模块测量数据、立体视觉模块的图像数据及卫星导航原始测量数据,通过卫星导航原始测量数据结合立体视觉模块的图像数据对惯性导航模块的漂移误差进行校正,从而可辅助惯性导航模块限制漂移误差,图2为系统紧耦合模块的具体流程图。在该图中,惯性传感器用于获取车辆在固定坐标系下的3轴加速度和3轴角速度输入到惯性导航模块中;双目摄像头采集图片用于直接方法的输入;GPS接收机用于获取卫星的原始测量数据。将双目摄像头直接方法的重投影误差,根据卫星原始测量数据得到的卫星导航误差和惯性导航的时间误差结合,进行总体的最优化估计来校正惯性导航模块的漂移误差,最后输出最优的位姿。(4) The data input of the three parts of the stereo vision image processing module, the inertial navigation module and the satellite navigation module are tightly coupled. The data includes the measurement data of the inertial navigation module, the image data of the stereo vision module and the original measurement data of the satellite navigation. The original measurement data is combined with the image data of the stereo vision module to correct the drift error of the inertial navigation module, which can assist the inertial navigation module to limit the drift error. Figure 2 is a specific flow chart of the tightly coupled module of the system. In this figure, the inertial sensor is used to obtain the 3-axis acceleration and 3-axis angular velocity of the vehicle in a fixed coordinate system and input to the inertial navigation module; the binocular camera collects pictures for direct method input; the GPS receiver is used to obtain satellites Raw measurement data. Combine the reprojection error of the direct method of the binocular camera, the satellite navigation error obtained according to the original satellite measurement data and the time error of the inertial navigation, and perform the overall optimization estimation to correct the drift error of the inertial navigation module, and finally output the optimal position posture.
当进行紧耦合时,紧耦合由立体视觉的加权重投影误差、卫星导航误差和来自惯性导航的时间误差构成成本函数来进行表达,表达式如下:When performing tight coupling, tight coupling is expressed by the cost function of weighted reprojection error of stereo vision, satellite navigation error, and time error from inertial navigation. The expression is as follows:
Figure PCTCN2020102996-appb-000050
Figure PCTCN2020102996-appb-000050
其中,e r为立体视觉的加权重投影误差,e g为卫星导航误差,e s为惯性导航的时间误差项,i表示相机的标号,k表示摄像机帧 标号,j表示地标标号。在第k帧和第i个相机中可见的地标标号被写为集合J(i,k)。此外,
Figure PCTCN2020102996-appb-000051
表示相应下标测量的信息矩阵,并且
Figure PCTCN2020102996-appb-000052
表示对应第k帧图片时惯性导航误差信息,
Figure PCTCN2020102996-appb-000053
表示对应t时刻的第s颗卫星时,卫星对应的误差信息矩阵。
Among them, e r is the weighted reprojection error of stereo vision, e g is the satellite navigation error, e s is the time error term of the inertial navigation, i is the camera index, k is the camera frame index, and j is the landmark index. The landmark labels visible in the k-th frame and the i-th camera are written as the set J(i,k). also,
Figure PCTCN2020102996-appb-000051
Represents the information matrix measured by the corresponding subscript, and
Figure PCTCN2020102996-appb-000052
Indicates the error information of the inertial navigation corresponding to the k-th frame picture,
Figure PCTCN2020102996-appb-000053
Represents the error information matrix corresponding to the satellite corresponding to the sth satellite at time t.
再者,e r为立体视觉的加权重投影误差,立体视觉的重投影误差的表现形式为: Furthermore, e r is the weighted re-projection error of stereo vision, and the manifestation of the re-projection error of stereo vision is:
Figure PCTCN2020102996-appb-000054
Figure PCTCN2020102996-appb-000054
其中,h i表示相机的投影模型,z i,j,k表示特征的图像坐标。
Figure PCTCN2020102996-appb-000055
表示优化系统的姿态,
Figure PCTCN2020102996-appb-000056
表示惯性导航和相机的外参,
Figure PCTCN2020102996-appb-000057
表示特征坐标。
Wherein, h i denotes a projection camera model, z i, j, k represents the coordinates of the image features.
Figure PCTCN2020102996-appb-000055
Indicates the attitude of the optimized system,
Figure PCTCN2020102996-appb-000056
Represents the external parameters of inertial navigation and camera,
Figure PCTCN2020102996-appb-000057
Represents feature coordinates.
e g为卫星导航误差,每一个时刻t每一颗导航卫星s的误差包含三个部分的误差,表现形式为: e g is the satellite navigation error, the error of each navigation satellite s at each time t contains three parts of error, and the expression is as follows:
Figure PCTCN2020102996-appb-000058
Figure PCTCN2020102996-appb-000058
其中,e p表示伪距误差,ed表示多普勒误差,e c表示载波相位误差。 Wherein, e p represents a pseudorange error, ed denotes the Doppler error, e c represents the carrier phase error.
再者,相应的误差信息矩阵W g的形式就变成了如下的形式: Furthermore, the form of the corresponding error information matrix W g becomes the following form:
Figure PCTCN2020102996-appb-000059
Figure PCTCN2020102996-appb-000059
通过线性或非线性优化的方法使得成本函数J(x)的值达到最 小,从而完成立体视觉图像处理模块、惯性导航模块及卫星导航模块三者之间的紧耦合,通过卫星导航原始测量数据结合立体视觉模块的图像数据对惯性导航模块的漂移误差进行校正。当卫星数小于设定数量时,无法对接收机进行定位,通过卫星导航原始测量数据结合立体视觉模块的图像数据对惯性导航模块的系统误差进行校正,从而提高了导航的精度,大大增强了导航系统的鲁棒性。The value of the cost function J(x) is minimized by linear or non-linear optimization methods, thereby completing the tight coupling between the stereo vision image processing module, the inertial navigation module and the satellite navigation module, and combining the original measurement data of the satellite navigation The image data of the stereo vision module corrects the drift error of the inertial navigation module. When the number of satellites is less than the set number, the receiver cannot be positioned. The system error of the inertial navigation module is corrected through the original measurement data of satellite navigation combined with the image data of the stereo vision module, thereby improving the accuracy of navigation and greatly enhancing the navigation Robustness of the system.
在给车辆定位的同时,以上步骤还通过立体视觉直接方法,实现了环境三维点云描述,而不需要昂贵的激光扫描雷达。本环境感知系统再通过深度学习神经网络判断环境中的目标和障碍物。对象探测的中心问题是把不一样的物体在环境三维点云的信息中分割出来,并判断物体的性质。这个问题交语义分割。本系统产生三维环境信息是在图像中有强度变化的点构成的稀疏点云。而且在彩色摄像头下,每个点还带有黄绿蓝强度信息。能够处理三维点云、实现语义分割,目标检测的深度神经网络模块通常分为三维稀疏卷积神经网络(3D CNN)、点网神经网络(PointNet)或者是两者的组合构成。三维稀疏卷积神经网络可以高效处理格点近似下的三维点云,和通常的二维卷积神经网络一样,它由多个卷积层和池化层工程。图2为三维卷积神经网络的卷积层,它由一个3x3x3的卷积核以步长为1对3维数据进行卷积运算。图2显示了对于一个3x3x3的三维卷积核,当它的输入为9x9x9它输出一个新的7x7x7的3维数据。更多情况下,以0填充输入数据边沿,可以使得输入和输出有同样的尺度。While positioning the vehicle, the above steps also use the direct method of stereo vision to realize the description of the three-dimensional point cloud of the environment without the need for expensive laser scanning radar. The environment perception system then judges the targets and obstacles in the environment through the deep learning neural network. The central problem of object detection is to segment different objects from the environmental three-dimensional point cloud information, and to determine the nature of the object. This question is handed to semantic segmentation. The three-dimensional environmental information generated by this system is a sparse point cloud composed of points with varying intensity in the image. And under the color camera, each point also carries the intensity information of yellow, green and blue. The deep neural network module that can process three-dimensional point clouds, realize semantic segmentation, and target detection is usually divided into three-dimensional sparse convolutional neural network (3D CNN), point network neural network (PointNet) or a combination of the two. The three-dimensional sparse convolutional neural network can efficiently process the three-dimensional point cloud under the grid point approximation. Like the usual two-dimensional convolutional neural network, it is engineered by multiple convolutional layers and pooling layers. Figure 2 is the convolutional layer of the three-dimensional convolutional neural network, which consists of a 3x3x3 convolution kernel with a step size of 1 to perform convolution operations on the three-dimensional data. Figure 2 shows that for a 3x3x3 three-dimensional convolution kernel, when its input is 9x9x9, it outputs a new 7x7x7 three-dimensional data. In more cases, filling the input data edge with 0 can make the input and output have the same scale.
完整的三维卷积神经网络由多个三维卷积层和多个三维池化层构成(图3),它有多个卷积内核,产生出不一样的特征。然后 最大池化层取池化单元中的最大值作为单元的输出,如果以2x2x2作为池化单元大小,步长为2,池化处理后的三维数据将减小在每个方向减少一半。经过多级卷积层、池化层后数据在经过多层全连接层,最后把分类判断的结果从输出层,通常为归一化指数函数(Softmax)输出。A complete three-dimensional convolutional neural network is composed of multiple three-dimensional convolutional layers and multiple three-dimensional pooling layers (Figure 3). It has multiple convolution kernels, which produce different features. Then the maximum pooling layer takes the maximum value of the pooling unit as the output of the unit. If 2x2x2 is used as the size of the pooling unit and the step size is 2, the three-dimensional data after pooling will be reduced by half in each direction. After the multi-level convolutional layer and the pooling layer, the data passes through the multi-layer fully connected layer, and finally the result of the classification judgment is output from the output layer, usually a normalized exponential function (Softmax).
尽管三维卷积神经网络可以高效地处理格点化以后的三维点云数据,格点化丢失了部分细微位置信息,造成误差。点网(PointNet)方法并不需要格点化三维数据,它可以直接在使用三维点云,作为输入,实现三维环境点云的语义分割。Although the three-dimensional convolutional neural network can efficiently process the three-dimensional point cloud data after gridding, the gridding loses part of the subtle position information, causing errors. The PointNet method does not require gridded 3D data. It can directly use 3D point clouds as input to achieve semantic segmentation of 3D environmental point clouds.
点网的基本单元称为集合抽象(Set Abstraction)。它由抽样结合层和点网层构成。抽样结合层使用迭代最远点抽样(iterative farthest point sampling)方法,把三维空间的点的点分割成N个质心区域附件的点的结合,然后再使用多层感知器网络提取点云特征。点网一般由多层结合抽象单元,形成多级结合抽象功能,最后在使用全连接层和输出层。输出层使用归一化指数(Softmax)全连接输出分类结果(图4)。The basic unit of the dot network is called Set Abstraction. It is composed of sampling combined layer and dot network layer. The sampling combination layer uses the iterative farthest point sampling method to divide the points in the three-dimensional space into a combination of points near the centroid region, and then use the multi-layer perceptron network to extract the point cloud features. The dot network is generally composed of multiple layers of combined abstract units to form a multi-level combined abstract function, and finally a fully connected layer and an output layer are used. The output layer uses the normalized index (Softmax) fully connected to output the classification results (Figure 4).
该紧耦合的自动驾驶感知系统可以帮助对环境提供高精度测量和建立地图,不再借助于昂贵的激光扫描雷达,降低了自动驾驶汽车的成本。The tightly coupled autonomous driving perception system can help provide high-precision measurements of the environment and build maps without resorting to expensive laser scanning radar, reducing the cost of autonomous vehicles.
基于上述紧耦合的自动驾驶感知方法,本发明还提供了一种自动驾驶对象探测和定位系统100,图3为本发明实施例的自动驾驶对象探测和定位系统100的结构示意图,如图5所示,该自动驾驶对象探测和定位系统100包含双目直接方法的立体视觉图像处理模块10、卫星导航模块20、惯性导航模块30、系统紧耦合模块40及深度神经网络模块50,判断环境中的目标和障碍物。Based on the aforementioned tightly coupled automatic driving perception method, the present invention also provides an automatic driving object detection and positioning system 100. FIG. 3 is a schematic structural diagram of the automatic driving object detection and positioning system 100 according to an embodiment of the present invention, as shown in FIG. 5. As shown, the automatic driving object detection and positioning system 100 includes the stereo vision image processing module 10 of the binocular direct method, the satellite navigation module 20, the inertial navigation module 30, the system tight coupling module 40 and the deep neural network module 50, which can judge the environment in the environment. Goals and obstacles.
具体的,惯性导航模块30用于采用惯性传感器获取惯性导航模块测量数据;具体如上。立体视觉图像处理模块10采用双目直接方法获取立体视觉模块的图像数据;具体如上。卫星导航模块20用于通过接收机获取导航卫星的卫星导航原始测量数据;具体如上。系统紧耦合模块40用于将惯性导航模块测量数据、立体视觉模块的图像数据及卫星导航原始测量数据进行紧耦合处理;具体如上。各个模块之间的连接关系为:立体视觉图像处理模块10、卫星导航模块20及惯性导航模块30均与系统紧耦合模块40连接。Specifically, the inertial navigation module 30 is used to obtain the measurement data of the inertial navigation module by using an inertial sensor; the details are as above. The stereo vision image processing module 10 adopts the binocular direct method to obtain the image data of the stereo vision module; the details are as above. The satellite navigation module 20 is used to obtain the original satellite navigation measurement data of the navigation satellite through the receiver; the details are as above. The system tight coupling module 40 is used to perform tight coupling processing on the measurement data of the inertial navigation module, the image data of the stereo vision module, and the original measurement data of the satellite navigation; the details are as above. The connection relationship between the various modules is: the stereo vision image processing module 10, the satellite navigation module 20, and the inertial navigation module 30 are all connected to the system tight coupling module 40.
其中,立体视觉图像处理模块10包含双目摄像头,双目摄像头安装于自动驾驶车辆上;具体如上所述。再者,惯性导航模块30包含惯性传感器,惯性传感器固定于自动驾驶车辆上;具体如上所述。Among them, the stereo vision image processing module 10 includes a binocular camera, and the binocular camera is installed on an autonomous vehicle; the details are as described above. Furthermore, the inertial navigation module 30 includes inertial sensors, and the inertial sensors are fixed on the autonomous vehicle; the details are as described above.
其中,深度神经网络模块使用三维卷积神经网络或者是三维点网实现对象探测和语义分割。具体如上所述。Among them, the deep neural network module uses a three-dimensional convolutional neural network or a three-dimensional dot network to realize object detection and semantic segmentation. The details are as described above.
本发明将惯性导航模块测量数据、立体视觉模块的图像数据及卫星导航原始测量数据三者进行紧耦合,对惯性导航模块测量数据的误差进行修正,从而提高了定位精度,不再借助于昂贵的激光扫描雷达,从而降低自动驾驶汽车的成本。The invention tightly couples the measurement data of the inertial navigation module, the image data of the stereo vision module, and the original measurement data of the satellite navigation to correct the error of the measurement data of the inertial navigation module, thereby improving the positioning accuracy without resorting to expensive Laser scanning radar, thereby reducing the cost of self-driving cars.
以上所述,仅为本发明较佳的具体实施方式,但本发明的保护范围并不局限于此,任何熟悉本技术领域的技术人员在本发明披露的技术范围内,根据本发明的技术方案及其发明构思加以等同替换或改变,都应涵盖在本发明的保护范围之内。The above are only the preferred specific embodiments of the present invention, but the scope of protection of the present invention is not limited to this. Anyone familiar with the technical field within the technical scope disclosed in the present invention, according to the technical solution of the present invention Equivalent replacements or changes to its inventive concept should all fall within the protection scope of the present invention.

Claims (9)

  1. 一种自动驾驶对象探测和定位系统,其特征在于,包括立体视觉图像处理模块、卫星导航模块、惯性导航模块及系统紧耦合模块,其中:An automatic driving object detection and positioning system, which is characterized by comprising a stereo vision image processing module, a satellite navigation module, an inertial navigation module and a system tight coupling module, wherein:
    立体视觉图像处理模块,采用双目或者多目摄像头获取立体视觉模块的图像数据;Stereo vision image processing module, using binocular or multi-eye cameras to obtain image data of the stereo vision module;
    卫星导航模块,用于通过接收机获取卫星导航原始测量数据;The satellite navigation module is used to obtain the original measurement data of satellite navigation through the receiver;
    惯性导航模块,用于采用惯性传感器获取惯性导航模块的测量数据;The inertial navigation module is used to obtain the measurement data of the inertial navigation module by using inertial sensors;
    系统紧耦合模块,用于将所述惯性导航模块测量数据、所述立体视觉模块的图像数据及所述卫星导航原始测量数据进行紧耦合处理,并建立三维环境信息,最后利用深度神经网络模块检测环境中的目标和障碍物;The system tight coupling module is used to tightly couple the measurement data of the inertial navigation module, the image data of the stereo vision module and the original measurement data of the satellite navigation, and establish three-dimensional environment information, and finally use the deep neural network module to detect Targets and obstacles in the environment;
    所述深度神经网络由三维稀疏卷积神经网络、点网神经网络和二者的组合构成;The deep neural network is composed of a three-dimensional sparse convolutional neural network, a dot network neural network, and a combination of the two;
    所述立体视觉图像处理模块、所述卫星导航模块及所述惯性导航模块均与所述系统紧耦合模块连接。The stereo vision image processing module, the satellite navigation module and the inertial navigation module are all connected to the system tight coupling module.
  2. 如权利要求1所述的自动驾驶对象探测和定位系统,其特征在于,所述立体视觉图像处理模块包含双目或者多目摄像头。The automatic driving object detection and positioning system according to claim 1, wherein the stereo vision image processing module comprises a binocular or multi-eye camera.
  3. 一种如权利要求1所述的自动驾驶对象探测和定位的方法,其特征在于,包括以下步骤:A method for detecting and locating an autonomous driving object according to claim 1, characterized in that it comprises the following steps:
    S1:获取立体视觉模块的图像数据、惯性导航模块的测量数据及卫星导航原始测量数据;S1: Obtain the image data of the stereo vision module, the measurement data of the inertial navigation module and the original measurement data of the satellite navigation;
    S2:将立体视觉模块的图像数据、卫星导航原始测量数据及所述惯性导航模块的测量数据进行紧耦合,对惯性导航模块的漂移误差进行修正;S2: tightly couple the image data of the stereo vision module, the original measurement data of the satellite navigation and the measurement data of the inertial navigation module to correct the drift error of the inertial navigation module;
    S3:将立体视觉模块的图像数据、卫星导航原始测量数据及所 述惯性导航模块的测量数据进行紧耦合,建立三维环境信息,再使用深度神经网络模块判断和分类环境中的目标和障碍物;S3: Tightly couple the image data of the stereo vision module, the original measurement data of the satellite navigation and the measurement data of the inertial navigation module to establish three-dimensional environment information, and then use the deep neural network module to judge and classify the targets and obstacles in the environment;
    S4:深度神经网络使用三维稀疏卷积神经网络、点网神经网络或者是二者的组合。S4: The deep neural network uses a three-dimensional sparse convolutional neural network, a dot network neural network, or a combination of the two.
  4. 如权利要求3所述的自动驾驶对象探测和定位的方法,其特征在于,所述立体视觉模块采用多目,包括双目摄像头直接方法进行处理,所述图像数据包括在同一时刻多目或者双目立体视觉摄像头之间的视觉差和每一个摄像头在不同时间、位置拍摄到的图像信息来建立环境的三维信息。The method for detecting and positioning an automatic driving object according to claim 3, wherein the stereo vision module adopts a multi-eye, including binocular camera direct method for processing, and the image data includes multi-eye or dual-eye at the same time. The visual difference between the three-dimensional vision cameras and the image information captured by each camera at different times and positions are used to establish the three-dimensional information of the environment.
  5. 如权利要求3所述的自动驾驶对象探测和定位的方法,其特征在于,所述紧耦合由立体视觉的加权重投影误差、卫星导航误差和来自惯性导航的状态误差构成成本函数。The method for detecting and positioning an automatic driving object according to claim 3, wherein the tight coupling is composed of a weighted reprojection error of stereo vision, satellite navigation error, and state error from inertial navigation to form a cost function.
  6. 如权利要求3所述的自动驾驶对象探测和定位的方法,其特征在于,所述S1具体包括如下步骤:The method for detecting and locating an autonomous driving object according to claim 3, wherein the S1 specifically includes the following steps:
    S11:采用多目或者是双目摄像头获取立体视觉模块的图像数据、图像数据包括在同一时刻多目或者双目立体视觉摄像头之间的视觉差和每一个摄像头在不同时间、位置拍摄到的图像信息;S11: Use multi-eye or binocular cameras to obtain image data of the stereo vision module. The image data includes the visual difference between the multi-eye or binocular stereo vision cameras at the same time and the images captured by each camera at different times and positions information;
    S12:采用惯性传感器获取惯性导航模块的测量数据;S12: Use inertial sensors to obtain measurement data of the inertial navigation module;
    S13:通过接收机获取卫星导航原始测量数据;S13: Obtain the original measurement data of satellite navigation through the receiver;
    S14:将所述惯性导航模块测量数据、所述图像数据及所述卫星导航原始测量数据进行紧耦合处理。S14: Perform tight coupling processing on the measurement data of the inertial navigation module, the image data, and the original measurement data of the satellite navigation.
  7. 如权利要求6所述的自动驾驶对象探测和定位的方法,其特征在于,所述S11具体包括如下步骤:The method for detecting and locating an autonomous driving object according to claim 6, wherein the S11 specifically includes the following steps:
    S111:用多目或者双目摄像头采集环境图像信号;S111: Use a multi-eye or binocular camera to collect environmental image signals;
    S112:组合多个不同摄像头在同一个时刻拍摄到的视觉差和每 一个摄像头在不同时间、位置拍摄到的图像,形成立体视觉观测的直接方法;S112: Combine the visual difference captured by multiple different cameras at the same time and the images captured by each camera at different times and positions to form a direct method of stereo vision observation;
    S113:并组合惯性导航、卫星导航的测量数据,建立三维环境信息,最后使用深度神经网络模块分类和判断环境中的目标和障碍物。S113: Combine the measurement data of inertial navigation and satellite navigation to establish three-dimensional environment information, and finally use the deep neural network module to classify and judge the targets and obstacles in the environment.
  8. 如权利要求6所述的自动驾驶对象探测和定位的方法,其特征在于,所述S12具体包括如下步骤:The method for detecting and positioning an autonomous driving object according to claim 6, wherein the S12 specifically includes the following steps:
    S121:利用惯性传感器测量自动驾驶车辆在固定坐标系下的3轴加速度和3轴角速度;S121: Use inertial sensors to measure the 3-axis acceleration and 3-axis angular velocity of the autonomous vehicle in a fixed coordinate system;
    S122:将所述加速度和所述角速度转动到导航坐标系下,求解惯性导航机械编排方程并计算出自动驾驶车辆的位置和姿态角;S122: Rotate the acceleration and the angular velocity to the navigation coordinate system, solve the inertial navigation mechanical arrangement equation, and calculate the position and attitude angle of the autonomous vehicle;
    S123:图像数据、惯性导航数据和卫星导航数据组合,建立三维环境信息,再利用深度神经网络模块,分类和判断环境中的目标和障碍物。S123: Combine image data, inertial navigation data and satellite navigation data to establish three-dimensional environment information, and then use deep neural network modules to classify and judge objects and obstacles in the environment.
  9. 如权利要求6所述的自动驾驶对象探测和定位的方法,其特征在于,所述S14具体包括:通过所述卫星导航原始测量数据结合所述立体视觉模块的图像数据对所述惯性导航模块的漂移误差进行校正。The method for detecting and positioning an automatic driving object according to claim 6, wherein said S14 specifically comprises: comparing the original measurement data of the satellite navigation with the image data of the stereo vision module to the inertial navigation module. The drift error is corrected.
PCT/CN2020/102996 2020-06-12 2020-07-20 System and method for detecting and positioning autonomous driving object WO2021248636A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202010538035.6 2020-06-12
CN202010538035.6A CN111929718A (en) 2020-06-12 2020-06-12 Automatic driving object detection and positioning system and method

Publications (1)

Publication Number Publication Date
WO2021248636A1 true WO2021248636A1 (en) 2021-12-16

Family

ID=73316190

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2020/102996 WO2021248636A1 (en) 2020-06-12 2020-07-20 System and method for detecting and positioning autonomous driving object

Country Status (2)

Country Link
CN (1) CN111929718A (en)
WO (1) WO2021248636A1 (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20220107184A1 (en) * 2020-08-13 2022-04-07 Invensense, Inc. Method and system for positioning using optical sensor and motion sensors
CN114485632A (en) * 2021-12-31 2022-05-13 深圳市易成自动驾驶技术有限公司 Vehicle positioning method, system and computer readable storage medium
CN114543842A (en) * 2022-02-28 2022-05-27 重庆长安汽车股份有限公司 Positioning precision evaluation system and method of multi-sensor fusion positioning system
CN115560757A (en) * 2022-09-01 2023-01-03 中国人民解放军战略支援部队信息工程大学 Neural network-based unmanned aerial vehicle direct positioning correction method under random attitude error condition
CN115790282A (en) * 2022-10-11 2023-03-14 西安岳恒机电工程有限责任公司 Direction control system and control method for unmanned target vehicle
CN116363205A (en) * 2023-03-30 2023-06-30 中国科学院西安光学精密机械研究所 Space target pose resolving method based on deep learning and computer program product
CN116778101A (en) * 2023-06-26 2023-09-19 北京道仪数慧科技有限公司 Map generation method and system based on camping carrier

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112731497A (en) * 2020-12-04 2021-04-30 广西融科科技有限公司 Method for improving satellite positioning precision by using field vision field analysis
WO2022120733A1 (en) * 2020-12-10 2022-06-16 Beijing Didi Infinity Technology And Development Co., Ltd. Systems and methods for constructing map
CN113240917B (en) * 2021-05-08 2022-11-08 广州隧华智慧交通科技有限公司 Traffic management system applying deep neural network to intelligent traffic
CN113406682B (en) * 2021-06-22 2024-03-12 腾讯科技(深圳)有限公司 Positioning method, positioning device, electronic equipment and storage medium
CN113465598B (en) * 2021-08-04 2024-02-09 北京云恒科技研究院有限公司 Inertial integrated navigation system suitable for unmanned aerial vehicle

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100109949A1 (en) * 2008-04-11 2010-05-06 Samsung Electronics Co., Ltd. Mobile terminal having a hybrid navigation system and method for determining a location thereof
CN103983263A (en) * 2014-05-30 2014-08-13 东南大学 Inertia/visual integrated navigation method adopting iterated extended Kalman filter and neural network
CN108171796A (en) * 2017-12-25 2018-06-15 燕山大学 A kind of inspection machine human visual system and control method based on three-dimensional point cloud
CN109643125A (en) * 2016-06-28 2019-04-16 柯尼亚塔有限公司 For training the 3D virtual world true to nature of automated driving system to create and simulation
CN109683629A (en) * 2019-01-09 2019-04-26 燕山大学 Unmanned plane electric stringing system based on integrated navigation and computer vision
CN109725339A (en) * 2018-12-20 2019-05-07 东莞市普灵思智能电子有限公司 A kind of tightly coupled automatic Pilot cognitive method and system
CN109991636A (en) * 2019-03-25 2019-07-09 启明信息技术股份有限公司 Map constructing method and system based on GPS, IMU and binocular vision
CN111239790A (en) * 2020-01-13 2020-06-05 上海师范大学 Vehicle navigation system based on 5G network machine vision

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8509965B2 (en) * 2006-12-12 2013-08-13 American Gnc Corporation Integrated collision avoidance system for air vehicle
CN105116431A (en) * 2015-09-08 2015-12-02 中国人民解放军装备学院 Inertial navigation platform and Beidou satellite-based high-precision and ultra-tightly coupled navigation method
CN110221328A (en) * 2019-07-23 2019-09-10 广州小鹏汽车科技有限公司 A kind of Combinated navigation method and device
CN110332945B (en) * 2019-08-01 2021-06-04 北京眸星科技有限公司 Vehicle navigation method and device based on traffic road marking visual identification

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100109949A1 (en) * 2008-04-11 2010-05-06 Samsung Electronics Co., Ltd. Mobile terminal having a hybrid navigation system and method for determining a location thereof
CN103983263A (en) * 2014-05-30 2014-08-13 东南大学 Inertia/visual integrated navigation method adopting iterated extended Kalman filter and neural network
CN109643125A (en) * 2016-06-28 2019-04-16 柯尼亚塔有限公司 For training the 3D virtual world true to nature of automated driving system to create and simulation
CN108171796A (en) * 2017-12-25 2018-06-15 燕山大学 A kind of inspection machine human visual system and control method based on three-dimensional point cloud
CN109725339A (en) * 2018-12-20 2019-05-07 东莞市普灵思智能电子有限公司 A kind of tightly coupled automatic Pilot cognitive method and system
CN109683629A (en) * 2019-01-09 2019-04-26 燕山大学 Unmanned plane electric stringing system based on integrated navigation and computer vision
CN109991636A (en) * 2019-03-25 2019-07-09 启明信息技术股份有限公司 Map constructing method and system based on GPS, IMU and binocular vision
CN111239790A (en) * 2020-01-13 2020-06-05 上海师范大学 Vehicle navigation system based on 5G network machine vision

Cited By (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20220107184A1 (en) * 2020-08-13 2022-04-07 Invensense, Inc. Method and system for positioning using optical sensor and motion sensors
US11875519B2 (en) * 2020-08-13 2024-01-16 Medhat Omr Method and system for positioning using optical sensor and motion sensors
CN114485632A (en) * 2021-12-31 2022-05-13 深圳市易成自动驾驶技术有限公司 Vehicle positioning method, system and computer readable storage medium
CN114485632B (en) * 2021-12-31 2024-03-22 深圳市易成自动驾驶技术有限公司 Vehicle positioning method, system and computer readable storage medium
CN114543842B (en) * 2022-02-28 2023-07-28 重庆长安汽车股份有限公司 Positioning accuracy evaluation system and method for multi-sensor fusion positioning system
CN114543842A (en) * 2022-02-28 2022-05-27 重庆长安汽车股份有限公司 Positioning precision evaluation system and method of multi-sensor fusion positioning system
CN115560757A (en) * 2022-09-01 2023-01-03 中国人民解放军战略支援部队信息工程大学 Neural network-based unmanned aerial vehicle direct positioning correction method under random attitude error condition
CN115560757B (en) * 2022-09-01 2023-08-22 中国人民解放军战略支援部队信息工程大学 Unmanned aerial vehicle direct positioning correction method based on neural network under random attitude error condition
CN115790282B (en) * 2022-10-11 2023-08-22 西安岳恒机电工程有限责任公司 Unmanned target vehicle direction control system and control method
CN115790282A (en) * 2022-10-11 2023-03-14 西安岳恒机电工程有限责任公司 Direction control system and control method for unmanned target vehicle
CN116363205A (en) * 2023-03-30 2023-06-30 中国科学院西安光学精密机械研究所 Space target pose resolving method based on deep learning and computer program product
CN116778101A (en) * 2023-06-26 2023-09-19 北京道仪数慧科技有限公司 Map generation method and system based on camping carrier
CN116778101B (en) * 2023-06-26 2024-04-09 北京道仪数慧科技有限公司 Map generation method and system based on camping carrier

Also Published As

Publication number Publication date
CN111929718A (en) 2020-11-13

Similar Documents

Publication Publication Date Title
WO2021248636A1 (en) System and method for detecting and positioning autonomous driving object
CN110926474B (en) Satellite/vision/laser combined urban canyon environment UAV positioning and navigation method
WO2020124624A1 (en) Autonomous driving sensing method and system employing close coupling
CN107850673A (en) Vision inertia ranging attitude drift is calibrated
CN110873570B (en) Method and apparatus for sourcing, generating and updating a map representing a location
US11875519B2 (en) Method and system for positioning using optical sensor and motion sensors
Mostafa et al. A novel GPS/RAVO/MEMS-INS smartphone-sensor-integrated method to enhance USV navigation systems during GPS outages
CN101782642B (en) Method and device for absolutely positioning measurement target by multi-sensor fusion
CN114019552A (en) Bayesian multi-sensor error constraint-based location reliability optimization method
Li et al. Multi-GNSS PPP/INS/Vision/LiDAR tightly integrated system for precise navigation in urban environments
JPH0772244A (en) Interference-type synthetic aperture radar equipment and topographic change observation system
Bai et al. Real-time GNSS NLOS detection and correction aided by sky-pointing camera and 3D LiDAR
Hsu et al. Hong Kong UrbanNav: An open-source multisensory dataset for benchmarking urban navigation algorithms
Khoshelham et al. Vehicle positioning in the absence of GNSS signals: Potential of visual-inertial odometry
Grejner-Brzezinska et al. From Mobile Mapping to Telegeoinformatics
Aggarwal GPS-based localization of autonomous vehicles
Park et al. Precise and reliable positioning based on the integration of navigation satellite system and vision system
CN116105729A (en) Multi-sensor fusion positioning method for reconnaissance of forest environment of field cave
CN115900732A (en) Combined navigation method and system based on roadside camera and vehicle-mounted unit
Wei Multi-sources fusion based vehicle localization in urban environments under a loosely coupled probabilistic framework
CN109341685B (en) Fixed wing aircraft vision auxiliary landing navigation method based on homography transformation
Zhang et al. 3D mapping aided gnss positioning using doppler frequency for urban areas
CN115574816B (en) Bionic vision multi-source information intelligent perception unmanned platform
Adeel et al. Research and performance analysis of tightly coupled vision, INS and GNSS system for land vehicle applications
Huang Wheel Odometry Aided Visual-Inertial Odometry in Winter Urban Environments

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20940278

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20940278

Country of ref document: EP

Kind code of ref document: A1