WO2022141914A1 - Multi-target vehicle detection and re-identification method based on radar and video fusion - Google Patents

Multi-target vehicle detection and re-identification method based on radar and video fusion Download PDF

Info

Publication number
WO2022141914A1
WO2022141914A1 PCT/CN2021/085150 CN2021085150W WO2022141914A1 WO 2022141914 A1 WO2022141914 A1 WO 2022141914A1 CN 2021085150 W CN2021085150 W CN 2021085150W WO 2022141914 A1 WO2022141914 A1 WO 2022141914A1
Authority
WO
WIPO (PCT)
Prior art keywords
vehicle
target
millimeter
video camera
data
Prior art date
Application number
PCT/CN2021/085150
Other languages
French (fr)
Chinese (zh)
Inventor
杜豫川
许军
赵聪
覃伯豪
暨育雄
沈煜
静 曹
Original Assignee
杜豫川
许军
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 杜豫川, 许军 filed Critical 杜豫川
Priority to CN202180010955.3A priority Critical patent/CN115943439A/en
Priority to GB2313217.8A priority patent/GB2619196A/en
Publication of WO2022141914A1 publication Critical patent/WO2022141914A1/en

Links

Images

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S7/00Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
    • G01S7/02Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S13/00
    • G01S7/40Means for monitoring or calibrating
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S17/00Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
    • G01S17/87Combinations of systems using electromagnetic waves other than radio waves
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S17/00Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
    • G01S17/88Lidar systems specially adapted for specific applications
    • G01S17/89Lidar systems specially adapted for specific applications for mapping or imaging
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S17/00Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
    • G01S17/88Lidar systems specially adapted for specific applications
    • G01S17/93Lidar systems specially adapted for specific applications for anti-collision purposes
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S17/00Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
    • G01S17/88Lidar systems specially adapted for specific applications
    • G01S17/93Lidar systems specially adapted for specific applications for anti-collision purposes
    • G01S17/931Lidar systems specially adapted for specific applications for anti-collision purposes of land vehicles
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S7/00Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
    • G01S7/003Transmission of data between radar, sonar or lidar systems and remote stations
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S7/00Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
    • G01S7/48Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S17/00
    • G01S7/4808Evaluating distance, position or velocity data
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S7/00Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
    • G01S7/48Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S17/00
    • G01S7/497Means for monitoring or calibrating
    • G01S7/4972Alignment of sensor
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/56Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
    • GPHYSICS
    • G08SIGNALLING
    • G08GTRAFFIC CONTROL SYSTEMS
    • G08G1/00Traffic control systems for road vehicles
    • G08G1/01Detecting movement of traffic to be counted or controlled
    • G08G1/0104Measuring and analyzing of parameters relative to traffic conditions
    • G08G1/0108Measuring and analyzing of parameters relative to traffic conditions based on the source of data
    • G08G1/0116Measuring and analyzing of parameters relative to traffic conditions based on the source of data from roadside infrastructure, e.g. beacons
    • GPHYSICS
    • G08SIGNALLING
    • G08GTRAFFIC CONTROL SYSTEMS
    • G08G1/00Traffic control systems for road vehicles
    • G08G1/01Detecting movement of traffic to be counted or controlled
    • G08G1/0104Measuring and analyzing of parameters relative to traffic conditions
    • G08G1/0125Traffic data processing
    • G08G1/0133Traffic data processing for classifying traffic situation
    • GPHYSICS
    • G08SIGNALLING
    • G08GTRAFFIC CONTROL SYSTEMS
    • G08G1/00Traffic control systems for road vehicles
    • G08G1/01Detecting movement of traffic to be counted or controlled
    • G08G1/0104Measuring and analyzing of parameters relative to traffic conditions
    • G08G1/0137Measuring and analyzing of parameters relative to traffic conditions for specific applications
    • G08G1/0141Measuring and analyzing of parameters relative to traffic conditions for specific applications for traffic information dissemination
    • GPHYSICS
    • G08SIGNALLING
    • G08GTRAFFIC CONTROL SYSTEMS
    • G08G1/00Traffic control systems for road vehicles
    • G08G1/01Detecting movement of traffic to be counted or controlled
    • G08G1/0104Measuring and analyzing of parameters relative to traffic conditions
    • G08G1/0137Measuring and analyzing of parameters relative to traffic conditions for specific applications
    • G08G1/0145Measuring and analyzing of parameters relative to traffic conditions for specific applications for active traffic flow control
    • GPHYSICS
    • G08SIGNALLING
    • G08GTRAFFIC CONTROL SYSTEMS
    • G08G1/00Traffic control systems for road vehicles
    • G08G1/01Detecting movement of traffic to be counted or controlled
    • G08G1/04Detecting movement of traffic to be counted or controlled using optical or ultrasonic detectors
    • GPHYSICS
    • G08SIGNALLING
    • G08GTRAFFIC CONTROL SYSTEMS
    • G08G1/00Traffic control systems for road vehicles
    • G08G1/01Detecting movement of traffic to be counted or controlled
    • G08G1/048Detecting movement of traffic to be counted or controlled with provision for compensation of environmental or other condition, e.g. snow, vehicle stopped at detector
    • GPHYSICS
    • G08SIGNALLING
    • G08GTRAFFIC CONTROL SYSTEMS
    • G08G1/00Traffic control systems for road vehicles
    • G08G1/16Anti-collision systems
    • G08G1/164Centralised systems, e.g. external to vehicles
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/44Arrangements for executing specific programs
    • G06F9/451Execution arrangements for user interfaces
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/194Segmentation; Edge detection involving foreground-background segmentation

Definitions

  • the second category based on vehicle license plate information.
  • the main method is to obtain multiple regional segmentation results of the target vehicle from the image to be recognized; use the convolutional neural network CNN to extract regional feature vectors from the multiple regional segmentation results, and fuse them with the global feature vectors to obtain the target vehicle's segmentation results. Appearance feature vector. Finally, the fused feature vector is used for vehicle re-identification and retrieval.
  • this scheme considers the influence of posture on vehicle re-identification, the accuracy of the model is limited by the diversity of the dataset, which must include vehicle images from various angles. , and the scale is large enough, in real scenarios, it is difficult to collect all vehicle images from different angles and the number of datasets reaches hundreds of thousands.
  • Video cameras can extract the image features of multi-target vehicles, and can also use camera calibration technology and target detection algorithms to obtain positioning data of multi-target vehicles.
  • Millimeter-wave radar can also detect the positioning data of multi-target vehicles, and the data fusion algorithm can obtain higher-precision positioning data of multi-target vehicles.
  • millimeter-wave radar can accurately capture geospatial information such as vehicle speed and driving direction, which can add a new dimension to vehicle attributes. Combined with video image data, it can provide a solution to the above-mentioned vehicle re-identification problem.
  • Single video camera a single video camera, multi-target vehicles can drive into and out of the field of view of the video camera.
  • Vehicle image features The vehicle image HSV value, LBP and HOG are all vehicle image features. In addition, it also includes vehicle color, model, size, and license plate information.
  • HOG Histogram of Oriented Gradient feature is a feature descriptor used for object detection in computer vision and image processing. It constructs features by calculating and counting the gradient direction histograms of local regions of the image. HOG features combined with SVM classifiers have been widely used in image recognition.
  • Targets of different classes have high similarity in feature reflection.
  • vehicle re-identification it is reflected as follows: for different vehicles, because their models, colors, sizes are similar or belong to the same brand and the same model, they are reflected in the image features with high similarity.
  • Data set The data set obtained by the data acquisition equipment, this patent refers to the positioning data of the experimental test vehicle collected by the video camera, the millimeter wave radar and the vehicle RTK equipment.
  • Target vehicle A vehicle that enters the field of view of the sensor (video camera and millimeter-wave radar).
  • Multi-target vehicle A vehicle set containing at least 1 target vehicle.
  • Multi-sensor data fusion algorithm Synthesize the integrated multi-source data to form the best consistent estimate of the measured object and its properties.
  • Kalman filtering algorithm is used for data fusion.
  • Mahalanobis distance was proposed by Indian statistician P.C. Mahalanobis, which represents the distance between a point and a distribution. It is an efficient method to calculate the similarity between two unknown sample sets. Unlike Euclidean distance, it takes into account the connections between various properties (eg: a piece of information about height leads to an information about weight because the two are related) and is scale-independent ( scale-invariant), i.e. independent of the measurement scale.
  • the sensors in this patent refer to instruments that can obtain data, including video cameras, millimeter-wave radars, and RTK positioning equipment (handheld and vehicle-mounted).
  • Fig. 7 is the flow chart of the time synchronization optimization model between the target detection frame data and the vehicle RTK positioning data
  • the downward tilt angle is 10° and the scene has no other obstructions, its field of view can reach 100-150m.
  • Both types of sensors collect data at a frequency of 25Hz and access the central server for data storage and processing.
  • video image data use the pre-trained deep learning target detection algorithm to identify the multi-target vehicles within the camera's field of view, and extract the image features of each target vehicle (including: vehicle color, model, size, license plate information (non- necessary)) and geospatial features (world geographic coordinates).
  • geospatial features world geographic coordinates.
  • millimeter wave radar data extract the geospatial features of the target vehicle (including: vehicle world geographic coordinates, speed, heading angle, etc.).
  • the present invention regards the world geographic coordinates obtained through RTK as a relative true value.
  • the pixel coordinate system represents the position of the image pixel in the image. Usually, the pixel in the upper left corner of the image is used as the origin, and the right direction is the positive direction of the x-axis, and the downward direction is the positive direction of the y-axis.
  • the horizontal and vertical coordinates of the pixels represent the distance of the pixel respectively. The number of pixels in the y-axis and x-axis.
  • the target detection frame coordinate calibration is to establish the mapping relationship between the pixel coordinates of the vehicle target detection frame in the camera image and the vehicle world geographic coordinates.
  • the pixel coordinates of the vehicle target detection frame refer to the pixel coordinates of the midpoint pixel of the lower bottom edge of the detection frame. (The camera is located directly above the road coming from the road)
  • Variant B The camera is located on the road to the upper right, and the pixel coordinates of the vehicle target detection frame refer to the pixel coordinates of the vertex pixels at the lower left of the detection frame.
  • the target detection frame is obtained by the deep learning target detection algorithm, the data frequency is 25Hz, and the vehicle world geographic coordinates are obtained by the vehicle RTK, and the data frequency is 5Hz.
  • Install on-board RTK for the experimental test vehicle to obtain the real-time world geographic coordinate data of the vehicle, drive the vehicle into the field of view of the camera, and use the deep learning target detection algorithm to mark and detect the experimental vehicle in the video image to obtain its pixel coordinates.
  • the calculated homography transformation matrix corresponding to the pixel coordinates of the camera detection target and the world geographical coordinates, the world geographical coordinates of the detected target vehicle are calculated. It is a common phenomenon that the clocks between different sensors are out of synchronization.
  • x, y and z represent the coordinates of the radar detection target in the radar coordinate system
  • longitude, latitude and height represent the transformation of the coordinates of the radar detection target in the radar coordinate system into the world geographic coordinates in the world geographic coordinate system.
  • longitude, latitude and height as the world geographic coordinates of the detected target calculated from the radar coordinates, and at the same time obtain the vehicle world geographic coordinates obtained by the vehicle RTK, denoted as longitude rtk , latitude rtk and height rtk .
  • Two types of space-time matching optimization models of world geographic coordinates are established, including two parts: time synchronization optimization model and spatial error calibration.
  • Step 4 Feature matching and data fusion
  • the video and the target vehicle data obtained by the millimeter-wave radar are matched according to the geospatial features, and the multi-sensor data fusion method is used to fuse the video and the vehicle speed and geographic location data obtained by the millimeter-wave radar to improve the data accuracy.
  • the fusion algorithm the high-precision positioning coordinates obtained by the on-board RTK of the experimental vehicle are used as the reference truth value, and sensor data fusion methods such as Kalman filtering, multi-Bayesian estimation, fuzzy logic reasoning and deep neural network are used to The geographic location data of the target vehicle is fused to calculate.
  • Kalman filter algorithm is one of the algorithms widely used in multi-source data fusion.
  • Kalman filtering is a recursive filtering algorithm, which is characterized by no need to save past historical information, and the new data is combined with the estimated value obtained in the previous frame (or previous moment) and the state equation of the system itself according to a certain method. Find a new estimate.
  • the principle of Kalman filter can be expressed by the following five formulas:
  • the state variable at this time is inferred from time t-1, and the observation value at this time has not been corrected;
  • the state variable at this moment is estimated from the previous moment, and the observed value at this moment has been corrected
  • Cascade matching algorithm A tracker is assigned to each detector, and each tracker is set with a timer parameter. If the tracker is matched and updated, the timer parameter is reset to 0, otherwise it is incremented by 1 unit. In cascading matching, the trackers will be sorted according to the timer parameters. Smaller parameters will be matched first, and larger parameters will be matched later. That is, the tracker that matches the first frame in the previous frame is given high priority, and the tracker that has not been matched for several frames has a lower priority.
  • Feature matching Based on the Mahalanobis distance and cosine distance features, target features are matched between different image frames to complete the process of target re-identification (ID transfer).

Abstract

A multi-target vehicle detection and re-identification method based on radar and video fusion. A roadside sensing device mainly used in the method comprises a roadside video camera and a millimeter-wave radar. In the method, multi-target vehicle detection is performed on a travelling vehicle in a road by using two types of sensors, i.e. a video camera and a millimeter-wave radar, and data fusion is performed on video camera image data and millimeter-wave radar detection target data by using a multi-sensor data fusion method, such that the precision of target detection data is improved. In the method, two scenes, i.e. a single-video camera performing continuous tracking on a multi-target vehicle and a cross-video camera performing continuous tracking on the multi-target vehicle, are taken into consideration; on the basis of a multi-target vehicle image feature extracted on the basis of a video image, in combination with data detected by a roadside millimeter-wave radar, a geospatial feature of the multi-target vehicle is acquired, and a new dimension is added to a vehicle attribute; and the present invention is applied to a multi-target vehicle re-identification algorithm, such that the accuracy of multi-target vehicle re-identification can be effectively improved, thereby improving the vehicle supervision level in a management and control region.

Description

一种基于雷视融合的多目标车辆检测及重识别方法A multi-target vehicle detection and re-identification method based on Raivision fusion 技术领域technical field
本发明属于移动车辆目标检测、多传感器数据融合和车辆重识别技术领域,涉及运用视频摄像机图像数据与毫米波雷达数据融合检测与重识别多目标车辆的方法。The invention belongs to the technical field of moving vehicle target detection, multi-sensor data fusion and vehicle re-identification, and relates to a method for detecting and re-identifying multi-target vehicles using video camera image data and millimeter-wave radar data fusion.
背景技术Background technique
近年来,随着我国经济的不断发展,人们的生活水平日益改善,开车出行已成为许多人出行的首选方式,开车通勤也成为较为普遍的现象。机动车辆的快速增长给交通监管部门带来更大的压力。随着视频监控在公共安全领域占据着越来越重要的地位,车辆相关的任务越来越受到人们的关注,如车辆目标检测、车辆分类、车辆追踪和驾驶员行为分析等。如何判断单个视频摄像机视频不同图像帧之间的车辆是否为同一车辆,以及判断跨视频摄像机之间的视频图像中的车辆是否为同一车辆成为车辆管控的重要需求,在此基础上才能更好地开展车辆轨迹追踪、驾驶员行为分析等后续工作。In recent years, with the continuous development of my country's economy and the improvement of people's living standards, driving has become the preferred way for many people to travel, and commuting by car has also become a relatively common phenomenon. The rapid growth of motor vehicles puts more pressure on traffic regulators. As video surveillance plays an increasingly important role in the field of public safety, vehicle-related tasks such as vehicle object detection, vehicle classification, vehicle tracking, and driver behavior analysis have received increasing attention. How to judge whether the vehicle between different image frames of a single video camera is the same vehicle, and whether the vehicle in the video images between cross-video cameras is the same vehicle has become an important requirement for vehicle management and control. Carry out follow-up work such as vehicle trajectory tracking and driver behavior analysis.
车辆重识别是指在一个特定范围内的交通管控场景中,判断不同时刻或非重叠区域的不同位置的路侧传感器(视频摄像机、毫米波雷达等)捕捉到的车辆数据是否为同一辆车的车辆身份识别问题。解决这一问题对区域内车辆精准管控、区域安防、车路协同等多方面具有重要意义。Vehicle re-identification refers to judging whether the vehicle data captured by roadside sensors (video cameras, millimeter-wave radar, etc.) Vehicle identification problem. Solving this problem is of great significance to the precise management and control of vehicles in the region, regional security, and vehicle-road coordination.
目前,现有车辆重识别方法主要分为四类:At present, the existing vehicle re-identification methods are mainly divided into four categories:
第一类:基于传感器的方法Category 1: Sensor-Based Approaches
使用各种传感器对车辆进行探测并推断身份是最基本的、最早出现的车辆重识别方法。对于每个被传感器探测到的车辆来说,这类方法通常利用特定的手段提取车辆特征,通过特征匹配来判断车辆身份。最早出现的车辆重识别方法基于各种硬件探测器(如红外线、超声波等)提取车辆的特征信息。之后,许多利用其他传感器或感应器的车辆识别方法被提出,如利用三维磁感应器来探测车辆多维特征并能够从感应器中获取时间信息,用于训练高斯极大似然分类器。感应线圈是在交通场景中最常用的获取数据的工具,可以监测多种车辆属性(如速度、体积和车辆占地面积),城市的主干道路和高速公路上一般都部署了感应线圈。基于感应线圈的实时车辆重识别方法利用感应线圈提供的数据提取车辆特征,并估计车辆行程时间,从而完成车辆重识别任务。随着一些新兴传感器或技术(如全球定位系统(GPS)、无线射频识别(RFID)和手机)的面世,一些方法探索了基于信标的车辆追踪和监视系统,以完成车辆重识别任务。如基于无线射频识别标签的车辆重识别算法,适用于各类收费站,基于GPS的车辆行驶时间估计方法,用于解决车辆重识别问题。基于传感器的方法大多需要安装大量的硬件设备,实验环境较为苛刻,难以复现。此外,许多方法易受客观环境影响,如天气情况、信号强弱、交通拥挤程度和车辆行驶速度等,这些都将不同程度地降低传感器的灵敏度。同一时期也没有统一的性能评估标准,因此这类方法不能算是理想的车辆重识别方法。Using various sensors to detect vehicles and infer their identities is the most basic and earliest vehicle re-identification method. For each vehicle detected by the sensor, such methods usually use specific means to extract vehicle features, and determine the vehicle identity through feature matching. The earliest vehicle re-identification methods are based on various hardware detectors (such as infrared, ultrasonic, etc.) to extract the characteristic information of vehicles. After that, many vehicle recognition methods using other sensors or sensors have been proposed, such as the use of three-dimensional magnetic sensors to detect multi-dimensional features of vehicles and to obtain temporal information from the sensors for training Gaussian maximum likelihood classifiers. Induction coils are the most commonly used data acquisition tools in traffic scenarios, which can monitor various vehicle attributes (such as speed, volume, and vehicle footprint). Induction coils are generally deployed on urban arterial roads and highways. The real-time vehicle re-identification method based on the induction coil uses the data provided by the induction coil to extract the vehicle characteristics and estimate the travel time of the vehicle, so as to complete the vehicle re-identification task. With the advent of some emerging sensors or technologies such as Global Positioning System (GPS), Radio Frequency Identification (RFID), and cell phones, some approaches have explored beacon-based vehicle tracking and surveillance systems to accomplish vehicle re-identification tasks. For example, the vehicle re-identification algorithm based on radio frequency identification tags is suitable for various toll stations, and the GPS-based vehicle travel time estimation method is used to solve the problem of vehicle re-identification. Most of the sensor-based methods require the installation of a large number of hardware devices, and the experimental environment is harsh and difficult to reproduce. In addition, many methods are easily affected by the objective environment, such as weather conditions, signal strength, traffic congestion and vehicle speed, which will reduce the sensitivity of the sensor to varying degrees. There is also no unified performance evaluation standard in the same period, so this kind of method cannot be regarded as an ideal vehicle re-identification method.
第二类:基于车辆车牌信息。The second category: based on vehicle license plate information.
随着计算机视觉技术的发展,通过图像或视频数据识别车辆身份信息成为可能,且硬件设 施的部署量大大减少,可节约大量的硬件成本和人工安装成本。初始阶段,基于计算机视觉技术的车辆重识别方法主要通过车牌定位、字符分割与识别的方法提取车辆车牌信息。定位功能主要运用灰度信息、颜色与纹理信息等方法实现,分割与识别功能主要采用模板匹配和神经网络等方法实现。基于车牌的车辆重识别方法准确性高,但缺点也十分明显:交通监控系统存在拍摄视角变化、天气影响、光照变化和图像分辨率低等问题,一旦目标车辆车牌信息丢失,该方法便失效。而在实际复杂的交通环境中,因车牌被遮挡、摄像机分辨率低、拍摄距离远和拍摄角度等因素导致车牌信息难以获取是时常发生的事件,而这会在很大程度上降低车牌识别的准确率。对于车辆假车牌、套车牌甚至无车牌的现象更是无能为力。因此,虽然车牌识别是区分不同车辆最简单直接的方法,但是在很多情况下仅依靠车牌信息无法完成重识别任务。With the development of computer vision technology, it is possible to identify vehicle identity information through image or video data, and the deployment of hardware facilities is greatly reduced, which can save a lot of hardware costs and labor installation costs. In the initial stage, the vehicle re-identification method based on computer vision technology mainly extracts the vehicle license plate information through the methods of license plate location, character segmentation and recognition. The positioning function is mainly realized by grayscale information, color and texture information, and the segmentation and recognition functions are mainly realized by template matching and neural network methods. The vehicle re-recognition method based on license plate has high accuracy, but its shortcomings are also very obvious: the traffic monitoring system has problems such as the change of shooting angle, the influence of weather, the change of illumination and the low image resolution. Once the license plate information of the target vehicle is lost, the method will fail. In the actual complex traffic environment, it is a common occurrence that the license plate information is difficult to obtain due to factors such as the occlusion of the license plate, the low resolution of the camera, the long shooting distance and the shooting angle, which will greatly reduce the license plate recognition. Accuracy. For the phenomenon of fake license plates, sets of license plates or even no license plates, it is powerless to do anything. Therefore, although license plate recognition is the simplest and most direct method to distinguish different vehicles, in many cases, only relying on license plate information cannot complete the re-recognition task.
第三类:基于车辆图像特征。The third category: based on vehicle image features.
不完全依赖车牌,综合其他非车牌信息完成车辆重识别工作,继而提升车辆重识别的稳定性和准确性。传统的无车牌信息车辆重识别方法主要是通过提取车辆图像的HSV特征、LBP特征以及HOG等特征来进行图像特征匹配过程,之后通过对车辆颜色、车型、以及车辆挡风玻璃等分类实现车辆重识别工作。基于无车牌信息的车辆重识别方法可解释性强,但易受视角变化和遮挡影响,重识别准确率仍然较低。It does not completely rely on the license plate, and integrates other non-license plate information to complete the vehicle re-identification work, thereby improving the stability and accuracy of the vehicle re-identification. The traditional vehicle re-recognition method without license plate information mainly performs the image feature matching process by extracting the HSV features, LBP features and HOG features of the vehicle image, and then realizes the vehicle re-identification process by classifying the vehicle color, model, and vehicle windshield. Identify work. The vehicle re-identification method based on no license plate information has strong interpretability, but is easily affected by the change of viewing angle and occlusion, and the re-identification accuracy is still low.
第四类:基于机器学习和深度神经网络。The fourth category: based on machine learning and deep neural networks.
近年来,随着人工智能算法和深度神经网络在计算机视觉领域的发展,车辆重识别有了新的技术路线。越来越多的基于机器学习和深度学习的车辆重识别方法被提出,该类方法是的车辆重识别准确性大大提高,逐渐成为实现车辆重识别工作的主流方法。In recent years, with the development of artificial intelligence algorithms and deep neural networks in the field of computer vision, vehicle re-identification has a new technical route. More and more vehicle re-identification methods based on machine learning and deep learning have been proposed, which greatly improve the accuracy of vehicle re-identification and gradually become the mainstream method for vehicle re-identification work.
卷积神经网络将图像作为输入,不用预先提取复杂的人工特征,通过不断前向学习、后向反馈的过程进行特征提取。卷积神经网络的每一层主要包含了特征提取和特征映射操作。在特征提取操作中,神经元的输入为上一层的输出,使用卷积核对输入进行卷积操作,来得到局部特征,每一层可使用多个卷积核,表示针对输入提取多个特征。由于卷积核的权值共享,极大的减少了网络的参数。在特征映射操作中,使用sigmod或tanh函数作为卷积网络的激活函数,使得提取的特征具有位移不变性。卷积神经网络特征提取操作对训练数据进行自动的学习,避免了使用人工定义的特征提取方法来固定的提取特征,而是隐式地从训练数据中自主学习,并且由于卷积核权值共享,这样可并行的学习,提高了计算效率。The convolutional neural network takes the image as input, and does not need to extract complex artificial features in advance, and performs feature extraction through the process of continuous forward learning and backward feedback. Each layer of the convolutional neural network mainly includes feature extraction and feature mapping operations. In the feature extraction operation, the input of the neuron is the output of the previous layer, and the convolution kernel is used to perform the convolution operation on the input to obtain local features. Each layer can use multiple convolution kernels, which means that multiple features are extracted for the input. . Due to the weight sharing of the convolution kernel, the parameters of the network are greatly reduced. In the feature mapping operation, the sigmod or tanh function is used as the activation function of the convolutional network, so that the extracted features have displacement invariance. The convolutional neural network feature extraction operation automatically learns the training data, avoiding the use of artificially defined feature extraction methods to extract features fixedly, but implicitly learns from the training data autonomously, and because the convolution kernel weights are shared , so that the parallel learning can improve the computational efficiency.
基于卷积神经网络并考虑了姿态对识别准确率的影响。主要做法是从待识别图像中,获取目标车辆的多个区域分割结果;使用卷积神经网络CNN对多个区域分割结果提取区域特征向量,并与全局特征向量进行融合,得到所述目标车辆的外观特征向量。最后使用融合后的特征向量进行车辆重识别与检索,该方案虽然考虑姿态对车辆重识别的影响,但是模型的准确度受限于数据集的多样性,数据集必须包括各种角度的车辆图,且规模足够大,在现实场景下,收集所有车辆在不同角度下的车辆图片并且数量达到几十万量级的数据集困难较大。此外,在收集的数据集上,要针对不同角度的车辆图片进行关键点的标注,不同图片的角度不同,因而标注的关键点的数量和位置不同,导致工作量巨大。因而从可行性和工作量上分析,该方法较为复杂。Based on convolutional neural network and considering the influence of pose on recognition accuracy. The main method is to obtain multiple regional segmentation results of the target vehicle from the image to be recognized; use the convolutional neural network CNN to extract regional feature vectors from the multiple regional segmentation results, and fuse them with the global feature vectors to obtain the target vehicle's segmentation results. Appearance feature vector. Finally, the fused feature vector is used for vehicle re-identification and retrieval. Although this scheme considers the influence of posture on vehicle re-identification, the accuracy of the model is limited by the diversity of the dataset, which must include vehicle images from various angles. , and the scale is large enough, in real scenarios, it is difficult to collect all vehicle images from different angles and the number of datasets reaches hundreds of thousands. In addition, on the collected data set, it is necessary to mark the key points of the vehicle pictures from different angles. The angles of different pictures are different, so the number and position of the marked key points are different, resulting in a huge workload. Therefore, the method is more complicated in terms of feasibility and workload.
此外,由于摄像机视角和光照变化等导致同一辆车在不同视角或光照情况下产生的类内差异大或不同车辆因型号相同形成类间相似度高,这一严峻挑战仍然是限制车辆重识别准确率的 主要原因。In addition, due to changes in camera perspective and illumination, the same vehicle has large intra-class differences under different perspectives or lighting conditions, or different vehicles have the same model to form high inter-class similarity. This severe challenge still limits the accuracy of vehicle re-identification. the main reason for the rate.
上述各类车辆重识别方法都是基于路侧摄像机拍摄的图像或视频数据,在车辆外观层面对多目标车辆进行重识别,其流程图如图1所示,包含车辆图像或视频的获取、车辆检测、特征提取与表达、相似性度量计算和检测结果的展示。然而,不同的车辆除了车牌以外,在外观上完全一致的情况也十分常见。对于这种情况,仅从外观层面对目标车辆进行重识别,误判率将大大增加。The above types of vehicle re-identification methods are based on images or video data captured by roadside cameras, and re-identify multi-target vehicles at the vehicle appearance level. The flow chart is shown in Figure 1, including vehicle image or video acquisition, vehicle Detection, feature extraction and expression, similarity metric calculation and display of detection results. However, it is not uncommon for different vehicles to be identical in appearance, except for the license plate. In this case, only re-identifying the target vehicle from the appearance level will greatly increase the misjudgment rate.
毫米波雷达,是工作在毫米波波段探测的雷达。通常毫米波是指30~300GHz频域(波长为1~10mm)的。毫米波的波长介于微波和厘米波之间,因此毫米波雷达兼有微波雷达和光电雷达的一些优点。同厘米波导引头相比,毫米波导引头具有体积小、质量轻和空间分辨率高的特点。与红外、激光、电视等光学导引头相比,毫米波导引头穿透雾、烟、灰尘的能力强,具有全天候(大雨天除外)全天时的特点。光波在大气中传播衰减严重,器件加工精度要求高。毫米波与光波相比,它们利用大气窗口(毫米波与亚毫米波在大气中传播时,由于气体分子谐振吸收所致的某些衰减为极小值的频率)传播时的衰减小,受自然光和热辐射源影响小。为此,它们在通信、雷达、制导、遥感技术、射电天文学和波谱学方面都有重大的意义。其优势主要有以下几点:Millimeter wave radar is a radar that works in the millimeter wave band. Usually millimeter wave refers to the 30-300GHz frequency domain (wavelength is 1-10mm). The wavelength of millimeter wave is between microwave and centimeter wave, so millimeter wave radar has some advantages of microwave radar and photoelectric radar. Compared with the centimeter-wave seeker, the millimeter-wave seeker has the characteristics of small size, light weight and high spatial resolution. Compared with optical seekers such as infrared, laser, and TV, the millimeter-wave seeker has a strong ability to penetrate fog, smoke, and dust, and has the characteristics of all-weather (except heavy rain) all day. The propagation of light waves in the atmosphere is seriously attenuated, and the processing precision of the device is required to be high. Compared with light waves, millimeter waves use the atmospheric window (when millimeter waves and submillimeter waves propagate in the atmosphere, some attenuation due to the resonance absorption of gas molecules is a minimum frequency), the attenuation is small when they propagate, and they are affected by natural light. and thermal radiation sources have little impact. To this end, they are of great significance in communications, radar, guidance, remote sensing technology, radio astronomy and spectroscopy. Its advantages are mainly as follows:
(1)小天线口径、窄波束:高追踪和引导精度;易于进行低仰角追踪,抗地面多径和杂波干扰;对近空目标具有高横向分辨力;对区域成像和目标监视具备高角分辨力;窄波束的高抗干扰性能;高天线增益;容易检测小目标等。(1) Small antenna aperture and narrow beam: high tracking and guidance accuracy; easy to perform low-elevation tracking, anti-ground multipath and clutter interference; high lateral resolution for near-air targets; high angular resolution for regional imaging and target monitoring high anti-jamming performance with narrow beam; high antenna gain; easy to detect small targets, etc.
(2)大带宽:具有高信息速率,容易采用窄脉冲或宽带调频信号获得目标的细节结构特征;具有宽的扩谱能力,减少多径、杂波并增强抗干扰能力;相邻频率的雷达或毫米波识别器工作,易克服相互干扰;高距离分辨力,易得到精确的目标追踪和识别能力。(2) Large bandwidth: With high information rate, it is easy to use narrow pulse or wideband FM signal to obtain the detailed structural characteristics of the target; it has wide spectrum spreading ability, reduces multipath, clutter and enhances anti-jamming ability; adjacent frequency radar Or millimeter wave identifiers work, easy to overcome mutual interference; high distance resolution, easy to obtain accurate target tracking and identification capabilities.
(3)高多普勒频率:慢目标和振动目标的良好检测和识别能力;易于利用目标多普勒频率特性进行目标特征识别;对干性大气污染的穿透特性,提供在尘埃、烟尘和干雪条件下的良好检测能力。(3) High Doppler frequency: good detection and identification capabilities of slow targets and vibrating targets; easy to use target Doppler frequency characteristics for target feature identification; penetration characteristics of dry air pollution, providing high performance in dust, smoke and dust Good detection capability in dry snow conditions.
视频摄像机可提取多目标车辆的图像特征,同时也可以采用摄像机标定技术与目标检测算法获取多目标车辆的定位数据。毫米波雷达也可探测多目标车辆的定位数据,采用数据融合算法可获取多目标车辆更高精度的定位数据。此外,毫米波雷达能精准捕捉车辆速度与行驶方向等地理空间信息,可为车辆属性增添新的维度,结合视频图像数据可以为上述车辆重识别难题提供解决的途径。Video cameras can extract the image features of multi-target vehicles, and can also use camera calibration technology and target detection algorithms to obtain positioning data of multi-target vehicles. Millimeter-wave radar can also detect the positioning data of multi-target vehicles, and the data fusion algorithm can obtain higher-precision positioning data of multi-target vehicles. In addition, millimeter-wave radar can accurately capture geospatial information such as vehicle speed and driving direction, which can add a new dimension to vehicle attributes. Combined with video image data, it can provide a solution to the above-mentioned vehicle re-identification problem.
现有技术current technology
专利文件CN108875754APatent document CN108875754A
专利文件CN111582178APatent document CN111582178A
专利文件CN111553205APatent document CN111553205A
专利文件CN109508731APatent document CN109508731A
专利文件CN111435421APatent document CN111435421A
相关术语说明Description of related terms
1.单视频摄像机:单一视频摄像机,多目标车辆可驶入与驶出该视频摄像机视野范围。1. Single video camera: a single video camera, multi-target vehicles can drive into and out of the field of view of the video camera.
2.跨视频摄像机:两个或两个以上视野范围不同的视频摄像机集合,多目标车辆可驶出某一视频摄像机视野范围后进入另一视频摄像机视野范围。2. Cross-video camera: a set of two or more video cameras with different fields of view. A multi-target vehicle can drive out of the field of view of one video camera and enter the field of view of another video camera.
3.车辆图像特征:车辆图像HSV值、LBP与HOG均属于车辆图像特征,此外,还包括车辆颜色、车型、尺寸、车牌信息。3. Vehicle image features: The vehicle image HSV value, LBP and HOG are all vehicle image features. In addition, it also includes vehicle color, model, size, and license plate information.
4.车辆地理空间特征:指车辆的经纬度坐标,速度,航向角信息。4. Vehicle geospatial features: refers to the longitude and latitude coordinates, speed, and heading angle information of the vehicle.
5.HSV:HSV(Hue,Saturation,Value)是根据颜色的直观特性由A.R.Smith在1978年创建的一种颜色空间,也称六角锥体模型。这个模型中颜色的参数分别是:色调(H),饱和度(S),明度(V)。5.HSV: HSV (Hue, Saturation, Value) is a color space created by A.R. Smith in 1978 based on the intuitive characteristics of color, also known as the hexagonal pyramid model. The parameters of the color in this model are: Hue (H), Saturation (S), Lightness (V).
6.LBP:LBP(Local Binary Pattern)算子是一种有效的纹理描述算子,它具有旋转不变性和灰度不变性等显著优点。其基本思想是用其中心像素的灰度值作为阈值,与它的邻域相比较得到的二进制码来表述局部纹理特征。6. LBP: The LBP (Local Binary Pattern) operator is an effective texture description operator, which has significant advantages such as rotation invariance and grayscale invariance. The basic idea is to use the gray value of the central pixel as a threshold and compare the binary code with its neighborhood to express local texture features.
7.HOG:方向梯度直方图(Histogram of Oriented Gradient)特征是一种在计算机视觉和图像处理中用来进行物体检测的特征描述子。它通过计算和统计图像局部区域的梯度方向直方图来构成特征,HOG特征结合SVM分类器已经被广泛应用于图像识别中。7.HOG: Histogram of Oriented Gradient feature is a feature descriptor used for object detection in computer vision and image processing. It constructs features by calculating and counting the gradient direction histograms of local regions of the image. HOG features combined with SVM classifiers have been widely used in image recognition.
8.类内差异大:同一类别的目标在特征反映上具有较大差异。在车辆重识别上反映为:对于同一车辆,由于视频摄像机拍摄角度不同、光照强度不同等因素影响,导致其图像特征具有较大差异。8. Large intra-class differences: Targets in the same category have large differences in feature reflection. In vehicle re-identification, it is reflected as follows: for the same vehicle, due to factors such as different shooting angles of video cameras and different light intensities, the image features of the same vehicle are quite different.
9.类间相似度高:不同类别的目标在特征反映上具有较高相似度。在车辆重识别上反映为:对于不同车辆,由于其车型、颜色、尺寸相似或属于同一品牌同一型号,在图像特征中反映出具有较高的相似度。9. High similarity between classes: Targets of different classes have high similarity in feature reflection. In vehicle re-identification, it is reflected as follows: for different vehicles, because their models, colors, sizes are similar or belong to the same brand and the same model, they are reflected in the image features with high similarity.
10.视频摄像机数据:视频图像数据,对摄像机进行标定与对图像进行目标检测后还可得到图像检测目标像素坐标与世界地理坐标。10. Video camera data: video image data, after calibrating the camera and performing target detection on the image, the pixel coordinates of the image detection target and the world geographic coordinates can be obtained.
11.视频摄像机定位数据:由经过标定的视频摄像机输出的目标定位数据,即由像素坐标系转换而得的世界地理坐标系中的坐标。11. Video camera positioning data: the target positioning data output by the calibrated video camera, that is, the coordinates in the world geographic coordinate system converted from the pixel coordinate system.
12.毫米波雷达数据:车辆相对于毫米波雷达的距离、方位角、速度与车辆航向角等,可根据以上数据计算车辆在雷达坐标系下的坐标,对毫米波雷达进行标定后可计算车辆在世界坐标系下的坐标。12. Millimeter-wave radar data: the distance, azimuth, speed and vehicle heading angle of the vehicle relative to the millimeter-wave radar. The coordinates of the vehicle in the radar coordinate system can be calculated according to the above data, and the vehicle can be calculated after calibrating the millimeter-wave radar. The coordinates in the world coordinate system.
13.毫米波雷达定位数据:毫米波雷达数据中的目标定位坐标数据。13. Millimeter-wave radar positioning data: target positioning coordinate data in the millimeter-wave radar data.
14.RTK:高精度的GPS测量必须采用载波相位观测值,RTK设备采用差分定位技术,即基于载波相位观测值的实时动态定位技术,能够实时地提供测站点在指定坐标系中的三维定位结果,并达到厘米级精度。RTK根据用途可分为手持型与车载型。手持型RTK可用于手持进行单点坐标测量。车载型可安装于测试车辆上,用于连续的坐标测量。14. RTK: High-precision GPS measurement must use carrier phase observations. RTK equipment adopts differential positioning technology, that is, real-time dynamic positioning technology based on carrier phase observations, which can provide real-time three-dimensional positioning results of the station in the specified coordinate system. , and achieve centimeter-level accuracy. RTK can be divided into handheld type and vehicle type according to the application. Hand-held RTK can be used for hand-held single-point coordinate measurement. The vehicle type can be installed on the test vehicle for continuous coordinate measurement.
15.RTK定位数据:手持型与车载型RTK测量得到的经纬度坐标。15. RTK positioning data: latitude and longitude coordinates obtained by handheld and vehicle-mounted RTK measurements.
16.数据集:数据采集设备获取的数据集合,本专利中指代视频摄像机、毫米波雷达与车载RTK设备采集的实验测试车辆的定位数据。16. Data set: The data set obtained by the data acquisition equipment, this patent refers to the positioning data of the experimental test vehicle collected by the video camera, the millimeter wave radar and the vehicle RTK equipment.
17.中心计算服务器:用于接收、处理与储存各数据采集设备获取的数据(包括视频摄像机图像数据与毫米波雷达定位数据)的计算机设备。17. Central computing server: computer equipment used to receive, process and store data (including video camera image data and millimeter-wave radar positioning data) acquired by each data acquisition device.
18.时空匹配:指通过最优化匹配方法,使得不同设备获取的数据集对于同一目标采集的定位数据在时间上达到同步,在空间上达到一致,即误差在可接受范围内(0.5米)。18. Space-time matching: refers to the optimization of the matching method, so that the data sets obtained by different devices are synchronized in time for the positioning data collected by the same target, and are consistent in space, that is, the error is within an acceptable range (0.5 meters).
19.世界地理坐标系:本专利中所述世界地理坐标系均为地理坐标系中的WGS-84坐标系。19. World geographic coordinate system: The world geographic coordinate system described in this patent is the WGS-84 coordinate system in the geographic coordinate system.
20.像素坐标系:像素坐标系表示图像像素在图像中的位置,通常情况下,以图像左上角像素为原点,规定右方向为x轴正方向,下方向为y轴正方向,像素横纵坐标分别表示该像素距离y轴与x轴的像素个数。20. Pixel coordinate system: The pixel coordinate system represents the position of the image pixel in the image. Usually, the pixel in the upper left corner of the image is used as the origin, and the right direction is the positive direction of the x-axis, the downward direction is the positive direction of the y-axis, and the pixels are horizontal and vertical. The coordinates represent the number of pixels from the pixel to the y-axis and the x-axis, respectively.
21.雷达坐标系:即毫米波雷达坐标系,以毫米波雷达自身作为坐标系原点的三维空间坐标系。21. Radar coordinate system: the millimeter-wave radar coordinate system, a three-dimensional space coordinate system with the millimeter-wave radar itself as the origin of the coordinate system.
22.单应性变换矩阵:不同坐标系下的坐标转换矩阵,可通过选取不同坐标系下对应的关键点(至少4对不共线的点)计算得到。22. Homography transformation matrix: The coordinate transformation matrix in different coordinate systems can be calculated by selecting the corresponding key points (at least 4 pairs of non-collinear points) in different coordinate systems.
23.多次:至少2次。23. Multiple times: at least 2 times.
24.目标车辆:进入传感器(视频摄像机与毫米波雷达)视野范围内的车辆。24. Target vehicle: A vehicle that enters the field of view of the sensor (video camera and millimeter-wave radar).
25.多目标车辆:包含至少1个目标车辆的车辆集合。25. Multi-target vehicle: A vehicle set containing at least 1 target vehicle.
26.多目标车辆数据:至少1辆车的定位数据,由视频摄像机与毫米波雷达获取得到。26. Multi-target vehicle data: The positioning data of at least one vehicle is obtained by video cameras and millimeter-wave radars.
27.深度学习目标检测算法:本专利中采用的深度学习目标检测算法为YOLO v5目标检测算法对道路中的车辆进行目标检测。27. Deep learning target detection algorithm: The deep learning target detection algorithm used in this patent is the YOLO v5 target detection algorithm to detect targets on vehicles on the road.
28.bb:Bounding box,目标检测矩形框,由目标检测算法计算结果返回,用于框选视频图像中检测到的目标轮廓。28.bb: Bounding box, a rectangular box for target detection, which is returned by the calculation result of the target detection algorithm and is used to frame the detected target contour in the video image.
29.多传感器数据融合算法:将经过集成处理的多源数据进行合成,形成对被测对象及其性质的最佳一致估计。本专利中采用卡尔曼滤波算法进行数据融合。29. Multi-sensor data fusion algorithm: Synthesize the integrated multi-source data to form the best consistent estimate of the measured object and its properties. In this patent, Kalman filtering algorithm is used for data fusion.
30.马氏距离:马氏距离(Mahalanobis distance)由印度统计学家马哈拉诺比斯(P.C.Mahalanobis)提出,表示点与一个分布之间的距离。它是一种有效的计算两个未知样本集的相似度的方法。与欧氏距离不同的是,它考虑到各种特性之间的联系(例如:一条关于身高的信息会带来一条关于体重的信息,因为两者是有关联的),并且是尺度无关的(scale-invariant),即独立于测量尺度。30. Mahalanobis distance: Mahalanobis distance was proposed by Indian statistician P.C. Mahalanobis, which represents the distance between a point and a distribution. It is an efficient method to calculate the similarity between two unknown sample sets. Unlike Euclidean distance, it takes into account the connections between various properties (eg: a piece of information about height leads to an information about weight because the two are related) and is scale-independent ( scale-invariant), i.e. independent of the measurement scale.
31.余弦距离:余弦距离又称余弦相似度,或余弦相似性,通过计算两个向量的夹角余弦值来评估两向量的相似度。31. Cosine distance: Cosine distance, also known as cosine similarity, or cosine similarity, evaluates the similarity of two vectors by calculating the cosine value of the angle between the two vectors.
32.路侧:位于不属于道路路面本身的其他位置,如道路侧边、道路上方(由立杆或龙门架支持)等。32. Roadside: located in other positions that do not belong to the road surface itself, such as the side of the road, above the road (supported by poles or gantry), etc.
33.丢失车辆:曾在所述传感器设备(视频摄像机、毫米波雷达)集合视野范围内出现,而当前未有一台设备捕捉到的车辆。33. Lost vehicles: vehicles that have appeared within the collective field of view of the sensor devices (video cameras, millimeter-wave radars), but are currently not captured by a single device.
34.传感器设备:本专利传感器指代可获取数据的仪器,包括视频摄像机、毫米波雷达、RTK定位设备(手持型与车载型)。34. Sensor equipment: The sensors in this patent refer to instruments that can obtain data, including video cameras, millimeter-wave radars, and RTK positioning equipment (handheld and vehicle-mounted).
35.车辆重识别系统:用于对丢失车辆进行重识别的系统框架。35. Vehicle re-identification system: a system framework for re-identification of lost vehicles.
36.车辆丢失数据库:记录车辆丢失前最后n帧,(n的取值可根据效果进行调节)的特征,包括车辆图像特征与地理空间特征。36. Vehicle Loss Database: Record the last n frames before the vehicle is lost, (the value of n can be adjusted according to the effect) features, including vehicle image features and geospatial features.
37.ID重赋予:当车辆重识别系统找回丢失车辆时,将该车辆的历史ID赋予找回的车辆上。37. ID re-assignment: When the vehicle re-identification system retrieves the lost vehicle, it assigns the vehicle's historical ID to the retrieved vehicle.
【发明内容】[Content of the invention]
为了解决目前基于视频图像的车辆重识别技术因摄像机视角和光照变化等原因导致同一辆车在不同视角或光照情况下产生的类内差异大或不同车辆因型号相同形成类间相似度高而导致识别准确率不高的问题,本发明采用的技术方案思路是:In order to solve the problem that the current vehicle re-identification technology based on video images may cause large intra-class differences for the same vehicle under different viewing angles or lighting conditions due to changes in camera angle of view and illumination, or high inter-class similarity between different vehicles due to the same model. The problem that the recognition accuracy is not high, the technical solution idea adopted by the present invention is:
一种基于雷视融合的多目标车辆检测及重识别方法,通过加入毫米波雷达检测数据为车辆特征增加地理空间信息维度。A multi-target vehicle detection and re-identification method based on Raivision fusion, which adds the dimension of geospatial information to vehicle features by adding millimeter wave radar detection data.
具体方案为:The specific plans are:
1)对视频摄像机S 1与毫米波雷达S 2进行标定,以车载RTK定位数据D 0作为相对真值(因为车载RTK定位数据误差级别为厘米级,视频摄像机定位数据D 1与毫米波雷达定位数据D 2误差级别为米级,车载RTK定位数据精度远高于视频摄像机与毫米波雷达定位数据精度),分别建立定位数据D 1和D 2与定位数据D 0的时空匹配优化模型,采用最优化算法对定位数据D 1和D 2进行时空匹配,完成传感器标定工作。 1) Calibrate the video camera S 1 and the millimeter-wave radar S 2 , and use the on-board RTK positioning data D 0 as the relative true value (because the error level of the on-board RTK positioning data is centimeter-level, the video camera positioning data D 1 and the millimeter-wave radar positioning The error level of data D2 is meter level, and the accuracy of vehicle RTK positioning data is much higher than the accuracy of video camera and millimeter wave radar positioning data ) . The optimization algorithm performs space-time matching on the positioning data D 1 and D 2 to complete the sensor calibration.
2)利用视频摄像机与毫米波雷达获取的多目标车辆数据,采用深度学习目标检测算法对视频摄像机的视频图像内多目标车辆进行目标检测并分别提取目标车辆的图像特征与地理空间特征。以车载RTK定位数据D 0作为相对真值,采用多传感器数据融合算法对视频摄像机与毫米波雷达获取的多目标车辆数据进行数据融合,得到多目标车辆定位精度的最佳估计,融合后的定位数据D 3相比于D 1与D 2,具有更高的定位精度; 2) Using the multi-target vehicle data obtained by the video camera and the millimeter-wave radar, the deep learning target detection algorithm is used to detect the multi-target vehicle in the video image of the video camera and extract the image features and geospatial features of the target vehicle respectively. Taking the vehicle RTK positioning data D 0 as the relative true value, the multi-sensor data fusion algorithm is used to fuse the multi-target vehicle data obtained by the video camera and the millimeter-wave radar to obtain the best estimate of the multi-target vehicle positioning accuracy. Compared with D 1 and D 2 , data D 3 has higher positioning accuracy;
利用融合后的多目标车辆数据,对单视频摄像机不同图像帧间多目标车辆与跨视频摄像机图像多目标车辆进行连续追踪与重识别判断。Using the fused multi-target vehicle data, continuous tracking and re-identification judgment of multi-target vehicles between different image frames of a single video camera and multi-target vehicles across video camera images are performed.
本发明具体主要涉及技术问题包括以下几个方面,分别为:The present invention specifically mainly relates to technical problems including the following aspects, which are respectively:
1.RTK差分定位技术;1. RTK differential positioning technology;
2.摄相机标定技术;2. Camera calibration technology;
3.图像数据多目标车辆检测技术;3. Image data multi-target vehicle detection technology;
4.毫米波雷达标定技术;4. Millimeter wave radar calibration technology;
5.多传感器数据时空匹配优化技术;5. Multi-sensor data spatiotemporal matching optimization technology;
6.多传感器数据融合技术;6. Multi-sensor data fusion technology;
7.单视频摄像机图像画面多目标车辆连续追踪与重识别技术;7. Continuous tracking and re-identification technology for multi-target vehicles in a single video camera image;
8.跨视频摄像机图像画面多目标车辆连续追踪与重识别技术;8. Continuous tracking and re-identification technology for multi-target vehicles across video camera images;
本发明解决其技术问题所采用的技术方案具体流程如下所述:The specific flow of the technical solution adopted by the present invention to solve the technical problem is as follows:
1.标定:1. Calibration:
1)摄像机标定:使用手持RTK设备对摄像机图像画面内关键点进行世界地理坐标精准定位,至少选取4个不共线的关键点进行定位。确定关键点在摄像机图像像素坐标系中的像素坐标。计算像素坐标系与世界地理坐标系的单应性变换矩阵,完成摄像机标定。1) Camera calibration: Use a handheld RTK device to accurately locate the key points in the camera image with world geographic coordinates, and select at least 4 non-collinear key points for positioning. Determines the pixel coordinates of the keypoint in the camera image pixel coordinate system. Calculate the homography transformation matrix between the pixel coordinate system and the world geographic coordinate system to complete the camera calibration.
2)目标检测框坐标标定:为实验测试车辆加装车载RTK设备,获取车辆实时定位数 据。将实验测试车辆驶入摄像机视野范围内,使用深度学习目标检测算法在视频图像中为实验车辆标注检测框。以检测框下底边中点像素坐标作为实验测试车辆的像素坐标,根据已计算好的视频摄像机图像像素坐标系与世界地理坐标系对应的单应性变换矩阵计算实验车辆的世界地理坐标,即视频摄像机检测的实验测试车辆定位数据。采集一段时间内实验车辆(时长需大于2分钟,实验车辆多次重复出现在视频摄像机视野范围内)的视频摄像机输出的实验车辆定位数据与同步的车载RTK定位数据D 0,建立D 0与D 1两数据集之间的时空匹配优化模型,求解参数完成视频摄像机目标检测框的坐标标定。3)毫米波雷达标定:根据毫米波雷达原始数据(包括目标相对于毫米波雷达的距离与方位角数据)计算实验车辆相对于毫米波雷达的相对坐标,即实验车辆在雷达坐标系下的坐标。采集一段时间内实验车辆(时长需大于2分钟,实验车辆多次重复出现在毫米波雷达视野范围内)的毫米波雷达获取的实验车辆的定位数据D 2与同步的车载RTK定位数据D 0,计算D 0与D 2两类数据集的单应性变换矩阵与建立两类数据集之间的时空匹配优化模型,求解参数完成毫米波雷达标定。毫米波雷达标定采集数据与目标检测框坐标标定采集数据可同时进行,也可依次进行(顺序不分先后)。 2) Target detection frame coordinate calibration: install on-board RTK equipment for the experimental test vehicle to obtain real-time vehicle positioning data. Drive the experimental test vehicle into the field of view of the camera, and use the deep learning target detection algorithm to mark the detection frame for the experimental vehicle in the video image. Take the pixel coordinates of the midpoint of the bottom edge of the detection frame as the pixel coordinates of the experimental test vehicle, and calculate the world geographic coordinates of the experimental vehicle according to the calculated homography transformation matrix corresponding to the pixel coordinate system of the video camera image and the world geographic coordinate system, that is, Experimental test vehicle localization data for video camera detection. Collect the experimental vehicle positioning data and the synchronized vehicle RTK positioning data D 0 output by the video camera of the experimental vehicle for a period of time (the duration must be greater than 2 minutes, and the experimental vehicle repeatedly appears in the field of view of the video camera), and establish D 0 and D 0 1. The space-time matching optimization model between the two data sets, solve the parameters to complete the coordinate calibration of the target detection frame of the video camera. 3) Millimeter-wave radar calibration: Calculate the relative coordinates of the experimental vehicle relative to the millimeter-wave radar according to the original data of the millimeter-wave radar (including the distance and azimuth data of the target relative to the millimeter-wave radar), that is, the coordinates of the experimental vehicle in the radar coordinate system . Collect the positioning data D 2 of the experimental vehicle and the synchronized vehicle RTK positioning data D 0 obtained by the millimeter-wave radar of the experimental vehicle for a period of time (the duration must be greater than 2 minutes, and the experimental vehicle repeatedly appears in the field of view of the millimeter-wave radar), Calculate the homography transformation matrix of the two data sets D 0 and D 2 and establish a space-time matching optimization model between the two data sets, and solve the parameters to complete the millimeter wave radar calibration. The millimeter wave radar calibration acquisition data and the target detection frame coordinate calibration acquisition data can be performed simultaneously or sequentially (in no particular order).
2.数据获取:利用路侧布设的视频摄像机获取摄像机视野范围内的视频图像数据,利用路侧布设的毫米波雷达获取雷达视野范围内多目标车辆数据,包括各车辆目标ID、毫米波雷达数据(在世界地理坐标系下的定位坐标、速度、航向角等)。将视频摄像机和毫米波雷达获取的数据(包括视频图像数据与毫米波雷达定位数据)传输至中心计算服务器上进行后续计算。2. Data acquisition: use the video camera deployed on the roadside to acquire the video image data within the field of view of the camera, and use the millimeter-wave radar deployed on the roadside to acquire the multi-target vehicle data within the radar field of view, including the target ID of each vehicle and the millimeter-wave radar data (Positioning coordinates, speed, heading angle, etc. in the world geographic coordinate system). The data obtained by the video camera and the millimeter-wave radar (including video image data and millimeter-wave radar positioning data) are transmitted to the central computing server for subsequent calculation.
3.车辆目标识别与特征提取:中心计算服务器接收视频摄像机与毫米波雷达上传的数据(包括视频图像数据与毫米波雷达定位数据)后,进行车辆目标识别与特征提取步骤。对于视频图像数据:使用预训练后的深度学习目标检测算法对摄像机视野范围内的多目标车辆进行目标识别,提取置信度较高(置信度大于0.6)的目标车辆检测框图像(目标检测框框选范围内的视频图像部分,属于完整视频图像的子集),并计算每一个目标车辆图像特征(包括:车辆颜色、车型、尺寸、车牌信息(非必要))与地理空间特征,即定位数据(世界地理坐标)。对于毫米波雷达数据:提取目标车辆地理空间特征,包括:车辆世界地理坐标、速度、航向角等。3. Vehicle target recognition and feature extraction: After the central computing server receives the data uploaded by the video camera and millimeter-wave radar (including video image data and millimeter-wave radar positioning data), it performs the steps of vehicle target recognition and feature extraction. For video image data: use the pre-trained deep learning target detection algorithm to identify the multi-target vehicles within the camera's field of view, and extract the target vehicle detection frame image (target detection frame selection) with high confidence (confidence degree greater than 0.6). The video image part within the range, which belongs to a subset of the complete video image), and calculates the image features of each target vehicle (including: vehicle color, model, size, license plate information (optional)) and geospatial features, that is, positioning data ( world geographic coordinates). For millimeter wave radar data: extract the geospatial features of the target vehicle, including: vehicle world geographic coordinates, speed, heading angle, etc.
4.特征匹配与数据融合:对视频摄像机与毫米波雷达获取的多目标车辆地理空间特征进行匹配,采用多传感器数据融合方法对视频摄像机与毫米波雷达获取的定位数据进行融合,提高定位数据精度。4. Feature matching and data fusion: Match the geospatial features of multi-target vehicles obtained by video cameras and millimeter-wave radars, and use multi-sensor data fusion methods to fuse the positioning data obtained by video cameras and millimeter-wave radars to improve the accuracy of positioning data. .
5.车辆重追踪:5. Vehicle re-tracking:
1)单视频摄像机视野内车辆重追踪:使用预训练好的深度学习目标检测算法识别单视频摄像机视野内车辆并标记目标框,使用Deepsort多目标追踪算法对多目标车辆进行ID赋予,对不同帧之间的目标框进行重识别匹配,实现单视频摄像机视野内车辆 重追踪。1) Re-tracking of vehicles in the field of view of a single video camera: Use the pre-trained deep learning target detection algorithm to identify vehicles in the field of view of a single video camera and mark the target frame, use the Deepsort multi-target tracking algorithm to assign IDs to multi-target vehicles, and use the Deepsort multi-target tracking algorithm to assign IDs to multi-target vehicles. The target frame between them is re-identified and matched to realize the re-tracking of the vehicle within the field of view of the single video camera.
2)跨视频摄像机车辆重追踪:当某一目标车辆V x在摄像机1视野内丢失时,车辆重识别系统记录该目标车辆V x丢失前车辆图像特征与地理空间特征,并传入丢失车辆数据库,所述丢失的目标车辆V x进入摄像机2视野内时,车辆重识别系统捕捉车辆特征(包括图像特征与地理空间特征)并在车辆丢失数据库内进行特征匹配,对相似度高(相似度大于0.5)的车辆进行ID重赋予,在ID重赋予时,取与丢失数据库中相似度最高的丢失车辆ID进行ID重赋予,实现车辆重追踪。若车辆丢失数据库内未发现匹配车辆(即与车辆丢失数据库所有丢失车辆特征相似度均低于0.5),则对目标车辆V x赋予新的ID。 2) Vehicle re-tracking across video cameras: when a target vehicle V x is lost in the field of view of camera 1, the vehicle re-identification system records the vehicle image features and geospatial features before the target vehicle V x is lost, and transfers it to the lost vehicle database , when the lost target vehicle V x enters the field of view of the camera 2, the vehicle re-identification system captures vehicle features (including image features and geospatial features) and performs feature matching in the vehicle lost database. 0.5), the ID re-assignment is performed. When the ID is re-assigned, the lost vehicle ID with the highest similarity in the lost database is taken for ID re-assignment to realize the vehicle re-tracking. If no matching vehicle is found in the vehicle lost database (that is, the similarity of all lost vehicle features with the vehicle lost database is lower than 0.5), a new ID is assigned to the target vehicle V x .
附图简要说明Brief Description of Drawings
图1传统车辆重识别技术路线流程图;Figure 1 Flow chart of traditional vehicle re-identification technology route;
图2本发明车辆重识别技术路线流程图;2 is a flow chart of the vehicle re-identification technology route of the present invention;
图3为传感器布设示意图;Figure 3 is a schematic diagram of sensor layout;
图4为摄像机标定示意图;4 is a schematic diagram of camera calibration;
图5为摄像机标定图像示意图;5 is a schematic diagram of a camera calibration image;
图6为摄像机图像检车目标框坐标转换示意图;6 is a schematic diagram of coordinate conversion of a camera image vehicle inspection target frame;
图7为目标检测框数据与车载RTK定位数据时间同步优化模型流程图;Fig. 7 is the flow chart of the time synchronization optimization model between the target detection frame data and the vehicle RTK positioning data;
图8为目标检测框数据与车载RTK定位数据空间误差校准流程图;Figure 8 is a flow chart of spatial error calibration between target detection frame data and vehicle RTK positioning data;
图9为毫米波雷达旋转偏移角与目标检测示意图;Figure 9 is a schematic diagram of the rotation offset angle and target detection of the millimeter-wave radar;
图10为毫米波雷达检测目标世界地理坐标计算示意图;Figure 10 is a schematic diagram of the calculation of the world geographic coordinates of the target detected by the millimeter wave radar;
图11为毫米波雷达检测数据与车载RTK定位数据时间同步优化模型流程图;Figure 11 is the flow chart of the optimization model for time synchronization between millimeter-wave radar detection data and vehicle RTK positioning data;
图12为毫米波雷达检测数据与车载RTK定位数据空间校准流程图;Figure 12 is a flow chart of spatial calibration of millimeter-wave radar detection data and vehicle RTK positioning data;
图13为单视频摄像机车辆重识别示意图;13 is a schematic diagram of vehicle re-identification with a single video camera;
图14为跨视频摄像机车辆识别流程图;Figure 14 is a flow chart of vehicle recognition across video cameras;
图15为跨视频摄像机车辆重识别示意图。FIG. 15 is a schematic diagram of vehicle re-identification across video cameras.
具体实施方式Detailed ways
下面结合附图和具体实施方式对本发明作详细说明。The present invention will be described in detail below with reference to the accompanying drawings and specific embodiments.
本发明涉及一种基于雷视融合的多目标车辆检测及重识别方法,总体技术路线如附图2所示,具体过程分为五个步骤,包括:传感器的布设与数据采集、传感器标定(视频摄像机与毫米波雷达)、多传感器数据融合和车辆重识别方法(单视频摄像机与跨视频摄像机)。The invention relates to a multi-target vehicle detection and re-identification method based on Raivision fusion. The overall technical route is shown in Figure 2. The specific process is divided into five steps, including: sensor layout and data collection, sensor calibration (video). camera and mmWave radar), multi-sensor data fusion, and vehicle re-identification methods (single video camera vs. cross-video camera).
第一步:传感器布设与数据采集Step 1: Sensor layout and data collection
本发明依靠视频摄像机与毫米波雷达两类传感器数据实现多车辆目标检测、追踪、数据融合与车辆重识别。视频摄像机采用定向式摄像机,毫米波雷达采用79GHz频段的长距离雷达。基础布设方案如附图3所示,视频摄像机与毫米波雷达布设于路侧立杆同一位置(保证两传感器经纬度坐标距离小于0.5米),使得两者世界地理坐标保持一致。传感器探测范围可根据安装 高度与角度进行调整,当传感器安装高度为6米,向下倾斜俯角为10°且场景无其他遮挡物时,其视野范围可达100-150m。两类传感器均以25Hz的频率采集数据,接入中心服务器,进行数据保存与处理。对于视频图像数据:使用预训练后的深度学习目标检测算法对摄像机视野范围内的多目标车辆进行目标识别,并提取每一个目标车辆图像特征(包括:车辆颜色、车型、尺寸、车牌信息(非必要))与地理空间特征(世界地理坐标)。对于毫米波雷达数据:提取目标车辆地理空间特征(包括:车辆世界地理坐标、速度、航向角等)。The invention realizes multi-vehicle target detection, tracking, data fusion and vehicle re-identification by relying on two types of sensor data of video cameras and millimeter-wave radars. Video cameras use directional cameras, and millimeter-wave radars use long-range radars in the 79GHz band. The basic layout plan is shown in Figure 3. The video camera and the millimeter-wave radar are placed in the same position on the roadside pole (to ensure that the distance between the longitude and latitude coordinates of the two sensors is less than 0.5 meters), so that the world geographic coordinates of the two are consistent. The detection range of the sensor can be adjusted according to the installation height and angle. When the installation height of the sensor is 6 meters, the downward tilt angle is 10° and the scene has no other obstructions, its field of view can reach 100-150m. Both types of sensors collect data at a frequency of 25Hz and access the central server for data storage and processing. For video image data: use the pre-trained deep learning target detection algorithm to identify the multi-target vehicles within the camera's field of view, and extract the image features of each target vehicle (including: vehicle color, model, size, license plate information (non- necessary)) and geospatial features (world geographic coordinates). For millimeter wave radar data: extract the geospatial features of the target vehicle (including: vehicle world geographic coordinates, speed, heading angle, etc.).
变种方案A:毫米波雷达与视频摄像机布设于不同的路侧立杆上,保证两传感器视野重合率大于90%。该路段感知区域的感知范围为视频摄像机和毫米波雷达视野范围的交集。Variant A: The millimeter-wave radar and the video camera are arranged on different roadside poles to ensure that the overlap rate of the two sensors is greater than 90%. The sensing range of the sensing area of this road segment is the intersection of the visual range of the video camera and the millimeter-wave radar.
变种方案B:毫米波雷达与视频摄像机布设于道路上方同一龙门架处,保证两传感器视野重合率大于90%。该路段感知区域的感知范围为视频摄像机和毫米波雷达视野范围的交集。Variant B: The millimeter-wave radar and the video camera are arranged on the same gantry above the road to ensure that the overlap rate of the two sensors is greater than 90%. The sensing range of the sensing area of this road segment is the intersection of the visual range of the video camera and the millimeter-wave radar.
变种方案C:毫米波雷达与视频摄像机布设于道路上方不同龙门架处,保证两传感器视野重合率大于90%。该路段感知区域的感知范围为视频摄像机和毫米波雷达视野范围的交集。Variant C: The millimeter-wave radar and the video camera are arranged on different gantry above the road to ensure that the overlap rate of the two sensors is greater than 90%. The sensing range of the sensing area of this road segment is the intersection of the visual range of the video camera and the millimeter-wave radar.
第二步:标定Step 2: Calibration
(1)视频摄像机标定(1) Video camera calibration
本发明中视频摄像机的标定指建立摄像机图像画面中像素坐标系与世界地理坐标系的映射关系,即单应性变换矩阵H,如式(1)所示。世界地理坐标系为在环境中还选择一个参考坐标系来描述摄像机和物体的位置,该坐标系称为世界地理坐标系。本发明中规定世界地理坐标系为地理坐标系中WGS-84坐标系。本发明中,世界地理坐标由RTK设备获取。高精度的GPS测量必须采用载波相位观测值,RTK设备采用差分定位技术,即基于载波相位观测值的实时动态定位技术,它能够实时地提供测站点在指定坐标系中的三维定位结果,并达到厘米级精度。因其定位精度高,故本发明中将通过RTK获取的世界地理坐标视为相对真值。像素坐标系表示图像像素在图像中的位置,通常情况下,以图像左上角像素为原点,规定右方向为x轴正方向,下方向为y轴正方向,像素横纵坐标分别表示该像素距离y轴与x轴的像素个数。式(2)描述了世界地理坐标系与像素坐标系通过单应性矩阵变换得到,其中longitude代表经度,latitude代表纬度,x代表像素坐标横坐标,y代表像素坐标横坐标。式(3)与式(4)为转换计算公式。The calibration of the video camera in the present invention refers to establishing the mapping relationship between the pixel coordinate system in the camera image and the world geographic coordinate system, that is, the homography transformation matrix H, as shown in formula (1). The world geographic coordinate system is to select a reference coordinate system in the environment to describe the position of the camera and the object, which is called the world geographic coordinate system. The present invention stipulates that the world geographic coordinate system is the WGS-84 coordinate system in the geographic coordinate system. In the present invention, the world geographic coordinates are acquired by RTK equipment. High-precision GPS measurement must use carrier phase observations. RTK equipment uses differential positioning technology, that is, real-time dynamic positioning technology based on carrier phase observations. It can provide real-time three-dimensional positioning results of the station in the specified coordinate system, and achieve Centimeter-level accuracy. Because of its high positioning accuracy, the present invention regards the world geographic coordinates obtained through RTK as a relative true value. The pixel coordinate system represents the position of the image pixel in the image. Usually, the pixel in the upper left corner of the image is used as the origin, and the right direction is the positive direction of the x-axis, and the downward direction is the positive direction of the y-axis. The horizontal and vertical coordinates of the pixels represent the distance of the pixel respectively. The number of pixels in the y-axis and x-axis. Equation (2) describes that the world geographic coordinate system and the pixel coordinate system are obtained through homography matrix transformation, where longitude represents longitude, latitude represents latitude, x represents the abscissa of the pixel coordinate, and y represents the abscissa of the pixel coordinate. Formula (3) and formula (4) are conversion calculation formulas.
Figure PCTCN2021085150-appb-000001
Figure PCTCN2021085150-appb-000001
Figure PCTCN2021085150-appb-000002
Figure PCTCN2021085150-appb-000002
Figure PCTCN2021085150-appb-000003
Figure PCTCN2021085150-appb-000003
Figure PCTCN2021085150-appb-000004
Figure PCTCN2021085150-appb-000004
如附图4与附图5所示,将手持RTK移动站放入视频摄像机视野范围内,分别选取至少4个不共线的关键点采集其世界地理坐标。在摄像机图像中选取手持RTK移动站手持杆底部像素坐标作为定位点对应的像素坐标。最后,计算两坐标系之间的单应性变换矩阵,建立视频图像像素坐标与世界地理坐标的映射关系,完成视频摄像机参数标定。As shown in FIG. 4 and FIG. 5 , place the handheld RTK mobile station within the field of view of the video camera, and select at least four non-collinear key points to collect its world geographic coordinates. In the camera image, select the pixel coordinates of the bottom of the handheld RTK mobile station handheld stick as the pixel coordinates corresponding to the positioning point. Finally, the homography transformation matrix between the two coordinate systems is calculated, the mapping relationship between the pixel coordinates of the video image and the world geographic coordinates is established, and the parameter calibration of the video camera is completed.
(3)目标检测框坐标标定(3) Coordinate calibration of target detection frame
目标检测框坐标标定为建立摄像机图像中车辆目标检测框像素坐标与车辆世界地理坐标之间的映射关系。其中,车辆目标检测框像素坐标指检测框下底边中点像素的像素坐标。(摄像机位于道路来向正上方)The target detection frame coordinate calibration is to establish the mapping relationship between the pixel coordinates of the vehicle target detection frame in the camera image and the vehicle world geographic coordinates. The pixel coordinates of the vehicle target detection frame refer to the pixel coordinates of the midpoint pixel of the lower bottom edge of the detection frame. (The camera is located directly above the road coming from the road)
变种方案A:摄像机位于道路来向左上方,车辆目标检测框像素坐标指检测框右下方顶点像素的像素坐标。Variant A: The camera is located at the upper left of the road, and the pixel coordinates of the vehicle target detection frame refer to the pixel coordinates of the lower right vertex pixel of the detection frame.
变种方案B:摄像机位于道路来向右上方,车辆目标检测框像素坐标指检测框左下方顶点像素的像素坐标。Variant B: The camera is located on the road to the upper right, and the pixel coordinates of the vehicle target detection frame refer to the pixel coordinates of the vertex pixels at the lower left of the detection frame.
目标检测框由深度学习目标检测算法获取,数据频率为25Hz,车辆世界地理坐标由车载RTK获取,数据频率为5Hz。为实验测试车辆加装车载RTK获取车辆实时世界地理坐标数据,将车辆驶入摄像机视野范围内,使用深度学习目标检测算法在视频图像中为实验车辆标注检测,获取其像素坐标。根据已计算好的摄像机检测目标像素坐标与世界地理坐标对应的单应性变换矩阵计算检测目标车辆世界地理坐标。不同传感器之间时钟不同步是普遍存在的现象,此外,目标检测框像素坐标转换后的世界地理坐标与车辆真实世界地理坐标存在差异,需要建立两类世界地理坐标的时空匹配优化模型,包括时间同步优化模型与空间误差校准两部分。The target detection frame is obtained by the deep learning target detection algorithm, the data frequency is 25Hz, and the vehicle world geographic coordinates are obtained by the vehicle RTK, and the data frequency is 5Hz. Install on-board RTK for the experimental test vehicle to obtain the real-time world geographic coordinate data of the vehicle, drive the vehicle into the field of view of the camera, and use the deep learning target detection algorithm to mark and detect the experimental vehicle in the video image to obtain its pixel coordinates. According to the calculated homography transformation matrix corresponding to the pixel coordinates of the camera detection target and the world geographical coordinates, the world geographical coordinates of the detected target vehicle are calculated. It is a common phenomenon that the clocks between different sensors are out of synchronization. In addition, the world geographic coordinates converted from the pixel coordinates of the target detection frame are different from the real world geographic coordinates of the vehicle. It is necessary to establish a space-time matching optimization model for two types of world geographic coordinates, including time There are two parts of synchronous optimization model and spatial error calibration.
如附图7所示,在时间同步优化模型,首先需要统一两类数据采集频率。采用线性插值的方法对车载RTK获取的世界地理坐标进行上采样,使其采样频率与目标检测框数据频率一致。建立时间同步优化模型目标函数如式(5)所示,采用最优化算法寻找一组合适的参数Δt使两世界地理坐标在时间序列上的欧式距离达到最小。As shown in FIG. 7 , in the time synchronization optimization model, it is first necessary to unify the two types of data collection frequencies. The linear interpolation method is used to upsample the world geographic coordinates obtained by the vehicle RTK, so that the sampling frequency is consistent with the data frequency of the target detection frame. The objective function of the time synchronization optimization model is established as shown in formula (5). The optimization algorithm is used to find a set of suitable parameters Δt to minimize the Euclidean distance between the two world geographic coordinates in the time series.
Figure PCTCN2021085150-appb-000005
Figure PCTCN2021085150-appb-000005
其中,longitude bb_t和latitude bb_t代表t时刻目标检测框像素坐标对应的世界地理坐标,longitude rtk_t+Δt和latitude rtk_t+Δt代表t+Δt时刻车载RTK测量的世界地理坐标。 Among them, longitude bb_t and latitude bb_t represent the world geographic coordinates corresponding to the pixel coordinates of the target detection frame at time t, and longitude rtk_t+Δt and latitude rtk_t+Δt represent the world geographic coordinates measured by the vehicle RTK at time t+Δt.
如附图8所示,在空间误差校准中,建立两类世界地理坐标在像素坐标系上的空间误差分布,误差计算公式如式(6)所示,通过曲面拟合的方法计算误差空间分布曲面,按照式(7)对目标检测框计算的世界地理坐标进行校正。As shown in Figure 8, in the spatial error calibration, the spatial error distribution of two types of world geographic coordinates on the pixel coordinate system is established. The error calculation formula is shown in formula (6), and the error spatial distribution is calculated by the method of surface fitting. Curved surface, the world geographic coordinates calculated by the target detection frame are corrected according to formula (7).
Error=(longitude bb_t,latitude bb_t)-(longitude rtk_t+Δt,latitude rtk_t+Δt)   (6) Error=(longitude bb_t , latitude bb_t )-(longitude rtk_t+Δt , latitude rtk_t+Δt ) (6)
coordinate correction=(longitude bb_t,latitude bb_t)-Error Curced Surface   (7) coordinate correction = (longitude bb_t , latitude bb_t )-Error Curced Surface (7)
(3)毫米波雷达标定(3) Millimeter wave radar calibration
根据毫米波雷达适合检测移动目标的特性,使用安装了车载RTK的实验车辆对毫米波雷达进行标定。雷达返回数据为目标级数据,包含检测目标的距离,方位角,速度等。附图9展示了毫米波雷达目标检测示意图,根据毫米波雷达原始数据计算实验车辆相对于雷达的相对坐标,计算公式如式(8)~式(10)所示。According to the characteristics of the millimeter-wave radar suitable for detecting moving targets, the millimeter-wave radar was calibrated using an experimental vehicle equipped with an on-board RTK. The data returned by the radar is target-level data, including the distance, azimuth, speed, etc. of the detected target. Figure 9 shows a schematic diagram of millimeter-wave radar target detection. The relative coordinates of the experimental vehicle relative to the radar are calculated according to the original data of the millimeter-wave radar. The calculation formulas are shown in equations (8) to (10).
x=dis·cosθ 1·cosθ 2       (8) x=dis·cosθ 1 ·cosθ 2 (8)
y=dis·cosθ 1·cosθ 2       (9) y=dis·cosθ 1 ·cosθ 2 (9)
z=-dis·cosθ 1               (10) z=-dis·cosθ 1 (10)
其中,distance表示目标车辆与雷达之间的距离,θ 1和θ 2分别表示目标车辆与雷达之间的俯仰角和水平角,x,y和z表示雷达检测目标在雷达坐标系中的坐标。 Among them, distance represents the distance between the target vehicle and the radar, θ 1 and θ 2 represent the pitch angle and horizontal angle between the target vehicle and the radar, respectively, and x, y and z represent the coordinates of the radar detection target in the radar coordinate system.
如附图10所示,雷达坐标系与世界地理坐标系之间存在位姿角度偏差与空间位置偏差,其中位姿角度偏差包含3个方向的偏差角,分别记为α,β和γ。空间位置偏差为毫米波雷达坐标系原点与世界地理坐标系原点之间的偏差,即为毫米波雷达在世界地理坐标系中的坐标,分别记为longitude radar,latitude radar和height radarAs shown in FIG. 10 , there are pose angle deviations and spatial position deviations between the radar coordinate system and the world geographic coordinate system, where the pose angle deviations include deviation angles in three directions, denoted as α, β and γ, respectively. The spatial position deviation is the deviation between the origin of the millimeter-wave radar coordinate system and the origin of the world geographic coordinate system, that is, the coordinates of the millimeter-wave radar in the world geographic coordinate system, which are respectively recorded as longitude radar , latitude radar and height radar .
3个方向的位姿偏差角修正矩阵分别如式(11)、式(12)和式(13)所示。The pose deviation angle correction matrices of the three directions are shown in Equation (11), Equation (12) and Equation (13), respectively.
Figure PCTCN2021085150-appb-000006
Figure PCTCN2021085150-appb-000006
Figure PCTCN2021085150-appb-000007
Figure PCTCN2021085150-appb-000007
Figure PCTCN2021085150-appb-000008
Figure PCTCN2021085150-appb-000008
雷达坐标系转换为世界地理坐标系的转换公式如式(14)所示。The conversion formula from the radar coordinate system to the world geographic coordinate system is shown in formula (14).
Figure PCTCN2021085150-appb-000009
Figure PCTCN2021085150-appb-000009
其中,x,y和z表示雷达检测目标在雷达坐标系中的坐标,longitude,latitude和height表示雷达检测目标在雷达坐标系中的坐标转换成世界地理坐标系中的世界地理坐标。Among them, x, y and z represent the coordinates of the radar detection target in the radar coordinate system, and longitude, latitude and height represent the transformation of the coordinates of the radar detection target in the radar coordinate system into the world geographic coordinates in the world geographic coordinate system.
记longitude,latitude和height为由雷达坐标计算得到的检测目标的世界地理坐标,同时得到车载RTK获取的车辆世界地理坐标,记为longitude rtk,latitude rtk和height rtk。建立两类世界地理坐标的时空匹配优化模型,包括时间同步优化模型与空间误差校准两部分。 Denote longitude, latitude and height as the world geographic coordinates of the detected target calculated from the radar coordinates, and at the same time obtain the vehicle world geographic coordinates obtained by the vehicle RTK, denoted as longitude rtk , latitude rtk and height rtk . Two types of space-time matching optimization models of world geographic coordinates are established, including two parts: time synchronization optimization model and spatial error calibration.
如附图11所示,在时间同步优化模型,首先需要统一两类数据采集频率。采用线性插值的方法对车载RTK获取的世界地理坐标进行上采样,使其采样频率与目标检测框数据频率一致。建立目标函数如式(5)所示,采用最优化算法寻找一组合适的参数α,β,γ和Δt,使两世界地理坐标在时间序列上的欧式距离达到最小。As shown in FIG. 11 , in the time synchronization optimization model, it is first necessary to unify the two types of data collection frequencies. The linear interpolation method is used to upsample the world geographic coordinates obtained by the vehicle RTK, so that the sampling frequency is consistent with the data frequency of the target detection frame. The objective function is established as shown in formula (5), and the optimization algorithm is used to find a set of suitable parameters α, β, γ and Δt to minimize the Euclidean distance between the two world geographic coordinates in the time series.
Figure PCTCN2021085150-appb-000010
Figure PCTCN2021085150-appb-000010
其中,longitude t和latitude t表示t时刻雷达检测目标在雷达坐标系中的坐标转换成世界地理坐标系中的世界地理坐标,longitude rtk_t+Δt和lattitude rtk_t+Δt表示t+Δt时刻车载RTK获取的车辆的世界地理坐标。 Among them, longitude t and latitude t represent that the coordinates of the radar detection target in the radar coordinate system at time t are converted into world geographic coordinates in the world geographic coordinate system, and longitude rtk_t+Δt and lattitude rtk_t+Δt represent the time t+Δt obtained by the vehicle RTK The world geographic coordinates of the vehicle.
如附图12所示,在空间误差校准中,建立两类世界地理坐标在像素坐标系上的空间误差分布,误差计算公式如式(14)所示,通过曲面拟合的方法计算误差空间分布曲面,按照式(15)对目标检测框计算的世界地理坐标进行校正。As shown in Figure 12, in the spatial error calibration, the spatial error distribution of two types of world geographic coordinates on the pixel coordinate system is established. The error calculation formula is shown in formula (14), and the error spatial distribution is calculated by the method of surface fitting. Curved surface, the world geographic coordinates calculated by the target detection frame are corrected according to formula (15).
Error=(longitude t,latitude t)-(longitude rtk_t+Δt,latitude rtk_t+Δt)   (16) Error=(longitude t , latitude t )-(longitude rtk_t+Δt , latitude rtk_t+Δt ) (16)
coordinate correction=(longitude t,latitude t)-Error Curved Surface  (17) coordinate correction = (longitude t , latitude t )-Error Curved Surface (17)
第三步:视频图像目标识别与特征提取Step 3: Video image target recognition and feature extraction
使用预训练后的深度学习目标检测算法对摄像机视野范围内的多目标车辆进行目标识别,并提取每一个目标车辆图像特征(包括:车辆颜色、车型、尺寸、车牌信息(非必要))与地理空间特征(世界地理坐标),当车牌信息无法获取时,该栏填为空值。对于毫米波雷达数据:提取目标车辆地理空间特征(包括:车辆世界地理坐标、速度、航向角等)。Use the pre-trained deep learning target detection algorithm to perform target recognition on multi-target vehicles within the camera's field of view, and extract the image features of each target vehicle (including: vehicle color, model, size, license plate information (optional)) and geographic location. Spatial features (world geographic coordinates), when the license plate information cannot be obtained, this column is filled with a blank value. For millimeter wave radar data: extract the geospatial features of the target vehicle (including: vehicle world geographic coordinates, speed, heading angle, etc.).
第四步:特征匹配与数据融合Step 4: Feature matching and data fusion
对视频与毫米波雷达获取的目标车辆数据根据地理空间特征进行匹配,采用多传感器数据融合方法对视频与毫米波雷达获取的车辆速度、地理位置数据进行融合,提高数据精度。其中,融合算法中,以实验车辆的车载RTK获取的高精度定位坐标作为参考真值,采用如卡尔曼滤波、多贝叶斯估计、模糊逻辑推理与深度神经网络等传感器数据融合的方法对多目标车辆的地理位置数据进行融合计算。The video and the target vehicle data obtained by the millimeter-wave radar are matched according to the geospatial features, and the multi-sensor data fusion method is used to fuse the video and the vehicle speed and geographic location data obtained by the millimeter-wave radar to improve the data accuracy. Among them, in the fusion algorithm, the high-precision positioning coordinates obtained by the on-board RTK of the experimental vehicle are used as the reference truth value, and sensor data fusion methods such as Kalman filtering, multi-Bayesian estimation, fuzzy logic reasoning and deep neural network are used to The geographic location data of the target vehicle is fused to calculate.
其中,卡尔曼滤波算法是在多源数据融合中广泛应用的算法之一。卡尔曼滤波是一种递推 式滤波算法,其特点为不需要保存过去的历史信息,新数据结合前一帧(或前一时刻)已求得的估计值及系统本身的状态方程按一定方式求得新的估计值。卡尔曼滤波原理可以用以下5个公式表达:Among them, the Kalman filter algorithm is one of the algorithms widely used in multi-source data fusion. Kalman filtering is a recursive filtering algorithm, which is characterized by no need to save past historical information, and the new data is combined with the estimated value obtained in the previous frame (or previous moment) and the state equation of the system itself according to a certain method. Find a new estimate. The principle of Kalman filter can be expressed by the following five formulas:
预测:predict:
Figure PCTCN2021085150-appb-000011
Figure PCTCN2021085150-appb-000011
Figure PCTCN2021085150-appb-000012
Figure PCTCN2021085150-appb-000012
更新:renew:
Figure PCTCN2021085150-appb-000013
Figure PCTCN2021085150-appb-000013
Figure PCTCN2021085150-appb-000014
Figure PCTCN2021085150-appb-000014
Figure PCTCN2021085150-appb-000015
Figure PCTCN2021085150-appb-000015
其中,各参数含义如下:The meaning of each parameter is as follows:
F:状态转移矩阵;F: state transition matrix;
B:控制矩阵;B: control matrix;
P:状态协方差矩阵;P: state covariance matrix;
Q:状态转移协方差矩阵;Q: state transition covariance matrix;
H:观测矩阵;H: observation matrix;
R:观测噪声方差;R: observation noise variance;
Figure PCTCN2021085150-appb-000016
由t-1时刻推测此时刻的状态变量,还未更具此时刻观测值做修正;
Figure PCTCN2021085150-appb-000016
The state variable at this time is inferred from time t-1, and the observation value at this time has not been corrected;
Figure PCTCN2021085150-appb-000017
由上一时刻推测此时刻的状态变量,已更具此时刻观测值做修正;
Figure PCTCN2021085150-appb-000017
The state variable at this moment is estimated from the previous moment, and the observed value at this moment has been corrected;
z:实际观测值;z: actual observed value;
k:卡尔曼系数;k: Kalman coefficient;
t:第t时刻。t: time t.
第五步:车辆重追踪Step 5: Vehicle re-tracking
1)单视频摄像机视野内车辆重追踪:使用预训练好的深度学习目标检测算法识别单视频摄像机视野内车辆并标记目标框,使用Deepsort多目标追踪算法对多目标车辆进行ID赋予与对不同帧之间的目标框进行重识别匹配,实现单视频摄像机视野内车辆重追踪。重识别结果示意图如附图13所示。1) Re-tracking of vehicles in the field of view of a single video camera: Use the pre-trained deep learning target detection algorithm to identify vehicles in the field of view of a single video camera and mark the target frame, and use the Deepsort multi-target tracking algorithm to assign IDs to multi-target vehicles and assign different frames to different frames. The target frame between them is re-identified and matched to realize the re-tracking of the vehicle within the field of view of the single video camera. A schematic diagram of the re-identification results is shown in FIG. 13 .
其中Deepsort算法由Nicolai Wojke等人提出,该算法在Sort算法改进而来。Sort与Deepsort算法是多目标追踪(Multiple Object Tracking MOT)中的常用算法,其中Sort算法使用简单的卡尔曼滤波处理逐帧数据的关联性以及使用匈牙利算法进行关联度量,这种简单的算法在高帧速率下获得了良好的性能。但由于Sort算法忽略了被检测物体的表面特征,因此只有在物体状态估计不确定性较低是才会准确,在Deepsort算法中,使用更加可靠的度量来代替关联度量,并使用CNN神经网络在大规模数据集进行训练,并提取特征,增加网络对遗失和障碍的鲁棒性。The Deepsort algorithm was proposed by Nicolai Wojke et al., which was improved from the Sort algorithm. Sort and Deepsort algorithms are commonly used algorithms in Multiple Object Tracking MOT. Good performance is obtained at frame rates. However, since the Sort algorithm ignores the surface features of the detected object, it is only accurate when the uncertainty of the state estimation of the object is low. Train on large-scale datasets and extract features that increase the network's robustness to loss and obstacles.
Deepsort算法中将构建检测器与追踪器用于目标重识别。其中检测器即由目标检测框输入,追踪器用于目标的追踪与重识别。Deepsort算法输入包括:目标检测框、目标检测置信度、目标特征(图像特征与运动特征)。目标检测置信度主要用于进行一部分的检测框的筛选,目标检测框与目标特征用于追踪器构建与后续追踪计算。在算法预测模块中,会对追踪器使用卡尔曼滤波器进行预测,其中卡尔曼滤波器构建的模型为匀速运动和线性观测模型。算法更新模块中包括匹配,追踪器更新与目标特征集更新。In the Deepsort algorithm, a detector and a tracker are constructed for object re-identification. The detector is input from the target detection frame, and the tracker is used for target tracking and re-identification. The input of the Deepsort algorithm includes: target detection frame, target detection confidence, target features (image features and motion features). The target detection confidence is mainly used to screen a part of the detection frame, and the target detection frame and target features are used for tracker construction and subsequent tracking calculation. In the algorithm prediction module, the Kalman filter is used to predict the tracker, and the model constructed by the Kalman filter is a uniform motion and linear observation model. The algorithm update module includes matching, tracker update and target feature set update.
级联匹配算法:针对每一个检测器都会分配一个追踪器,每个追踪器会设定一个计时器参数。如果追踪器完成匹配并进行更新,那么计时器参数会重置为0,否则就会递增1个单位值。在级联匹配中,会根据计时器参数来对追踪器分先后顺序,参数小的先来匹配,参数大的后匹配。也就是给上一帧最先匹配的追踪器高的优先权,给好几帧都没匹配上的追踪器降低优先权。Cascade matching algorithm: A tracker is assigned to each detector, and each tracker is set with a timer parameter. If the tracker is matched and updated, the timer parameter is reset to 0, otherwise it is incremented by 1 unit. In cascading matching, the trackers will be sorted according to the timer parameters. Smaller parameters will be matched first, and larger parameters will be matched later. That is, the tracker that matches the first frame in the previous frame is given high priority, and the tracker that has not been matched for several frames has a lower priority.
特征比对:引入马氏距离与余弦距离用于针对运动信息与图像信息的比对。马氏距离规避了欧氏距离中对于数据特征方差不同的风险,在计算中添加了协方差矩阵,其目的为对方差进行归一化,从而使距离值更加符合数据特征以及实际意义。马氏距离是对于差异度的衡量中,的一种距离度量方式,而不同于马氏距离,余弦距离则是一种相似度度量方式。前者是针对于位置进行区分,而后者则是针对于方向。使用余弦距离时,可以用来衡量不同个体在维度之间的差异。结合两类特征距离,在整体上可以使得计算结果相对全面地衡量不同目标特征差异性。Feature comparison: Mahalanobis distance and cosine distance are introduced for comparison between motion information and image information. Mahalanobis distance avoids the risk of different variances of data features in Euclidean distance. A covariance matrix is added in the calculation to normalize the variance, so that the distance value is more in line with the data features and practical significance. Mahalanobis distance is a distance measure in the measure of difference, and different from Mahalanobis distance, cosine distance is a similarity measure. The former is based on location, while the latter is based on direction. When using cosine distance, it can be used to measure the differences between different individuals in dimensions. Combining the two types of feature distances can make the calculation results relatively comprehensive to measure the difference of different target features.
特征匹配:基于马氏距离与余弦距离特征,对不同图像帧之间进行目标特征匹配,完成目标重识别(ID传递)的过程。Feature matching: Based on the Mahalanobis distance and cosine distance features, target features are matched between different image frames to complete the process of target re-identification (ID transfer).
2)跨视频摄像机车辆重追踪:车辆跨视频摄像机重追踪总体逻辑流程图如附图14所示。当目标车辆在摄像机1视野内丢失时,系统记录其丢失前n帧(n≥10)车辆图像特征、地理空间特征,并传入丢失车辆数据库,其中,地理空间特征除世界地理坐标外,包含毫米波雷达采集的车辆速度与行驶方向特征。车辆丢失后,采用线性拟合、卡尔曼滤波等方法计算车辆的预测车速,并根据车辆丢失前行驶方向计算车辆的预测位置,动态更新丢失数据库。目标车辆进入摄像机2视野内时,系统捕捉车辆目标出现时前n帧(n≥10)特征并在车辆丢失数据库内进行特征匹配,对相似度高的车辆进行ID重赋予,实现车辆的重追踪。若车辆丢失数据库内无匹配车辆,则对目标车辆赋予新的ID。在本发明中,车牌信息不是必须的,若在环境因素良好的情况(光照、拍摄角度等)时,摄像机能识别部分车牌信息(如关键字符、车牌颜色等),按照单摄像机视野范围内连续帧的识别结果置信度赋予其动态权重,可作为多目标车辆重追踪的校核依据。最终,实现多目标车辆的跨视频摄像机重追踪,其示意结果如附图15所示。2) Cross-video camera vehicle re-tracking: The overall logic flow chart of vehicle cross-video camera re-tracking is shown in FIG. 14 . When the target vehicle is lost within the field of view of the camera 1, the system records the vehicle image features and geospatial features of the n frames (n≥10) before the loss, and transmits it to the lost vehicle database. In addition to the world geographic coordinates, the geospatial features include Vehicle speed and driving direction characteristics collected by millimeter-wave radar. After the vehicle is lost, linear fitting, Kalman filtering and other methods are used to calculate the predicted speed of the vehicle, and the predicted position of the vehicle is calculated according to the driving direction before the vehicle is lost, and the lost database is dynamically updated. When the target vehicle enters the field of view of camera 2, the system captures the features of the first n frames (n≥10) when the vehicle target appears and performs feature matching in the vehicle loss database, and assigns IDs to vehicles with high similarity to realize re-tracking of vehicles . If there is no matching vehicle in the vehicle lost database, a new ID is assigned to the target vehicle. In the present invention, the license plate information is not necessary. If the environmental factors are good (illumination, shooting angle, etc.), the camera can recognize part of the license plate information (such as key characters, license plate color, etc.) The confidence of the recognition result of the frame is given its dynamic weight, which can be used as the verification basis for multi-target vehicle re-tracking. Finally, cross-video camera re-tracking of multi-target vehicles is achieved, and the schematic results are shown in Figure 15.

Claims (11)

  1. 一种基于雷视融合的多目标车辆检测及重识别方法,采用预先布设的至少两个视频摄像机和至少两个毫米波雷达,包括如下步骤:A multi-target vehicle detection and re-identification method based on Raivision fusion, using at least two video cameras and at least two millimeter-wave radars pre-arranged, including the following steps:
    1)标定:1) Calibration:
    1.1)视频摄像机标定;1.1) Video camera calibration;
    1.2)目标检测框坐标标定;1.2) Coordinate calibration of target detection frame;
    1.3)毫米波雷达标定;1.3) Millimeter wave radar calibration;
    2)数据获取:2) Data acquisition:
    2.1)利用路侧布设的视频摄像机获取摄像机视野范围内的视频图像数据;2.1) Use the video camera deployed on the roadside to obtain the video image data within the field of view of the camera;
    2.2)利用路侧布设的毫米波雷达获取雷达视野范围内多目标车辆数据;2.2) Use the millimeter-wave radar deployed on the roadside to obtain multi-target vehicle data within the radar field of view;
    3)车辆目标识别与特征提取:中心计算服务器接收到所述视频图像数据和所述多目标车辆数据后,进行车辆目标识别与特征提取;3) Vehicle target recognition and feature extraction: After receiving the video image data and the multi-target vehicle data, the central computing server performs vehicle target recognition and feature extraction;
    3.1)视频图像数据特征提取;3.1) Feature extraction of video image data;
    3.2)毫米波雷达数据特征提取;3.2) Feature extraction of millimeter wave radar data;
    4)特征匹配与数据融合:对视频摄像机与毫米波雷达获取的多目标车辆地理空间特征进行匹配,以实验车辆车载RTK定位数据(D 0)作为相对真值,采用多传感器数据融合方法对视频摄像机获取的定位数据(D 1)与毫米波雷达获取的定位数据(D 2)进行融合; 4) Feature matching and data fusion: The multi-target vehicle geospatial features obtained by the video camera and the millimeter wave radar are matched, and the on-board RTK positioning data (D 0 ) of the experimental vehicle is used as the relative truth value, and the multi-sensor data fusion method is used to analyze the video. The positioning data (D 1 ) obtained by the camera is fused with the positioning data (D 2 ) obtained by the millimeter-wave radar;
    5)车辆重追踪:包含如下两种情形,5) Vehicle re-tracking: includes the following two situations:
    5.1)单视频摄像机视野内车辆重追踪;5.1) Re-tracking of vehicles within the field of view of a single video camera;
    5.2)跨视频摄像机车辆重追踪。5.2) Vehicle re-tracking across video cameras.
  2. 如权利要求1所述的方法,其特征在于,所述视频摄像机采用定向式摄像机,毫米波雷达采用79GHz频段的长距离雷达;所述视频摄像机与所述毫米波雷达布设于路侧立杆同一位置,使得两者经纬度坐标保持一致。The method according to claim 1, wherein the video camera adopts a directional camera, and the millimeter-wave radar adopts a long-distance radar in a frequency band of 79 GHz; the video camera and the millimeter-wave radar are arranged on the same roadside pole position, so that the latitude and longitude coordinates of the two are consistent.
  3. 如权利要求1所述的方法,其特征在于,所述视频摄像机与所述毫米波雷达采用如下方案之一布设:The method of claim 1, wherein the video camera and the millimeter-wave radar are arranged by one of the following schemes:
    方案A:毫米波雷达与视频摄像机布设于不同的路侧立杆上,二者视野重合率大于90%;Scheme A: The millimeter-wave radar and the video camera are arranged on different roadside poles, and the overlap rate of the two fields of view is greater than 90%;
    方案B:毫米波雷达与视频摄像机布设于道路上方同一龙门架处,二者视野重合率大于90%;Scheme B: The millimeter-wave radar and the video camera are arranged on the same gantry above the road, and the overlap rate of the two fields of view is greater than 90%;
    方案C:毫米波雷达与视频摄像机布设于道路上方不同龙门架处,二者视野重合率大于90%。Scheme C: The millimeter-wave radar and the video camera are arranged on different gantry above the road, and the overlap rate of the two fields of view is greater than 90%.
  4. 如权利要求2或3所述的方法,其特征在于,所述视频摄像机与所述毫米波雷达的探测范围根据安装高度与角度进行调整:所述安装高度为6米,向下倾斜俯角为10°;所述视频摄像机与所述毫米波雷达均以25Hz的频率采集数据,接入中心服务器,进行数据保存与处理。The method according to claim 2 or 3, wherein the detection range of the video camera and the millimeter-wave radar is adjusted according to the installation height and angle: the installation height is 6 meters, and the downward tilt angle is 10 meters. °; Both the video camera and the millimeter-wave radar collect data at a frequency of 25 Hz, and access the central server for data storage and processing.
  5. 如权利要求1所述的方法,其特征在于,目标检测框坐标标定为建立摄像机图像中车辆目标检测框像素坐标与车辆世界地理坐标之间的映射关系;所述目标检测框坐标标定分如下三种情形:The method of claim 1, wherein the target detection frame coordinate calibration is to establish a mapping relationship between the pixel coordinates of the vehicle target detection frame in the camera image and the vehicle world geographic coordinates; the target detection frame coordinate calibration is divided into the following three a situation:
    1.2.1)当视频摄像机位于道路来向正上方时,车辆目标检测框像素坐标指检测框下底边 中点像素的像素坐标;1.2.1) When the video camera is located directly above the road, the pixel coordinates of the vehicle target detection frame refer to the pixel coordinates of the midpoint pixel of the bottom edge of the detection frame;
    1.2.2)当视频摄像机位于道路来向左上方,车辆目标检测框像素坐标指检测框右下方顶点像素的像素坐标;1.2.2) When the video camera is located on the road to the upper left, the pixel coordinates of the vehicle target detection frame refer to the pixel coordinates of the vertex pixels at the lower right of the detection frame;
    1.2.3)当视频摄像机位于道路来向右上方,车辆目标检测框像素坐标指检测框左下方顶点像素的像素坐标。1.2.3) When the video camera is located on the road to the upper right, the pixel coordinates of the vehicle target detection frame refer to the pixel coordinates of the lower left vertex pixel of the detection frame.
  6. 如权利要求1所述的方法,其特征在于,所述视频图像数据包含如下目标车辆图像特征:车辆颜色、车型、尺寸以及地理空间特征。The method of claim 1, wherein the video image data includes the following target vehicle image characteristics: vehicle color, model, size, and geospatial characteristics.
  7. 如权利要求4所述的方法,其特征在于,所述视频图像数据还包含车牌信息。The method of claim 4, wherein the video image data further includes license plate information.
  8. 如权利要求1所述的方法,其特征在于,所述毫米波雷达数据包含如下目标车辆地理空间特征:世界地理坐标、速度、航向角。The method of claim 1, wherein the millimeter-wave radar data includes the following geospatial features of the target vehicle: world geographic coordinates, speed, and heading angle.
  9. 如权利要求1所述的方法,其特征在于,所述数据融合采用卡尔曼滤波法。The method of claim 1, wherein the data fusion adopts a Kalman filtering method.
  10. 如权利要求1所述的方法,其特征在于,所述单视频摄像机视野内车辆重追踪包含如下步骤:5.1.1)使用预训练好的深度学习目标检测算法识别单视频摄像机视野内车辆并标记目标框;The method according to claim 1, wherein the re-tracking of vehicles within the field of view of the single video camera comprises the following steps: 5.1.1) Use a pre-trained deep learning target detection algorithm to identify vehicles within the field of view of the single video camera and mark them target box;
    5.1.2)使用Deepsort多目标追踪算法对多目标车辆进行ID赋予与对不同图像帧之间的目标框进行重识别匹配,实现单视频摄像机视野内车辆重追踪。5.1.2) Use the Deepsort multi-target tracking algorithm to assign IDs to multi-target vehicles and re-identify and match the target frames between different image frames to achieve re-tracking of vehicles within the field of view of a single video camera.
  11. 如权利要求1所述的方法,其特征在于,所述跨视频摄像机车辆重追踪包含如下步骤:The method of claim 1, wherein the cross-video camera vehicle re-tracking comprises the steps of:
    5.2.1)当某一目标车辆在摄像机1视野内丢失时,系统记录其丢失前车辆图像特征、地理空间特征,并传入丢失车辆数据库;5.2.1) When a target vehicle is lost in the field of view of camera 1, the system records the image characteristics and geospatial characteristics of the vehicle before it is lost, and transmits it to the lost vehicle database;
    5.2.2)当所述目标车辆进入摄像机2视野内时,系统捕捉车辆特征并在车辆丢失数据库内进行特征匹配,对相似度高的车辆进行ID重赋予,实现车辆的重追踪;5.2.2) When the target vehicle enters the field of view of the camera 2, the system captures the vehicle features and performs feature matching in the vehicle lost database, and re-assigns the ID to the vehicle with high similarity to realize the re-tracking of the vehicle;
    5.2.3)若在所述车辆丢失数据库内未发现匹配车辆,则对所述目标车辆赋予新的ID。5.2.3) If no matching vehicle is found in the vehicle lost database, assign a new ID to the target vehicle.
PCT/CN2021/085150 2021-01-01 2021-04-01 Multi-target vehicle detection and re-identification method based on radar and video fusion WO2022141914A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN202180010955.3A CN115943439A (en) 2021-03-01 2021-04-01 Multi-target vehicle detection and re-identification method based on radar vision fusion
GB2313217.8A GB2619196A (en) 2021-03-01 2021-04-01 Multi-target vehicle detection and re-identification method based on radar and video fusion

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
CN202110000327.9 2021-01-01
CN202110000327 2021-01-01
CN202110228419 2021-03-01
CN202110228419.2 2021-03-01

Publications (1)

Publication Number Publication Date
WO2022141914A1 true WO2022141914A1 (en) 2022-07-07

Family

ID=82260124

Family Applications (9)

Application Number Title Priority Date Filing Date
PCT/CN2021/085148 WO2022141912A1 (en) 2021-01-01 2021-04-01 Vehicle-road collaboration-oriented sensing information fusion representation and target detection method
PCT/CN2021/085147 WO2022141911A1 (en) 2021-01-01 2021-04-01 Roadside sensing unit-based method for quick recognition of dynamic target point cloud and point cloud segmentation
PCT/CN2021/085150 WO2022141914A1 (en) 2021-01-01 2021-04-01 Multi-target vehicle detection and re-identification method based on radar and video fusion
PCT/CN2021/085146 WO2022141910A1 (en) 2021-01-01 2021-04-01 Vehicle-road laser radar point cloud dynamic segmentation and fusion method based on driving safety risk field
PCT/CN2021/085149 WO2022141913A1 (en) 2021-01-01 2021-04-01 On-board positioning device-based roadside millimeter-wave radar calibration method
PCT/CN2022/084738 WO2022206942A1 (en) 2021-01-01 2022-04-01 Laser radar point cloud dynamic segmentation and fusion method based on driving safety risk field
PCT/CN2022/084929 WO2022206978A1 (en) 2021-01-01 2022-04-01 Roadside millimeter-wave radar calibration method based on vehicle-mounted positioning apparatus
PCT/CN2022/084912 WO2022206974A1 (en) 2021-01-01 2022-04-01 Roadside sensing unit-based static and non-static object point cloud recognition method
PCT/CN2022/084925 WO2022206977A1 (en) 2021-01-01 2022-04-01 Cooperative-vehicle-infrastructure-oriented sensing information fusion representation and target detection method

Family Applications Before (2)

Application Number Title Priority Date Filing Date
PCT/CN2021/085148 WO2022141912A1 (en) 2021-01-01 2021-04-01 Vehicle-road collaboration-oriented sensing information fusion representation and target detection method
PCT/CN2021/085147 WO2022141911A1 (en) 2021-01-01 2021-04-01 Roadside sensing unit-based method for quick recognition of dynamic target point cloud and point cloud segmentation

Family Applications After (6)

Application Number Title Priority Date Filing Date
PCT/CN2021/085146 WO2022141910A1 (en) 2021-01-01 2021-04-01 Vehicle-road laser radar point cloud dynamic segmentation and fusion method based on driving safety risk field
PCT/CN2021/085149 WO2022141913A1 (en) 2021-01-01 2021-04-01 On-board positioning device-based roadside millimeter-wave radar calibration method
PCT/CN2022/084738 WO2022206942A1 (en) 2021-01-01 2022-04-01 Laser radar point cloud dynamic segmentation and fusion method based on driving safety risk field
PCT/CN2022/084929 WO2022206978A1 (en) 2021-01-01 2022-04-01 Roadside millimeter-wave radar calibration method based on vehicle-mounted positioning apparatus
PCT/CN2022/084912 WO2022206974A1 (en) 2021-01-01 2022-04-01 Roadside sensing unit-based static and non-static object point cloud recognition method
PCT/CN2022/084925 WO2022206977A1 (en) 2021-01-01 2022-04-01 Cooperative-vehicle-infrastructure-oriented sensing information fusion representation and target detection method

Country Status (3)

Country Link
CN (5) CN116685873A (en)
GB (2) GB2620877A (en)
WO (9) WO2022141912A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116189116A (en) * 2023-04-24 2023-05-30 江西方兴科技股份有限公司 Traffic state sensing method and system

Families Citing this family (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114724362B (en) * 2022-03-23 2022-12-27 中交信息技术国家工程实验室有限公司 Vehicle track data processing method
CN115358530A (en) * 2022-07-26 2022-11-18 上海交通大学 Vehicle-road cooperative sensing roadside test data quality evaluation method
CN115113157B (en) * 2022-08-29 2022-11-22 成都瑞达物联科技有限公司 Beam pointing calibration method based on vehicle-road cooperative radar
CN115480243B (en) * 2022-09-05 2024-02-09 江苏中科西北星信息科技有限公司 Multi-millimeter wave radar end-edge cloud fusion calculation integration and application method thereof
CN115166721B (en) * 2022-09-05 2023-04-07 湖南众天云科技有限公司 Radar and GNSS information calibration fusion method and device in roadside sensing equipment
CN115272493B (en) * 2022-09-20 2022-12-27 之江实验室 Abnormal target detection method and device based on continuous time sequence point cloud superposition
CN115235478B (en) * 2022-09-23 2023-04-07 武汉理工大学 Intelligent automobile positioning method and system based on visual label and laser SLAM
CN115830860B (en) * 2022-11-17 2023-12-15 西部科学城智能网联汽车创新中心(重庆)有限公司 Traffic accident prediction method and device
CN115966084B (en) * 2023-03-17 2023-06-09 江西昂然信息技术有限公司 Holographic intersection millimeter wave radar data processing method and device and computer equipment
CN117452392B (en) * 2023-12-26 2024-03-08 哈尔滨工业大学(深圳)(哈尔滨工业大学深圳科技创新研究院) Radar data processing system and method for vehicle-mounted auxiliary driving system
CN117471461B (en) * 2023-12-26 2024-03-08 哈尔滨工业大学(深圳)(哈尔滨工业大学深圳科技创新研究院) Road side radar service device and method for vehicle-mounted auxiliary driving system

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107609522A (en) * 2017-09-19 2018-01-19 东华大学 A kind of information fusion vehicle detecting system based on laser radar and machine vision
CN110532896A (en) * 2019-08-06 2019-12-03 北京航空航天大学 A kind of road vehicle detection method merged based on trackside millimetre-wave radar and machine vision
WO2020009060A1 (en) * 2018-07-02 2020-01-09 ソニーセミコンダクタソリューションズ株式会社 Information processing device, information processing method, computer program, and moving body device
CN111862157A (en) * 2020-07-20 2020-10-30 重庆大学 Multi-vehicle target tracking method integrating machine vision and millimeter wave radar
CN111914664A (en) * 2020-07-06 2020-11-10 同济大学 Vehicle multi-target detection and track tracking method based on re-identification

Family Cites Families (42)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6661370B2 (en) * 2001-12-11 2003-12-09 Fujitsu Ten Limited Radar data processing apparatus and data processing method
US9562971B2 (en) * 2012-11-22 2017-02-07 Geosim Systems Ltd. Point-cloud fusion
KR101655606B1 (en) * 2014-12-11 2016-09-07 현대자동차주식회사 Apparatus for tracking multi object using lidar and method thereof
TWI597513B (en) * 2016-06-02 2017-09-01 財團法人工業技術研究院 Positioning system, onboard positioning device and positioning method thereof
CN105892471B (en) * 2016-07-01 2019-01-29 北京智行者科技有限公司 Automatic driving method and apparatus
WO2018126248A1 (en) * 2017-01-02 2018-07-05 Okeeffe James Micromirror array for feedback-based image resolution enhancement
KR102056147B1 (en) * 2016-12-09 2019-12-17 (주)엠아이테크 Registration method of distance data and 3D scan data for autonomous vehicle and method thereof
CN106846494A (en) * 2017-01-16 2017-06-13 青岛海大新星软件咨询有限公司 Oblique photograph three-dimensional building thing model automatic single-body algorithm
US10281920B2 (en) * 2017-03-07 2019-05-07 nuTonomy Inc. Planning for unknown objects by an autonomous vehicle
CN108629231B (en) * 2017-03-16 2021-01-22 百度在线网络技术(北京)有限公司 Obstacle detection method, apparatus, device and storage medium
CN107133966B (en) * 2017-03-30 2020-04-14 浙江大学 Three-dimensional sonar image background segmentation method based on sampling consistency algorithm
CN108932462B (en) * 2017-05-27 2021-07-16 华为技术有限公司 Driving intention determining method and device
FR3067495B1 (en) * 2017-06-08 2019-07-05 Renault S.A.S METHOD AND SYSTEM FOR IDENTIFYING AT LEAST ONE MOVING OBJECT
CN109509260B (en) * 2017-09-14 2023-05-26 阿波罗智能技术(北京)有限公司 Labeling method, equipment and readable medium of dynamic obstacle point cloud
CN108152831B (en) * 2017-12-06 2020-02-07 中国农业大学 Laser radar obstacle identification method and system
CN108010360A (en) * 2017-12-27 2018-05-08 中电海康集团有限公司 A kind of automatic Pilot context aware systems based on bus or train route collaboration
CN108639059B (en) * 2018-05-08 2019-02-19 清华大学 Driver based on least action principle manipulates behavior quantization method and device
CN109188379B (en) * 2018-06-11 2023-10-13 深圳市保途者科技有限公司 Automatic calibration method for driving auxiliary radar working angle
US10839530B1 (en) * 2018-09-04 2020-11-17 Apple Inc. Moving point detection
CN109297510B (en) * 2018-09-27 2021-01-01 百度在线网络技术(北京)有限公司 Relative pose calibration method, device, equipment and medium
CN111429739A (en) * 2018-12-20 2020-07-17 阿里巴巴集团控股有限公司 Driving assisting method and system
JP7217577B2 (en) * 2019-03-20 2023-02-03 フォルシアクラリオン・エレクトロニクス株式会社 CALIBRATION DEVICE, CALIBRATION METHOD
CN110296713A (en) * 2019-06-17 2019-10-01 深圳数翔科技有限公司 Trackside automatic driving vehicle Position Fixing Navigation System and single, multiple vehicle positioning and navigation methods
CN110220529B (en) * 2019-06-17 2023-05-23 深圳数翔科技有限公司 Positioning method for automatic driving vehicle at road side
CN110443978B (en) * 2019-08-08 2021-06-18 南京联舜科技有限公司 Tumble alarm device and method
CN110458112B (en) * 2019-08-14 2020-11-20 上海眼控科技股份有限公司 Vehicle detection method and device, computer equipment and readable storage medium
CN110850378B (en) * 2019-11-22 2021-11-19 深圳成谷科技有限公司 Automatic calibration method and device for roadside radar equipment
CN110850431A (en) * 2019-11-25 2020-02-28 盟识(上海)科技有限公司 System and method for measuring trailer deflection angle
CN110906939A (en) * 2019-11-28 2020-03-24 安徽江淮汽车集团股份有限公司 Automatic driving positioning method and device, electronic equipment, storage medium and automobile
CN111121849B (en) * 2020-01-02 2021-08-20 大陆投资(中国)有限公司 Automatic calibration method for orientation parameters of sensor, edge calculation unit and roadside sensing system
CN111999741B (en) * 2020-01-17 2023-03-14 青岛慧拓智能机器有限公司 Method and device for detecting roadside laser radar target
CN111157965B (en) * 2020-02-18 2021-11-23 北京理工大学重庆创新中心 Vehicle-mounted millimeter wave radar installation angle self-calibration method and device and storage medium
CN111476822B (en) * 2020-04-08 2023-04-18 浙江大学 Laser radar target detection and motion tracking method based on scene flow
CN111554088B (en) * 2020-04-13 2022-03-22 重庆邮电大学 Multifunctional V2X intelligent roadside base station system
CN111192295B (en) * 2020-04-14 2020-07-03 中智行科技有限公司 Target detection and tracking method, apparatus, and computer-readable storage medium
CN111537966B (en) * 2020-04-28 2022-06-10 东南大学 Array antenna error correction method suitable for millimeter wave vehicle-mounted radar field
CN111766608A (en) * 2020-06-12 2020-10-13 苏州泛像汽车技术有限公司 Environmental perception system based on laser radar
CN111880191B (en) * 2020-06-16 2023-03-28 北京大学 Map generation method based on multi-agent laser radar and visual information fusion
CN111880174A (en) * 2020-07-03 2020-11-03 芜湖雄狮汽车科技有限公司 Roadside service system for supporting automatic driving control decision and control method thereof
CN111985322B (en) * 2020-07-14 2024-02-06 西安理工大学 Road environment element sensing method based on laser radar
CN112019997A (en) * 2020-08-05 2020-12-01 锐捷网络股份有限公司 Vehicle positioning method and device
CN112509333A (en) * 2020-10-20 2021-03-16 智慧互通科技股份有限公司 Roadside parking vehicle track identification method and system based on multi-sensor sensing

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107609522A (en) * 2017-09-19 2018-01-19 东华大学 A kind of information fusion vehicle detecting system based on laser radar and machine vision
WO2020009060A1 (en) * 2018-07-02 2020-01-09 ソニーセミコンダクタソリューションズ株式会社 Information processing device, information processing method, computer program, and moving body device
CN110532896A (en) * 2019-08-06 2019-12-03 北京航空航天大学 A kind of road vehicle detection method merged based on trackside millimetre-wave radar and machine vision
CN111914664A (en) * 2020-07-06 2020-11-10 同济大学 Vehicle multi-target detection and track tracking method based on re-identification
CN111862157A (en) * 2020-07-20 2020-10-30 重庆大学 Multi-vehicle target tracking method integrating machine vision and millimeter wave radar

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116189116A (en) * 2023-04-24 2023-05-30 江西方兴科技股份有限公司 Traffic state sensing method and system
CN116189116B (en) * 2023-04-24 2024-02-23 江西方兴科技股份有限公司 Traffic state sensing method and system

Also Published As

Publication number Publication date
CN117441197A (en) 2024-01-23
WO2022206978A1 (en) 2022-10-06
GB2618936A (en) 2023-11-22
CN116685873A (en) 2023-09-01
CN117441113A (en) 2024-01-23
WO2022141913A1 (en) 2022-07-07
CN117836653A (en) 2024-04-05
WO2022206974A1 (en) 2022-10-06
WO2022141911A1 (en) 2022-07-07
GB202316625D0 (en) 2023-12-13
WO2022141910A1 (en) 2022-07-07
WO2022141912A1 (en) 2022-07-07
WO2022206977A1 (en) 2022-10-06
GB2620877A (en) 2024-01-24
CN117836667A (en) 2024-04-05
GB202313215D0 (en) 2023-10-11
WO2022206942A1 (en) 2022-10-06

Similar Documents

Publication Publication Date Title
WO2022141914A1 (en) Multi-target vehicle detection and re-identification method based on radar and video fusion
Liu et al. Robust target recognition and tracking of self-driving cars with radar and camera information fusion under severe weather conditions
Khan et al. A survey of advances in vision-based vehicle re-identification
CN109444911B (en) Unmanned ship water surface target detection, identification and positioning method based on monocular camera and laser radar information fusion
CN111369541B (en) Vehicle detection method for intelligent automobile under severe weather condition
JP2019527832A (en) System and method for accurate localization and mapping
CN115943439A (en) Multi-target vehicle detection and re-identification method based on radar vision fusion
CN113485441A (en) Distribution network inspection method combining unmanned aerial vehicle high-precision positioning and visual tracking technology
Fu et al. A camera–radar fusion method based on edge computing
Cui et al. 3D detection and tracking for on-road vehicles with a monovision camera and dual low-cost 4D mmWave radars
Harlow et al. A new wave in robotics: Survey on recent mmwave radar applications in robotics
Zhu et al. A simple outdoor environment obstacle detection method based on information fusion of depth and infrared
He et al. A novel multi-source vehicle detection algorithm based on deep learning
CN116266360A (en) Vehicle target detection tracking method based on multi-source information fusion
CN115690713A (en) Binocular camera-based radar-vision fusion event detection method
Li et al. Pedestrian liveness detection based on mmwave radar and camera fusion
Cao et al. Lane determination of vehicles based on a novel clustering algorithm for intelligent traffic monitoring
Jibrin et al. An object detection and classification method using radar and camera data fusion
Atiq et al. Vehicle detection and shape recognition using optical sensors: a review
CN115471526A (en) Automatic driving target detection and tracking method based on multi-source heterogeneous information fusion
Deng et al. Robust target detection, position deducing and tracking based on radar camera fusion in transportation scenarios
Lin et al. Identification and Tracking of Takeout Delivery Motorcycles using Low-channel Roadside LiDAR
Xu et al. Multiview Fusion 3D Target Information Perception Model in Nighttime Unmanned Intelligent Vehicles
CN111239761B (en) Method for indoor real-time establishment of two-dimensional map
Huang et al. An efficient multi-threshold selection method for lane detection based on lidar

Legal Events

Date Code Title Description
WPC Withdrawal of priority claims after completion of the technical preparations for international publication

Ref document number: 202110000327.9

Country of ref document: CN

Date of ref document: 20230528

Free format text: WITHDRAWN AFTER TECHNICAL PREPARATION FINISHED

NENP Non-entry into the national phase

Ref country code: DE

ENP Entry into the national phase

Ref document number: 202313217

Country of ref document: GB

Kind code of ref document: A

Free format text: PCT FILING DATE = 20210401