WO2020108647A1 - Target detection method, apparatus and system based on linkage between vehicle-mounted camera and vehicle-mounted radar - Google Patents

Target detection method, apparatus and system based on linkage between vehicle-mounted camera and vehicle-mounted radar Download PDF

Info

Publication number
WO2020108647A1
WO2020108647A1 PCT/CN2019/122171 CN2019122171W WO2020108647A1 WO 2020108647 A1 WO2020108647 A1 WO 2020108647A1 CN 2019122171 W CN2019122171 W CN 2019122171W WO 2020108647 A1 WO2020108647 A1 WO 2020108647A1
Authority
WO
WIPO (PCT)
Prior art keywords
target
image
radar
confidence
coordinate system
Prior art date
Application number
PCT/CN2019/122171
Other languages
French (fr)
Chinese (zh)
Inventor
邝宏武
方梓成
孙杰
Original Assignee
杭州海康威视数字技术股份有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 杭州海康威视数字技术股份有限公司 filed Critical 杭州海康威视数字技术股份有限公司
Publication of WO2020108647A1 publication Critical patent/WO2020108647A1/en

Links

Images

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S13/00Systems using the reflection or reradiation of radio waves, e.g. radar systems; Analogous systems using reflection or reradiation of waves whose nature or wavelength is irrelevant or unspecified
    • G01S13/86Combinations of radar systems with non-radar systems, e.g. sonar, direction finder
    • G01S13/867Combination of radar systems with cameras
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S13/00Systems using the reflection or reradiation of radio waves, e.g. radar systems; Analogous systems using reflection or reradiation of waves whose nature or wavelength is irrelevant or unspecified
    • G01S13/86Combinations of radar systems with non-radar systems, e.g. sonar, direction finder
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S7/00Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
    • G01S7/02Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S13/00
    • G01S7/41Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S13/00 using analysis of echo signal for target characterisation; Target signature; Target cross-section
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S7/00Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
    • G01S7/02Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S13/00
    • G01S7/41Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S13/00 using analysis of echo signal for target characterisation; Target signature; Target cross-section
    • G01S7/417Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S13/00 using analysis of echo signal for target characterisation; Target signature; Target cross-section involving the use of neural networks

Definitions

  • the present application relates to the field of intelligent transportation, and in particular, to a target detection method, device, and system in which a car camera and a car radar are linked.
  • Automated driving is an important application in intelligent transportation systems. Automated driving requires vehicles to detect the vehicles around them and avoid them according to the vehicles around them to avoid traffic accidents. Vehicles currently include cameras and radars, and vehicles can detect vehicles around them based on cameras and radars.
  • a method for detecting vehicles around the car which can be: detecting the first position information of each image target of the vehicle around the vehicle based on the camera, and detecting the second position of each radar target of the vehicle based on the radar information.
  • the information is mapped to the road coordinate system to obtain the fourth position information of each radar target.
  • At least one target pair is determined according to the third position information of each image target and the fourth position information of each radar target, and each target pair includes an image target and a radar target belonging to the same object. Since the image target and the radar target in the target pair are the targets detected by the camera and the radar, respectively, which may be the target of the vehicle, the target in the target pair may be the vehicle, and the target in the target pair is the detected vehicle.
  • the preset first perspective matrix is used to reflect the conversion relationship between the image coordinate system of the camera and the road coordinate system.
  • the conversion relationship may change due to vehicles driving under different road conditions.
  • the preset first perspective matrix cannot accurately reflect the current The conversion relationship between the image coordinate system of the camera and the road coordinate system, so the accuracy of the third position information obtained by mapping the first position information of the image target to the road coordinate system through the preset first perspective matrix will be reduced, thereby reducing the detection target Accuracy.
  • embodiments of the present application provide a vehicle detection method and device based on a camera and a radar.
  • the technical solution is as follows:
  • a target detection method in which a vehicle-mounted camera and a vehicle-mounted radar are linked.
  • the method includes:
  • Each radar target around the vehicle is detected from the speed and distance images provided by the radar, the position information of each radar target in the radar coordinate system, and the confidence of each radar target, the confidence is The probability that the target category of the real target corresponding to the image target or the radar target is a specified category;
  • the first perspective matrix is used to represent the image image The conversion relationship between the coordinate system and the preset road coordinate system;
  • target categories are detected from image targets whose confidence does not exceed the first preset threshold and radar targets whose confidence does not exceed the second preset threshold.
  • the method further includes:
  • the detected confidence level of each of the radar targets classify the radar target whose confidence level exceeds the second preset threshold to obtain the target category of the real target corresponding to the radar target, and output the target category.
  • the method before the acquiring the first perspective matrix according to the position information of the image target whose confidence exceeds the first preset threshold and the position information of the radar target whose confidence exceeds the second preset threshold, the method further includes :
  • N associated target pairs are determined from an image target with a confidence exceeding a first preset threshold and a radar target with a confidence exceeding a second preset threshold, any one of the associated target pairs includes a radar target and an image satisfying a preset association condition Target, the N is a positive integer greater than or equal to 1;
  • the obtaining the first perspective matrix according to the position information of the image target whose confidence exceeds the first preset threshold and the position information of the radar target whose confidence exceeds the second preset threshold includes:
  • the first perspective matrix is determined according to the position information of the radar target and the image target in the N associated target pairs.
  • the determining of N associated target pairs from the image target with the confidence level exceeding the first preset threshold and the radar target with the confidence level exceeding the second preset threshold includes:
  • the third position information of each of the first image targets and the fourth position information of each of the first radar targets positionally associate each of the first image targets and each of the first radar targets to obtain the N related target pairs.
  • the performing position association on each of the first image targets and each of the second radar targets to obtain the N associated target pairs includes:
  • the first radar target is at The projected area in the road coordinate system, and the overlapping projected area of the first image target and the first radar target in the road coordinate system;
  • the determining the first perspective matrix according to the position information of the radar target and the image target in the N associated target pairs includes:
  • the first perspective matrix is used to detect a target category from an image target whose confidence does not exceed the first preset threshold and a radar target whose confidence does not exceed the second preset threshold, include:
  • M feature fusion target pairs are determined from the image target whose confidence level does not exceed the first preset threshold and the radar target whose confidence level does not exceed the second preset threshold, any one of the feature fusion target pairs includes those satisfying the preset association condition One radar target and one image target, where M is a positive integer greater than or equal to 1;
  • the echo energy feature of the radar target and the image feature of the image target in the feature fusion target pair are respectively convoluted and spliced to obtain the feature fusion
  • the determining M feature fusion target pairs from the second image target and the second radar target includes:
  • the second image target is an image target whose confidence level does not exceed the first preset threshold
  • the second radar target is a radar target whose confidence level does not exceed the second preset threshold
  • the second radar target performs position correlation to obtain the M feature fusion target pairs.
  • the detecting the confidence of each image target around the vehicle from the video image provided by the on-board camera includes:
  • the detection of the confidence of each radar target around the vehicle from the speed and distance images collected by the radar includes:
  • a target detection device in which a car camera and a car radar are linked, and the device includes:
  • the first detection module is used to detect each image target around the vehicle from the video image provided by the onboard camera, the confidence of each image target, and the position information of each image target in the image coordinate system;
  • the second detection module is used to detect each radar target around the vehicle from the speed and distance images provided by the radar, the position information of each radar target in the radar coordinate system, and the confidence of each radar target Degree, the confidence is used to indicate the probability that the target category of the real target corresponding to the image target or the radar target is a specified category;
  • An obtaining module configured to obtain a first perspective matrix based on the position information of the image target with the confidence level exceeding the first preset threshold and the position information of the radar target with the confidence level exceeding the second preset threshold It represents the conversion relationship between the image coordinate system and the preset road coordinate system;
  • the third detection module is used to detect the target category from the image target whose confidence does not exceed the first preset threshold and the radar target whose confidence does not exceed the second preset threshold through the first perspective matrix .
  • the device further includes:
  • a classification module is used to classify image objects whose confidence exceeds the first preset threshold according to the detected confidence of each of the image objects, to obtain the target category of the real target corresponding to the image object, and to Target category output, and, based on the detected confidence of each of the radar targets, classify the radar targets with confidence exceeding the second preset threshold to obtain the target category of the real target corresponding to the radar target, and The target category is output.
  • the device further includes:
  • a determining module configured to determine N associated target pairs from an image target with a confidence exceeding a first preset threshold and a radar target with a confidence exceeding a second preset threshold, any one of the associated target pairs includes satisfying a preset associated condition Radar target and image target, N is a positive integer greater than or equal to 1;
  • the acquisition module is configured to determine the first perspective matrix according to the position information of the radar target and the image target in the N associated target pairs.
  • the determination module is used to:
  • the third position information of each of the first image targets and the fourth position information of each of the first radar targets positionally associate each of the first image targets and each of the first radar targets to obtain the N related target pairs.
  • the acquisition module is used to:
  • the third detection module is used to:
  • M feature fusion target pairs are determined from the image target whose confidence level does not exceed the first preset threshold and the radar target whose confidence level does not exceed the second preset threshold, any one of the feature fusion target pairs includes those satisfying the preset association condition One radar target and one image target, where M is a positive integer greater than or equal to 1;
  • the echo energy feature of the radar target and the image feature of the image target in the feature fusion target pair are respectively convoluted and spliced to obtain the feature fusion
  • the third detection module is used to:
  • the second image target is an image target whose confidence level does not exceed the first preset threshold
  • the second radar target is a radar target whose confidence level does not exceed the second preset threshold
  • the second radar target performs position correlation to obtain the M feature fusion target pairs.
  • the first detection module is used to:
  • the second detection module is used to:
  • a target detection system including a radar provided on a vehicle, an on-board camera provided on the vehicle, and a detection device communicating with the radar and the on-board camera ,
  • the on-board camera is used to photograph the surroundings of the vehicle to obtain a video image of the current frame, and provide the video image of the current frame to the detection device;
  • the radar is used to generate a current frame speed distance image based on the transmitted radar signal and the received echo signal, and provide the current frame speed distance image to the detection device;
  • the detection device is configured to detect each image target around the vehicle from the video image provided by the on-board camera, the confidence of each image target, and the position information of each image target in the image coordinate system Detecting each radar target around the vehicle from the speed and distance images provided by the radar, and the position information of each radar target in the radar coordinate system, the confidence level of the radar target, the confidence level is The probability that the target category of the real target corresponding to the image target or the radar target is a specified category; according to the location information and confidence level of the image target whose confidence exceeds the first preset threshold Obtain the first perspective matrix of the position information of the radar target, and the first perspective matrix is used to represent the conversion relationship between the image coordinate system and the preset road coordinate system; through the first perspective matrix, from the confidence level The target category is detected from the image target that does not exceed the first preset threshold and the radar target that does not exceed the second preset threshold.
  • the on-vehicle camera is provided on the front and rear and/or left and right sides of the vehicle, and the radar is provided on the front and rear of the vehicle.
  • the radar is a millimeter wave radar.
  • a non-volatile computer-readable storage medium for storing a computer program, which is loaded and executed by a processor to implement the first aspect or the first aspect Instructions for any alternative method.
  • the present application provides an electronic device, the electronic device comprising:
  • At least one processor At least one processor
  • At least one memory At least one memory
  • the at least one memory stores one or more instructions, and the one or more instructions are configured to be executed by the at least one processor to execute the following instructions:
  • Each radar target around the vehicle is detected from the speed and distance images provided by the radar, the position information of each radar target in the radar coordinate system, and the confidence of each radar target, the confidence is The probability that the target category of the real target corresponding to the image target or the radar target is a specified category;
  • the first perspective matrix is used to represent the image image The conversion relationship between the coordinate system and the preset road coordinate system;
  • target categories are detected from image targets whose confidence does not exceed the first preset threshold and radar targets whose confidence does not exceed the second preset threshold.
  • the position information of the image target around the vehicle in the image coordinate system and the confidence of the image target are detected by the on-board camera, and the position information of the radar target around the vehicle in the radar coordinate system and the confidence of each radar target are detected by radar;
  • Image targets with confidence exceeding the first preset threshold and radar targets with confidence exceeding the second preset threshold are real targets in the specified category, so image targets with confidence exceeding the first preset threshold and confidence exceeding the second preset
  • the first perspective matrix acquired by the radar target with a threshold can accurately reflect the conversion relationship between the image coordinate system of the vehicle's on-board camera and the road coordinate relationship under the current road conditions, so that the confidence level does not exceed the first preset according to the first perspective matrix
  • the accuracy of the target target detected in the image target of the threshold and the radar target whose confidence level does not exceed the second preset threshold is the real target of the specified category is high, thereby improving the accuracy of target detection.
  • FIG. 1 is a schematic structural diagram of a target detection system in which an in-vehicle camera and an in-vehicle radar are linked according to an embodiment of the present application;
  • FIG. 2 is a flowchart of a method for detecting a target in which an onboard camera and an onboard radar are linked according to an embodiment of the present application;
  • FIG. 3 is a flowchart of a target detection method in which an onboard camera and an onboard radar are linked according to an embodiment of the present application;
  • FIG. 4 is a block diagram of a target fusion of a vehicle-mounted camera and a vehicle-mounted radar provided by an embodiment of the present application;
  • FIG. 5 is a flowchart of a method for obtaining confidence of an image target provided by an embodiment of the present application
  • FIG. 6 is a schematic structural diagram of a first convolutional neural network provided by an embodiment of the present application.
  • FIG. 7 is a flow chart of a method for obtaining confidence of a radar target provided by an embodiment of the present application.
  • FIG. 8 is a flowchart of a method for obtaining a first perspective matrix provided by an embodiment of the present application.
  • FIG. 9 is a flowchart of a method for detecting a target of a specified category provided by an embodiment of the present application.
  • FIG. 10 is a flowchart of detecting a target of a specified category by a convolutional neural network according to an embodiment of the present application
  • FIG. 11 is a schematic structural diagram of a target detection device in which an in-vehicle camera and an in-vehicle radar are linked according to an embodiment of the present application;
  • FIG. 12 is a schematic structural diagram of an apparatus provided by an embodiment of the present application.
  • FIG. 13 is a schematic structural diagram of a target detection system provided by an embodiment of the present application.
  • FIG. 14 is a schematic diagram of various image targets around a vehicle detected from a video image provided by an embodiment of the present application.
  • 15 is a schematic diagram of detecting various radar targets around a vehicle from a speed and distance image provided by a radar provided by an embodiment of the present application.
  • the vehicle can detect other vehicles around it while driving on the road, so that the driver can prompt the driver when driving the vehicle to assist the driver to drive the vehicle, or, when autonomous driving, the vehicle can be based on its The surrounding vehicles plan the driving lane or avoid other vehicles.
  • an embodiment of the present application provides a target detection system in which an on-board camera and an on-board radar are linked, including an on-board camera on a vehicle 1, an on-board radar 2 on a vehicle, and an on-board radar 2 and on-board camera Communication detection equipment 3; vehicle-mounted camera 1 is arranged in front and rear and/or left and right sides of the vehicle, and vehicle-mounted radar 2 is arranged in front and rear of the vehicle;
  • Vehicle-mounted camera used to capture the video image of the surrounding environment of the vehicle
  • the vehicle-mounted radar 2 is used to send radar waves around the vehicle, receive the reflected waves corresponding to the radar waves and obtain the echo energy intensity of the reflected waves to obtain a speed distance image;
  • the detection device 3 is used to detect the position information and confidence level of each image target around the vehicle in the image coordinate system based on the video image, the confidence level is used to indicate the probability that the target category of the real target corresponding to the image target is a specified category, And detecting the position information and confidence of the radar target around the vehicle in the radar coordinate system based on the speed distance image, the confidence being used to indicate the probability that the target category of the real target corresponding to the radar target is the specified category;
  • the detection device 3 is further configured to acquire the first perspective matrix based on the position information of the image target whose confidence exceeds the first preset threshold and the position information of the image target whose confidence exceeds the second preset threshold. Represents the conversion relationship between the image image coordinate system and the preset road coordinate system; through the first perspective matrix, from the image target whose confidence does not exceed the first preset threshold and the radar target whose confidence does not exceed the second preset threshold Detection target category.
  • the vehicle includes on-board equipment, and the vehicle detects surrounding vehicles through the on-board equipment, on-board camera, and on-board radar.
  • the vehicle-mounted device can detect each image target around the vehicle through the vehicle-mounted camera, and each radar target around the vehicle through the vehicle-mounted radar.
  • Each image target and each radar target may be real targets in a specified category, and the real targets in a specified category may be vehicles and the like.
  • the in-vehicle device may be the detection device 3 described above.
  • the vehicle-mounted radar 2 may be a millimeter wave radar or a laser radar.
  • the in-vehicle device may be the above-mentioned detection device 3.
  • each image target detected by the on-board camera and each radar target detected by radar can be fused to improve the detection accuracy.
  • the radar target and the fusion please refer to the description of any of the following embodiments, which will not be described here.
  • an embodiment of the present application provides a target detection method in which a car camera and a car radar are linked.
  • the method includes:
  • Step 201 Detect each image object around the vehicle, the confidence of each image object, and the position information of each image object in the image coordinate system from the video image provided by the on-board camera.
  • Step 202 Detect each radar target around the vehicle from the speed and distance images provided by the onboard radar, and the position information of each radar target in the radar coordinate system, the confidence of the radar target, which is used to represent the image target or the radar target
  • the target category of the corresponding real target is the probability of the specified category.
  • Step 203 Obtain a first perspective matrix according to the position information of the image target whose confidence exceeds the first preset threshold and the position information of the radar target whose confidence exceeds the second preset threshold, and the first perspective matrix is used to represent the image image coordinates The conversion relationship between the system and the preset road coordinate system.
  • Step 204 Using the first perspective matrix, detect the target category from the image target whose confidence does not exceed the first preset threshold and the radar target whose confidence does not exceed the second preset threshold.
  • step 205 is further included, according to the detected confidence of each image target, the image target whose confidence exceeds the first preset threshold is classified to obtain the target category of the real target corresponding to the image target And output the target category, and classify the radar targets whose confidence exceeds the second preset threshold according to the detected confidence of each radar target to obtain the target category of the real target corresponding to the radar target, and Output the target category.
  • the targets of the desired target category can be classified and output.
  • each image object around the vehicle detected from the video image provided by the on-board camera, and the confidence of each image object, and the position information of each image object in the image coordinate system are illustrated.
  • the target in the superimposed frame is an image target with a confidence exceeding the first preset threshold.
  • FIG. 15 to illustrate the detection of various radar targets around the vehicle from the speed and distance images provided by the radar, and the radar targets in the radar coordinate system
  • the position information in FIG. 3 is the confidence of the radar target, where the target in the superimposed frame is a radar target whose confidence exceeds the second preset threshold.
  • step 206 is further included, and N associated target pairs are determined from the image target whose confidence exceeds the first preset threshold and the radar target whose confidence exceeds the second preset threshold, any one of the associations
  • the target pair includes a radar target and an image target satisfying a preset association condition, and N is a positive integer greater than or equal to 1.
  • the first perspective matrix can be determined according to the position information of the radar target and the image target in the N associated target pairs.
  • step 206 may include:
  • the second position information of the first radar target and the stored third perspective matrix map the first radar target from the radar coordinate system to the road coordinate system to obtain the corresponding fourth position of the first radar target in the road coordinate system Information; where the first radar target is a radar target whose confidence exceeds a second preset threshold, and the second position information is the position information of the first radar target in the radar coordinate system;
  • N associated target pairs can be obtained.
  • the high-confidence image target detected by the vehicle camera and the high-confidence radar target detected by the radar can be dynamically aligned.
  • the dynamic alignment between the image coordinate system, road coordinate system and radar coordinate system in 4.
  • step 2063 may be:
  • the projected area of the first image target in the road coordinate system determines the associated cost between each first image target and each second radar target ;
  • association cost between each first image target and each second radar target is determined, a first radar target and a first image target with the smallest association cost are determined from the first image target and the second image target as the association target pair , which can improve the accuracy of the associated target pair.
  • the above step 203 may include:
  • the above step 204 may include:
  • M feature fusion target pairs are determined from the image target whose confidence level does not exceed the first preset threshold and the radar target whose confidence level does not exceed the second preset threshold, any one of the feature fusion target pairs includes those satisfying the preset association condition
  • M is a positive integer greater than or equal to 1.
  • the echo energy feature of the radar target in the feature fusion target pair and the image feature of the image target are respectively convoluted and stitched together to obtain the fusion feature corresponding to the feature fusion target pair Figure;
  • step 2041 may be:
  • the second image object is mapped from the image coordinate system to the road coordinate system to obtain the corresponding position information of the second image object in the road coordinate system
  • the The second radar target is mapped from the radar coordinate system to the road coordinate system to obtain the corresponding position information of the second radar target in the road coordinate system.
  • the second image target is an image target whose confidence level does not exceed the first preset threshold.
  • the second radar target is a radar target whose confidence level does not exceed the second preset threshold;
  • position correlation is performed for each second image target and each second radar target to obtain M Feature fusion target pair.
  • the above step 201 may include:
  • step 202 may be:
  • the confidence of the radar target is determined according to the intensity of the echo energy of any radar target in the current frame speed distance image, the distance from the vehicle, and the duration of the radar target in the multi-frame historical frame speed distance image.
  • an embodiment of the present application provides a target detection method in which a car camera and a car radar are linked.
  • This method can be applied to the architecture shown in FIG. 1, and the execution subject of the method can be a car device.
  • the vehicle-mounted device may be the detection device 3 in the system shown in FIG. 1, including:
  • Step 301 Detect the position information, target area and confidence level of each image target around the vehicle in the image coordinate system from the video image provided by the vehicle's on-board camera.
  • the confidence level is used to represent the target of the real target corresponding to the image target
  • the category is the probability of the specified category.
  • the on-board camera can capture a frame of video image of the environment around the vehicle. Whenever a frame of video image is captured, for convenience of description, the current frame of video image is called the first video image.
  • the video image is input to the in-vehicle device; the in-vehicle device includes the first convolutional neural network, which can detect the position information of each image target in the first video image in the specified category according to the first video image , Target area and confidence.
  • the image objects whose confidence exceeds the first preset threshold may be classified according to the detected confidence of each image object, to obtain the target category of the real target corresponding to the image object, and output the target category.
  • the vehicle-mounted camera includes an image coordinate system.
  • the position information and target area of the image target are the position information and area of the image target in the image coordinate system.
  • the vehicle-mounted device may detect the position information, target area, and confidence of each image target through operations from 3011 to 3015, which are:
  • the first convolutional neural network includes a convolutional neural network (Convolutional Neural Network, CNN), a regional candidate network (Region Proposal Network, RPN), and a region-based convolutional neural network (Fast Region-based Convolution Neural Network) , RCNN) and other components
  • a first category set is set in RCNN in advance
  • the category of the first category set may include a specified category, for example, the specified category may be a vehicle
  • the first category set may also include houses, trees, flower stands, Street lights and other categories.
  • This step may be: first, when the vehicle-mounted device receives the first video image input from the vehicle-mounted camera, the first image is input to the first convolutional neural network, and each of the first video images output by the first convolutional neural network is acquired.
  • the location information, area and probability of each object in the image coordinate system belong to each category in the first category set.
  • the vehicle-mounted device may input the first video image to the CNN, and the CNN convolves the first video image and extracts features to obtain a first feature map corresponding to the first video image, and converts the first feature
  • the graph is input to RPN and RCNN respectively.
  • RPN determines at least one first candidate box area in the first feature map to obtain a second feature map (each first candidate box area is a second feature map), and also inputs the second feature map to the RCNN.
  • RCNN performs regression on the target position and target category of any second feature map according to the first feature map, that is, the target feature in any first candidate frame area, and finally obtains the position of the target object in the first candidate frame area Information and target area, and the probability that each target object belongs to each category in the first category set.
  • the vehicle-mounted device obtains the position information and target area of each target output by the RCNN and the probability that each target belongs to each category in the first category set.
  • the vehicle-mounted device selects the maximum probability from the probability of each category corresponding to the target, and takes the category corresponding to the maximum probability as the category of the target, so that the category of each target can be obtained, and then from each target Select the target with the specified category as the image target.
  • the target area of the image target may be different from the actual area of the image target.
  • the RCNN may also output an image target frame corresponding to each target, that is, each image target may also have a corresponding image target frame, and the image target frame corresponding to the image target includes the image of the image target.
  • the K-frame video images taken most recently before the current can be obtained and constitute a video image set.
  • a to-be-processed image target for convenience of description, it is called a to-be-processed image target, and each video image in the video image set is determined.
  • the confidence of the target of the image to be processed in the video image set is obtained according to the confidence of the target of the image to be processed in each video image in the video image set, and K is a preset value.
  • the classification confidence of the image object to be processed is used to indicate the probability that the image object to be processed is a real vehicle on the image texture.
  • the video image collection includes the first video image, the second video image, ..., the Kth video image, the first video image is the first video image captured in the video image collection, and the Kth video image is the most Video images taken late.
  • the implementation manner may be:
  • the confidence of the image target to be processed in the video image has been obtained at present, and for the video image that does not include the image target to be processed, the The confidence level is set to a preset reliability, for example, the preset reliability may be a value of 0, 1, or the like.
  • the correspondence between the target and the number of images is also stored in the vehicle-mounted terminal.
  • a target and the number of images including the target are saved in the record.
  • the image target in the first video image and the image target in the K-th video image are the same target, the number of images saved in the record including the image target in the correspondence relationship is increased by one.
  • the image target may be a newly emerging target, set the number of images including the image target to 1, and set the image target Correspond to 1 and stored in the corresponding relationship.
  • the classification confidence of the image target to be processed may be calculated according to the following first formula.
  • the first formula is:
  • C c is the classification confidence of the image target to be processed
  • C c[k] is the confidence level of the image target to be processed in the k-th video image
  • the confidence of the tracking frame number of the image target to be processed is used to indicate the stability of the image target to be processed as a real target of a preset category.
  • the confidence of the tracking frame number of the image target to be processed can be obtained according to the second formula below.
  • the second formula is:
  • CT is the confidence of the number of tracking frames of the image target to be processed
  • T is the number of second video images.
  • any image object in the first video image for convenience of description, it is referred to as an image object to be processed, and the distance between the image object to be processed and the vehicle is obtained according to the first position information of the image object to be processed, from Obtain the image height corresponding to the image target to be processed and the blocked ratio of the target frame corresponding to the image target to be processed from the first video image, the target frame including the image corresponding to the image target to be processed, according to the distance, the image height and the blocked Proportionally obtain the position confidence of the image target to be processed.
  • the position confidence of the image target to be processed can be calculated according to the following third formula.
  • C p is the position confidence of the image target to be processed
  • h is the image height corresponding to the image target to be processed
  • y is the distance between the image target to be processed and the vehicle
  • is the image to be processed The occlusion ratio of the target frame corresponding to the target.
  • the confidence of the position of the image object depends on the ratio of the position information of the image object in the first video image to the image object being blocked: when the position information of the image object is closer to the bottom of the first video image, the position of the image object is confident The higher the degree; the smaller the proportion of the image target blocked by other targets, the higher the confidence of the position of the image target.
  • the confidence of the image object in the first video image is calculated according to the following fourth formula.
  • C s is the confidence of the image target
  • W c is the weight coefficient of the classification confidence of the image target
  • W p is the weight coefficient of the location confidence of the image target
  • W T is the image Weight coefficients for the confidence of the tracking frame number of the target.
  • the three weight coefficients are all preset values.
  • the image target detected by the on-board camera with a confidence level exceeding the first preset threshold is higher in the accuracy of the specified category of real targets, and the image target whose confidence level does not exceed the first preset threshold is the actual target of the designated category The accuracy is low. Therefore, in this embodiment, the image target whose confidence exceeds the first preset threshold may be determined as the detected real target of the specified category.
  • Step 302 Detect the position information of each radar target around the vehicle in the radar coordinate system from the speed distance image provided by the radar of the vehicle and the confidence of each radar target.
  • the radar has a radar coordinate system, and the detected position information of each radar target is the position information in the radar coordinate system.
  • the detected confidence of each radar target classify the radar target whose confidence exceeds the second preset threshold to obtain the target category of the real target corresponding to the radar target, and output the target category.
  • this step can be implemented by the following operations from 3021 to 3022, which are:
  • the vehicle's radar detects the target category around the vehicle as the position information, area, and echo energy intensity of each radar target of the specified category.
  • the radar can send radar waves to the surroundings of the vehicle.
  • the radar waves can be reflected by objects around the car to form reflected waves.
  • the radar can receive the reflected waves and obtain the return energy intensity of the reflected waves.
  • the position of each reflection point can be calculated according to the echo energy intensity of each reflected wave, and an energy map can be drawn according to the position of each reflection point and the echo energy intensity.
  • the energy map can be a speed distance image.
  • the energy map includes At least one energy block, each energy block includes a plurality of continuous reflection points corresponding to the energy intensity and position of the echo, and each energy block is a target.
  • the target category and area of the target corresponding to each energy block can be determined.
  • the target with the target category as the specified category is determined as the radar target, the extreme point of the echo energy intensity is found from the energy block of the radar target, and the position of the extreme point is determined as the position information of the radar target in the radar coordinate system.
  • the average value of the energy intensity of each echo in the energy block of the radar target can be used as the energy intensity of the echo of the radar target, or the energy intensity of each echo in the energy block of the radar target can be sorted to select the echo in the middle position
  • the energy intensity is taken as the energy intensity of the echo of the radar target.
  • the area of the radar target detected by the radar is the actual area of the radar target.
  • the longest distance that the radar can detect and the maximum echo energy intensity of the radar are obtained.
  • the distance between the radar target and the vehicle is calculated according to the position information of the radar target.
  • the maximum echo energy intensity, the calculated distance, and the echo energy intensity of the radar target is calculated according to the following fifth formula.
  • C is the confidence of the radar target
  • a is the weight coefficient of the radar target position
  • a can be greater than 0 and less than 1
  • d is the calculated distance
  • d max is the furthest distance
  • b is The weight coefficient of the radar target's echo energy intensity
  • b can be greater than 0 and less than 1
  • p is the radar target's echo energy intensity
  • p max is the maximum echo energy intensity.
  • the accuracy of the radar target detected by the radar with a confidence level exceeding the second preset threshold is a specified category of real targets
  • the accuracy of the radar target with a confidence level not exceeding the second preset threshold is a designated category of real targets.
  • the radar target whose confidence exceeds the second preset threshold may be determined as the detected real target of the specified category.
  • step 301 can be executed before step 302
  • step 301 can be executed before step 301
  • step 301 and step 302 can be executed simultaneously.
  • Step 303 Obtain the first perspective matrix according to the position information of the image target with confidence exceeding the first preset threshold and the position information of the radar target with confidence exceeding the second preset threshold, the first perspective matrix is used to represent the image coordinate system The conversion relationship with the preset road coordinate system.
  • the first preset threshold and the second preset threshold may be the same or different.
  • N associated target pairs are determined from the image targets whose confidence exceeds the first preset threshold and the radar targets whose confidence exceeds the second preset threshold. Any one of the associated target pairs includes Suppose the radar target and the image target of the association condition, N is a positive integer greater than or equal to 1.
  • the process of determining N associated target pairs may include the following operations from 3031 to 3033.
  • the operations from 3031 to 3033 are as follows:
  • the first image object is an image object whose confidence exceeds a first preset threshold
  • the first position information of the first image object is position information of the first image object in the image coordinate system.
  • the in-vehicle terminal locally stores the second perspective matrix obtained last time, and the second perspective matrix is used to reflect the conversion relationship between the image coordinate system of the in-vehicle camera and the preset road coordinate system.
  • the road coordinate system is different from the image coordinate system of the on-board camera and the radar coordinate system of the radar.
  • the road coordinate system can take the position of a certain point in the vehicle as the coordinate origin and the direction of the vehicle's forward axis as the horizontal axis.
  • the vertical axis is perpendicular to the direction of vehicle advancement.
  • the position of the center point of the front bumper of the vehicle can be used as the coordinate origin of the road coordinate system.
  • the first position information of each first image object whose confidence exceeds the first preset threshold constitutes a first matrix
  • a second matrix is obtained according to the following sixth formula.
  • the second matrix includes the confidence exceeding the first
  • the third position information of each first image object in the road coordinate system with a preset threshold is obtained according to the following sixth formula.
  • the second perspective matrix can also be used to convert the target area of each first image target whose confidence exceeds a preset threshold to obtain the projected area of each first image target in the road coordinate system.
  • the projected area of the image object in the road coordinate system is the actual area.
  • the first radar target is a radar target whose confidence exceeds a second preset threshold
  • the second position information is the position information of the first radar target in the radar coordinate system.
  • the second position information of each first radar target whose confidence exceeds the second preset threshold constitutes a third matrix, and a fourth matrix is obtained according to the following seventh formula.
  • the fourth matrix includes the confidence exceeding the first The fourth position information of each second radar target in the road coordinate system with two preset thresholds.
  • the area of each radar target whose confidence exceeds the second preset threshold is the actual area of the radar target, which is equal to the projected area of each first radar target in the road coordinate system, so there is no need to convert each The area of the first radar target.
  • the N associated target pairs can be determined through the first and second steps as follows:
  • Step 1 Establish a cost matrix based on the third position information and projected area of the first image target and the fourth position information and projected area of each radar target, each first image target corresponds to a row in the cost matrix, and each A radar target corresponds to a column of the cost matrix, and a row corresponding to the first image target includes cost coefficients between the first image target and each first radar target, and the cost coefficients of the first image target and the first radar target represent The probability that the first image target and the first radar target are the same target.
  • the cost matrix thus established includes N rows and X columns.
  • i 1, 2, ..., N
  • j 1, 2, ..., X
  • the fourth position information and projected area of the jth first radar target in the road coordinate system Calculate the projected overlap area S ij between the i-th first image target and the j-th first radar target.
  • the cost between the i-th first image target and the j-th first radar target is calculated according to the following eighth formula Coefficient d ij .
  • the cost coefficient d ij between the i-th first image target and the j-th first radar target is used as the element of the i-th row and j-th column of the cost matrix.
  • the eighth formula is:
  • Step 2 Select a maximum cost coefficient from a row of cost coefficients corresponding to the first image target, and form a pair of associated target pairs with the first image target and the first radar target corresponding to the maximum cost coefficient.
  • the first perspective matrix is determined according to the position information of the radar target and the image target in the N associated target pairs.
  • the position information of the first image target in the associated target pair is corrected according to the position information of the first radar target in the associated target pair;
  • the corrected position information of each of the first image objects in the N associated target pairs corrects the second perspective matrix to obtain a first perspective matrix.
  • the probability that the first radar target detected by the radar is a real target in a specified category is higher than the probability that the first image target detected by a vehicle camera is a real target in a specified category. Therefore, in this step, for each associated target pair, the first image target and the first radar target included in the associated target pair are the same target, and the third position information of the first image target can be corrected to the first The fourth position information of the radar target, so that the fourth position information of each first image target is obtained.
  • a fifth matrix is constructed from the fourth position information of each first image object, and the first perspective matrix is obtained according to the following ninth formula according to the fifth matrix and the first matrix composed of the first position information of each first image object.
  • the first image target with the confidence level exceeding the first preset threshold and the first radar target with the confidence level exceeding the second preset threshold are both detected real targets of the specified category. Therefore, in this step, according to the position information of the first image target whose confidence exceeds the first preset threshold and the position information of the second radar target whose confidence exceeds the second preset threshold, acquiring the first perspective matrix can make the first
  • the perspective matrix reflects the conversion relationship between the image coordinate system of the vehicle-mounted camera and the road coordinate system when the vehicle is driving under the current road conditions.
  • the second perspective matrix saved in the vehicle-mounted device may also be updated to the first perspective matrix.
  • Step 304 Using the first perspective matrix, detect the real target whose category is the specified category from the image target whose confidence level does not exceed the first preset threshold and the radar target whose confidence level does not exceed the second preset threshold.
  • this step may be implemented by operations from 3041 to 3043, respectively:
  • the position information of each second image object whose confidence level does not exceed the first preset threshold constitutes a sixth matrix, and the seventh matrix is obtained according to the following tenth formula.
  • the seventh matrix includes the confidence level not exceeding the first The position information of each second image object in the road coordinate system with a preset threshold.
  • the first perspective matrix can also be used to convert the target area of each second image target whose confidence level does not exceed the first preset threshold, to obtain the projected area of each second image target in the road coordinate system.
  • the projected area of a second image object in the road coordinate system is the actual area.
  • the position information of each second radar target whose confidence level does not exceed the second preset threshold constitutes an eighth matrix
  • a ninth matrix is obtained according to the following eleventh formula, which includes the confidence level not exceeding The position information of each second radar target in the road coordinate system of the second preset threshold.
  • the area of each second radar target whose confidence level does not exceed the second preset threshold is the actual area of the radar target, which is equal to the projected area of each second radar target in the road coordinate system, so there is no need to pass the first perspective
  • the matrix converts the area of each second radar target.
  • the radar target pair includes the image target and the radar target that are the same target.
  • M feature fusion target pairs can be determined through the first and second steps as follows:
  • Step 1 According to the position information and projection area of each second image target whose confidence level does not exceed the first preset threshold in the road coordinate system and each second radar target whose confidence level does not exceed the second preset threshold on the road The position information and the projected area of the coordinate system establish a second cost matrix, each second image object whose confidence level does not exceed the first preset threshold corresponds to a row in the second cost matrix, and each confidence level where the confidence level does not exceed the second preset threshold A second radar target corresponds to a column of a second cost matrix, and a row corresponding to the second image target includes a second cost coefficient between the image target and each radar target whose confidence level does not exceed a preset threshold, the second The second cost coefficient of the image target and the second radar target represents the probability that the second image target whose confidence level does not exceed the first preset threshold and the second radar target whose confidence level does not exceed the second preset threshold are the same target.
  • the cost matrix thus established includes M rows and Y columns.
  • p 1, 2, ..., M
  • q 1, 2, ..., Y
  • the projected overlap area S pq between the second radar targets calculates the cost coefficient d pq between the p-th second image target and the q-th second radar target according to the twelfth formula as follows.
  • the cost coefficient d pq between the p-th second image target and the q-th second radar target is taken as the element of the p-th row and the q-th column of the cost matrix.
  • d pq S pq /(S p C +S q R -S pq ).
  • Step 2 Select a maximum second cost coefficient from a row of second cost coefficients corresponding to the second image target, and form a feature fusion target pair with the second image target and the second radar target corresponding to the maximum second cost coefficient.
  • the second convolutional neural network detects from the M feature fusion target pairs that the target category is the true target of the specified category.
  • the vehicle-mounted device includes a second convolutional neural network, and a second category set is set in the second convolutional neural network in advance, and the categories of the second category set may include designated categories, non-designated categories, and other categories.
  • the specified category is a vehicle or a motor vehicle
  • the non-preset category may be a non-motor vehicle.
  • the feature fusion target pair includes a second image target and a second radar target
  • the in-vehicle device may use the corresponding image of the second image target in the first video image and the corresponding image of the second radar target
  • the energy block is input to the second convolutional neural network.
  • the image features extracted from the image of the second image target and the energy features extracted from the energy block of the second radar target through the second convolutional neural network will be Image features and energy features are stitched together to obtain feature sequences.
  • the probability of the feature fusion target pair belonging to each category in the second category set is output; the maximum probability is selected Is used as the target category of the feature fusion target pair.
  • the target category of the feature fusion target pair is the specified category
  • the second radar target in the feature fusion target pair is used as the real target with the detected category as the specified category.
  • the second convolutional network can also be composed of multiple sub-networks, each of which completes the following processes: extracting image features from the image of the second image target, and extracting echo energy features from the energy block of the second radar target ; Splicing the image features and energy features to obtain a feature sequence; according to the feature sequence, multi-layer convolution calculation and fully connected layer calculation are performed to output the feature fusion target for each category in the second category set Probability.
  • the category of the feature fusion target pair When the category of the feature fusion target pair is the specified category, it indicates that the target category of the second image target and the second radar target in the feature fusion target pair are both specified categories, that is to say that the vehicle camera and the radar simultaneously detect the same
  • the target is a real target of the specified category, and because the accuracy of radar detection is higher than that of the on-board camera detection, the second radar target in the feature fusion target pair is used as the real target with the target category detected as the specified category.
  • the second convolutional neural network is obtained by training through a sample set in advance, and the sample set includes a plurality of preset feature fusion target pairs and a target category corresponding to each feature fusion target.
  • the sample set is input to the second convolutional neural network for training.
  • the beneficial effects of the embodiments of the present application are: the image target and the image target's confidence around the vehicle are detected by the on-board camera, and the radar target and the radar target's confidence around the vehicle are detected by the radar; because the confidence exceeds the first preset
  • the first image target with a threshold and the first radar target with a confidence exceeding the second preset threshold are real targets of the specified category, so the first image target with a confidence exceeding the first preset threshold and the confidence exceeding the second preset
  • the first perspective matrix acquired by the threshold first radar target can reflect the conversion relationship between the image coordinate system of the vehicle's on-board camera and the road coordinate system under the current road conditions, so that the confidence level does not exceed the first preset according to the first perspective matrix
  • the accuracy of the target target detected in the second image target of the threshold and the second radar target whose confidence level does not exceed the second preset threshold is a real target of the specified category is high, thereby improving the accuracy of target detection.
  • an embodiment of the present application provides a target detection device 400 in which a vehicle camera and a vehicle radar are linked.
  • the device 400 includes:
  • the first detection module 401 is used to detect each image target around the vehicle from the video image provided by the on-board camera, the confidence of each image target, and the position information of each image target in the image coordinate system;
  • the second detection module 402 is used to detect each radar target around the vehicle from the speed distance image provided by the radar, the position information of each radar target in the radar coordinate system, and the Confidence, the confidence is used to indicate the probability that the target category of the real target corresponding to the image target or the radar target is a specified category;
  • the obtaining module 403 is configured to obtain the first perspective matrix based on the position information of the image target whose confidence exceeds the first preset threshold and the position information of the radar target whose confidence exceeds the second preset threshold. Representing the conversion relationship between the image coordinate system and the preset road coordinate system;
  • the third detection module 404 is configured to detect the target from the image target whose confidence does not exceed the first preset threshold and the radar target whose confidence does not exceed the second preset threshold through the first perspective matrix category.
  • the device 400 further includes:
  • a classification module is used to classify image objects whose confidence exceeds the first preset threshold according to the detected confidence of each of the image objects, to obtain the target category of the real target corresponding to the image object, and to Target category output, and, based on the detected confidence of each of the radar targets, classify the radar targets with confidence exceeding the second preset threshold to obtain the target category of the real target corresponding to the radar target, and The target category is output.
  • the device 400 further includes:
  • a determining module configured to determine N associated target pairs from an image target with a confidence exceeding a first preset threshold and a radar target with a confidence exceeding a second preset threshold, any one of the associated target pairs includes satisfying a preset associated condition Radar target and image target, N is a positive integer greater than or equal to 1;
  • the acquisition module is configured to determine the first perspective matrix according to the position information of the radar target and the image target in the N associated target pairs.
  • the determination module is used to:
  • the determination module is used to:
  • the obtaining module 403 is used to:
  • the third detection module 404 is used to:
  • M feature fusion target pairs are determined from the image target whose confidence level does not exceed the first preset threshold and the radar target whose confidence level does not exceed the second preset threshold, any one of the feature fusion target pairs includes those satisfying the preset association condition One radar target and one image target, where M is a positive integer greater than or equal to 1;
  • the echo energy feature of the radar target and the image feature of the image target in the feature fusion target pair are respectively convoluted and spliced to obtain the feature fusion
  • the third detection module 404 is used to:
  • the second image target is an image target whose confidence level does not exceed the first preset threshold
  • the second radar target is a radar target whose confidence level does not exceed the second preset threshold
  • the second radar target performs position correlation to obtain the M feature fusion target pairs.
  • the first detection module 401 is used to:
  • the second detection module 402 is used to:
  • the first detection module detects the image target and the confidence of the image target around the vehicle through the on-board camera
  • the second detection module detects the radar target around the vehicle and the confidence of each radar target through the radar Degree
  • the acquisition module since the image target with confidence exceeding the first preset threshold and the radar target with confidence exceeding the second preset threshold are real targets in the specified category, the acquisition module according to the image target and confidence with the confidence exceeding the first preset threshold
  • the radar target whose degree exceeds the second preset threshold is the real perspective of the specified category.
  • the first perspective matrix can reflect the conversion relationship between the image coordinate system of the vehicle's on-board camera and the road coordinate system under the current road conditions.
  • a perspective matrix detects the target class from the image target whose confidence level does not exceed the first preset threshold and the radar target whose confidence level does not exceed the second preset threshold. Accuracy.
  • FIG. 12 shows a structural block diagram of a terminal 500 provided by an exemplary embodiment of the present application.
  • the terminal 500 may be an in-vehicle terminal, and the terminal 500 may also be called other names such as user equipment and portable terminal.
  • the terminal 500 includes a processor 501 and a memory 502.
  • the processor 501 may include one or more processing cores, such as a 4-core processor, an 8-core processor, and so on.
  • the processor 501 may adopt at least one hardware form of DSP (Digital Signal Processing, digital signal processing), FPGA (Field-Programmable Gate Array), PLA (Programmable Logic Array). achieve.
  • the processor 501 may also include a main processor and a coprocessor.
  • the main processor is a processor for processing data in a wake-up state, also known as a CPU (Central Processing Unit, central processor); the coprocessor is A low-power processor for processing data in the standby state.
  • the processor 501 may be integrated with a GPU (Graphics Processing Unit, image processor), and the GPU is used to render and draw content that needs to be displayed on the display screen.
  • the processor 501 may further include an AI (Artificial Intelligence, artificial intelligence) processor, which is used to process computing operations related to machine learning.
  • AI Artificial Intelligence, artificial intelligence
  • the memory 502 may include one or more computer-readable storage media, which may be non-transitory.
  • the memory 502 may also include high-speed random access memory, as well as non-volatile memory, such as one or more magnetic disk storage devices, flash memory storage devices.
  • the non-transitory computer-readable storage medium in the memory 502 is used to store at least one instruction, which is executed by the processor 501 to implement the camera-based camera provided by the method embodiment in the present application And radar vehicle detection methods.
  • the terminal 500 may optionally further include: a peripheral device interface 503 and at least one peripheral device.
  • the processor 501, the memory 502 and the peripheral device interface 503 may be connected by a bus or a signal line.
  • Each peripheral device may be connected to the peripheral device interface 503 through a bus, a signal line, or a circuit board.
  • the peripheral device includes at least one of a radio frequency circuit 504, a touch display screen 505, a camera 506, an audio circuit 507, a positioning component 508, and a power supply 509.
  • the peripheral device interface 503 may be used to connect at least one peripheral device related to I/O (Input/Output) to the processor 501 and the memory 502.
  • the processor 501, the memory 502, and the peripheral device interface 503 are integrated on the same chip or circuit board; in some other embodiments, any one of the processor 501, the memory 502, and the peripheral device interface 503 or Both can be implemented on a separate chip or circuit board, which is not limited in this embodiment.
  • the radio frequency circuit 504 is used to receive and transmit RF (Radio Frequency) signals, also called electromagnetic signals.
  • the radio frequency circuit 504 communicates with a communication network and other communication devices through electromagnetic signals.
  • the radio frequency circuit 504 converts the electrical signal into an electromagnetic signal for transmission, or converts the received electromagnetic signal into an electrical signal.
  • the radio frequency circuit 504 includes: an antenna system, an RF transceiver, one or more amplifiers, a tuner, an oscillator, a digital signal processor, a codec chipset, a user identity module card, and so on.
  • the radio frequency circuit 504 can communicate with other terminals through at least one wireless communication protocol.
  • the wireless communication protocol includes but is not limited to: World Wide Web, Metropolitan Area Network, Intranet, various generations of mobile communication networks (2G, 3G, 4G, and 5G), wireless local area networks, and/or WiFi (Wireless Fidelity) networks.
  • the radio frequency circuit 504 may further include NFC (Near Field Communication) related circuits, which is not limited in this application.
  • the display screen 505 is used to display UI (User Interface).
  • the UI may include graphics, text, icons, video, and any combination thereof.
  • the display screen 505 also has the ability to collect touch signals on or above the surface of the display screen 505.
  • the touch signal can be input to the processor 501 as a control signal for processing.
  • the display screen 505 can also be used to provide virtual buttons and/or virtual keyboards, also called soft buttons and/or soft keyboards.
  • the display screen 505 there may be one display screen 505, which is provided with the front panel of the terminal 500; in other embodiments, there may be at least two display screens 505, which are respectively provided on different surfaces of the terminal 500 or have a folded design; In still other embodiments, the display screen 505 may be a flexible display screen, which is disposed on the curved surface or the folding surface of the terminal 500. Even, the display screen 505 can also be set as a non-rectangular irregular figure, that is, a special-shaped screen.
  • the display screen 505 may be made of LCD (Liquid Crystal Display), OLED (Organic Light-Emitting Diode, organic light emitting diode) and other materials.
  • the camera component 506 is used to collect images or videos.
  • the camera assembly 506 includes a front camera and a rear camera.
  • the front camera is set on the front panel of the terminal, and the rear camera is set on the back of the terminal.
  • the camera assembly 506 may also include a flash.
  • the flash can be a single-color flash or a dual-color flash. Dual color temperature flash refers to the combination of warm light flash and cold light flash, which can be used for light compensation at different color temperatures.
  • the audio circuit 507 may include a microphone and a speaker.
  • the microphone is used to collect sound waves of the user and the environment, and convert the sound waves into electrical signals and input them to the processor 501 for processing, or input them to the radio frequency circuit 504 to implement voice communication.
  • the microphone can also be an array microphone or an omnidirectional acquisition microphone.
  • the speaker is used to convert the electrical signal from the processor 501 or the radio frequency circuit 504 into sound waves.
  • the speaker can be a traditional thin-film speaker or a piezoelectric ceramic speaker.
  • the speaker When the speaker is a piezoelectric ceramic speaker, it can not only convert electrical signals into sound waves audible by humans, but also convert electrical signals into sound waves inaudible to humans for distance measurement and other purposes.
  • the audio circuit 507 may also include a headphone jack.
  • the positioning component 508 is used to locate the current geographic location of the terminal 500 to implement navigation or LBS (Location Based Service).
  • the positioning component 508 may be a positioning component based on the GPS (Global Positioning System) of the United States, the Beidou system of China, or the Galileo system of Russia.
  • the power supply 509 is used to supply power to various components in the terminal 500.
  • the power source 509 may be alternating current, direct current, disposable batteries, or rechargeable batteries.
  • the rechargeable battery may be a wired rechargeable battery or a wireless rechargeable battery.
  • the wired rechargeable battery is a battery charged through a wired line
  • the wireless rechargeable battery is a battery charged through a wireless coil.
  • the rechargeable battery can also be used to support fast charging technology.
  • the terminal 500 further includes one or more sensors 510.
  • the one or more sensors 510 include, but are not limited to: an acceleration sensor 511, a gyro sensor 512, a pressure sensor 513, a fingerprint sensor 514, an optical sensor 515, and a proximity sensor 516.
  • the acceleration sensor 511 can detect the magnitude of acceleration on the three coordinate axes of the coordinate system established with the terminal 500.
  • the acceleration sensor 511 can be used to detect components of gravity acceleration on three coordinate axes.
  • the processor 501 may control the touch display screen 505 to display the user interface in a landscape view or a portrait view according to the gravity acceleration signal collected by the acceleration sensor 511.
  • the acceleration sensor 511 can also be used for game or user movement data collection.
  • the gyro sensor 512 can detect the body direction and rotation angle of the terminal 500, and the gyro sensor 512 can cooperate with the acceleration sensor 511 to collect a 3D motion of the user on the terminal 500. Based on the data collected by the gyro sensor 512, the processor 501 can realize the following functions: motion sensing (such as changing the UI according to the user's tilt operation), image stabilization during shooting, game control, and inertial navigation.
  • the pressure sensor 513 may be disposed on the side frame of the terminal 500 and/or the lower layer of the touch display screen 505.
  • the processor 501 can perform left-right hand recognition or shortcut operation according to the grip signal collected by the pressure sensor 513.
  • the processor 501 controls the operability control on the UI interface according to the user's pressure operation on the touch display screen 505.
  • the operability control includes at least one of a button control, a scroll bar control, an icon control, and a menu control.
  • the fingerprint sensor 514 is used to collect the user's fingerprint, and the processor 501 identifies the user's identity according to the fingerprint collected by the fingerprint sensor 514, or the fingerprint sensor 514 identifies the user's identity based on the collected fingerprint. When the user's identity is recognized as a trusted identity, the processor 501 authorizes the user to perform related sensitive operations, including unlocking the screen, viewing encrypted information, downloading software, paying, and changing settings.
  • the fingerprint sensor 514 may be provided on the front, back, or side of the terminal 500. When a physical button or manufacturer logo is provided on the terminal 500, the fingerprint sensor 514 may be integrated with the physical button or manufacturer logo.
  • the optical sensor 515 is used to collect the ambient light intensity.
  • the processor 501 can control the display brightness of the touch display 505 according to the ambient light intensity collected by the optical sensor 515. Specifically, when the ambient light intensity is high, the display brightness of the touch display screen 505 is increased; when the ambient light intensity is low, the display brightness of the touch display screen 505 is reduced.
  • the processor 501 can also dynamically adjust the shooting parameters of the camera assembly 506 according to the ambient light intensity collected by the optical sensor 515.
  • the proximity sensor 516 also called a distance sensor, is usually provided on the front panel of the terminal 500.
  • the proximity sensor 516 is used to collect the distance between the user and the front of the terminal 500.
  • the processor 501 controls the touch display 505 to switch from the bright screen state to the breathing screen state; when the proximity sensor 516 detects When the distance from the user to the front of the terminal 500 gradually becomes larger, the processor 501 controls the touch display 505 to switch from the breath-hold state to the bright-screen state.
  • FIG. 12 does not constitute a limitation on the terminal 500, and may include more or fewer components than those illustrated, or combine certain components, or adopt different component arrangements.
  • an embodiment of the present application provides a target detection system 600, including a radar 601 provided on a vehicle, a vehicle-mounted camera 602 provided on the vehicle, and communicating with the radar 601 and the vehicle-mounted camera 602 Of detection device 603,
  • the on-board camera 602 is used to photograph the surroundings of the vehicle to obtain a video image of the current frame and provide the video image of the current frame to the detection device 603;
  • the radar 601 is configured to generate a current frame speed distance image based on the transmitted radar signal and the received echo signal, and provide the current frame speed distance image to the detection device 603;
  • the detection device 603 is configured to detect each image object around the vehicle, the confidence of each image object, and the position of each image object in the image coordinate system from the video image provided by the vehicle-mounted camera 602 Information; each radar target around the vehicle is detected from the speed distance image provided by the radar 601, the position information of each radar target in the radar coordinate system, and the confidence of each radar target, the The confidence level is used to indicate the probability that the target category of the real target corresponding to the image target or the radar target is the specified category; according to the position information and confidence level of the image target whose confidence level exceeds the first preset threshold value Position information of the radar target with a threshold value to obtain a first perspective matrix, where the first perspective matrix is used to represent a conversion relationship between the image coordinate system and a preset road coordinate system; through the first perspective matrix, A target category is detected from an image target whose confidence level does not exceed the first preset threshold and a radar target whose confidence level does not exceed the second preset threshold.
  • the in-vehicle camera 602 is provided in front and rear and/or both sides of the vehicle, and the radar 601 is provided in front and rear of the vehicle.
  • the radar 601 is a millimeter wave radar.

Abstract

A target detection method, apparatus and system based on the linkage between a vehicle-mounted camera and a vehicle-mounted radar. The method comprises: detecting, from a video image provided by a vehicle-mounted camera, all image targets around a vehicle, the degrees of confidence of all the image targets, and position information of all the image targets in an image coordinate system (201); detecting, from a speed distance image provided by a radar, all radar targets around the vehicle, and position information of all the radar targets in a radar coordinate system (202); acquiring a first perspective matrix according to position information of image targets, the degree of confidence of which exceeds a first preset threshold, and position information of radar targets, the degree of confidence of which exceeds a second preset threshold (203); and detecting, by means of the first perspective matrix, a target category from image targets, the degree of confidence of which does not exceed the first preset threshold, and radar targets, the degree of confidence of which does not exceed the second preset threshold (204). The method can improve the precision of detecting a target.

Description

车载摄像头和车载雷达联动的目标检测方法、装置及系统Target detection method, device and system linked by vehicle camera and vehicle radar
本申请要求于2018年11月30日提交的申请号为201811459743.X、发明名称为“车载摄像头和车载雷达联动的目标检测方法、装置及系统”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。This application requires the priority of the Chinese patent application filed on November 30, 2018 with the application number 201811459743.X and the invention titled "Target Detection Method, Device and System for Vehicle Camera and Vehicle Radar Linkage", all of which are approved by The reference is incorporated in this application.
技术领域Technical field
本申请涉及智能交通领域,特别涉及一种车载摄像头和车载雷达联动的目标检测方法、装置及系统。The present application relates to the field of intelligent transportation, and in particular, to a target detection method, device, and system in which a car camera and a car radar are linked.
背景技术Background technique
自动驾驶是智能交通系统中的一项重要应用,自动驾驶需要车辆检测到其周围的车辆,根据其周围的车辆进行避让,以免交通事故发生。目前车辆中包括摄像头和雷达,车辆可以基于摄像头和雷达检测其周围的车辆。Automated driving is an important application in intelligent transportation systems. Automated driving requires vehicles to detect the vehicles around them and avoid them according to the vehicles around them to avoid traffic accidents. Vehicles currently include cameras and radars, and vehicles can detect vehicles around them based on cameras and radars.
目前存在一种检测汽车周围车辆的方法,可以为:基于摄像头检测车辆周围可能是车辆的每个图像目标的第一位置信息,基于雷达检测车辆周围可能是车辆的每个雷达目标的第二位置信息。通过预设第一透视矩阵将每个图像目标的第一位置信息映射至道路坐标系中得到每个图像目标的第三位置信息,通过预设第二透视矩阵将每个雷达目标的第二位置信息映射至道路坐标系中得到每个雷达目标的第四位置信息。根据每个图像目标的第三位置信息和每个雷达目标的第四位置信息确定至少一个目标对,每个目标对包括属于同一物体的图像目标和雷达目标。由于目标对中的图像目标和雷达目标分别为摄像机和雷达同时检测出的可能是车辆的目标,因此目标对中的目标为车辆的可能较大,将目标对中的目标作为检测出的车辆。At present, there is a method for detecting vehicles around the car, which can be: detecting the first position information of each image target of the vehicle around the vehicle based on the camera, and detecting the second position of each radar target of the vehicle based on the radar information. Map the first position information of each image target to the road coordinate system by presetting the first perspective matrix to obtain the third position information of each image target, and preset the second position of each radar target by presetting the second perspective matrix The information is mapped to the road coordinate system to obtain the fourth position information of each radar target. At least one target pair is determined according to the third position information of each image target and the fourth position information of each radar target, and each target pair includes an image target and a radar target belonging to the same object. Since the image target and the radar target in the target pair are the targets detected by the camera and the radar, respectively, which may be the target of the vehicle, the target in the target pair may be the vehicle, and the target in the target pair is the detected vehicle.
发明人在实现本申请的过程中,发现上述方式至少存在如下缺陷:During the process of implementing the present application, the inventor found that the above-mentioned method has at least the following defects:
预设第一透视矩阵用于反应摄像头的图像坐标系和道路坐标系之间的转换关系,由于车辆在不同路况下行驶,该转换关系可能会发生变化,预设第一透视矩阵不能准确反应当前摄像头的图像坐标系和道路坐标系之间的转换关系,所以通过预设第一透视矩阵将图像目标的第一位置信息映射道路坐标系得到的第三位置信息的精度会降低,进而降低检测目标的精度。The preset first perspective matrix is used to reflect the conversion relationship between the image coordinate system of the camera and the road coordinate system. The conversion relationship may change due to vehicles driving under different road conditions. The preset first perspective matrix cannot accurately reflect the current The conversion relationship between the image coordinate system of the camera and the road coordinate system, so the accuracy of the third position information obtained by mapping the first position information of the image target to the road coordinate system through the preset first perspective matrix will be reduced, thereby reducing the detection target Accuracy.
发明内容Summary of the invention
为了提高检测目标的精度,本申请实施例提供了一种基于摄像头和雷达的车辆检测方法及装置。所述技术方案如下:In order to improve the accuracy of detecting targets, embodiments of the present application provide a vehicle detection method and device based on a camera and a radar. The technical solution is as follows:
根据本申请实施例的第一方面,提供了一种车载摄像头和车载雷达联动的目标检测方法,所述方法包括:According to a first aspect of the embodiments of the present application, there is provided a target detection method in which a vehicle-mounted camera and a vehicle-mounted radar are linked. The method includes:
从车载摄像头提供的视频图像中检测出车辆周围的各个图像目标,各个所述图像目标的置信度,以及各个所述图像目标在图像坐标系中的位置信息;Detecting each image object around the vehicle from the video image provided by the onboard camera, the confidence of each image object, and the position information of each image object in the image coordinate system;
从所述雷达提供的速度距离图像中检测出所述车辆周围的各个雷达目标,各个所述雷达目标在雷达坐标系中的位置信息,以及各个所述雷达目标的置信度,所述置信度用于表示所述图像目标或所述雷达目标所对应的真实目标的目标类别为指定类别的概率;Each radar target around the vehicle is detected from the speed and distance images provided by the radar, the position information of each radar target in the radar coordinate system, and the confidence of each radar target, the confidence is The probability that the target category of the real target corresponding to the image target or the radar target is a specified category;
根据置信度超过第一预设阈值的图像目标的位置信息和置信度超过第二预设阈值的雷达目标的位置信息,获取第一透视矩阵,所述第一透视矩阵用于表示所述图像像坐标系与预设道路坐标系之间的转换关系;Obtain a first perspective matrix according to the position information of the image target whose confidence exceeds the first preset threshold and the position information of the radar target whose confidence exceeds the second preset threshold, the first perspective matrix is used to represent the image image The conversion relationship between the coordinate system and the preset road coordinate system;
通过所述第一透视矩阵,从置信度未超过所述第一预设阈值的图像目标和置信度未超过所述第二预设阈值的雷达目标中检测出目标类别。Through the first perspective matrix, target categories are detected from image targets whose confidence does not exceed the first preset threshold and radar targets whose confidence does not exceed the second preset threshold.
可选的,所述方法还包括:Optionally, the method further includes:
根据检测出的各个所述图像目标的置信度,对置信度超过所述第一预设阈值的图像目标进行分类,得到该图像目标对应的真实目标的目标类别,并将该目标类别输出,以及Classify image objects whose confidence exceeds the first preset threshold according to the detected confidence of each of the image objects, obtain the target category of the real target corresponding to the image object, and output the target category, and
根据检测出的各个所述雷达目标的置信度,对置信度超过所述第二预设阈值的雷达目标进行分类,得到该雷达目标对应的真实目标的目标类别,并将该目标类别输出。According to the detected confidence level of each of the radar targets, classify the radar target whose confidence level exceeds the second preset threshold to obtain the target category of the real target corresponding to the radar target, and output the target category.
可选的,在所述根据置信度超过第一预设阈值的图像目标的位置信息和置信度超过第二预设阈值的雷达目标的位置信息,获取第一透视矩阵之前,所述方法还包括:Optionally, before the acquiring the first perspective matrix according to the position information of the image target whose confidence exceeds the first preset threshold and the position information of the radar target whose confidence exceeds the second preset threshold, the method further includes :
从置信度超过第一预设阈值的图像目标和置信度超过第二预设阈值的雷达 目标中确定N个关联目标对,任意一个所述关联目标对包括满足预设关联条件的雷达目标和图像目标,所述N为大于或等于1的正整数;N associated target pairs are determined from an image target with a confidence exceeding a first preset threshold and a radar target with a confidence exceeding a second preset threshold, any one of the associated target pairs includes a radar target and an image satisfying a preset association condition Target, the N is a positive integer greater than or equal to 1;
所述根据置信度超过第一预设阈值的图像目标的位置信息和置信度超过第二预设阈值的雷达目标的位置信息,获取第一透视矩阵,包括:The obtaining the first perspective matrix according to the position information of the image target whose confidence exceeds the first preset threshold and the position information of the radar target whose confidence exceeds the second preset threshold includes:
根据所述N个关联目标对中的雷达目标和图像目标的位置信息,确定所述第一透视矩阵。The first perspective matrix is determined according to the position information of the radar target and the image target in the N associated target pairs.
可选的,所述从置信度超过第一预设阈值的图像目标和置信度超过第二预设阈值的雷达目标中确定N个关联目标对,包括:Optionally, the determining of N associated target pairs from the image target with the confidence level exceeding the first preset threshold and the radar target with the confidence level exceeding the second preset threshold includes:
根据第一图像目标的第一位置信息和存储的第二透视矩阵,将所述第一图像目标从所述图像坐标系映射至所述道路坐标系,得到所述第一图像目标在所述道路坐标系中对应的第三位置信息;其中所述第一图像目标为置信度超过第一预设阈值的图像目标,所述第一位置信息为所述第一图像目标在所述图像坐标系中的位置信息;Mapping the first image object from the image coordinate system to the road coordinate system according to the first position information of the first image object and the stored second perspective matrix, to obtain the first image object on the road Corresponding third position information in the coordinate system; wherein the first image object is an image object whose confidence exceeds a first preset threshold, and the first position information is the first image object in the image coordinate system Location information
根据第一雷达目标的第二位置信息和存储的第三透视矩阵,将所述第一雷达目标从所述雷达坐标系映射至所述道路坐标系,得到所述第一雷达目标在所述道路坐标系中对应的第四位置信息;其中所述第一雷达目标为置信度超过第二预设阈值的雷达目标,所述第二位置信息为所述第一雷达目标在所述雷达坐标系中的位置信息;Mapping the first radar target from the radar coordinate system to the road coordinate system according to the second position information of the first radar target and the stored third perspective matrix, to obtain the first radar target on the road Corresponding fourth position information in the coordinate system; wherein the first radar target is a radar target whose confidence exceeds a second preset threshold, and the second position information is the first radar target in the radar coordinate system Location information
根据各个所述第一图像目标的第三位置信息和各个所述第一雷达目标的第四位置信息,对各个所述第一图像目标与各个所述第一雷达目标进行位置关联,得到所述N个关联目标对。According to the third position information of each of the first image targets and the fourth position information of each of the first radar targets, positionally associate each of the first image targets and each of the first radar targets to obtain the N related target pairs.
可选的,所述对各个所述第一图像目标与各个所述第二雷达目标进行位置关联,得到所述N个关联目标对,包括:Optionally, the performing position association on each of the first image targets and each of the second radar targets to obtain the N associated target pairs includes:
根据所述第一图像目标的第三位置信息和所述第一雷达目标的第四位置信息,确定所述第一图像目标在所述道路坐标系中的投影面积,所述第一雷达目标在所述道路坐标系中的投影面积,以及所述第一图像目标和所述第一雷达目标在所述道路坐标系中的重叠投影面积;Determine the projected area of the first image target in the road coordinate system according to the third position information of the first image target and the fourth position information of the first radar target, the first radar target is at The projected area in the road coordinate system, and the overlapping projected area of the first image target and the first radar target in the road coordinate system;
根据所述第一图像目标在所述道路坐标系中的投影面积,所述第一雷达目标在所述道路坐标系中的投影面积,以及所述重叠投影面积,确定各个所述第 一图像目标与各个所述第二雷达目标之间的关联代价;Determine each of the first image targets according to the projected area of the first image target in the road coordinate system, the projected area of the first radar target in the road coordinate system, and the overlapping projected area The associated cost with each of the second radar targets;
从所述第一图像目标和所述第二图像目标中确定关联代价最小的一个所述第一雷达目标和一个所述第一图像目标为关联目标对,进而得到所述N个关联目标对。It is determined from the first image target and the second image target that one of the first radar target and one of the first image target with the smallest associated cost are the associated target pairs, and then the N associated target pairs are obtained.
可选的,所述根据所述N个关联目标对中的雷达目标和图像目标的位置信息,确定所述第一透视矩阵,包括:Optionally, the determining the first perspective matrix according to the position information of the radar target and the image target in the N associated target pairs includes:
针对所述N个关联目标对中的任一个关联目标对,根据所述任一个关联目标对中的所述第一雷达目标的位置信息,修正所述任一个关联目标对中的所述第一图像目标的位置信息;For any one of the N associated target pairs, according to the position information of the first radar target in the any associated target pair, modify the first of the any associated target pair Image target location information;
根据所述N个关联目标对中的各个所述第一图像目标修正后的位置信息,修正所述第二透视矩阵,得到所述第一透视矩阵。Modify the second perspective matrix according to the corrected position information of each of the first image targets in the N associated target pairs to obtain the first perspective matrix.
可选的,所述通过所述第一透视矩阵,从置信度未超过所述第一预设阈值的图像目标和置信度未超过所述第二预设阈值的雷达目标中检测出目标类别,包括:Optionally, the first perspective matrix is used to detect a target category from an image target whose confidence does not exceed the first preset threshold and a radar target whose confidence does not exceed the second preset threshold, include:
从置信度未超过第一预设阈值的图像目标和置信度未超过第二预设阈值的雷达目标中确定M个特征融合目标对,任意一个所述特征融合目标对包括满足预设关联条件的一个雷达目标和一个图像目标,所述M为大于或等于1的正整数;M feature fusion target pairs are determined from the image target whose confidence level does not exceed the first preset threshold and the radar target whose confidence level does not exceed the second preset threshold, any one of the feature fusion target pairs includes those satisfying the preset association condition One radar target and one image target, where M is a positive integer greater than or equal to 1;
针对任一所述特征融合目标对,将所述特征融合目标对中的所述雷达目标的回波能量特征和所述图像目标的图像特征分别进行卷积计算后进行拼接,得到所述特征融合目标对所对应的融合特征图;For any of the feature fusion target pairs, the echo energy feature of the radar target and the image feature of the image target in the feature fusion target pair are respectively convoluted and spliced to obtain the feature fusion The fusion feature map corresponding to the target pair;
将所述融合特征图进行卷积和全连接计算后输入到分类网络进行目标分类,得到所述融合特征图对应的目标类别。After performing convolution and full connection calculation on the fusion feature map, input it to a classification network for target classification to obtain a target category corresponding to the fusion feature map.
可选的,所述从第二图像目标和第二雷达目标中确定M个特征融合目标对,包括:Optionally, the determining M feature fusion target pairs from the second image target and the second radar target includes:
通过所述第一透视矩阵,将第二图像目标从所述图像坐标系映射至所述道路坐标系中,得到所述第二图像目标在所述道路坐标系中对应的位置信息,以及,通过预先存储的第三透视矩阵,将第二雷达目标从所述雷达坐标系映射至 所述道路坐标系中,得到所述第二雷达目标在所述道路坐标系中对应的位置信息,所述第二图像目标为置信度未超过第一预设阈值的图像目标,所述第二雷达目标为置信度未超过第二预设阈值的雷达目标;Mapping the second image object from the image coordinate system to the road coordinate system through the first perspective matrix to obtain the corresponding position information of the second image object in the road coordinate system, and, by A pre-stored third perspective matrix, mapping the second radar target from the radar coordinate system to the road coordinate system, to obtain the corresponding position information of the second radar target in the road coordinate system, the first The second image target is an image target whose confidence level does not exceed the first preset threshold, and the second radar target is a radar target whose confidence level does not exceed the second preset threshold;
根据各个所述第二图像目标在所述道路坐标系中对应的位置信息和各个所述第二雷达目标在所述道路坐标系中对应的位置信息,对各个所述第二图像目标和各个所述第二雷达目标进行位置关联,得到所述M个特征融合目标对。According to the corresponding position information of each second image target in the road coordinate system and the corresponding position information of each second radar target in the road coordinate system, for each second image target and each position The second radar target performs position correlation to obtain the M feature fusion target pairs.
可选的,所述从车载摄像头提供的视频图像中检测出车辆周围的各个图像目标的置信度,包括:Optionally, the detecting the confidence of each image target around the vehicle from the video image provided by the on-board camera includes:
根据所述车载摄像头提供的当前帧视频图像,以及与所述当前帧视频图像接近的多帧历史帧视频图像,获取所述任意一个所述图像目标的分类置信度、跟踪帧数置信度和位置置信度;Obtain the classification confidence, tracking frame number confidence and position of any one of the image objects according to the current frame video image provided by the vehicle camera and the multi-frame historical frame video image close to the current frame video image Confidence;
根据所述分类置信度、位置置信度和跟踪帧数置信度中的一个或多个,确定所述图像目标的置信度;Determine the confidence of the image target according to one or more of the classification confidence, position confidence, and tracking frame number confidence;
所述从所述雷达采集的速度距离图像中检测出所述车辆周围的各个雷达目标的置信度,包括:The detection of the confidence of each radar target around the vehicle from the speed and distance images collected by the radar includes:
根据所述当前帧速度距离图像中任一个所述雷达目标的回波能量强度、距离所述车辆的距离,以及所述雷达目标在多帧历史帧速度距离图像中的持续时间,确定所述雷达目标的置信度。Determine the radar according to the intensity of the echo energy of any of the radar targets in the current frame speed distance image, the distance from the vehicle, and the duration of the radar target in the multi-frame historical frame speed distance image The confidence of the goal.
根据本申请实施例的第二方面,提供了一种车载摄像头和车载雷达联动的目标检测装置,所述装置包括:According to a second aspect of the embodiments of the present application, there is provided a target detection device in which a car camera and a car radar are linked, and the device includes:
第一检测模块,用于从车载摄像头提供的视频图像中检测出车辆周围的各个图像目标,各个所述图像目标的置信度,以及各个所述图像目标在图像坐标系中的位置信息;The first detection module is used to detect each image target around the vehicle from the video image provided by the onboard camera, the confidence of each image target, and the position information of each image target in the image coordinate system;
第二检测模块,用于从所述雷达提供的速度距离图像中检测出所述车辆周围的各个雷达目标,各个所述雷达目标在雷达坐标系中的位置信息,以及各个所述雷达目标的置信度,所述置信度用于表示所述图像目标或所述雷达目标所对应的真实目标的目标类别为指定类别的概率;The second detection module is used to detect each radar target around the vehicle from the speed and distance images provided by the radar, the position information of each radar target in the radar coordinate system, and the confidence of each radar target Degree, the confidence is used to indicate the probability that the target category of the real target corresponding to the image target or the radar target is a specified category;
获取模块,用于根据置信度超过第一预设阈值的图像目标的位置信息和置信度超过第二预设阈值的雷达目标的位置信息,获取第一透视矩阵,所述第一 透视矩阵用于表示所述图像像坐标系与预设道路坐标系之间的转换关系;An obtaining module, configured to obtain a first perspective matrix based on the position information of the image target with the confidence level exceeding the first preset threshold and the position information of the radar target with the confidence level exceeding the second preset threshold It represents the conversion relationship between the image coordinate system and the preset road coordinate system;
第三检测模块,用于通过所述第一透视矩阵,从置信度未超过所述第一预设阈值的图像目标和置信度未超过所述第二预设阈值的雷达目标中检测出目标类别。The third detection module is used to detect the target category from the image target whose confidence does not exceed the first preset threshold and the radar target whose confidence does not exceed the second preset threshold through the first perspective matrix .
可选的,所述装置还包括:Optionally, the device further includes:
分类模块,用于根据检测出的各个所述图像目标的置信度,对置信度超过所述第一预设阈值的图像目标进行分类,得到该图像目标对应的真实目标的目标类别,并将该目标类别输出,以及,根据检测出的各个所述雷达目标的置信度,对置信度超过所述第二预设阈值的雷达目标进行分类,得到该雷达目标对应的真实目标的目标类别,并将该目标类别输出。A classification module is used to classify image objects whose confidence exceeds the first preset threshold according to the detected confidence of each of the image objects, to obtain the target category of the real target corresponding to the image object, and to Target category output, and, based on the detected confidence of each of the radar targets, classify the radar targets with confidence exceeding the second preset threshold to obtain the target category of the real target corresponding to the radar target, and The target category is output.
可选的,所述装置还包括:Optionally, the device further includes:
确定模块,用于从置信度超过第一预设阈值的图像目标和置信度超过第二预设阈值的雷达目标中确定N个关联目标对,任意一个所述关联目标对包括满足预设关联条件的雷达目标和图像目标,所述N为大于或等于1的正整数;A determining module, configured to determine N associated target pairs from an image target with a confidence exceeding a first preset threshold and a radar target with a confidence exceeding a second preset threshold, any one of the associated target pairs includes satisfying a preset associated condition Radar target and image target, N is a positive integer greater than or equal to 1;
所述获取模块,用于根据所述N个关联目标对中的雷达目标和图像目标的位置信息,确定所述第一透视矩阵。The acquisition module is configured to determine the first perspective matrix according to the position information of the radar target and the image target in the N associated target pairs.
可选的,所述确定模块,用于:Optionally, the determination module is used to:
根据第一图像目标的第一位置信息和存储的第二透视矩阵,将所述第一图像目标从所述图像坐标系映射至所述道路坐标系,得到所述第一图像目标在所述道路坐标系中对应的第三位置信息;其中所述第一图像目标为置信度超过第一预设阈值的图像目标,所述第一位置信息为所述第一图像目标在所述图像坐标系中的位置信息;Mapping the first image object from the image coordinate system to the road coordinate system according to the first position information of the first image object and the stored second perspective matrix, to obtain the first image object on the road Corresponding third position information in the coordinate system; wherein the first image object is an image object whose confidence exceeds a first preset threshold, and the first position information is the first image object in the image coordinate system Location information
根据第一雷达目标的第二位置信息和存储的第三透视矩阵,将所述第一雷达目标从所述雷达坐标系映射至所述道路坐标系,得到所述第一雷达目标在所述道路坐标系中对应的第四位置信息;其中所述第一雷达目标为置信度超过第二预设阈值的雷达目标,所述第二位置信息为所述第一雷达目标在所述雷达坐标系中的位置信息;Mapping the first radar target from the radar coordinate system to the road coordinate system according to the second position information of the first radar target and the stored third perspective matrix, to obtain the first radar target on the road Corresponding fourth position information in the coordinate system; wherein the first radar target is a radar target whose confidence exceeds a second preset threshold, and the second position information is the first radar target in the radar coordinate system Location information
根据各个所述第一图像目标的第三位置信息和各个所述第一雷达目标的第 四位置信息,对各个所述第一图像目标与各个所述第一雷达目标进行位置关联,得到所述N个关联目标对。According to the third position information of each of the first image targets and the fourth position information of each of the first radar targets, positionally associate each of the first image targets and each of the first radar targets to obtain the N related target pairs.
可选的,所述获取模块,用于:Optionally, the acquisition module is used to:
针对所述N个关联目标对中的任一个关联目标对,根据所述任一个关联目标对中的所述第一雷达目标的位置信息,修正所述任一个关联目标对中的所述第一图像目标的位置信息;For any one of the N associated target pairs, according to the position information of the first radar target in the any associated target pair, modify the first of the any associated target pair Image target location information;
根据所述N个关联目标对中的各个所述第一图像目标修正后的位置信息,修正所述第二透视矩阵,得到所述第一透视矩阵。Modify the second perspective matrix according to the corrected position information of each of the first image targets in the N associated target pairs to obtain the first perspective matrix.
可选的,所述第三检测模块,用于:Optionally, the third detection module is used to:
从置信度未超过第一预设阈值的图像目标和置信度未超过第二预设阈值的雷达目标中确定M个特征融合目标对,任意一个所述特征融合目标对包括满足预设关联条件的一个雷达目标和一个图像目标,所述M为大于或等于1的正整数;M feature fusion target pairs are determined from the image target whose confidence level does not exceed the first preset threshold and the radar target whose confidence level does not exceed the second preset threshold, any one of the feature fusion target pairs includes those satisfying the preset association condition One radar target and one image target, where M is a positive integer greater than or equal to 1;
针对任一所述特征融合目标对,将所述特征融合目标对中的所述雷达目标的回波能量特征和所述图像目标的图像特征分别进行卷积计算后进行拼接,得到所述特征融合目标对所对应的融合特征图;For any of the feature fusion target pairs, the echo energy feature of the radar target and the image feature of the image target in the feature fusion target pair are respectively convoluted and spliced to obtain the feature fusion The fusion feature map corresponding to the target pair;
将所述融合特征图进行卷积和全连接计算后输入到分类网络进行目标分类,得到所述融合特征图对应的目标类别。After performing convolution and full connection calculation on the fusion feature map, input it to a classification network for target classification to obtain a target category corresponding to the fusion feature map.
可选的,所述第三检测模块,用于:Optionally, the third detection module is used to:
通过所述第一透视矩阵,将第二图像目标从所述图像坐标系映射至所述道路坐标系中,得到所述第二图像目标在所述道路坐标系中对应的位置信息,以及,通过预先存储的第三透视矩阵,将第二雷达目标从所述雷达坐标系映射至所述道路坐标系中,得到所述第二雷达目标在所述道路坐标系中对应的位置信息,所述第二图像目标为置信度未超过第一预设阈值的图像目标,所述第二雷达目标为置信度未超过第二预设阈值的雷达目标;Mapping the second image object from the image coordinate system to the road coordinate system through the first perspective matrix to obtain the corresponding position information of the second image object in the road coordinate system, and, by A pre-stored third perspective matrix, mapping the second radar target from the radar coordinate system to the road coordinate system, to obtain the corresponding position information of the second radar target in the road coordinate system, the first The second image target is an image target whose confidence level does not exceed the first preset threshold, and the second radar target is a radar target whose confidence level does not exceed the second preset threshold;
根据各个所述第二图像目标在所述道路坐标系中对应的位置信息和各个所述第二雷达目标在所述道路坐标系中对应的位置信息,对各个所述第二图像目标和各个所述第二雷达目标进行位置关联,得到所述M个特征融合目标对。According to the corresponding position information of each second image target in the road coordinate system and the corresponding position information of each second radar target in the road coordinate system, for each second image target and each position The second radar target performs position correlation to obtain the M feature fusion target pairs.
可选的,所述第一检测模块,用于:Optionally, the first detection module is used to:
根据所述车载摄像头提供的当前帧视频图像,以及与所述当前帧视频图像接近的多帧历史帧视频图像,获取所述任意一个所述图像目标的分类置信度、跟踪帧数置信度和位置置信度;Obtain the classification confidence, tracking frame number confidence and position of any one of the image objects according to the current frame video image provided by the vehicle camera and the multi-frame historical frame video image close to the current frame video image Confidence;
根据所述分类置信度、位置置信度和跟踪帧数置信度中的一个或多个,确定所述图像目标的置信度;Determine the confidence of the image target according to one or more of the classification confidence, position confidence, and tracking frame number confidence;
所述第二检测模块,用于:The second detection module is used to:
根据所述当前帧速度距离图像中任一个所述雷达目标的回波能量强度、距离所述车辆的距离,以及所述雷达目标在多帧历史帧速度距离图像中的持续时间,确定所述雷达目标的置信度。Determine the radar according to the intensity of the echo energy of any of the radar targets in the current frame speed distance image, the distance from the vehicle, and the duration of the radar target in the multi-frame historical frame speed distance image The confidence of the goal.
根据本申请实施例的第三方面,提供了一种目标检测系统,包括设置在车辆上的雷达、设置在所述车辆上的车载摄像头,以及与所述雷达和所述车载摄像头通信的检测装置,According to a third aspect of the embodiments of the present application, there is provided a target detection system, including a radar provided on a vehicle, an on-board camera provided on the vehicle, and a detection device communicating with the radar and the on-board camera ,
所述车载摄像头,用于对所述车辆周围进行拍摄,得到当前帧视频图像,并向所述检测装置提供拍摄的当前帧视频图像;The on-board camera is used to photograph the surroundings of the vehicle to obtain a video image of the current frame, and provide the video image of the current frame to the detection device;
所述雷达,用于根据发射的雷达信号和接收的回波信号生成当前帧速度距离图像,并向所述检测装置提供所述当前帧速度距离图像;The radar is used to generate a current frame speed distance image based on the transmitted radar signal and the received echo signal, and provide the current frame speed distance image to the detection device;
所述检测装置,用于从所述车载摄像头提供的视频图像中检测出车辆周围的各个图像目标,以及各个所述图像目标的置信度,以及各个所述图像目标在图像坐标系中的位置信息;从所述雷达提供的速度距离图像中检测出所述车辆周围的各个雷达目标,以及各个所述雷达目标在雷达坐标系中的位置信息,所述雷达目标的置信度,所述置信度用于表示所述图像目标或所述雷达目标所对应的真实目标的目标类别为指定类别的概率;根据置信度超过第一预设阈值的图像目标的位置信息和置信度超过第二预设阈值的雷达目标的位置信息,获取第一透视矩阵,所述第一透视矩阵用于表示所述图像像坐标系与预设道路坐标系之间的转换关系;通过所述第一透视矩阵,从置信度未超过所述第一预设阈值的图像目标和置信度未超过所述第二预设阈值的雷达目标中检测出目标类别。The detection device is configured to detect each image target around the vehicle from the video image provided by the on-board camera, the confidence of each image target, and the position information of each image target in the image coordinate system Detecting each radar target around the vehicle from the speed and distance images provided by the radar, and the position information of each radar target in the radar coordinate system, the confidence level of the radar target, the confidence level is The probability that the target category of the real target corresponding to the image target or the radar target is a specified category; according to the location information and confidence level of the image target whose confidence exceeds the first preset threshold Obtain the first perspective matrix of the position information of the radar target, and the first perspective matrix is used to represent the conversion relationship between the image coordinate system and the preset road coordinate system; through the first perspective matrix, from the confidence level The target category is detected from the image target that does not exceed the first preset threshold and the radar target that does not exceed the second preset threshold.
可选的,所述车载摄像头设置在所述车辆的前后方和/或左右两侧,所述雷达设置在所述车辆的前后方。Optionally, the on-vehicle camera is provided on the front and rear and/or left and right sides of the vehicle, and the radar is provided on the front and rear of the vehicle.
可选的,所述雷达为毫米波雷达。Optionally, the radar is a millimeter wave radar.
根据本申请实施例的第四方面,提供一种非易失性计算机可读存储介质,用于存储计算机程序,所述计算机程序被处理器加载并执行,以实现第一方面或第一方面的任一种可选的方法的指令。According to a fourth aspect of the embodiments of the present application, there is provided a non-volatile computer-readable storage medium for storing a computer program, which is loaded and executed by a processor to implement the first aspect or the first aspect Instructions for any alternative method.
根据本申请实施例的第五方面,本申请提供了一种电子设备,所述电子设备包括:According to a fifth aspect of the embodiments of the present application, the present application provides an electronic device, the electronic device comprising:
至少一个处理器;和At least one processor; and
至少一个存储器;At least one memory;
所述至少一个存储器存储有一个或多个指令,所述一个或多个指令被配置成由所述至少一个处理器执行,以执行以下指令:The at least one memory stores one or more instructions, and the one or more instructions are configured to be executed by the at least one processor to execute the following instructions:
从车载摄像头提供的视频图像中检测出车辆周围的各个图像目标,各个所述图像目标的置信度,以及各个所述图像目标在图像坐标系中的位置信息;Detecting each image object around the vehicle from the video image provided by the onboard camera, the confidence of each image object, and the position information of each image object in the image coordinate system;
从所述雷达提供的速度距离图像中检测出所述车辆周围的各个雷达目标,各个所述雷达目标在雷达坐标系中的位置信息,以及各个所述雷达目标的置信度,所述置信度用于表示所述图像目标或所述雷达目标所对应的真实目标的目标类别为指定类别的概率;Each radar target around the vehicle is detected from the speed and distance images provided by the radar, the position information of each radar target in the radar coordinate system, and the confidence of each radar target, the confidence is The probability that the target category of the real target corresponding to the image target or the radar target is a specified category;
根据置信度超过第一预设阈值的图像目标的位置信息和置信度超过第二预设阈值的雷达目标的位置信息,获取第一透视矩阵,所述第一透视矩阵用于表示所述图像像坐标系与预设道路坐标系之间的转换关系;Obtain a first perspective matrix according to the position information of the image target whose confidence exceeds the first preset threshold and the position information of the radar target whose confidence exceeds the second preset threshold, the first perspective matrix is used to represent the image image The conversion relationship between the coordinate system and the preset road coordinate system;
通过所述第一透视矩阵,从置信度未超过所述第一预设阈值的图像目标和置信度未超过所述第二预设阈值的雷达目标中检测出目标类别。Through the first perspective matrix, target categories are detected from image targets whose confidence does not exceed the first preset threshold and radar targets whose confidence does not exceed the second preset threshold.
本申请实施例提供的技术方案可以包括以下有益效果:The technical solutions provided by the embodiments of the present application may include the following beneficial effects:
通过车载摄像头检测出车辆周围的图像目标在图像坐标系的位置信息和图像目标的置信度,通过雷达检测出车辆周围在雷达坐标系的雷达目标的位置信息和每个雷达目标的置信度;由于置信度超过第一预设阈值的图像目标和置信 度超过第二预设阈值的雷达目标为指定类别的真实目标,所以根据置信度超过第一预设阈值的图像目标和置信度超过第二预设阈值的雷达目标获取的第一透视矩阵能够精准反映当前路况下车辆的车载摄像头的图像坐标系与道路坐标关系之间的转换关系,这样根据第一透视矩阵从置信度未超过第一预设阈值的图像目标和置信度未超过第二预设阈值的雷达目标中检测出的目标类别为指定类别的真实目标的精度较高,从而提高目标检测的精度。The position information of the image target around the vehicle in the image coordinate system and the confidence of the image target are detected by the on-board camera, and the position information of the radar target around the vehicle in the radar coordinate system and the confidence of each radar target are detected by radar; Image targets with confidence exceeding the first preset threshold and radar targets with confidence exceeding the second preset threshold are real targets in the specified category, so image targets with confidence exceeding the first preset threshold and confidence exceeding the second preset The first perspective matrix acquired by the radar target with a threshold can accurately reflect the conversion relationship between the image coordinate system of the vehicle's on-board camera and the road coordinate relationship under the current road conditions, so that the confidence level does not exceed the first preset according to the first perspective matrix The accuracy of the target target detected in the image target of the threshold and the radar target whose confidence level does not exceed the second preset threshold is the real target of the specified category is high, thereby improving the accuracy of target detection.
应当理解的是,以上的一般描述和后文的细节描述仅是示例性和解释性的,并不能限制本申请。It should be understood that the above general description and the following detailed description are only exemplary and explanatory, and cannot limit the present application.
附图说明BRIEF DESCRIPTION
此处的附图被并入说明书中并构成本说明书的一部分,示出了符合本申请的实施例,并与说明书一起用于解释本申请的原理。The drawings herein are incorporated into the specification and constitute a part of the specification, show embodiments consistent with the application, and are used together with the specification to explain the principles of the application.
图1是本申请实施例提供的一种车载摄像头和车载雷达联动的目标检测系统结构示意图;FIG. 1 is a schematic structural diagram of a target detection system in which an in-vehicle camera and an in-vehicle radar are linked according to an embodiment of the present application;
图2是本申请实施例提供的一种车载摄像头和车载雷达联动的目标检测方法流程图;2 is a flowchart of a method for detecting a target in which an onboard camera and an onboard radar are linked according to an embodiment of the present application;
图3是本申请实施例提供的一种车载摄像头和车载雷达联动的目标检测方法流程图;FIG. 3 is a flowchart of a target detection method in which an onboard camera and an onboard radar are linked according to an embodiment of the present application;
图4是本申请实施例提供的一种车载摄像头和车载雷达联动的目标融合框图;4 is a block diagram of a target fusion of a vehicle-mounted camera and a vehicle-mounted radar provided by an embodiment of the present application;
图5是本申请实施例提供的一种获取图像目标置信度的方法流程图;5 is a flowchart of a method for obtaining confidence of an image target provided by an embodiment of the present application;
图6是本申请实施例提供的一种第一卷积神经网络结构示意图;6 is a schematic structural diagram of a first convolutional neural network provided by an embodiment of the present application;
图7是本申请实施例提供的一种获取雷达目标置信度的方法流程图;7 is a flow chart of a method for obtaining confidence of a radar target provided by an embodiment of the present application;
图8是本申请实施例提供的一种获取第一透视矩阵的方法流程图;8 is a flowchart of a method for obtaining a first perspective matrix provided by an embodiment of the present application;
图9是本申请实施例提供的一种检测类别为指定类别的目标的方法流程图;9 is a flowchart of a method for detecting a target of a specified category provided by an embodiment of the present application;
图10是本申请实施例提供的通过卷积神经网络检测类别为指定类别的目标的流程图;10 is a flowchart of detecting a target of a specified category by a convolutional neural network according to an embodiment of the present application;
图11是本申请实施例提供的一种车载摄像头和车载雷达联动的目标检测装置结构示意图;FIG. 11 is a schematic structural diagram of a target detection device in which an in-vehicle camera and an in-vehicle radar are linked according to an embodiment of the present application;
图12是本申请实施例提供的一种装置结构示意图;12 is a schematic structural diagram of an apparatus provided by an embodiment of the present application;
图13是本申请实施例提供的一种目标检测系统的结构示意图;13 is a schematic structural diagram of a target detection system provided by an embodiment of the present application;
图14是本申请实施例提供的一种从视频图像中检测出的车辆周围的各个图像目标的示意图;14 is a schematic diagram of various image targets around a vehicle detected from a video image provided by an embodiment of the present application;
图15是本申请实施例提供的一种从雷达提供的速度距离图像中检测出车辆周围的各个雷达目标的示意图。15 is a schematic diagram of detecting various radar targets around a vehicle from a speed and distance image provided by a radar provided by an embodiment of the present application.
通过上述附图,已示出本申请明确的实施例,后文中将有更详细的描述。这些附图和文字描述并不是为了通过任何方式限制本申请构思的范围,而是通过参考特定实施例为本领域技术人员说明本申请的概念。Through the above drawings, a clear embodiment of the present application has been shown, which will be described in more detail later. These drawings and text descriptions are not intended to limit the scope of the concept of the present application in any way, but to explain the concept of the present application to those skilled in the art by referring to specific embodiments.
具体实施方式detailed description
这里将详细地对示例性实施例进行说明,其示例表示在附图中。下面的描述涉及附图时,除非另有表示,不同附图中的相同数字表示相同或相似的要素。以下示例性实施例中所描述的实施方式并不代表与本申请相一致的所有实施方式。相反,它们仅是与如所附权利要求书中所详述的、本申请的一些方面相一致的装置和方法的例子。Exemplary embodiments will be described in detail here, examples of which are shown in the drawings. When referring to the drawings below, unless otherwise indicated, the same numerals in different drawings represent the same or similar elements. The embodiments described in the following exemplary embodiments do not represent all embodiments consistent with this application. Rather, they are merely examples of devices and methods consistent with some aspects of the application as detailed in the appended claims.
在智能交通系统中,车辆行驶在道路上可以检测其周围的其他车辆,这样司机在驾驶车辆时可以向司机提示周围的其他车辆,辅助司机驾驶车辆,或者,在自动驾驶时,车辆可以根据其周围的车辆规划行驶的车道或对其他车辆进行避让。In the intelligent transportation system, the vehicle can detect other vehicles around it while driving on the road, so that the driver can prompt the driver when driving the vehicle to assist the driver to drive the vehicle, or, when autonomous driving, the vehicle can be based on its The surrounding vehicles plan the driving lane or avoid other vehicles.
参见图1,本申请实施例提供了一种车载摄像头和车载雷达联动的目标检测系统,包括设置在车辆上的车载摄像头1、设置在车辆上的车载雷达2和与车载雷达2和车载摄像头1通信的检测设备3;车载摄像头1设置在车辆的前后方和/或左右两侧,车载雷达2设置在车辆的前后方;Referring to FIG. 1, an embodiment of the present application provides a target detection system in which an on-board camera and an on-board radar are linked, including an on-board camera on a vehicle 1, an on-board radar 2 on a vehicle, and an on-board radar 2 and on-board camera Communication detection equipment 3; vehicle-mounted camera 1 is arranged in front and rear and/or left and right sides of the vehicle, and vehicle-mounted radar 2 is arranged in front and rear of the vehicle;
车载摄像头1,用于对车辆周围环境进行拍摄得到视频图像;Vehicle-mounted camera 1, used to capture the video image of the surrounding environment of the vehicle;
车载雷达2,用于向车辆周围发送雷达波,接收该雷达波对应的反射波并获取该反射波的回波能量强度,得到速度距离图像;The vehicle-mounted radar 2 is used to send radar waves around the vehicle, receive the reflected waves corresponding to the radar waves and obtain the echo energy intensity of the reflected waves to obtain a speed distance image;
检测设备3,用于基于该视频图像检测车辆周围的各个图像目标在图像坐标系中的位置信息和置信度,该置信度用于表示图像目标对应的真实目标的目标类别为指定类别的概率,以及基于该速度距离图像检测车辆周围的雷达目标在雷达坐标系中的位置信息和置信度,该置信度用于表示雷达目标对应的真实目 标的目标类别为指定类别的概率;The detection device 3 is used to detect the position information and confidence level of each image target around the vehicle in the image coordinate system based on the video image, the confidence level is used to indicate the probability that the target category of the real target corresponding to the image target is a specified category, And detecting the position information and confidence of the radar target around the vehicle in the radar coordinate system based on the speed distance image, the confidence being used to indicate the probability that the target category of the real target corresponding to the radar target is the specified category;
检测设备3,还用于根据置信度超过第一预设阈值的图像目标的位置信息和置信度超过第二预设阈值的图像目标的位置信息,获取第一透视矩阵,第一透视矩阵用于表示图像像坐标系与预设道路坐标系之间的转换关系;通过第一透视矩阵,从置信度未超过第一预设阈值的图像目标和置信度未超过第二预设阈值的雷达目标中检测目标类别。The detection device 3 is further configured to acquire the first perspective matrix based on the position information of the image target whose confidence exceeds the first preset threshold and the position information of the image target whose confidence exceeds the second preset threshold. Represents the conversion relationship between the image image coordinate system and the preset road coordinate system; through the first perspective matrix, from the image target whose confidence does not exceed the first preset threshold and the radar target whose confidence does not exceed the second preset threshold Detection target category.
可选的,车辆中包括车载设备,车辆通过车载设备、车载摄像头和车载雷达检测周围的车辆。车载设备可以通过车载摄像头检测到车辆周围的每个图像目标,以及通过车载雷达检测到车辆周围的每个雷达目标。每个图像目标和每个雷达目标可能为指定类别的真实目标,指定类别的真实目标可以为车辆等。可选的,该车载设备可以是上述检测设备3。Optionally, the vehicle includes on-board equipment, and the vehicle detects surrounding vehicles through the on-board equipment, on-board camera, and on-board radar. The vehicle-mounted device can detect each image target around the vehicle through the vehicle-mounted camera, and each radar target around the vehicle through the vehicle-mounted radar. Each image target and each radar target may be real targets in a specified category, and the real targets in a specified category may be vehicles and the like. Optionally, the in-vehicle device may be the detection device 3 described above.
可选的,该车载雷达2可以为毫米波雷达或激光雷达等。车载设备可以为上述检测设备3。Optionally, the vehicle-mounted radar 2 may be a millimeter wave radar or a laser radar. The in-vehicle device may be the above-mentioned detection device 3.
需要说明的是:如果一个目标同时被车载摄像头检测出为指定类别的真实目标,以及被雷达检测出为指定类别的真实目标时,该目标为指定类别的一个真实目标的可能性较大。因此为了提高检测汽车周围目标的精度,可以将通过车载摄像头检测出的每个图像目标和通过雷达检测出的每个雷达目标融合,以提高检测精度。其中,上述检测图像目标、雷达目标以及融合的过程可以参见如下任一实施例的描述,在此先不说明。It should be noted that if a target is simultaneously detected as a real target in the specified category by the on-board camera and detected as a real target in the specified category by the radar, the target is more likely to be a real target in the specified category. Therefore, in order to improve the accuracy of detecting targets around the car, each image target detected by the on-board camera and each radar target detected by radar can be fused to improve the detection accuracy. Wherein, for the process of detecting the image target, the radar target and the fusion, please refer to the description of any of the following embodiments, which will not be described here.
参见图2,本申请实施例提供了一种车载摄像头和车载雷达联动的目标检测方法,该方法包括:Referring to FIG. 2, an embodiment of the present application provides a target detection method in which a car camera and a car radar are linked. The method includes:
步骤201:从车载摄像头提供的视频图像中检测出车辆周围的各个图像目标,以及各个图像目标的置信度,以及各个图像目标在图像坐标系中的位置信息。Step 201: Detect each image object around the vehicle, the confidence of each image object, and the position information of each image object in the image coordinate system from the video image provided by the on-board camera.
步骤202:从车载雷达提供的速度距离图像中检测出车辆周围的各个雷达目标,以及各个雷达目标在雷达坐标系中的位置信息,雷达目标的置信度,置信度用于表示图像目标或雷达目标所对应的真实目标的目标类别为指定类别的概率。Step 202: Detect each radar target around the vehicle from the speed and distance images provided by the onboard radar, and the position information of each radar target in the radar coordinate system, the confidence of the radar target, which is used to represent the image target or the radar target The target category of the corresponding real target is the probability of the specified category.
步骤203:根据置信度超过第一预设阈值的图像目标的位置信息和置信度超过第二预设阈值的雷达目标的位置信息,获取第一透视矩阵,第一透视矩阵用 于表示图像像坐标系与预设道路坐标系之间的转换关系。Step 203: Obtain a first perspective matrix according to the position information of the image target whose confidence exceeds the first preset threshold and the position information of the radar target whose confidence exceeds the second preset threshold, and the first perspective matrix is used to represent the image image coordinates The conversion relationship between the system and the preset road coordinate system.
步骤204:通过第一透视矩阵,从置信度未超过第一预设阈值的图像目标和置信度未超过第二预设阈值的雷达目标中检测出目标类别。Step 204: Using the first perspective matrix, detect the target category from the image target whose confidence does not exceed the first preset threshold and the radar target whose confidence does not exceed the second preset threshold.
可选的,步骤204之后,还包括步骤205,根据检测出的各个图像目标的置信度,对置信度超过第一预设阈值的图像目标进行分类,得到该图像目标对应的真实目标的目标类别,并将该目标类别输出,以及根据检测出的各个雷达目标的置信度,对置信度超过所述第二预设阈值的雷达目标进行分类,得到该雷达目标对应的真实目标的目标类别,并将该目标类别输出。通过对高置信度的图像目标和雷达目标进行分类可以分类出所需要的目标类别的目标并输出。通过此步骤,可以实现图4中计算图像目标的置信度和目标分类,以及计算雷达坐标的置信度和目标分类,以及进行高置信度目标的决策级融合。Optionally, after step 204, step 205 is further included, according to the detected confidence of each image target, the image target whose confidence exceeds the first preset threshold is classified to obtain the target category of the real target corresponding to the image target And output the target category, and classify the radar targets whose confidence exceeds the second preset threshold according to the detected confidence of each radar target to obtain the target category of the real target corresponding to the radar target, and Output the target category. By categorizing high-confidence image targets and radar targets, the targets of the desired target category can be classified and output. Through this step, it is possible to realize the calculation of the confidence and target classification of the image target in FIG. 4, the calculation of the confidence and target classification of the radar coordinates, and the decision-level fusion of high-confidence targets.
其中,参见图14,示意出了从车载摄像头提供的视频图像中检测出的车辆周围的各个图像目标,以及各个图像目标的置信度,以及各个图像目标在图像坐标系中的位置信息。叠加框内的目标是置信度超过第一预设阈值的图像目标,参见图15,示意出了从雷达提供的速度距离图像中检测出车辆周围的各个雷达目标,以及各个雷达目标在雷达坐标系中的位置信息,雷达目标的置信度,其中,叠加框内的目标是置信度超过所述第二预设阈值的雷达目标。Among them, referring to FIG. 14, each image object around the vehicle detected from the video image provided by the on-board camera, and the confidence of each image object, and the position information of each image object in the image coordinate system are illustrated. The target in the superimposed frame is an image target with a confidence exceeding the first preset threshold. Refer to FIG. 15 to illustrate the detection of various radar targets around the vehicle from the speed and distance images provided by the radar, and the radar targets in the radar coordinate system The position information in FIG. 3 is the confidence of the radar target, where the target in the superimposed frame is a radar target whose confidence exceeds the second preset threshold.
可选的,在步骤203之后,还包括步骤206,从置信度超过第一预设阈值的图像目标和置信度超过第二预设阈值的雷达目标中确定N个关联目标对,任意一个该关联目标对包括满足预设关联条件的雷达目标和图像目标,N为大于或等于1的正整数;如此可以根据N个关联目标对中的雷达目标和图像目标的位置信息,确定第一透视矩阵。Optionally, after step 203, step 206 is further included, and N associated target pairs are determined from the image target whose confidence exceeds the first preset threshold and the radar target whose confidence exceeds the second preset threshold, any one of the associations The target pair includes a radar target and an image target satisfying a preset association condition, and N is a positive integer greater than or equal to 1. In this way, the first perspective matrix can be determined according to the position information of the radar target and the image target in the N associated target pairs.
可选的,步骤206可以包括:Optionally, step 206 may include:
2061:根据第一图像目标的第一位置信息和存储的第二透视矩阵,将第一图像目标从所述图像坐标系映射至道路坐标系,得到第一图像目标在道路坐标系中对应的第三位置信息;其中第一图像目标为置信度超过第一预设阈值的图像目标,第一位置信息为第一图像目标在所述图像坐标系中的位置信息;2061: Based on the first position information of the first image object and the stored second perspective matrix, map the first image object from the image coordinate system to the road coordinate system to obtain the corresponding first image object in the road coordinate system. Three position information; wherein the first image object is an image object with a confidence exceeding a first preset threshold, and the first position information is the position information of the first image object in the image coordinate system;
2062:根据第一雷达目标的第二位置信息和存储的第三透视矩阵,将第一雷达目标从雷达坐标系映射至道路坐标系,得到第一雷达目标在道路坐标系中对应的第四位置信息;其中第一雷达目标为置信度超过第二预设阈值的雷达目标,第二位置信息为第一雷达目标在雷达坐标系中的位置信息;2062: According to the second position information of the first radar target and the stored third perspective matrix, map the first radar target from the radar coordinate system to the road coordinate system to obtain the corresponding fourth position of the first radar target in the road coordinate system Information; where the first radar target is a radar target whose confidence exceeds a second preset threshold, and the second position information is the position information of the first radar target in the radar coordinate system;
2063:根据各个第一图像目标的第三位置信息和各个第一雷达目标的第四位置信息,对各个第一图像目标与各个第一雷达目标进行位置关联,得到N个关联目标对。2063: Based on the third position information of each first image target and the fourth position information of each first radar target, position-associate each first image target with each first radar target to obtain N associated target pairs.
通过上述步骤2061至步骤2063,可以得到N个关联目标对,基于该N个关联目标对可以将车载摄像头检测的高置信度图像目标和雷达检测的高置信度雷达目标进行动态对齐,可以实现图4中的图像坐标系、道路坐标系和雷达坐标系间的动态对齐。Through the above steps 2061 to 2063, N associated target pairs can be obtained. Based on the N associated target pairs, the high-confidence image target detected by the vehicle camera and the high-confidence radar target detected by the radar can be dynamically aligned. The dynamic alignment between the image coordinate system, road coordinate system and radar coordinate system in 4.
可选的,上述步骤2063可以为:Optionally, the above step 2063 may be:
根据第一图像目标的第三位置信息和第一雷达目标的所述第四位置信息,确定第一图像目标在道路坐标系中的投影面积,第一雷达目标在道路坐标系中的投影面积,以及第一图像目标和第一雷达目标在道路坐标系中的重叠投影面积;Determine the projected area of the first image target in the road coordinate system according to the third position information of the first image target and the fourth position information of the first radar target, the projected area of the first radar target in the road coordinate system, And the overlapping projected area of the first image target and the first radar target in the road coordinate system;
根据第一图像目标在道路坐标系中的投影面积,第一雷达目标在道路坐标系中的投影面积,以及该重叠投影面积,确定各个第一图像目标与各个第二雷达目标之间的关联代价;According to the projected area of the first image target in the road coordinate system, the projected area of the first radar target in the road coordinate system, and the overlapping projected area, determine the associated cost between each first image target and each second radar target ;
从第一图像目标和第二图像目标中确定关联代价最小的一个第一雷达目标和一个第一图像目标为关联目标对,进而得到N个关联目标对。From the first image target and the second image target, it is determined that a first radar target and a first image target with the smallest associated cost are associated target pairs, and then N associated target pairs are obtained.
由于确定各个第一图像目标与各个第二雷达目标之间的关联代价,从第一图像目标和第二图像目标中确定关联代价最小的一个第一雷达目标和一个第一图像目标为关联目标对,从而可以提高关联目标对的准确性。Since the association cost between each first image target and each second radar target is determined, a first radar target and a first image target with the smallest association cost are determined from the first image target and the second image target as the association target pair , Which can improve the accuracy of the associated target pair.
可选的,上述步骤203,可以包括:Optionally, the above step 203 may include:
2031:针对N个关联目标对中的任一个关联目标对,根据该关联目标对中的第一雷达目标的位置信息,修正该关联目标对中的第一图像目标的位置信息;2031: For any associated target pair of the N associated target pairs, modify the position information of the first image target in the associated target pair according to the position information of the first radar target in the associated target pair;
2032:根据N个关联目标对中的各个第一图像目标修正后的位置信息,修正第二透视矩阵,得到第一透视矩阵。由于雷达坐标系的位置信息相对图像坐标系的位置信息更为准确,因此,对于一个关联目标对,使用雷达目标的位置信息,修正图像目标的位置信息,可以使得图像目标的位置信息与真实目标的位置信息更接近,按照修正后的各个图像目标的位置信息将第二透视矩阵修正为第一透视矩阵,从而得到的第一透视矩阵能够精准地反映当前图像坐标系与道路坐标系之间的转换关系,这有利于将低置信度图像目标映射到道路坐标系时的位置信息也进行相应的校正。2032: Correct the second perspective matrix according to the corrected position information of each first image target in the N associated target pairs to obtain the first perspective matrix. Because the position information of the radar coordinate system is more accurate than the position information of the image coordinate system, for a related target pair, using the position information of the radar target to modify the position information of the image target can make the position information of the image target and the real target The location information is closer, the second perspective matrix is corrected to the first perspective matrix according to the corrected location information of each image target, and the resulting first perspective matrix can accurately reflect the current image coordinate system and the road coordinate system. Conversion relationship, which is conducive to correct the position information when mapping low-confidence image objects to the road coordinate system.
可选的,上述步骤204可以包括:Optionally, the above step 204 may include:
2041:从置信度未超过第一预设阈值的图像目标和置信度未超过第二预设阈值的雷达目标中确定M个特征融合目标对,任意一个特征融合目标对包括满足预设关联条件的一个雷达目标和一个图像目标,M为大于或等于1的正整数。2041: M feature fusion target pairs are determined from the image target whose confidence level does not exceed the first preset threshold and the radar target whose confidence level does not exceed the second preset threshold, any one of the feature fusion target pairs includes those satisfying the preset association condition For a radar target and an image target, M is a positive integer greater than or equal to 1.
2042:针对任一特征融合目标对,将特征融合目标对中的雷达目标的回波能量特征和图像目标的图像特征分别进行卷积计算后进行拼接,得到该特征融合目标对所对应的融合特征图;2042: For any feature fusion target pair, the echo energy feature of the radar target in the feature fusion target pair and the image feature of the image target are respectively convoluted and stitched together to obtain the fusion feature corresponding to the feature fusion target pair Figure;
2043:将该融合特征图进行卷积和全连接计算后输入到分类网络进行目标分类,得到融合特征图对应的目标类别。2043: After performing convolution and full connection calculation on the fusion feature map, input it to the classification network for target classification, and obtain the target category corresponding to the fusion feature map.
通过上述步骤2041至步骤2043可以实现图4中的低置信度目标的特征级融合,具体的步骤2042和步骤2043可以参见图10所示的流程。Through the above steps 2041 to 2043, the feature level fusion of the low-confidence target in FIG. 4 can be achieved. For the specific steps 2042 and 2043, refer to the process shown in FIG. 10.
可选的,上述步骤2041可以为:Optionally, the above step 2041 may be:
通过第一透视矩阵,将第二图像目标从图像坐标系映射至道路坐标系中,得到第二图像目标在道路坐标系中对应的位置信息,以及,通过预先存储的第三透视矩阵,将第二雷达目标从雷达坐标系映射至道路坐标系中,得到第二雷达目标在道路坐标系中对应的位置信息,第二图像目标为置信度未超过第一预设阈值的图像目标,第二雷达目标为置信度未超过第二预设阈值的雷达目标;Through the first perspective matrix, the second image object is mapped from the image coordinate system to the road coordinate system to obtain the corresponding position information of the second image object in the road coordinate system, and, through the pre-stored third perspective matrix, the The second radar target is mapped from the radar coordinate system to the road coordinate system to obtain the corresponding position information of the second radar target in the road coordinate system. The second image target is an image target whose confidence level does not exceed the first preset threshold. The second radar target The target is a radar target whose confidence level does not exceed the second preset threshold;
根据各个第二图像目标在道路坐标系中对应的位置信息和各个第二雷达目标在道路坐标系中对应的位置信息,对各个第二图像目标和各个第二雷达目标进行位置关联,得到M个特征融合目标对。According to the corresponding position information of each second image target in the road coordinate system and the corresponding position information of each second radar target in the road coordinate system, position correlation is performed for each second image target and each second radar target to obtain M Feature fusion target pair.
可选的,上述步骤201,可以包括:Optionally, the above step 201 may include:
2011:根据车载摄像头提供的当前帧视频图像,以及与当前帧视频图像接近的多帧历史帧视频图像,获取任意一个图像目标的分类置信度、跟踪帧数置信度和位置置信度;2011: According to the current frame video image provided by the car camera, and the multi-frame historical frame video image close to the current frame video image, obtain the classification confidence, tracking frame number confidence and position confidence of any image target;
2012:根据该分类置信度、位置置信度和跟踪帧数置信度中的一个或多个,确定图像目标的置信度。2012: Determine the confidence of the image target based on one or more of the classification confidence, position confidence, and tracking frame number confidence.
可选的,上述步骤202,可以为:Optionally, the above step 202 may be:
根据当前帧速度距离图像中任一个雷达目标的回波能量强度、距离车辆的距离,以及雷达目标在多帧历史帧速度距离图像中的持续时间,确定雷达目标的置信度。The confidence of the radar target is determined according to the intensity of the echo energy of any radar target in the current frame speed distance image, the distance from the vehicle, and the duration of the radar target in the multi-frame historical frame speed distance image.
参见图3和图4,本申请实施例提供了一种车载摄像头和车载雷达联动的目标检测方法,该方法可以应用于图1所示的架构中,且该方法的执行主体可以为车载设备,该车载设备可以为图1所示系统中的检测设备3,包括:Referring to FIGS. 3 and 4, an embodiment of the present application provides a target detection method in which a car camera and a car radar are linked. This method can be applied to the architecture shown in FIG. 1, and the execution subject of the method can be a car device. The vehicle-mounted device may be the detection device 3 in the system shown in FIG. 1, including:
步骤301:从车辆的车载摄像头提供的视频图像中检测车辆周围的每个图像目标在图像坐标系中的位置信息、目标面积和置信度,该置信度用于表示图像目标对应的真实目标的目标类别为指定类别的概率。Step 301: Detect the position information, target area and confidence level of each image target around the vehicle in the image coordinate system from the video image provided by the vehicle's on-board camera. The confidence level is used to represent the target of the real target corresponding to the image target The category is the probability of the specified category.
车载摄像头可以对车辆周围的环境进行拍摄得到一帧帧的视频图像,每当拍摄得到的一帧视频图像时,为了便于说明将当前拍摄的一帧视频图像称为第一视频图像,将第一视频图像输入到车载设备;车载设备中包括第一卷积神经网络,可以根据第一视频图像通过第一卷积神经网络检测出第一视频图像中类别为指定类别的每个图像目标的位置信息、目标面积和置信度。The on-board camera can capture a frame of video image of the environment around the vehicle. Whenever a frame of video image is captured, for convenience of description, the current frame of video image is called the first video image. The video image is input to the in-vehicle device; the in-vehicle device includes the first convolutional neural network, which can detect the position information of each image target in the first video image in the specified category according to the first video image , Target area and confidence.
可选的,可以根据检测出的各个图像目标的置信度,对置信度超过第一预设阈值的图像目标进行分类,得到该图像目标对应的真实目标的目标类别,并将该目标类别输出。Optionally, the image objects whose confidence exceeds the first preset threshold may be classified according to the detected confidence of each image object, to obtain the target category of the real target corresponding to the image object, and output the target category.
其中,车载摄像头包括一个图像坐标系,对于每个图像目标,该图像目标的位置信息和目标面积是该图像目标在图像坐标系中的位置信息和面积。The vehicle-mounted camera includes an image coordinate system. For each image target, the position information and target area of the image target are the position information and area of the image target in the image coordinate system.
可选的,参见图5,车载设备可以通过3011至3015的操作检测每个图像目标的位置信息、目标面积和置信度,分别为:Optionally, referring to FIG. 5, the vehicle-mounted device may detect the position information, target area, and confidence of each image target through operations from 3011 to 3015, which are:
3011:通过第一卷积神经网络检测车辆的车载摄像头当前拍摄的第一视频图像,得到类别为指定类别的每个图像目标的位置信息和目标面积。3011: Detect the first video image currently captured by the vehicle's on-board camera through the first convolutional neural network to obtain the position information and target area of each image target of the specified category.
参见图6,第一卷积神经网络包括卷积神经网络(Convolutional Neural Network,CNN)、区域侯选网络(Region Proposal Network,RPN)和基于区域的卷积神经网络(Fast Region-based Convolution Neural Network,RCNN)等组成部分,事先在RCNN中设置有第一类别集合,该第一类别集合的类别可以包括指定类别,例如指定类别可以为车辆、第一类别集合还可以包括房子、树木、花台、路灯等其他类别。Referring to FIG. 6, the first convolutional neural network includes a convolutional neural network (Convolutional Neural Network, CNN), a regional candidate network (Region Proposal Network, RPN), and a region-based convolutional neural network (Fast Region-based Convolution Neural Network) , RCNN) and other components, a first category set is set in RCNN in advance, the category of the first category set may include a specified category, for example, the specified category may be a vehicle, the first category set may also include houses, trees, flower stands, Street lights and other categories.
本步骤可以为:首先车载设备在接收到车载摄像头输入的第一视频图像时,将第一图像输入到第一卷积神经网络,获取第一卷积神经网络输出的第一视频图像中的每个目标在图像坐标系的位置信息、面积和每个目标属于第一类别集合中的每个类别的概率。This step may be: first, when the vehicle-mounted device receives the first video image input from the vehicle-mounted camera, the first image is input to the first convolutional neural network, and each of the first video images output by the first convolutional neural network is acquired The location information, area and probability of each object in the image coordinate system belong to each category in the first category set.
在实现时,参照图6,车载设备可以将第一视频图像输入到CNN,由CNN 对第一视频图像进行卷积以及提取特征,得到第一视频图像对应的第一特征图,将第一特征图分别输入到RPN和RCNN。RPN在第一特征图中确定出至少一个第一侯选框区域,得到第二特征图(每个第一侯选框区域为一个第二特征图),将第二特征图也输入到RCNN。RCNN根据第一特征图对任意一个第二特征图,即对任一第一侯选框区域中的目标特征进行目标位置和目标类别的回归,最终得到第一候选框区域内的目标对象的位置信息和目标面积,以及每个目标对象属于第一类别集合中的每个类别的概率。车载设备获取RCNN输出的每个目标的位置信息和目标面积以及每个目标属于第一类别集合中的每个类别的概率。During implementation, referring to FIG. 6, the vehicle-mounted device may input the first video image to the CNN, and the CNN convolves the first video image and extracts features to obtain a first feature map corresponding to the first video image, and converts the first feature The graph is input to RPN and RCNN respectively. RPN determines at least one first candidate box area in the first feature map to obtain a second feature map (each first candidate box area is a second feature map), and also inputs the second feature map to the RCNN. RCNN performs regression on the target position and target category of any second feature map according to the first feature map, that is, the target feature in any first candidate frame area, and finally obtains the position of the target object in the first candidate frame area Information and target area, and the probability that each target object belongs to each category in the first category set. The vehicle-mounted device obtains the position information and target area of each target output by the RCNN and the probability that each target belongs to each category in the first category set.
然后对于每个目标,车载设备从该目标对应的每个类别的概率中选择最大概率,将最大概率对应的类别作为该目标的类别,如此可以得到每个目标的类别,再从每个目标中选择类别为指定类别的目标作为图像目标。Then for each target, the vehicle-mounted device selects the maximum probability from the probability of each category corresponding to the target, and takes the category corresponding to the maximum probability as the category of the target, so that the category of each target can be obtained, and then from each target Select the target with the specified category as the image target.
其中,图像目标的目标面积可能与图像目标的实际面积不同。Among them, the target area of the image target may be different from the actual area of the image target.
可选的,RCNN还可以输出每个目标对应的图像目标框,即每个图像目标还可以有对应的图像目标框,图像目标对应的图像目标框包括图像目标的图像。Optionally, the RCNN may also output an image target frame corresponding to each target, that is, each image target may also have a corresponding image target frame, and the image target frame corresponding to the image target includes the image of the image target.
3012:根据车载摄像头在当前之前拍摄的视频图像获取每个图像目标的分类置信度。3012: Obtain the classification confidence of each image target according to the video image taken by the vehicle camera before the current one.
在本步骤中,可以获取在当前之前最近拍摄的K帧视频图像并组成视频图像集合,对于任一个图像目标,为了便于说明称为待处理图像目标,确定该视频图像集合中的每个视频图像中的待处理图像目标的置信度,根据该视频图像集合中的每个视频图像中的待处理图像目标的置信度,获取待处理图像目标的分类置信度,K为预设数值。其中,待处理图像目标的分类置信度用于表示待处理图像目标在图像纹理上为真实车辆的概率。In this step, the K-frame video images taken most recently before the current can be obtained and constitute a video image set. For any image target, for convenience of description, it is called a to-be-processed image target, and each video image in the video image set is determined. The confidence of the target of the image to be processed in the video image set is obtained according to the confidence of the target of the image to be processed in each video image in the video image set, and K is a preset value. The classification confidence of the image object to be processed is used to indicate the probability that the image object to be processed is a real vehicle on the image texture.
视频图像集合包括第一个视频图像、第二个视频图像、……、第K个视频图像,第一个视频图像是该视频图像集合中最先拍摄的视频图像,第K个视频图像是最晚拍摄的视频图像。The video image collection includes the first video image, the second video image, ..., the Kth video image, the first video image is the first video image captured in the video image collection, and the Kth video image is the most Video images taken late.
可选的,对于确定该视频图像集合中的每个视频图像中的待处理图像目标的置信度的操作,实现方式可以为:Optionally, for the operation of determining the confidence of the image target to be processed in each video image in the video image set, the implementation manner may be:
根据第一视频图像中的每个图像目标的第一位置信息和第K个视频图像中的每个图像目标的第一位置信息,计算第一视频图像中的每个图像目标和第K个视频图像中的每个图像目标之间的距离,将距离小于预设距离阈值的两个图像目标确定为该两个视频图像中的同一目标。在当前之前已确定第k个视频图 像和第k-1个视频图像中哪些目标是同一目标,k=2、3、……、K,因此对于第一视频图像中的待处理图像目标,可以确定视频图像集合中的哪些视频图像中包括待处理图像目标,哪些视频图像不包括待处理图像目标。对于包括待处理图像目标的视频图像,在当前已获取该视频图像中的待处理图像目标的置信度,对于不包括待处理图像目标的视频图像,可以将该视频图像中的待处理图像目标的置信度设置为预设置信度,例如预设置信度可以为0、1等数值。According to the first position information of each image object in the first video image and the first position information of each image object in the Kth video image, calculate each image object and Kth video in the first video image For the distance between each image target in the image, two image targets whose distance is less than a preset distance threshold are determined as the same target in the two video images. Before the current, it has been determined which of the kth video image and the k-1th video image are the same target, k=2, 3, ..., K, so for the image target to be processed in the first video image, you can Determine which video images in the video image set include the image target to be processed and which video images do not include the image target to be processed. For the video image that includes the image target to be processed, the confidence of the image target to be processed in the video image has been obtained at present, and for the video image that does not include the image target to be processed, the The confidence level is set to a preset reliability, for example, the preset reliability may be a value of 0, 1, or the like.
可选的,车载终端中还保存有目标与图像数目的对应关系。对于该对应关系中的每条记录,该记录中保存了一个目标和包括该目标的图像数目。Optionally, the correspondence between the target and the number of images is also stored in the vehicle-mounted terminal. For each record in the correspondence, a target and the number of images including the target are saved in the record.
相应的,在确定第一视频图像中的图像目标和第K个视频图像中的图像目标是同一目标时,将该对应关系中的包括该图像目标的记录中保存的图像数目加1。在确定第一视频图像中的图像目标和第K个视频图像中的任何一个图像目标不同时,该图像目标可能是新出现的目标,设置包括该图像目标的图像数目为1,将该图像目标和1对应保存在该对应关系中。Correspondingly, when it is determined that the image target in the first video image and the image target in the K-th video image are the same target, the number of images saved in the record including the image target in the correspondence relationship is increased by one. When it is determined that the image target in the first video image is different from any image target in the Kth video image, the image target may be a newly emerging target, set the number of images including the image target to 1, and set the image target Correspond to 1 and stored in the corresponding relationship.
可选的,对于获取待处理图像目标的分类置信度的操作,可以按如下第一公式计算待处理图像目标的分类置信度。Optionally, for the operation of obtaining the classification confidence of the image target to be processed, the classification confidence of the image target to be processed may be calculated according to the following first formula.
第一公式为:
Figure PCTCN2019122171-appb-000001
The first formula is:
Figure PCTCN2019122171-appb-000001
其中,在上述第一公式中,C c为待处理图像目标的分类置信度,C c[k]为第k个视频图像中的待处理图像目标的置信度,γ为阻尼系数,为预设数值,例如γ=0.9等数值。 In the above first formula, C c is the classification confidence of the image target to be processed, C c[k] is the confidence level of the image target to be processed in the k-th video image, and γ is the damping coefficient, which is the preset Numerical value, for example, γ=0.9.
3013:根据车载摄像头在当前之前拍摄的视频图像获取每个图像目标的跟踪帧数置信度。3013: Obtain the confidence of the tracking frame number of each image target according to the video image taken by the vehicle camera before the current one.
从目标与图像数目的对应关系中获取待处理图像目标对应的图像数目,将该图像数目减1,得到车载摄像头在当前之前拍摄的包括待处理图像目标的第二视频图像的数目,根据车载摄像头在当前之前拍摄的包括待处理图像目标的第二视频图像的数目,获取待处理图像目标的跟踪帧数置信度。Obtain the number of images corresponding to the image target to be processed from the correspondence between the target and the number of images, and decrement the number of images by 1 to obtain the number of second video images including the image target to be processed captured by the vehicle camera before the current date, according to For the number of second video images captured before the current including the target of the image to be processed, obtain the confidence of the number of tracking frames of the target of the image to be processed.
其中,待处理图像目标的跟踪帧数置信度用于表示待处理图像目标为预设类别的真实目标的稳定性。The confidence of the tracking frame number of the image target to be processed is used to indicate the stability of the image target to be processed as a real target of a preset category.
可选的,可以按如下第二公式获取待处理图像目标的跟踪帧数置信度。Optionally, the confidence of the tracking frame number of the image target to be processed can be obtained according to the second formula below.
第二公式为:
Figure PCTCN2019122171-appb-000002
The second formula is:
Figure PCTCN2019122171-appb-000002
其中,在上述第二公式中,C T为待处理图像目标的跟踪帧数置信度,T为第二视频图像的数目。在第二公式中,当第二视频图像的数目T小于或等于8 时,待处理图像目标的跟踪帧数置信度C T=0.2+T/10,当第二视频图像的数目T大于8时,待处理图像目标的跟踪帧数置信度C T=1.0。 In the above second formula, CT is the confidence of the number of tracking frames of the image target to be processed, and T is the number of second video images. In the second formula, when the number T of the second video image is less than or equal to 8, the confidence of the number of tracking frames of the image target to be processed C T =0.2+T/10, when the number T of the second video image is greater than 8, , The confidence of the tracking frame number of the image target to be processed C T = 1.0.
3014:根据第一视频图像获取每个图像目标的位置置信度。3014: Acquire the position confidence of each image target according to the first video image.
在本步骤中,对于第一视频图像中的任一图像目标,为了便于说明称为待处理图像目标,根据待处理图像目标的第一位置信息获取待处理图像目标与车辆之间的距离,从第一视频图像中获取待处理图像目标对应的图像高度和待处理图像目标对应的目标框的被遮挡比例,目标框包括待处理图像目标对应的图像,根据该距离、该图像高度和该被遮挡比例获取待处理图像目标的位置置信度。In this step, for any image object in the first video image, for convenience of description, it is referred to as an image object to be processed, and the distance between the image object to be processed and the vehicle is obtained according to the first position information of the image object to be processed, from Obtain the image height corresponding to the image target to be processed and the blocked ratio of the target frame corresponding to the image target to be processed from the first video image, the target frame including the image corresponding to the image target to be processed, according to the distance, the image height and the blocked Proportionally obtain the position confidence of the image target to be processed.
可选的,可以按如下第三公式计算待处理图像目标的位置置信度。Optionally, the position confidence of the image target to be processed can be calculated according to the following third formula.
第三公式为:C p=σ×y/h; The third formula is: C p =σ×y/h;
其中,在上述第三公式中,C p为待处理图像目标的位置置信度,h为待处理图像目标对应的图像高度,y为待处理图像目标与车辆之间的距离,σ为待处理图像目标对应的目标框的被遮挡比例。 In the above third formula, C p is the position confidence of the image target to be processed, h is the image height corresponding to the image target to be processed, y is the distance between the image target to be processed and the vehicle, and σ is the image to be processed The occlusion ratio of the target frame corresponding to the target.
图像目标的位置置信度取决于图像目标的在第一视频图像中的位置信息与图像目标被遮挡的比例:当图像目标的位置信息越靠近第一视频图像的底端时,图像目标的位置置信度更高;当图像目标被其他目标遮挡的比例越小,图像目标的位置置信度更高。The confidence of the position of the image object depends on the ratio of the position information of the image object in the first video image to the image object being blocked: when the position information of the image object is closer to the bottom of the first video image, the position of the image object is confident The higher the degree; the smaller the proportion of the image target blocked by other targets, the higher the confidence of the position of the image target.
3015:根据每个图像目标的分类置信度、位置置信度和跟踪帧数置信度中的至少一个,分别获取每个图像目标的置信度。3015: Obtain the confidence of each image object according to at least one of the classification confidence, position confidence, and tracking frame number confidence of each image object.
对于每个图像目标,根据该图像目标的分类置信度、位置置信度和跟踪帧数置信度,按如下第四公式计算在第一视频图像中该图像目标的置信度。For each image object, according to the classification confidence, position confidence and tracking frame number confidence of the image object, the confidence of the image object in the first video image is calculated according to the following fourth formula.
第四公式为:C s=f(C c,C p,C T)=W c×C c+W p×C p+W T×C TThe fourth formula is: C s =f(C c ,C p ,C T )=W c ×C c +W p ×C p +W T ×C T ;
其中,在上述第四公式中,C s为图像目标的置信度,W c为图像目标的分类置信度的权值系数,W p为图像目标的位置置信度的权值系数,W T为图像目标的跟踪帧数置信度的权值系数,该三个权值系数均为预设数值。 In the above fourth formula, C s is the confidence of the image target, W c is the weight coefficient of the classification confidence of the image target, W p is the weight coefficient of the location confidence of the image target, and W T is the image Weight coefficients for the confidence of the tracking frame number of the target. The three weight coefficients are all preset values.
其中,通过车载摄像头检测出的置信度超过第一预设阈值的图像目标为指定类别的真实目标的准确性较高,置信度未超过第一预设阈值的图像目标为指定类别的真实目标的准确性较低。因此在本实施例中,可以将置信度超过第一预设阈值的图像目标确定为检测出的指定类别的真实目标。Among them, the image target detected by the on-board camera with a confidence level exceeding the first preset threshold is higher in the accuracy of the specified category of real targets, and the image target whose confidence level does not exceed the first preset threshold is the actual target of the designated category The accuracy is low. Therefore, in this embodiment, the image target whose confidence exceeds the first preset threshold may be determined as the detected real target of the specified category.
步骤302:从车辆的雷达提供的速度距离图像中检测车辆周围的每个雷达目 标在雷达坐标系中的位置信息和每个雷达目标的置信度。Step 302: Detect the position information of each radar target around the vehicle in the radar coordinate system from the speed distance image provided by the radar of the vehicle and the confidence of each radar target.
雷达具有雷达坐标系,检测出的每个雷达目标的位置信息为雷达坐标系中的位置信息。The radar has a radar coordinate system, and the detected position information of each radar target is the position information in the radar coordinate system.
可选的,根据检测出的各个雷达目标的置信度,对置信度超过第二预设阈值的雷达目标进行分类,得到该雷达目标对应的真实目标的目标类别,并将该目标类别输出。Optionally, according to the detected confidence of each radar target, classify the radar target whose confidence exceeds the second preset threshold to obtain the target category of the real target corresponding to the radar target, and output the target category.
可选的,参见图7,本步骤可以通过如下3021至3022的操作来实现,分别为:Optionally, referring to FIG. 7, this step can be implemented by the following operations from 3021 to 3022, which are:
3021:通过车辆的雷达检测车辆周围的目标类别为指定类别的每个雷达目标的位置信息、面积和回波能量强度。3021: The vehicle's radar detects the target category around the vehicle as the position information, area, and echo energy intensity of each radar target of the specified category.
雷达可以向车辆周围发送雷达波,该雷达波可以被汽车周围的物体反射形成反射波,雷达可以接收各反射波并获取各反射波的回波能量强度。可以根据各反射波的回波能量强度计算出各反射点的位置,根据各反射点的位置和回波能量强度绘制出一个能量图,该能量图可以为速度距离图像,该能量图中的包括至少一个能量块,每个能量块包括多个位置连续的反射点对应的回波能量强度和位置,每个能量块为一个目标。The radar can send radar waves to the surroundings of the vehicle. The radar waves can be reflected by objects around the car to form reflected waves. The radar can receive the reflected waves and obtain the return energy intensity of the reflected waves. The position of each reflection point can be calculated according to the echo energy intensity of each reflected wave, and an energy map can be drawn according to the position of each reflection point and the echo energy intensity. The energy map can be a speed distance image. The energy map includes At least one energy block, each energy block includes a plurality of continuous reflection points corresponding to the energy intensity and position of the echo, and each energy block is a target.
基于每个能量块可以确定每个能量块对应的目标的目标类别和面积。将目标类别为指定类别的目标确定为雷达目标,从雷达目标的能量块中找出回波能量强度极值点,将该极值点的位置确定为雷达目标在雷达坐标系中的位置信息。可以将雷达目标的能量块中的各回波能量强度的平均值作为雷达目标的回波能量强度,或者,对雷达目标的能量块中的各回波能量强度进行排序,选择排在中间位置的回波能量强度作为雷达目标的回波能量强度。Based on each energy block, the target category and area of the target corresponding to each energy block can be determined. The target with the target category as the specified category is determined as the radar target, the extreme point of the echo energy intensity is found from the energy block of the radar target, and the position of the extreme point is determined as the position information of the radar target in the radar coordinate system. The average value of the energy intensity of each echo in the energy block of the radar target can be used as the energy intensity of the echo of the radar target, or the energy intensity of each echo in the energy block of the radar target can be sorted to select the echo in the middle position The energy intensity is taken as the energy intensity of the echo of the radar target.
通过雷达检测出的雷达目标的面积为雷达目标的实际面积。The area of the radar target detected by the radar is the actual area of the radar target.
3022:根据每个雷达目标的位置信息和回波能量强度,分别获取每个雷达目标的置信度。3022: Obtain the confidence of each radar target separately according to the position information of each radar target and the intensity of the echo energy.
在本步骤中,获取雷达能够探测的最远距离和雷达的最大回波能量强度,对于每个雷达目标,根据该雷达目标的位置信息计算该雷达目标与车辆之间的距离,根据该最远距离、该最大回波能量强度、计算的距离和该雷达目标的回波能量强度,按如下第五公式计算该雷达目标的置信度。In this step, the longest distance that the radar can detect and the maximum echo energy intensity of the radar are obtained. For each radar target, the distance between the radar target and the vehicle is calculated according to the position information of the radar target. For the distance, the maximum echo energy intensity, the calculated distance, and the echo energy intensity of the radar target, the confidence of the radar target is calculated according to the following fifth formula.
第五公式为:C=a×d/d max+b×p/p maxThe fifth formula is: C=a×d/d max +b×p/p max ;
其中,在第五公式中,C为雷达目标的置信度,a为雷达目标位置的权值系 数,a可以大于0且小于1,d为计算的距离,d max为该最远距离,b为雷达目标回波能量强度的权值系数,b可以大于0且小于1,p为雷达目标的回波能量强度,p max为该最大回波能量强度。 Among them, in the fifth formula, C is the confidence of the radar target, a is the weight coefficient of the radar target position, a can be greater than 0 and less than 1, d is the calculated distance, d max is the furthest distance, b is The weight coefficient of the radar target's echo energy intensity, b can be greater than 0 and less than 1, p is the radar target's echo energy intensity, and p max is the maximum echo energy intensity.
其中,雷达检测出的置信度超过第二预设阈值的雷达目标为指定类别的真实目标的准确性较高,置信度未超过第二预设阈值的雷达目标为指定类别的真实目标的准确性较低。因此在本实施例中,可以将置信度超过第二预设阈值的雷达目标确定为检测出的指定类别的真实目标。Among them, the accuracy of the radar target detected by the radar with a confidence level exceeding the second preset threshold is a specified category of real targets, and the accuracy of the radar target with a confidence level not exceeding the second preset threshold is a designated category of real targets. Lower. Therefore, in this embodiment, the radar target whose confidence exceeds the second preset threshold may be determined as the detected real target of the specified category.
其中,上述步骤301和步骤302的之间执行顺序没有先后,可以先执行步骤301再执行步骤302,也可以先执行步骤301再执行步骤301,也可以同时执行步骤301和步骤302。There is no sequence between step 301 and step 302. Step 301 can be executed before step 302, step 301 can be executed before step 301, or step 301 and step 302 can be executed simultaneously.
步骤303:根据置信度超过第一预设阈值的图像目标的位置信息和置信度超过第二预设阈值的雷达目标的位置信息,获取第一透视矩阵,第一透视矩阵用于表示图像坐标系与预设道路坐标系之间的转换关系。第一预设阈值和第二预设阈值可以相同也可以不同。Step 303: Obtain the first perspective matrix according to the position information of the image target with confidence exceeding the first preset threshold and the position information of the radar target with confidence exceeding the second preset threshold, the first perspective matrix is used to represent the image coordinate system The conversion relationship with the preset road coordinate system. The first preset threshold and the second preset threshold may be the same or different.
可选的,在执行本步骤之前,从置信度超过第一预设阈值的图像目标和置信度超过第二预设阈值的雷达目标中确定N个关联目标对,任意一个关联目标对包括满足预设关联条件的雷达目标和图像目标,N为大于或等于1的正整数。Optionally, before performing this step, N associated target pairs are determined from the image targets whose confidence exceeds the first preset threshold and the radar targets whose confidence exceeds the second preset threshold. Any one of the associated target pairs includes Suppose the radar target and the image target of the association condition, N is a positive integer greater than or equal to 1.
参见图8,确定N个关联目标对的过程可以包括如下3031至3033的操作,该3031至3033的操作分别为:Referring to FIG. 8, the process of determining N associated target pairs may include the following operations from 3031 to 3033. The operations from 3031 to 3033 are as follows:
3031:根据第一图像目标的第一位置信息和存储的第二透视矩阵,将第一图像目标从图像坐标系映射至道路坐标系,得到第一图像目标在道路坐标系中对应的第三位置信息。3031: Map the first image object from the image coordinate system to the road coordinate system according to the first position information of the first image object and the stored second perspective matrix, to obtain the corresponding third position of the first image object in the road coordinate system information.
其中,第一图像目标为置信度超过第一预设阈值的图像目标,第一图像目标的第一位置信息为第一图像目标在图像坐标系中的位置信息。Wherein, the first image object is an image object whose confidence exceeds a first preset threshold, and the first position information of the first image object is position information of the first image object in the image coordinate system.
车载终端本地存储有在最近上一次获取的第二透视矩阵,第二透视矩阵用于反映车载摄像头的图像坐标系和预设道路坐标系之间的转换关系。The in-vehicle terminal locally stores the second perspective matrix obtained last time, and the second perspective matrix is used to reflect the conversion relationship between the image coordinate system of the in-vehicle camera and the preset road coordinate system.
道路坐标系是不同于车载摄像头的图像坐标系和雷达的雷达坐标系以外的另一坐标系,道路坐标系可以以车辆中的某一点的位置为坐标原点,以车辆前进方向为横轴的方向,纵轴垂直车辆前进方向。The road coordinate system is different from the image coordinate system of the on-board camera and the radar coordinate system of the radar. The road coordinate system can take the position of a certain point in the vehicle as the coordinate origin and the direction of the vehicle's forward axis as the horizontal axis. , The vertical axis is perpendicular to the direction of vehicle advancement.
可选的,可以将车辆的前保险杠的中心点位置作为道路坐标系的坐标原点。Optionally, the position of the center point of the front bumper of the vehicle can be used as the coordinate origin of the road coordinate system.
在本步骤中,将置信度超过第一预设阈值的每个第一图像目标的第一位置 信息构成第一矩阵,按如下第六公式得到第二矩阵,第二矩阵中包括置信度超过第一预设阈值的每个第一图像目标在道路坐标系中的第三位置信息。In this step, the first position information of each first image object whose confidence exceeds the first preset threshold constitutes a first matrix, and a second matrix is obtained according to the following sixth formula. The second matrix includes the confidence exceeding the first The third position information of each first image object in the road coordinate system with a preset threshold.
第六公式为:A1*B1=C1;其中,A1为第一矩阵,B1为第二透视矩阵,C1为第二矩阵。The sixth formula is: A1*B1=C1; where A1 is the first matrix, B1 is the second perspective matrix, and C1 is the second matrix.
可选的,还可以通过第二透视矩阵,转换置信度超过预设阈值的每个第一图像目标的目标面积,得到每个第一图像目标在道路坐标系中的投影面积,每个第一图像目标在道路坐标系中的投影面积为实际面积。Optionally, the second perspective matrix can also be used to convert the target area of each first image target whose confidence exceeds a preset threshold to obtain the projected area of each first image target in the road coordinate system. The projected area of the image object in the road coordinate system is the actual area.
3032:根据第一雷达目标的第二位置信息和存储的第三透视矩阵,将第一雷达目标从雷达坐标系映射至道路坐标系,得到第一雷达目标在道路坐标系中对应的第四位置信息。3032: According to the second position information of the first radar target and the stored third perspective matrix, map the first radar target from the radar coordinate system to the road coordinate system to obtain the corresponding fourth position of the first radar target in the road coordinate system information.
其中,第一雷达目标为置信度超过第二预设阈值的雷达目标,第二位置信息为第一雷达目标在雷达坐标系中的位置信息。Wherein, the first radar target is a radar target whose confidence exceeds a second preset threshold, and the second position information is the position information of the first radar target in the radar coordinate system.
在本步骤中,将置信度超过第二预设阈值的每个第一雷达目标的第二位置信息构成第三矩阵,按如下第七公式得到第四矩阵,第四矩阵中包括置信度超过第二预设阈值的每个第二雷达目标在道路坐标系中的第四位置信息。In this step, the second position information of each first radar target whose confidence exceeds the second preset threshold constitutes a third matrix, and a fourth matrix is obtained according to the following seventh formula. The fourth matrix includes the confidence exceeding the first The fourth position information of each second radar target in the road coordinate system with two preset thresholds.
第七公式为:A2*B2=C2;其中,A2为第三矩阵,B2为第三透视矩阵,C2为第四矩阵。The seventh formula is: A2*B2=C2; where A2 is the third matrix, B2 is the third perspective matrix, and C2 is the fourth matrix.
其中,置信度超过第二预设阈值的每个雷达目标的面积为雷达目标的实际面积,等于每个第一雷达目标在道路坐标系中的投影面积,所以不需要通过第三透视矩阵转换每个第一雷达目标的面积。Among them, the area of each radar target whose confidence exceeds the second preset threshold is the actual area of the radar target, which is equal to the projected area of each first radar target in the road coordinate system, so there is no need to convert each The area of the first radar target.
3033:根据各个第一图像目标的第三位置信息和投影面积以及各个第一雷达目标的第四位置信息和投影面积,对各个第一图像目标与各个第一雷达目标进行位置关联,得到N个关联目标对。3033: According to the third position information and the projection area of each first image target and the fourth position information and the projection area of each first radar target, positionally associate each first image target with each first radar target to obtain N Associated target pair.
可选的,可以通过如下第一和第二两步骤确定N个关联目标对,分别为:Optionally, the N associated target pairs can be determined through the first and second steps as follows:
第一步:根据第一图像目标的第三位置信息和投影面积以及每个雷达目标的第四位置信息和投影面积建立代价矩阵,每个第一图像目标对应代价矩阵中的一行,每个第一雷达目标对应代价矩阵的一列,该第一图像目标对应的一行包括该第一图像目标分别与每个第一雷达目标之间的代价系数,第一图像目标与第一雷达目标的代价系数表示第一图像目标和第一雷达目标为同一目标的概率。Step 1: Establish a cost matrix based on the third position information and projected area of the first image target and the fourth position information and projected area of each radar target, each first image target corresponds to a row in the cost matrix, and each A radar target corresponds to a column of the cost matrix, and a row corresponding to the first image target includes cost coefficients between the first image target and each first radar target, and the cost coefficients of the first image target and the first radar target represent The probability that the first image target and the first radar target are the same target.
例如,假设置信度超过第一预设阈值的第一图像目标为N个,置信度超过 第二预设阈值的第一雷达目标为X个,这样建立的代价矩阵包括N行X列。对于第i个第一图像目标和第j个第一雷达目标,i=1、2、……、N,j=1、2、……、X,根据第i个第一图像目标在道路坐标系中的第三位置信息和投影面积
Figure PCTCN2019122171-appb-000003
以及第j个第一雷达目标在道路坐标系中的第四位置信息和投影面积
Figure PCTCN2019122171-appb-000004
计算第i个第一图像目标和第j个第一雷达目标之间的投影重叠面积S ij
For example, suppose that there are N first image targets whose reliability exceeds the first preset threshold, and X first radar targets whose confidence exceeds the second preset threshold. The cost matrix thus established includes N rows and X columns. For the i-th first image target and the j-th first radar target, i = 1, 2, ..., N, j = 1, 2, ..., X, according to the i-th first image target at the road coordinates The third position information and projected area in the system
Figure PCTCN2019122171-appb-000003
And the fourth position information and projected area of the jth first radar target in the road coordinate system
Figure PCTCN2019122171-appb-000004
Calculate the projected overlap area S ij between the i-th first image target and the j-th first radar target.
然后根据第i个第一图像目标在道路坐标系中的投影面积
Figure PCTCN2019122171-appb-000005
第j个第一雷达目标在道路坐标系中的投影面积
Figure PCTCN2019122171-appb-000006
以及第i个第一图像目标和第j个第一雷达目标之间的投影重叠面积S ij,按如下第八公式计算第i个第一图像目标与第j个第一雷达目标之间的代价系数d ij。将第i个第一图像目标与第j个第一雷达目标之间的代价系数d ij作为代价矩阵的第i行第j列的元素。
Then according to the projected area of the i-th first image target in the road coordinate system
Figure PCTCN2019122171-appb-000005
Projected area of the jth first radar target in the road coordinate system
Figure PCTCN2019122171-appb-000006
And the projected overlap area S ij between the i-th first image target and the j-th first radar target, the cost between the i-th first image target and the j-th first radar target is calculated according to the following eighth formula Coefficient d ij . The cost coefficient d ij between the i-th first image target and the j-th first radar target is used as the element of the i-th row and j-th column of the cost matrix.
第八公式为:
Figure PCTCN2019122171-appb-000007
The eighth formula is:
Figure PCTCN2019122171-appb-000007
第二步:从第一图像目标对应的一行代价系数中选择最大代价系数,将第一图像目标和最大代价系数对应的第一雷达目标组成一对关联目标对。Step 2: Select a maximum cost coefficient from a row of cost coefficients corresponding to the first image target, and form a pair of associated target pairs with the first image target and the first radar target corresponding to the maximum cost coefficient.
可选的,在得到关联目标对后,根据该N个关联目标对中的雷达目标和图像目标的位置信息,确定第一透视矩阵。Optionally, after obtaining the associated target pair, the first perspective matrix is determined according to the position information of the radar target and the image target in the N associated target pairs.
可选的,针对该N个关联目标对中的任一个关联目标对,根据该关联目标对中的第一雷达目标的位置信息,修正该关联目标对中的第一图像目标的位置信息;根据该N个关联目标对中的各个所述第一图像目标修正后的位置信息,修正第二透视矩阵,得到第一透视矩阵。Optionally, for any one of the N associated target pairs, the position information of the first image target in the associated target pair is corrected according to the position information of the first radar target in the associated target pair; The corrected position information of each of the first image objects in the N associated target pairs corrects the second perspective matrix to obtain a first perspective matrix.
雷达检测到的第一雷达目标为指定类别的真实目标的可能性高于车载摄像头检测到的第一图像目标为指定类别的真实目标的可能性。因此在本步骤中,对于每个关联目标对,该关联目标对中包括的第一图像目标和第一雷达目标为同一目标,可以将该第一图像目标的第三位置信息修正为该第一雷达目标的第四位置信息,这样得到每个第一图像目标的第四位置信息。将每个第一图像目标的第四位置信息构建第五矩阵,根据第五矩阵和由每个第一图像目标的第一位置信息构成的第一矩阵按如下第九公式获取第一透视矩阵。The probability that the first radar target detected by the radar is a real target in a specified category is higher than the probability that the first image target detected by a vehicle camera is a real target in a specified category. Therefore, in this step, for each associated target pair, the first image target and the first radar target included in the associated target pair are the same target, and the third position information of the first image target can be corrected to the first The fourth position information of the radar target, so that the fourth position information of each first image target is obtained. A fifth matrix is constructed from the fourth position information of each first image object, and the first perspective matrix is obtained according to the following ninth formula according to the fifth matrix and the first matrix composed of the first position information of each first image object.
第九公式为:A1*B3=C3;其中,C3为第五矩阵,B3为第一透视矩阵。The ninth formula is: A1*B3=C3; where C3 is the fifth matrix and B3 is the first perspective matrix.
由于置信度超过第一预设阈值的第一图像目标和置信度超过第二预设阈值的第一雷达目标都为检测出来的指定类别的真实目标。因此在本步骤中根据置信度超过第一预设阈值的第一图像目标的位置信息和置信度超过第二预设阈值的第二雷达目标的位置信息,获取第一透视矩阵,可以使第一透视矩阵反映车 辆在当前路况下行驶车载摄像头的图像坐标系和道路坐标系之间的转换关系。Since the first image target with the confidence level exceeding the first preset threshold and the first radar target with the confidence level exceeding the second preset threshold are both detected real targets of the specified category. Therefore, in this step, according to the position information of the first image target whose confidence exceeds the first preset threshold and the position information of the second radar target whose confidence exceeds the second preset threshold, acquiring the first perspective matrix can make the first The perspective matrix reflects the conversion relationship between the image coordinate system of the vehicle-mounted camera and the road coordinate system when the vehicle is driving under the current road conditions.
可选的,还可以将车载设备保存的第二透视矩阵更新为第一透视矩阵。Optionally, the second perspective matrix saved in the vehicle-mounted device may also be updated to the first perspective matrix.
步骤304:通过第一透视矩阵,从置信度未超过第一预设阈值的图像目标和置信度未超过第二预设阈值的雷达目标中检测类别为指定类别的真实目标。Step 304: Using the first perspective matrix, detect the real target whose category is the specified category from the image target whose confidence level does not exceed the first preset threshold and the radar target whose confidence level does not exceed the second preset threshold.
可选的,参见图9和图10,本步骤可以3041至3043的操作来实现,分别为:Optionally, referring to FIG. 9 and FIG. 10, this step may be implemented by operations from 3041 to 3043, respectively:
3041:通过第一透视矩阵,将置信度未超过第一预设阈值的每个第二图像目标的位置信息映射至道路坐标系中,得到每个第二图像目标在道路坐标系中对应的位置信息。3041: Map the position information of each second image object whose confidence level does not exceed the first preset threshold to the road coordinate system through the first perspective matrix to obtain the corresponding position of each second image object in the road coordinate system information.
在本步骤中,将置信度未超过第一预设阈值的每个第二图像目标的位置信息构成第六矩阵,按如下第十公式得到第七矩阵,第七矩阵中包括置信度未超过第一预设阈值的每个第二图像目标在道路坐标系中的位置信息。In this step, the position information of each second image object whose confidence level does not exceed the first preset threshold constitutes a sixth matrix, and the seventh matrix is obtained according to the following tenth formula. The seventh matrix includes the confidence level not exceeding the first The position information of each second image object in the road coordinate system with a preset threshold.
第十公式为:A5*B3=C4;其中,A5为第六矩阵,B3为第一透视矩阵,C4为第七矩阵。The tenth formula is: A5*B3=C4; where A5 is the sixth matrix, B3 is the first perspective matrix, and C4 is the seventh matrix.
可选的,还可以通过第一透视矩阵,转换置信度未超过第一预设阈值的每第二个图像目标的目标面积,得到每个第二图像目标在道路坐标系中的投影面积,每个第二图像目标在道路坐标系中的投影面积为实际面积。Optionally, the first perspective matrix can also be used to convert the target area of each second image target whose confidence level does not exceed the first preset threshold, to obtain the projected area of each second image target in the road coordinate system. The projected area of a second image object in the road coordinate system is the actual area.
3042:通过第三透视矩阵将置信度未超过第二预设阈值的每个第二雷达目标映射至道路坐标系中,得到每个第二雷达目标在道路坐标系中对应的位置信息。3042: Map each second radar target whose confidence level does not exceed the second preset threshold into the road coordinate system through the third perspective matrix to obtain corresponding position information of each second radar target in the road coordinate system.
在本步骤中,将置信度未超过第二预设阈值的每个第二雷达目标的位置信息构成第八矩阵,按如下第十一公式得到第九矩阵,第九矩阵中包括置信度未超过第二预设阈值的每个第二雷达目标在道路坐标系中的位置信息。In this step, the position information of each second radar target whose confidence level does not exceed the second preset threshold constitutes an eighth matrix, and a ninth matrix is obtained according to the following eleventh formula, which includes the confidence level not exceeding The position information of each second radar target in the road coordinate system of the second preset threshold.
第十一公式为:A6*B2=C5;其中,A6为第八矩阵,B2为第三透视矩阵,C2为第九矩阵。The eleventh formula is: A6*B2=C5; where A6 is the eighth matrix, B2 is the third perspective matrix, and C2 is the ninth matrix.
其中,置信度未超过第二预设阈值的每个第二雷达目标的面积为雷达目标的实际面积,等于每个第二雷达目标在道路坐标系中的投影面积,所以不需要通过第一透视矩阵转换每个第二雷达目标的面积。The area of each second radar target whose confidence level does not exceed the second preset threshold is the actual area of the radar target, which is equal to the projected area of each second radar target in the road coordinate system, so there is no need to pass the first perspective The matrix converts the area of each second radar target.
3043:根据置信度未超过第一预设阈值的每个第二图像目标在道路坐标系的位置信息和投影面积以及置信度未超过第二预设阈值的每个第二雷达目标在道路坐标系的位置信息和投影面积,确定M个特征融合目标对,雷达目标对包 括为同一目标的图像目标和雷达目标。3043: According to the position information and projection area of each second image target in the road coordinate system whose confidence level does not exceed the first preset threshold and in the road coordinate system each second radar target whose confidence level does not exceed the second preset threshold Position information and projected area, M feature fusion target pairs are determined. The radar target pair includes the image target and the radar target that are the same target.
可选的,可以通过如下第一和第二两步骤确定M个特征融合目标对,分别为:Optionally, M feature fusion target pairs can be determined through the first and second steps as follows:
第一步:根据置信度未超过第一预设阈值的每个第二图像目标在道路坐标系的位置信息和投影面积以及置信度未超过第二预设阈值的每个第二雷达目标在道路坐标系的位置信息和投影面积建立第二代价矩阵,置信度未超过第一预设阈值的每个第二图像目标对应第二代价矩阵中的一行,置信度未超过第二预设阈值的每个第二雷达目标对应第二代价矩阵的一列,该第二图像目标对应的一行包括该图像目标分别与置信度未超过预设阈值的每个雷达目标之间的第二代价系数,该第二图像目标与该第二雷达目标的第二代价系数表示置信度未超过第一预设阈值的第二图像目标和置信度未超过第二预设阈值的第二雷达目标为同一目标的概率。Step 1: According to the position information and projection area of each second image target whose confidence level does not exceed the first preset threshold in the road coordinate system and each second radar target whose confidence level does not exceed the second preset threshold on the road The position information and the projected area of the coordinate system establish a second cost matrix, each second image object whose confidence level does not exceed the first preset threshold corresponds to a row in the second cost matrix, and each confidence level where the confidence level does not exceed the second preset threshold A second radar target corresponds to a column of a second cost matrix, and a row corresponding to the second image target includes a second cost coefficient between the image target and each radar target whose confidence level does not exceed a preset threshold, the second The second cost coefficient of the image target and the second radar target represents the probability that the second image target whose confidence level does not exceed the first preset threshold and the second radar target whose confidence level does not exceed the second preset threshold are the same target.
例如,假设置信度未超过第一预设阈值的第二图像目标为M个,置信度超过第二预设阈值的第二雷达目标为Y个,这样建立的代价矩阵包括M行Y列。对于第p个第二图像目标和第q个第二雷达目标,p=1、2、……、M,q=1、2、……、Y,根据第p个第二图像目标在道路坐标系中的位置信息和投影面积S p C以及第q个第二雷达目标在道路坐标系中的位置信息和投影面积S q R,计算第p个第二图像目标和第q个第二雷达目标之间的投影重叠面积S pqFor example, suppose that there are M second image targets whose reliability does not exceed the first preset threshold, and Y second radar targets whose confidence exceeds the second preset threshold. The cost matrix thus established includes M rows and Y columns. For the p-th second image target and the q-th second radar target, p = 1, 2, ..., M, q = 1, 2, ..., Y, according to the p-th second image target at the road coordinates Position information and projected area S p C in the system and the position information and projected area S q R of the q-th second radar target in the road coordinate system, calculate the p-th second image target and the q-th second radar target The projected overlap area between S pq .
然后根据第p个第二图像目标在道路坐标系中的投影面积S p C、第q个第二雷达目标在道路坐标系中的投影面积S q R以及第p个第二图像目标和第q个第二雷达目标之间的投影重叠面积S pq,按如下第十二公式计算第p个第二图像目标与第q个第二雷达目标之间的代价系数d pq。将第p个第二图像目标与第q个第二雷达目标之间的代价系数d pq作为代价矩阵的第p行第q列的元素。 Then according to the projection area S p C of the p-th second image target in the road coordinate system, the projection area S q R of the q-th second radar target in the road coordinate system, and the p-th second image target and the q-th The projected overlap area S pq between the second radar targets calculates the cost coefficient d pq between the p-th second image target and the q-th second radar target according to the twelfth formula as follows. The cost coefficient d pq between the p-th second image target and the q-th second radar target is taken as the element of the p-th row and the q-th column of the cost matrix.
第十二公式为:d pq=S pq/(S p C+S q R-S pq)。 The twelfth formula is: d pq =S pq /(S p C +S q R -S pq ).
第二步:从第二图像目标对应的一行第二代价系数中选择最大第二代价系数,将该第二图像目标和最大第二代价系数对应的第二雷达目标组成特征融合目标对。Step 2: Select a maximum second cost coefficient from a row of second cost coefficients corresponding to the second image target, and form a feature fusion target pair with the second image target and the second radar target corresponding to the maximum second cost coefficient.
3044:通过第二卷积神经网络从M个特征融合目标对中检测出目标类别为指定类别的真实目标。3044: The second convolutional neural network detects from the M feature fusion target pairs that the target category is the true target of the specified category.
车载设备包括第二卷积神经网络,事先在第二卷积神经网络中设置有第二类别集合,第二类别集合的类别可以包括指定类别、非指定类别和其他类别。 例如假设指定类别为车辆或机动车、非预设类别可以为非机动车。The vehicle-mounted device includes a second convolutional neural network, and a second category set is set in the second convolutional neural network in advance, and the categories of the second category set may include designated categories, non-designated categories, and other categories. For example, assume that the specified category is a vehicle or a motor vehicle, and the non-preset category may be a non-motor vehicle.
对于每个特征融合目标对,该特征融合目标对包括第二图像目标和第二雷达目标,车载设备可以将该第二图像目标在第一视频图像中对应的图像和该第二雷达目标对应的能量块输入到第二卷积神经网络,如图10所示,通过第二卷积神经网络从第二图像目标的图像提取的图像特征和从第二雷达目标的能量块中提取能量特征,将图像特征和能量特征进行拼接,得到特征序列。按照拼接后得到特征序列,通过第二卷积神经网络,进行多层卷积计算和全连接层计算后,输出该特征融合目标对属于第二类别集合中的每个类别的概率;选择概率最大的类别作为该特征融合目标对的目标类别,在该特征融合目标对的目标类别为指定类别时,将该特征融合目标对中的该第二雷达目标作为检测出类别为指定类别的真实目标。当然,第二卷积网络也可以由多个子网络构成,多个子网络分别完成如下过程:从第二图像目标的图像中提取图像特征,以及从第二雷达目标的能量块中提取回波能量特征;将图像特征和能量特征进行拼接,拼接后得到特征序列;根据该特征序列,进行多层卷积计算和全连接层计算后输出该特征融合目标对属于第二类别集合中的每个类别的概率。For each feature fusion target pair, the feature fusion target pair includes a second image target and a second radar target, and the in-vehicle device may use the corresponding image of the second image target in the first video image and the corresponding image of the second radar target The energy block is input to the second convolutional neural network. As shown in FIG. 10, the image features extracted from the image of the second image target and the energy features extracted from the energy block of the second radar target through the second convolutional neural network will be Image features and energy features are stitched together to obtain feature sequences. According to the feature sequence after splicing, through the second convolutional neural network, after performing multi-layer convolution calculation and fully connected layer calculation, the probability of the feature fusion target pair belonging to each category in the second category set is output; the maximum probability is selected Is used as the target category of the feature fusion target pair. When the target category of the feature fusion target pair is the specified category, the second radar target in the feature fusion target pair is used as the real target with the detected category as the specified category. Of course, the second convolutional network can also be composed of multiple sub-networks, each of which completes the following processes: extracting image features from the image of the second image target, and extracting echo energy features from the energy block of the second radar target ; Splicing the image features and energy features to obtain a feature sequence; according to the feature sequence, multi-layer convolution calculation and fully connected layer calculation are performed to output the feature fusion target for each category in the second category set Probability.
在该特征融合目标对的类别为指定类别时,表明该特征融合目标对中的第二图像目标和第二雷达目标的目标类别均为指定类别,也就是说车载摄像头和雷达同时检测出同一个目标为指定类别的真实目标,又由于雷达检测的精度高于车载摄像头检测的精度,所以将该特征融合目标对中的该第二雷达目标作为检测出目标类别为指定类别的真实目标。When the category of the feature fusion target pair is the specified category, it indicates that the target category of the second image target and the second radar target in the feature fusion target pair are both specified categories, that is to say that the vehicle camera and the radar simultaneously detect the same The target is a real target of the specified category, and because the accuracy of radar detection is higher than that of the on-board camera detection, the second radar target in the feature fusion target pair is used as the real target with the target category detected as the specified category.
可选的,第二卷积神经网络是事先通过样本集合训练得到的,样本集合包括预设的多个特征融合目标对和每个特征融合目标对应的目标类别。在训练时,将该样本集合输入到第二卷积神经网络进行训练。Optionally, the second convolutional neural network is obtained by training through a sample set in advance, and the sample set includes a plurality of preset feature fusion target pairs and a target category corresponding to each feature fusion target. During training, the sample set is input to the second convolutional neural network for training.
本申请实施例的有益效果为:通过车载摄像头检测出车辆周围的图像目标和图像目标的置信度,通过雷达检测出车辆周围的雷达目标和雷达目标的置信度;由于置信度超过第一预设阈值的第一图像目标和置信度超过第二预设阈值的第一雷达目标为指定类别的真实目标,所以根据置信度超过第一预设阈值的第一图像目标和置信度超过第二预设阈值的第一雷达目标获取的第一透视矩阵能够反映当前路况下车辆的车载摄像头的图像坐标系和道路坐标系之间的转换关系,这样根据第一透视矩阵从置信度未超过第一预设阈值的第二图像目标和置信度未超过第二预设阈值的第二雷达目标中检测出的目标类别为指定类别的 真实目标的精度较高,从而提高目标检测的精度。The beneficial effects of the embodiments of the present application are: the image target and the image target's confidence around the vehicle are detected by the on-board camera, and the radar target and the radar target's confidence around the vehicle are detected by the radar; because the confidence exceeds the first preset The first image target with a threshold and the first radar target with a confidence exceeding the second preset threshold are real targets of the specified category, so the first image target with a confidence exceeding the first preset threshold and the confidence exceeding the second preset The first perspective matrix acquired by the threshold first radar target can reflect the conversion relationship between the image coordinate system of the vehicle's on-board camera and the road coordinate system under the current road conditions, so that the confidence level does not exceed the first preset according to the first perspective matrix The accuracy of the target target detected in the second image target of the threshold and the second radar target whose confidence level does not exceed the second preset threshold is a real target of the specified category is high, thereby improving the accuracy of target detection.
下述为本申请装置实施例,可以用于执行本申请方法实施例。对于本申请装置实施例中未披露的细节,请参照本申请方法实施例。The following is an embodiment of the device of the present application, which can be used to execute the method embodiment of the present application. For details not disclosed in the device embodiments of the present application, please refer to the method embodiments of the present application.
参见图11,本申请实施例提供了一种车载摄像头和车载雷达联动的目标检测装置400,装置400包括:Referring to FIG. 11, an embodiment of the present application provides a target detection device 400 in which a vehicle camera and a vehicle radar are linked. The device 400 includes:
第一检测模块401,用于从车载摄像头提供的视频图像中检测出车辆周围的各个图像目标,各个所述图像目标的置信度,以及各个所述图像目标在图像坐标系中的位置信息;The first detection module 401 is used to detect each image target around the vehicle from the video image provided by the on-board camera, the confidence of each image target, and the position information of each image target in the image coordinate system;
第二检测模块402,用于从所述雷达提供的速度距离图像中检测出所述车辆周围的各个雷达目标,各个所述雷达目标在雷达坐标系中的位置信息,以及各个所述雷达目标的置信度,所述置信度用于表示所述图像目标或所述雷达目标所对应的真实目标的目标类别为指定类别的概率;The second detection module 402 is used to detect each radar target around the vehicle from the speed distance image provided by the radar, the position information of each radar target in the radar coordinate system, and the Confidence, the confidence is used to indicate the probability that the target category of the real target corresponding to the image target or the radar target is a specified category;
获取模块403,用于根据置信度超过第一预设阈值的图像目标的位置信息和置信度超过第二预设阈值的雷达目标的位置信息,获取第一透视矩阵,所述第一透视矩阵用于表示所述图像像坐标系与预设道路坐标系之间的转换关系;The obtaining module 403 is configured to obtain the first perspective matrix based on the position information of the image target whose confidence exceeds the first preset threshold and the position information of the radar target whose confidence exceeds the second preset threshold. Representing the conversion relationship between the image coordinate system and the preset road coordinate system;
第三检测模块404,用于通过所述第一透视矩阵,从置信度未超过所述第一预设阈值的图像目标和置信度未超过所述第二预设阈值的雷达目标中检测出目标类别。The third detection module 404 is configured to detect the target from the image target whose confidence does not exceed the first preset threshold and the radar target whose confidence does not exceed the second preset threshold through the first perspective matrix category.
可选的,所述装置400还包括:Optionally, the device 400 further includes:
分类模块,用于根据检测出的各个所述图像目标的置信度,对置信度超过所述第一预设阈值的图像目标进行分类,得到该图像目标对应的真实目标的目标类别,并将该目标类别输出,以及,根据检测出的各个所述雷达目标的置信度,对置信度超过所述第二预设阈值的雷达目标进行分类,得到该雷达目标对应的真实目标的目标类别,并将该目标类别输出。A classification module is used to classify image objects whose confidence exceeds the first preset threshold according to the detected confidence of each of the image objects, to obtain the target category of the real target corresponding to the image object, and to Target category output, and, based on the detected confidence of each of the radar targets, classify the radar targets with confidence exceeding the second preset threshold to obtain the target category of the real target corresponding to the radar target, and The target category is output.
可选的,所述装置400还包括:Optionally, the device 400 further includes:
确定模块,用于从置信度超过第一预设阈值的图像目标和置信度超过第二预设阈值的雷达目标中确定N个关联目标对,任意一个所述关联目标对包括满足预设关联条件的雷达目标和图像目标,所述N为大于或等于1的正整数;A determining module, configured to determine N associated target pairs from an image target with a confidence exceeding a first preset threshold and a radar target with a confidence exceeding a second preset threshold, any one of the associated target pairs includes satisfying a preset associated condition Radar target and image target, N is a positive integer greater than or equal to 1;
所述获取模块,用于根据所述N个关联目标对中的雷达目标和图像目标的位置信息,确定所述第一透视矩阵。The acquisition module is configured to determine the first perspective matrix according to the position information of the radar target and the image target in the N associated target pairs.
可选的,所述确定模块,用于:Optionally, the determination module is used to:
根据第一图像目标的第一位置信息和存储的第二透视矩阵,将所述第一图像目标从所述图像坐标系映射至所述道路坐标系,得到所述第一图像目标在所述道路坐标系中对应的第三位置信息;其中所述第一图像目标为置信度超过第一预设阈值的图像目标,所述第一位置信息为所述第一图像目标在所述图像坐标系中的位置信息;Mapping the first image object from the image coordinate system to the road coordinate system according to the first position information of the first image object and the stored second perspective matrix, to obtain the first image object on the road Corresponding third position information in the coordinate system; wherein the first image object is an image object whose confidence exceeds a first preset threshold, and the first position information is the first image object in the image coordinate system Location information
根据第一雷达目标的第二位置信息和存储的第三透视矩阵,将所述第一雷达目标从所述雷达坐标系映射至所述道路坐标系,得到所述第一雷达目标在所述道路坐标系中对应的第四位置信息;其中所述第一雷达目标为置信度超过第二预设阈值的雷达目标,所述第二位置信息为所述第一雷达目标在所述雷达坐标系中的位置信息;Mapping the first radar target from the radar coordinate system to the road coordinate system according to the second position information of the first radar target and the stored third perspective matrix, to obtain the first radar target on the road Corresponding fourth position information in the coordinate system; wherein the first radar target is a radar target whose confidence exceeds a second preset threshold, and the second position information is the first radar target in the radar coordinate system Location information
根据各个所述第一图像目标的所述第三位置信息和各个所述第一雷达目标的所述第四位置信息,对各个所述第一图像目标与各个所述第一雷达目标进行位置关联,得到所述N个关联目标对。According to the third position information of each of the first image targets and the fourth position information of each of the first radar targets, position-associating each of the first image targets with each of the first radar targets To obtain the N associated target pairs.
可选的,所述确定模块,用于:Optionally, the determination module is used to:
根据所述第一图像目标的所述第三位置信息和所述第一雷达目标的所述第四位置信息,确定所述第一图像目标在所述道路坐标系中的投影面积,所述第一雷达目标在所述道路坐标系中的投影面积,以及所述第一图像目标和所述第一雷达目标在所述道路坐标系中的重叠投影面积;Determine the projected area of the first image target in the road coordinate system according to the third position information of the first image target and the fourth position information of the first radar target, the first A projected area of a radar target in the road coordinate system, and an overlapping projected area of the first image target and the first radar target in the road coordinate system;
根据所述第一图像目标在所述道路坐标系中的投影面积,所述第一雷达目标在所述道路坐标系中的投影面积,以及所述重叠投影面积,确定各个所述第一图像目标与各个所述第二雷达目标之间的关联代价;Determine each of the first image targets according to the projected area of the first image target in the road coordinate system, the projected area of the first radar target in the road coordinate system, and the overlapping projected area The associated cost with each of the second radar targets;
从所述第一图像目标和所述第二图像目标中确定关联代价最小的一个所述第一雷达目标和一个所述第一图像目标为关联目标对,进而得到所述N个关联目标对。It is determined from the first image target and the second image target that one of the first radar target and one of the first image target with the smallest associated cost are the associated target pairs, and then the N associated target pairs are obtained.
可选的,所述获取模块403,用于:Optionally, the obtaining module 403 is used to:
针对所述N个关联目标对中的任一个关联目标对,根据所述任一个关联目标对中的所述第一雷达目标的位置信息,修正所述任一个关联目标对中的所述第一图像目标的位置信息;For any one of the N associated target pairs, according to the position information of the first radar target in the any associated target pair, modify the first of the any associated target pair Image target location information;
根据所述N个关联目标对中的各个所述第一图像目标修正后的位置信息,修正所述第二透视矩阵,得到所述第一透视矩阵。Modify the second perspective matrix according to the corrected position information of each of the first image targets in the N associated target pairs to obtain the first perspective matrix.
可选的,所述第三检测模块404,用于:Optionally, the third detection module 404 is used to:
从置信度未超过第一预设阈值的图像目标和置信度未超过第二预设阈值的雷达目标中确定M个特征融合目标对,任意一个所述特征融合目标对包括满足预设关联条件的一个雷达目标和一个图像目标,所述M为大于或等于1的正整数;M feature fusion target pairs are determined from the image target whose confidence level does not exceed the first preset threshold and the radar target whose confidence level does not exceed the second preset threshold, any one of the feature fusion target pairs includes those satisfying the preset association condition One radar target and one image target, where M is a positive integer greater than or equal to 1;
针对任一所述特征融合目标对,将所述特征融合目标对中的所述雷达目标的回波能量特征和所述图像目标的图像特征分别进行卷积计算后进行拼接,得到所述特征融合目标对所对应的融合特征图;For any of the feature fusion target pairs, the echo energy feature of the radar target and the image feature of the image target in the feature fusion target pair are respectively convoluted and spliced to obtain the feature fusion The fusion feature map corresponding to the target pair;
将所述融合特征图进行卷积和全连接计算后输入到分类网络进行目标分类,得到所述融合特征图对应的目标类别。After performing convolution and full connection calculation on the fusion feature map, input it to a classification network for target classification to obtain a target category corresponding to the fusion feature map.
可选的,所述第三检测模块404,用于:Optionally, the third detection module 404 is used to:
通过所述第一透视矩阵,将第二图像目标从所述图像坐标系映射至所述道路坐标系中,得到所述第二图像目标在所述道路坐标系中对应的位置信息,以及,通过预先存储的第三透视矩阵,将第二雷达目标从所述雷达坐标系映射至所述道路坐标系中,得到所述第二雷达目标在所述道路坐标系中对应的位置信息,所述第二图像目标为置信度未超过第一预设阈值的图像目标,所述第二雷达目标为置信度未超过第二预设阈值的雷达目标;Mapping the second image object from the image coordinate system to the road coordinate system through the first perspective matrix to obtain the corresponding position information of the second image object in the road coordinate system, and, by A pre-stored third perspective matrix, mapping the second radar target from the radar coordinate system to the road coordinate system, to obtain the corresponding position information of the second radar target in the road coordinate system, the first The second image target is an image target whose confidence level does not exceed the first preset threshold, and the second radar target is a radar target whose confidence level does not exceed the second preset threshold;
根据各个所述第二图像目标在所述道路坐标系中对应的位置信息和各个所述第二雷达目标在所述道路坐标系中对应的位置信息,对各个所述第二图像目标和各个所述第二雷达目标进行位置关联,得到所述M个特征融合目标对。According to the corresponding position information of each second image target in the road coordinate system and the corresponding position information of each second radar target in the road coordinate system, for each second image target and each position The second radar target performs position correlation to obtain the M feature fusion target pairs.
可选的,所述第一检测模块401,用于:Optionally, the first detection module 401 is used to:
根据所述车载摄像头提供的当前帧视频图像,以及与所述当前帧视频图像接近的多帧历史帧视频图像,获取所述任意一个所述图像目标的分类置信度、跟踪帧数置信度和位置置信度;Obtain the classification confidence, tracking frame number confidence and position of any one of the image objects according to the current frame video image provided by the vehicle camera and the multi-frame historical frame video image close to the current frame video image Confidence;
根据所述分类置信度、位置置信度和跟踪帧数置信度中的一个或多个,确定所述图像目标的置信度;Determine the confidence of the image target according to one or more of the classification confidence, position confidence, and tracking frame number confidence;
所述第二检测模块402,用于:The second detection module 402 is used to:
根据所述当前帧速度距离图像中任一个所述雷达目标的回波能量强度、距离所述车辆的距离,以及所述雷达目标在多帧历史帧速度距离图像中的持续时间,确定所述雷达目标的置信度。Determine the radar according to the intensity of the echo energy of any of the radar targets in the current frame speed distance image, the distance from the vehicle, and the duration of the radar target in the multi-frame historical frame speed distance image The confidence of the goal.
本申请实施例的有益效果为:第一检测模块通过车载摄像头检测出车辆周 围的图像目标和图像目标的置信度,第二检测模块通过雷达检测出车辆周围的雷达目标和每个雷达目标的置信度;由于置信度超过第一预设阈值的图像目标和置信度超过第二预设阈值的雷达目标为指定类别的真实目标,所以获取模块根据置信度超过第一预设阈值的图像目标和置信度超过第二预设阈值的雷达目标为指定类别的真实目标获取的第一透视矩阵能够反映当前路况下车辆的车载摄像头的图像坐标系和道路坐标系的转换关系,这样第三检测模块根据第一透视矩阵从置信度未超过第一预设阈值的图像目标和置信度未超过第二预设阈值的雷达目标中检测出的目标类别为指定类别的真实目标的精度较高,从而提高目标检测的精度。The beneficial effects of the embodiments of the present application are as follows: the first detection module detects the image target and the confidence of the image target around the vehicle through the on-board camera, and the second detection module detects the radar target around the vehicle and the confidence of each radar target through the radar Degree; since the image target with confidence exceeding the first preset threshold and the radar target with confidence exceeding the second preset threshold are real targets in the specified category, the acquisition module according to the image target and confidence with the confidence exceeding the first preset threshold The radar target whose degree exceeds the second preset threshold is the real perspective of the specified category. The first perspective matrix can reflect the conversion relationship between the image coordinate system of the vehicle's on-board camera and the road coordinate system under the current road conditions. A perspective matrix detects the target class from the image target whose confidence level does not exceed the first preset threshold and the radar target whose confidence level does not exceed the second preset threshold. Accuracy.
关于上述实施例中的装置,其中各个模块执行操作的具体方式已经在有关该方法的实施例中进行了详细描述,此处将不做详细阐述说明。Regarding the device in the above embodiment, the specific manner in which each module performs operations has been described in detail in the embodiment related to the method, and will not be elaborated here.
图12示出了本申请一个示例性实施例提供的终端500的结构框图。该终端500可以是车载终端,终端500还可能被称为用户设备、便携式终端等其他名称。FIG. 12 shows a structural block diagram of a terminal 500 provided by an exemplary embodiment of the present application. The terminal 500 may be an in-vehicle terminal, and the terminal 500 may also be called other names such as user equipment and portable terminal.
通常,终端500包括有:处理器501和存储器502。Generally, the terminal 500 includes a processor 501 and a memory 502.
处理器501可以包括一个或多个处理核心,比如4核心处理器、8核心处理器等。处理器501可以采用DSP(Digital Signal Processing,数字信号处理)、FPGA(Field-Programmable Gate Array,现场可编程门阵列)、PLA(Programmable Logic Array,可编程逻辑阵列)中的至少一种硬件形式来实现。处理器501也可以包括主处理器和协处理器,主处理器是用于对在唤醒状态下的数据进行处理的处理器,也称CPU(Central Processing Unit,中央处理器);协处理器是用于对在待机状态下的数据进行处理的低功耗处理器。在一些实施例中,处理器501可以在集成有GPU(Graphics Processing Unit,图像处理器),GPU用于负责显示屏所需要显示的内容的渲染和绘制。一些实施例中,处理器501还可以包括AI(Artificial Intelligence,人工智能)处理器,该AI处理器用于处理有关机器学习的计算操作。The processor 501 may include one or more processing cores, such as a 4-core processor, an 8-core processor, and so on. The processor 501 may adopt at least one hardware form of DSP (Digital Signal Processing, digital signal processing), FPGA (Field-Programmable Gate Array), PLA (Programmable Logic Array). achieve. The processor 501 may also include a main processor and a coprocessor. The main processor is a processor for processing data in a wake-up state, also known as a CPU (Central Processing Unit, central processor); the coprocessor is A low-power processor for processing data in the standby state. In some embodiments, the processor 501 may be integrated with a GPU (Graphics Processing Unit, image processor), and the GPU is used to render and draw content that needs to be displayed on the display screen. In some embodiments, the processor 501 may further include an AI (Artificial Intelligence, artificial intelligence) processor, which is used to process computing operations related to machine learning.
存储器502可以包括一个或多个计算机可读存储介质,该计算机可读存储介质可以是非暂态的。存储器502还可包括高速随机存取存储器,以及非易失性存储器,比如一个或多个磁盘存储设备、闪存存储设备。在一些实施例中,存储器502中的非暂态的计算机可读存储介质用于存储至少一个指令,该至少 一个指令用于被处理器501所执行以实现本申请中方法实施例提供的基于摄像头和雷达的车辆检测方法。The memory 502 may include one or more computer-readable storage media, which may be non-transitory. The memory 502 may also include high-speed random access memory, as well as non-volatile memory, such as one or more magnetic disk storage devices, flash memory storage devices. In some embodiments, the non-transitory computer-readable storage medium in the memory 502 is used to store at least one instruction, which is executed by the processor 501 to implement the camera-based camera provided by the method embodiment in the present application And radar vehicle detection methods.
在一些实施例中,终端500还可选包括有:外围设备接口503和至少一个外围设备。处理器501、存储器502和外围设备接口503之间可以通过总线或信号线相连。各个外围设备可以通过总线、信号线或电路板与外围设备接口503相连。具体地,外围设备包括:射频电路504、触摸显示屏505、摄像头506、音频电路507、定位组件508和电源509中的至少一种。In some embodiments, the terminal 500 may optionally further include: a peripheral device interface 503 and at least one peripheral device. The processor 501, the memory 502 and the peripheral device interface 503 may be connected by a bus or a signal line. Each peripheral device may be connected to the peripheral device interface 503 through a bus, a signal line, or a circuit board. Specifically, the peripheral device includes at least one of a radio frequency circuit 504, a touch display screen 505, a camera 506, an audio circuit 507, a positioning component 508, and a power supply 509.
外围设备接口503可被用于将I/O(Input/Output,输入/输出)相关的至少一个外围设备连接到处理器501和存储器502。在一些实施例中,处理器501、存储器502和外围设备接口503被集成在同一芯片或电路板上;在一些其他实施例中,处理器501、存储器502和外围设备接口503中的任意一个或两个可以在单独的芯片或电路板上实现,本实施例对此不加以限定。The peripheral device interface 503 may be used to connect at least one peripheral device related to I/O (Input/Output) to the processor 501 and the memory 502. In some embodiments, the processor 501, the memory 502, and the peripheral device interface 503 are integrated on the same chip or circuit board; in some other embodiments, any one of the processor 501, the memory 502, and the peripheral device interface 503 or Both can be implemented on a separate chip or circuit board, which is not limited in this embodiment.
射频电路504用于接收和发射RF(Radio Frequency,射频)信号,也称电磁信号。射频电路504通过电磁信号与通信网络以及其他通信设备进行通信。射频电路504将电信号转换为电磁信号进行发送,或者,将接收到的电磁信号转换为电信号。可选地,射频电路504包括:天线系统、RF收发器、一个或多个放大器、调谐器、振荡器、数字信号处理器、编解码芯片组、用户身份模块卡等等。射频电路504可以通过至少一种无线通信协议来与其它终端进行通信。该无线通信协议包括但不限于:万维网、城域网、内联网、各代移动通信网络(2G、3G、4G及5G)、无线局域网和/或WiFi(Wireless Fidelity,无线保真)网络。在一些实施例中,射频电路504还可以包括NFC(Near Field Communication,近距离无线通信)有关的电路,本申请对此不加以限定。The radio frequency circuit 504 is used to receive and transmit RF (Radio Frequency) signals, also called electromagnetic signals. The radio frequency circuit 504 communicates with a communication network and other communication devices through electromagnetic signals. The radio frequency circuit 504 converts the electrical signal into an electromagnetic signal for transmission, or converts the received electromagnetic signal into an electrical signal. Optionally, the radio frequency circuit 504 includes: an antenna system, an RF transceiver, one or more amplifiers, a tuner, an oscillator, a digital signal processor, a codec chipset, a user identity module card, and so on. The radio frequency circuit 504 can communicate with other terminals through at least one wireless communication protocol. The wireless communication protocol includes but is not limited to: World Wide Web, Metropolitan Area Network, Intranet, various generations of mobile communication networks (2G, 3G, 4G, and 5G), wireless local area networks, and/or WiFi (Wireless Fidelity) networks. In some embodiments, the radio frequency circuit 504 may further include NFC (Near Field Communication) related circuits, which is not limited in this application.
显示屏505用于显示UI(User Interface,用户界面)。该UI可以包括图形、文本、图标、视频及其它们的任意组合。当显示屏505是触摸显示屏时,显示屏505还具有采集在显示屏505的表面或表面上方的触摸信号的能力。该触摸信号可以作为控制信号输入至处理器501进行处理。此时,显示屏505还可以用于提供虚拟按钮和/或虚拟键盘,也称软按钮和/或软键盘。在一些实施例中,显示屏505可以为一个,设置终端500的前面板;在另一些实施例中,显示屏505可以为至少两个,分别设置在终端500的不同表面或呈折叠设计;在再一些实施例中,显示屏505可以是柔性显示屏,设置在终端500的弯曲表面上或折叠面上。甚至,显示屏505还可以设置成非矩形的不规则图形,也即异形屏。 显示屏505可以采用LCD(Liquid Crystal Display,液晶显示屏)、OLED(Organic Light-Emitting Diode,有机发光二极管)等材质制备。The display screen 505 is used to display UI (User Interface). The UI may include graphics, text, icons, video, and any combination thereof. When the display screen 505 is a touch display screen, the display screen 505 also has the ability to collect touch signals on or above the surface of the display screen 505. The touch signal can be input to the processor 501 as a control signal for processing. At this time, the display screen 505 can also be used to provide virtual buttons and/or virtual keyboards, also called soft buttons and/or soft keyboards. In some embodiments, there may be one display screen 505, which is provided with the front panel of the terminal 500; in other embodiments, there may be at least two display screens 505, which are respectively provided on different surfaces of the terminal 500 or have a folded design; In still other embodiments, the display screen 505 may be a flexible display screen, which is disposed on the curved surface or the folding surface of the terminal 500. Even, the display screen 505 can also be set as a non-rectangular irregular figure, that is, a special-shaped screen. The display screen 505 may be made of LCD (Liquid Crystal Display), OLED (Organic Light-Emitting Diode, organic light emitting diode) and other materials.
摄像头组件506用于采集图像或视频。可选地,摄像头组件506包括前置摄像头和后置摄像头。通常,前置摄像头设置在终端的前面板,后置摄像头设置在终端的背面。在一些实施例中,后置摄像头为至少两个,分别为主摄像头、景深摄像头、广角摄像头、长焦摄像头中的任意一种,以实现主摄像头和景深摄像头融合实现背景虚化功能、主摄像头和广角摄像头融合实现全景拍摄以及VR(Virtual Reality,虚拟现实)拍摄功能或者其它融合拍摄功能。在一些实施例中,摄像头组件506还可以包括闪光灯。闪光灯可以是单色温闪光灯,也可以是双色温闪光灯。双色温闪光灯是指暖光闪光灯和冷光闪光灯的组合,可以用于不同色温下的光线补偿。The camera component 506 is used to collect images or videos. Optionally, the camera assembly 506 includes a front camera and a rear camera. Usually, the front camera is set on the front panel of the terminal, and the rear camera is set on the back of the terminal. In some embodiments, there are at least two rear cameras, each of which is a main camera, a depth-of-field camera, a wide-angle camera, and a telephoto camera, so as to realize the fusion of the main camera and the depth-of-field camera to realize the background blur function, the main camera Integrate with wide-angle camera to realize panoramic shooting and VR (Virtual Reality, virtual reality) shooting function or other fusion shooting functions. In some embodiments, the camera assembly 506 may also include a flash. The flash can be a single-color flash or a dual-color flash. Dual color temperature flash refers to the combination of warm light flash and cold light flash, which can be used for light compensation at different color temperatures.
音频电路507可以包括麦克风和扬声器。麦克风用于采集用户及环境的声波,并将声波转换为电信号输入至处理器501进行处理,或者输入至射频电路504以实现语音通信。出于立体声采集或降噪的目的,麦克风可以为多个,分别设置在终端500的不同部位。麦克风还可以是阵列麦克风或全向采集型麦克风。扬声器则用于将来自处理器501或射频电路504的电信号转换为声波。扬声器可以是传统的薄膜扬声器,也可以是压电陶瓷扬声器。当扬声器是压电陶瓷扬声器时,不仅可以将电信号转换为人类可听见的声波,也可以将电信号转换为人类听不见的声波以进行测距等用途。在一些实施例中,音频电路507还可以包括耳机插孔。The audio circuit 507 may include a microphone and a speaker. The microphone is used to collect sound waves of the user and the environment, and convert the sound waves into electrical signals and input them to the processor 501 for processing, or input them to the radio frequency circuit 504 to implement voice communication. For the purpose of stereo collection or noise reduction, there may be multiple microphones, which are respectively installed in different parts of the terminal 500. The microphone can also be an array microphone or an omnidirectional acquisition microphone. The speaker is used to convert the electrical signal from the processor 501 or the radio frequency circuit 504 into sound waves. The speaker can be a traditional thin-film speaker or a piezoelectric ceramic speaker. When the speaker is a piezoelectric ceramic speaker, it can not only convert electrical signals into sound waves audible by humans, but also convert electrical signals into sound waves inaudible to humans for distance measurement and other purposes. In some embodiments, the audio circuit 507 may also include a headphone jack.
定位组件508用于定位终端500的当前地理位置,以实现导航或LBS(Location Based Service,基于位置的服务)。定位组件508可以是基于美国的GPS(Global Positioning System,全球定位系统)、中国的北斗系统或俄罗斯的伽利略系统的定位组件。The positioning component 508 is used to locate the current geographic location of the terminal 500 to implement navigation or LBS (Location Based Service). The positioning component 508 may be a positioning component based on the GPS (Global Positioning System) of the United States, the Beidou system of China, or the Galileo system of Russia.
电源509用于为终端500中的各个组件进行供电。电源509可以是交流电、直流电、一次性电池或可充电电池。当电源509包括可充电电池时,该可充电电池可以是有线充电电池或无线充电电池。有线充电电池是通过有线线路充电的电池,无线充电电池是通过无线线圈充电的电池。该可充电电池还可以用于支持快充技术。The power supply 509 is used to supply power to various components in the terminal 500. The power source 509 may be alternating current, direct current, disposable batteries, or rechargeable batteries. When the power source 509 includes a rechargeable battery, the rechargeable battery may be a wired rechargeable battery or a wireless rechargeable battery. The wired rechargeable battery is a battery charged through a wired line, and the wireless rechargeable battery is a battery charged through a wireless coil. The rechargeable battery can also be used to support fast charging technology.
在一些实施例中,终端500还包括有一个或多个传感器510。该一个或多个传感器510包括但不限于:加速度传感器511、陀螺仪传感器512、压力传感器 513、指纹传感器514、光学传感器515以及接近传感器516。In some embodiments, the terminal 500 further includes one or more sensors 510. The one or more sensors 510 include, but are not limited to: an acceleration sensor 511, a gyro sensor 512, a pressure sensor 513, a fingerprint sensor 514, an optical sensor 515, and a proximity sensor 516.
加速度传感器511可以检测以终端500建立的坐标系的三个坐标轴上的加速度大小。比如,加速度传感器511可以用于检测重力加速度在三个坐标轴上的分量。处理器501可以根据加速度传感器511采集的重力加速度信号,控制触摸显示屏505以横向视图或纵向视图进行用户界面的显示。加速度传感器511还可以用于游戏或者用户的运动数据的采集。The acceleration sensor 511 can detect the magnitude of acceleration on the three coordinate axes of the coordinate system established with the terminal 500. For example, the acceleration sensor 511 can be used to detect components of gravity acceleration on three coordinate axes. The processor 501 may control the touch display screen 505 to display the user interface in a landscape view or a portrait view according to the gravity acceleration signal collected by the acceleration sensor 511. The acceleration sensor 511 can also be used for game or user movement data collection.
陀螺仪传感器512可以检测终端500的机体方向及转动角度,陀螺仪传感器512可以与加速度传感器511协同采集用户对终端500的3D动作。处理器501根据陀螺仪传感器512采集的数据,可以实现如下功能:动作感应(比如根据用户的倾斜操作来改变UI)、拍摄时的图像稳定、游戏控制以及惯性导航。The gyro sensor 512 can detect the body direction and rotation angle of the terminal 500, and the gyro sensor 512 can cooperate with the acceleration sensor 511 to collect a 3D motion of the user on the terminal 500. Based on the data collected by the gyro sensor 512, the processor 501 can realize the following functions: motion sensing (such as changing the UI according to the user's tilt operation), image stabilization during shooting, game control, and inertial navigation.
压力传感器513可以设置在终端500的侧边框和/或触摸显示屏505的下层。当压力传感器513设置在终端500的侧边框时,可以检测用户对终端500的握持信号,由处理器501根据压力传感器513采集的握持信号进行左右手识别或快捷操作。当压力传感器513设置在触摸显示屏505的下层时,由处理器501根据用户对触摸显示屏505的压力操作,实现对UI界面上的可操作性控件进行控制。可操作性控件包括按钮控件、滚动条控件、图标控件、菜单控件中的至少一种。The pressure sensor 513 may be disposed on the side frame of the terminal 500 and/or the lower layer of the touch display screen 505. When the pressure sensor 513 is disposed on the side frame of the terminal 500, it can detect the user's grip signal on the terminal 500, and the processor 501 can perform left-right hand recognition or shortcut operation according to the grip signal collected by the pressure sensor 513. When the pressure sensor 513 is disposed on the lower layer of the touch display screen 505, the processor 501 controls the operability control on the UI interface according to the user's pressure operation on the touch display screen 505. The operability control includes at least one of a button control, a scroll bar control, an icon control, and a menu control.
指纹传感器514用于采集用户的指纹,由处理器501根据指纹传感器514采集到的指纹识别用户的身份,或者,由指纹传感器514根据采集到的指纹识别用户的身份。在识别出用户的身份为可信身份时,由处理器501授权该用户执行相关的敏感操作,该敏感操作包括解锁屏幕、查看加密信息、下载软件、支付及更改设置等。指纹传感器514可以被设置终端500的正面、背面或侧面。当终端500上设置有物理按键或厂商Logo时,指纹传感器514可以与物理按键或厂商Logo集成在一起。The fingerprint sensor 514 is used to collect the user's fingerprint, and the processor 501 identifies the user's identity according to the fingerprint collected by the fingerprint sensor 514, or the fingerprint sensor 514 identifies the user's identity based on the collected fingerprint. When the user's identity is recognized as a trusted identity, the processor 501 authorizes the user to perform related sensitive operations, including unlocking the screen, viewing encrypted information, downloading software, paying, and changing settings. The fingerprint sensor 514 may be provided on the front, back, or side of the terminal 500. When a physical button or manufacturer logo is provided on the terminal 500, the fingerprint sensor 514 may be integrated with the physical button or manufacturer logo.
光学传感器515用于采集环境光强度。在一个实施例中,处理器501可以根据光学传感器515采集的环境光强度,控制触摸显示屏505的显示亮度。具体地,当环境光强度较高时,调高触摸显示屏505的显示亮度;当环境光强度较低时,调低触摸显示屏505的显示亮度。在另一个实施例中,处理器501还可以根据光学传感器515采集的环境光强度,动态调整摄像头组件506的拍摄参数。The optical sensor 515 is used to collect the ambient light intensity. In one embodiment, the processor 501 can control the display brightness of the touch display 505 according to the ambient light intensity collected by the optical sensor 515. Specifically, when the ambient light intensity is high, the display brightness of the touch display screen 505 is increased; when the ambient light intensity is low, the display brightness of the touch display screen 505 is reduced. In another embodiment, the processor 501 can also dynamically adjust the shooting parameters of the camera assembly 506 according to the ambient light intensity collected by the optical sensor 515.
接近传感器516,也称距离传感器,通常设置在终端500的前面板。接近传 感器516用于采集用户与终端500的正面之间的距离。在一个实施例中,当接近传感器516检测到用户与终端500的正面之间的距离逐渐变小时,由处理器501控制触摸显示屏505从亮屏状态切换为息屏状态;当接近传感器516检测到用户与终端500的正面之间的距离逐渐变大时,由处理器501控制触摸显示屏505从息屏状态切换为亮屏状态。The proximity sensor 516, also called a distance sensor, is usually provided on the front panel of the terminal 500. The proximity sensor 516 is used to collect the distance between the user and the front of the terminal 500. In one embodiment, when the proximity sensor 516 detects that the distance between the user and the front of the terminal 500 gradually becomes smaller, the processor 501 controls the touch display 505 to switch from the bright screen state to the breathing screen state; when the proximity sensor 516 detects When the distance from the user to the front of the terminal 500 gradually becomes larger, the processor 501 controls the touch display 505 to switch from the breath-hold state to the bright-screen state.
本领域技术人员可以理解,图12中示出的结构并不构成对终端500的限定,可以包括比图示更多或更少的组件,或者组合某些组件,或者采用不同的组件布置。Those skilled in the art may understand that the structure shown in FIG. 12 does not constitute a limitation on the terminal 500, and may include more or fewer components than those illustrated, or combine certain components, or adopt different component arrangements.
参见图13,本申请实施例提供了一种目标检测系统600,包括设置在车辆上的雷达601、设置在所述车辆上的车载摄像头602,以及与所述雷达601和所述车载摄像头602通信的检测装置603,Referring to FIG. 13, an embodiment of the present application provides a target detection system 600, including a radar 601 provided on a vehicle, a vehicle-mounted camera 602 provided on the vehicle, and communicating with the radar 601 and the vehicle-mounted camera 602 Of detection device 603,
所述车载摄像头602,用于对所述车辆周围进行拍摄,得到当前帧视频图像,并向所述检测装置603提供拍摄的当前帧视频图像;The on-board camera 602 is used to photograph the surroundings of the vehicle to obtain a video image of the current frame and provide the video image of the current frame to the detection device 603;
所述雷达601,用于根据发射的雷达信号和接收的回波信号生成当前帧速度距离图像,并向所述检测装置603提供所述当前帧速度距离图像;The radar 601 is configured to generate a current frame speed distance image based on the transmitted radar signal and the received echo signal, and provide the current frame speed distance image to the detection device 603;
所述检测装置603,用于从所述车载摄像头602提供的视频图像中检测出车辆周围的各个图像目标,各个所述图像目标的置信度,以及各个所述图像目标在图像坐标系中的位置信息;从所述雷达601提供的速度距离图像中检测出所述车辆周围的各个雷达目标,各个所述雷达目标在雷达坐标系中的位置信息,以及各个所述雷达目标的置信度,所述置信度用于表示所述图像目标或所述雷达目标所对应的真实目标的目标类别为指定类别的概率;根据置信度超过第一预设阈值的图像目标的位置信息和置信度超过第二预设阈值的雷达目标的位置信息,获取第一透视矩阵,所述第一透视矩阵用于表示所述图像像坐标系与预设道路坐标系之间的转换关系;通过所述第一透视矩阵,从置信度未超过所述第一预设阈值的图像目标和置信度未超过所述第二预设阈值的雷达目标中检测出目标类别。The detection device 603 is configured to detect each image object around the vehicle, the confidence of each image object, and the position of each image object in the image coordinate system from the video image provided by the vehicle-mounted camera 602 Information; each radar target around the vehicle is detected from the speed distance image provided by the radar 601, the position information of each radar target in the radar coordinate system, and the confidence of each radar target, the The confidence level is used to indicate the probability that the target category of the real target corresponding to the image target or the radar target is the specified category; according to the position information and confidence level of the image target whose confidence level exceeds the first preset threshold value Position information of the radar target with a threshold value to obtain a first perspective matrix, where the first perspective matrix is used to represent a conversion relationship between the image coordinate system and a preset road coordinate system; through the first perspective matrix, A target category is detected from an image target whose confidence level does not exceed the first preset threshold and a radar target whose confidence level does not exceed the second preset threshold.
可选的,所述车载摄像头602设置在所述车辆的前后方和/或左右两侧,所述雷达601设置在所述车辆的前后方。Optionally, the in-vehicle camera 602 is provided in front and rear and/or both sides of the vehicle, and the radar 601 is provided in front and rear of the vehicle.
可选的,所述雷达601为毫米波雷达。Optionally, the radar 601 is a millimeter wave radar.
本领域技术人员在考虑说明书及实践这里公开的申请后,将容易想到本申请的其它实施方案。本申请旨在涵盖本申请的任何变型、用途或者适应性变化,这些变型、用途或者适应性变化遵循本申请的一般性原理并包括本申请未公开的本技术领域中的公知常识或惯用技术手段。说明书和实施例仅被视为示例性的,本申请的真正范围和精神由下面的权利要求指出。After considering the description and practice of the application disclosed herein, those skilled in the art will easily think of other embodiments of the application. This application is intended to cover any variations, uses, or adaptations of this application, which follow the general principles of this application and include common general knowledge or customary technical means in the technical field not disclosed in this application . The description and examples are to be considered exemplary only, and the true scope and spirit of this application are pointed out by the following claims.
应当理解的是,本申请并不局限于上面已经描述并在附图中示出的精确结构,并且可以在不脱离其范围进行各种修改和改变。本申请的范围仅由所附的权利要求来限制。It should be understood that the present application is not limited to the precise structure that has been described above and shown in the drawings, and various modifications and changes can be made without departing from the scope thereof. The scope of this application is limited only by the appended claims.

Claims (21)

  1. 一种车载摄像头和车载雷达联动的目标检测方法,其特征在于,所述方法包括:A target detection method in which a car camera and a car radar are linked is characterized in that the method includes:
    从车载摄像头提供的视频图像中检测出车辆周围的各个图像目标,各个所述图像目标的置信度,以及各个所述图像目标在图像坐标系中的位置信息;Detecting each image object around the vehicle from the video image provided by the onboard camera, the confidence of each image object, and the position information of each image object in the image coordinate system;
    从所述雷达提供的速度距离图像中检测出所述车辆周围的各个雷达目标,各个所述雷达目标在雷达坐标系中的位置信息,以及各个所述雷达目标的置信度,所述置信度用于表示所述图像目标或所述雷达目标所对应的真实目标的目标类别为指定类别的概率;Each radar target around the vehicle is detected from the speed and distance images provided by the radar, the position information of each radar target in the radar coordinate system, and the confidence of each radar target, the confidence is The probability that the target category of the real target corresponding to the image target or the radar target is a specified category;
    根据置信度超过第一预设阈值的图像目标的位置信息和置信度超过第二预设阈值的雷达目标的位置信息,获取第一透视矩阵,所述第一透视矩阵用于表示所述图像像坐标系与预设道路坐标系之间的转换关系;Obtain a first perspective matrix according to the position information of the image target whose confidence exceeds the first preset threshold and the position information of the radar target whose confidence exceeds the second preset threshold, the first perspective matrix is used to represent the image image The conversion relationship between the coordinate system and the preset road coordinate system;
    通过所述第一透视矩阵,从置信度未超过所述第一预设阈值的图像目标和置信度未超过所述第二预设阈值的雷达目标中检测出目标类别。Through the first perspective matrix, target categories are detected from image targets whose confidence does not exceed the first preset threshold and radar targets whose confidence does not exceed the second preset threshold.
  2. 如权利要求1所述的方法,其特征在于,所述方法还包括:The method of claim 1, wherein the method further comprises:
    根据检测出的各个所述图像目标的置信度,对置信度超过所述第一预设阈值的图像目标进行分类,得到该图像目标对应的真实目标的目标类别,并将该目标类别输出,以及Classify image objects whose confidence exceeds the first preset threshold according to the detected confidence of each of the image objects, obtain the target category of the real target corresponding to the image object, and output the target category, and
    根据检测出的各个所述雷达目标的置信度,对置信度超过所述第二预设阈值的雷达目标进行分类,得到该雷达目标对应的真实目标的目标类别,并将该目标类别输出。According to the detected confidence level of each of the radar targets, classify the radar target whose confidence level exceeds the second preset threshold to obtain the target category of the real target corresponding to the radar target, and output the target category.
  3. 如权利要求1所述的方法,其特征在于,在所述根据置信度超过第一预设阈值的图像目标的位置信息和置信度超过第二预设阈值的雷达目标的位置信息,获取第一透视矩阵之前,所述方法还包括:The method according to claim 1, wherein the first position is acquired based on the position information of the image target whose confidence exceeds the first preset threshold and the position information of the radar target whose confidence exceeds the second preset threshold Before perspective matrix, the method further includes:
    从置信度超过第一预设阈值的图像目标和置信度超过第二预设阈值的雷达目标中确定N个关联目标对,任意一个所述关联目标对包括满足预设关联条件的雷达目标和图像目标,所述N为大于或等于1的正整数;N associated target pairs are determined from an image target with a confidence exceeding a first preset threshold and a radar target with a confidence exceeding a second preset threshold, any one of the associated target pairs includes a radar target and an image satisfying a preset association condition Target, the N is a positive integer greater than or equal to 1;
    所述根据置信度超过第一预设阈值的图像目标的位置信息和置信度超过第 二预设阈值的雷达目标的位置信息,获取第一透视矩阵,包括:The obtaining the first perspective matrix according to the position information of the image target whose confidence exceeds the first preset threshold and the position information of the radar target whose confidence exceeds the second preset threshold includes:
    根据所述N个关联目标对中的雷达目标和图像目标的位置信息,确定所述第一透视矩阵。The first perspective matrix is determined according to the position information of the radar target and the image target in the N associated target pairs.
  4. 如权利要求3所述的方法,其特征在于,所述从置信度超过第一预设阈值的图像目标和置信度超过第二预设阈值的雷达目标中确定N个关联目标对,包括:The method according to claim 3, wherein the determining of N associated target pairs from the image target whose confidence exceeds the first preset threshold and the radar target whose confidence exceeds the second preset threshold includes:
    根据第一图像目标的第一位置信息和存储的第二透视矩阵,将所述第一图像目标从所述图像坐标系映射至所述道路坐标系,得到所述第一图像目标在所述道路坐标系中对应的第三位置信息;其中所述第一图像目标为置信度超过第一预设阈值的图像目标,所述第一位置信息为所述第一图像目标在所述图像坐标系中的位置信息;Mapping the first image object from the image coordinate system to the road coordinate system according to the first position information of the first image object and the stored second perspective matrix, to obtain the first image object on the road Corresponding third position information in the coordinate system; wherein the first image object is an image object whose confidence exceeds a first preset threshold, and the first position information is the first image object in the image coordinate system Location information
    根据第一雷达目标的第二位置信息和存储的第三透视矩阵,将所述第一雷达目标从所述雷达坐标系映射至所述道路坐标系,得到所述第一雷达目标在所述道路坐标系中对应的第四位置信息;其中所述第一雷达目标为置信度超过第二预设阈值的雷达目标,所述第二位置信息为所述第一雷达目标在所述雷达坐标系中的位置信息;Mapping the first radar target from the radar coordinate system to the road coordinate system according to the second position information of the first radar target and the stored third perspective matrix, to obtain the first radar target on the road Corresponding fourth position information in the coordinate system; wherein the first radar target is a radar target whose confidence exceeds a second preset threshold, and the second position information is the first radar target in the radar coordinate system Location information
    根据各个所述第一图像目标的所述第三位置信息和各个所述第一雷达目标的所述第四位置信息,对各个所述第一图像目标与各个所述第一雷达目标进行位置关联,得到所述N个关联目标对。According to the third position information of each of the first image targets and the fourth position information of each of the first radar targets, position-associating each of the first image targets with each of the first radar targets To obtain the N associated target pairs.
  5. 如权利要求4所述的方法,其特征在于,所述对各个所述第一图像目标与各个所述第二雷达目标进行位置关联,得到所述N个关联目标对,包括:The method according to claim 4, wherein the performing position association on each of the first image targets and each of the second radar targets to obtain the N associated target pairs includes:
    根据所述第一图像目标的所述第三位置信息和所述第一雷达目标的所述第四位置信息,确定所述第一图像目标在所述道路坐标系中的投影面积,所述第一雷达目标在所述道路坐标系中的投影面积,以及所述第一图像目标和所述第一雷达目标在所述道路坐标系中的重叠投影面积;Determine the projected area of the first image target in the road coordinate system according to the third position information of the first image target and the fourth position information of the first radar target, the first A projected area of a radar target in the road coordinate system, and an overlapping projected area of the first image target and the first radar target in the road coordinate system;
    根据所述第一图像目标在所述道路坐标系中的投影面积,所述第一雷达目标在所述道路坐标系中的投影面积,以及所述重叠投影面积,确定各个所述第一图像目标与各个所述第二雷达目标之间的关联代价;Determine each of the first image targets according to the projected area of the first image target in the road coordinate system, the projected area of the first radar target in the road coordinate system, and the overlapping projected area The associated cost with each of the second radar targets;
    从所述第一图像目标和所述第二图像目标中确定关联代价最小的一个所述 第一雷达目标和一个所述第一图像目标为关联目标对,进而得到所述N个关联目标对。It is determined from the first image target and the second image target that the first radar target and the first image target with the smallest associated cost are the associated target pairs, and then the N associated target pairs are obtained.
  6. 如权利要求4所述的方法,其特征在于,所述根据所述N个关联目标对中的雷达目标和图像目标的位置信息,确定所述第一透视矩阵,包括:The method according to claim 4, wherein the determining the first perspective matrix according to the position information of the radar target and the image target in the N associated target pairs includes:
    针对所述N个关联目标对中的任一个关联目标对,根据所述任一个关联目标对中的所述第一雷达目标的位置信息,修正所述任一个关联目标对中的所述第一图像目标的位置信息;For any one of the N associated target pairs, according to the position information of the first radar target in the any associated target pair, modify the first of the any associated target pair Image target location information;
    根据所述N个关联目标对中的各个所述第一图像目标修正后的位置信息,修正所述第二透视矩阵,得到所述第一透视矩阵。Modify the second perspective matrix according to the corrected position information of each of the first image targets in the N associated target pairs to obtain the first perspective matrix.
  7. 如权利要求1所述的方法,其特征在于,所述通过所述第一透视矩阵,从置信度未超过所述第一预设阈值的图像目标和置信度未超过所述第二预设阈值的雷达目标中检测出目标类别,包括:The method according to claim 1, characterized in that, through the first perspective matrix, the image target whose confidence level does not exceed the first preset threshold and the confidence level does not exceed the second preset threshold The target category was detected in the radar target, including:
    从置信度未超过第一预设阈值的图像目标和置信度未超过第二预设阈值的雷达目标中确定M个特征融合目标对,任意一个所述特征融合目标对包括满足预设关联条件的一个雷达目标和一个图像目标,所述M为大于或等于1的正整数;M feature fusion target pairs are determined from the image target whose confidence level does not exceed the first preset threshold and the radar target whose confidence level does not exceed the second preset threshold, any one of the feature fusion target pairs includes those satisfying the preset association condition One radar target and one image target, where M is a positive integer greater than or equal to 1;
    针对任一所述特征融合目标对,将所述特征融合目标对中的所述雷达目标的回波能量特征和所述图像目标的图像特征分别进行卷积计算后进行拼接,得到所述特征融合目标对所对应的融合特征图;For any of the feature fusion target pairs, the echo energy feature of the radar target and the image feature of the image target in the feature fusion target pair are respectively convoluted and spliced to obtain the feature fusion The fusion feature map corresponding to the target pair;
    将所述融合特征图进行卷积和全连接计算后输入到分类网络进行目标分类,得到所述融合特征图对应的目标类别。After performing convolution and full connection calculation on the fusion feature map, input it to a classification network for target classification to obtain a target category corresponding to the fusion feature map.
  8. 如权利要求7所述的方法,其特征在于,所述从第二图像目标和第二雷达目标中确定M个特征融合目标对,包括:The method of claim 7, wherein the determining of M feature fusion target pairs from the second image target and the second radar target includes:
    通过所述第一透视矩阵,将第二图像目标从所述图像坐标系映射至所述道路坐标系中,得到所述第二图像目标在所述道路坐标系中对应的位置信息,以及,通过预先存储的第三透视矩阵,将第二雷达目标从所述雷达坐标系映射至所述道路坐标系中,得到所述第二雷达目标在所述道路坐标系中对应的位置信 息,所述第二图像目标为置信度未超过第一预设阈值的图像目标,所述第二雷达目标为置信度未超过第二预设阈值的雷达目标;Mapping the second image object from the image coordinate system to the road coordinate system through the first perspective matrix to obtain the corresponding position information of the second image object in the road coordinate system, and, by A pre-stored third perspective matrix, mapping the second radar target from the radar coordinate system to the road coordinate system, to obtain the corresponding position information of the second radar target in the road coordinate system, the first The second image target is an image target whose confidence level does not exceed the first preset threshold, and the second radar target is a radar target whose confidence level does not exceed the second preset threshold;
    根据各个所述第二图像目标在所述道路坐标系中对应的位置信息和各个所述第二雷达目标在所述道路坐标系中对应的位置信息,对各个所述第二图像目标和各个所述第二雷达目标进行位置关联,得到所述M个特征融合目标对。According to the corresponding position information of each second image target in the road coordinate system and the corresponding position information of each second radar target in the road coordinate system, for each second image target and each position The second radar target performs position correlation to obtain the M feature fusion target pairs.
  9. 如权利要求1所述的方法,其特征在于,所述从车载摄像头提供的视频图像中检测出车辆周围的各个图像目标的置信度,包括:The method according to claim 1, wherein the detecting the confidence of each image target around the vehicle from the video image provided by the on-board camera includes:
    根据所述车载摄像头提供的当前帧视频图像,以及与所述当前帧视频图像接近的多帧历史帧视频图像,获取所述任意一个所述图像目标的分类置信度、跟踪帧数置信度和位置置信度;Obtain the classification confidence, tracking frame number confidence and position of any one of the image objects according to the current frame video image provided by the vehicle camera and the multi-frame historical frame video image close to the current frame video image Confidence;
    根据所述分类置信度、位置置信度和跟踪帧数置信度中的一个或多个,确定所述图像目标的置信度;Determine the confidence of the image target according to one or more of the classification confidence, position confidence, and tracking frame number confidence;
    所述从所述雷达采集的速度距离图像中检测出所述车辆周围的各个雷达目标的置信度,包括:The detection of the confidence of each radar target around the vehicle from the speed and distance images collected by the radar includes:
    根据所述当前帧速度距离图像中任一个所述雷达目标的回波能量强度、距离所述车辆的距离,以及所述雷达目标在多帧历史帧速度距离图像中的持续时间,确定所述雷达目标的置信度。Determine the radar according to the intensity of the echo energy of any of the radar targets in the current frame speed distance image, the distance from the vehicle, and the duration of the radar target in the multi-frame historical frame speed distance image The confidence of the goal.
  10. 一种车载摄像头和车载雷达联动的目标检测装置,其特征在于,所述装置包括:A target detection device linked by a vehicle-mounted camera and a vehicle-mounted radar is characterized in that the device includes:
    第一检测模块,用于从车载摄像头提供的视频图像中检测出车辆周围的各个图像目标,各个所述图像目标的置信度,以及各个所述图像目标在图像坐标系中的位置信息;The first detection module is used to detect each image target around the vehicle from the video image provided by the onboard camera, the confidence of each image target, and the position information of each image target in the image coordinate system;
    第二检测模块,用于从所述雷达提供的速度距离图像中检测出所述车辆周围的各个雷达目标,各个所述雷达目标在雷达坐标系中的位置信息,以及各个所述雷达目标的置信度,所述置信度用于表示所述图像目标或所述雷达目标所对应的真实目标的目标类别为指定类别的概率;The second detection module is used to detect each radar target around the vehicle from the speed and distance images provided by the radar, the position information of each radar target in the radar coordinate system, and the confidence of each radar target Degree, the confidence is used to indicate the probability that the target category of the real target corresponding to the image target or the radar target is a specified category;
    获取模块,用于根据置信度超过第一预设阈值的图像目标的位置信息和置信度超过第二预设阈值的雷达目标的位置信息,获取第一透视矩阵,所述第一 透视矩阵用于表示所述图像像坐标系与预设道路坐标系之间的转换关系;An obtaining module, configured to obtain a first perspective matrix based on the position information of the image target with the confidence level exceeding the first preset threshold and the position information of the radar target with the confidence level exceeding the second preset threshold It represents the conversion relationship between the image coordinate system and the preset road coordinate system;
    第三检测模块,用于通过所述第一透视矩阵,从置信度未超过所述第一预设阈值的图像目标和置信度未超过所述第二预设阈值的雷达目标中检测出目标类别。The third detection module is used to detect the target category from the image target whose confidence does not exceed the first preset threshold and the radar target whose confidence does not exceed the second preset threshold through the first perspective matrix .
  11. 如权利要求10所述的装置,其特征在于,所述装置还包括:The device of claim 10, wherein the device further comprises:
    分类模块,用于根据检测出的各个所述图像目标的置信度,对置信度超过所述第一预设阈值的图像目标进行分类,得到该图像目标对应的真实目标的目标类别,并将该目标类别输出,以及,根据检测出的各个所述雷达目标的置信度,对置信度超过所述第二预设阈值的雷达目标进行分类,得到该雷达目标对应的真实目标的目标类别,并将该目标类别输出。A classification module is used to classify image objects whose confidence exceeds the first preset threshold according to the detected confidence of each of the image objects, to obtain the target category of the real target corresponding to the image object, and to Target category output, and, based on the detected confidence of each of the radar targets, classify the radar targets with confidence exceeding the second preset threshold to obtain the target category of the real target corresponding to the radar target, and The target category is output.
  12. 如权利要求10所述的装置,其特征在于,所述装置还包括:The device of claim 10, wherein the device further comprises:
    确定模块,用于从置信度超过第一预设阈值的图像目标和置信度超过第二预设阈值的雷达目标中确定N个关联目标对,任意一个所述关联目标对包括满足预设关联条件的雷达目标和图像目标,所述N为大于或等于1的正整数;A determining module, configured to determine N associated target pairs from an image target with a confidence exceeding a first preset threshold and a radar target with a confidence exceeding a second preset threshold, any one of the associated target pairs includes satisfying a preset associated condition Radar target and image target, N is a positive integer greater than or equal to 1;
    所述获取模块,用于根据所述N个关联目标对中的雷达目标和图像目标的位置信息,确定所述第一透视矩阵。The acquisition module is configured to determine the first perspective matrix according to the position information of the radar target and the image target in the N associated target pairs.
  13. 如权利要求12所述的装置,其特征在于,所述确定模块,用于:The apparatus according to claim 12, wherein the determination module is configured to:
    根据第一图像目标的第一位置信息和存储的第二透视矩阵,将所述第一图像目标从所述图像坐标系映射至所述道路坐标系,得到所述第一图像目标在所述道路坐标系中对应的第三位置信息;其中所述第一图像目标为置信度超过第一预设阈值的图像目标,所述第一位置信息为所述第一图像目标在所述图像坐标系中的位置信息;Mapping the first image object from the image coordinate system to the road coordinate system according to the first position information of the first image object and the stored second perspective matrix, to obtain the first image object on the road Corresponding third position information in the coordinate system; wherein the first image object is an image object whose confidence exceeds a first preset threshold, and the first position information is the first image object in the image coordinate system Location information
    根据第一雷达目标的第二位置信息和存储的第三透视矩阵,将所述第一雷达目标从所述雷达坐标系映射至所述道路坐标系,得到所述第一雷达目标在所述道路坐标系中对应的第四位置信息;其中所述第一雷达目标为置信度超过第二预设阈值的雷达目标,所述第二位置信息为所述第一雷达目标在所述雷达坐标系中的位置信息;Mapping the first radar target from the radar coordinate system to the road coordinate system according to the second position information of the first radar target and the stored third perspective matrix, to obtain the first radar target on the road Corresponding fourth position information in the coordinate system; wherein the first radar target is a radar target whose confidence exceeds a second preset threshold, and the second position information is the first radar target in the radar coordinate system Location information
    根据各个所述第一图像目标的所述第三位置信息和各个所述第一雷达目标的所述第四位置信息,对各个所述第一图像目标与各个所述第一雷达目标进行位置关联,得到所述N个关联目标对。According to the third position information of each of the first image targets and the fourth position information of each of the first radar targets, position-associating each of the first image targets with each of the first radar targets To obtain the N associated target pairs.
  14. 如权利要求13所述的装置,其特征在于,所述获取模块,用于:The apparatus according to claim 13, wherein the acquisition module is configured to:
    针对所述N个关联目标对中的任一个关联目标对,根据所述任一个关联目标对中的所述第一雷达目标的位置信息,修正所述任一个关联目标对中的所述第一图像目标的位置信息;For any one of the N associated target pairs, according to the position information of the first radar target in the any associated target pair, modify the first of the any associated target pair Image target location information;
    根据所述N个关联目标对中的各个所述第一图像目标修正后的位置信息,修正所述第二透视矩阵,得到所述第一透视矩阵。Modify the second perspective matrix according to the corrected position information of each of the first image targets in the N associated target pairs to obtain the first perspective matrix.
  15. 如权利要求10所述的装置,其特征在于,所述第三检测模块,用于:The device of claim 10, wherein the third detection module is configured to:
    从置信度未超过第一预设阈值的图像目标和置信度未超过第二预设阈值的雷达目标中确定M个特征融合目标对,任意一个所述特征融合目标对包括满足预设关联条件的一个雷达目标和一个图像目标,所述M为大于或等于1的正整数;M feature fusion target pairs are determined from the image target whose confidence level does not exceed the first preset threshold and the radar target whose confidence level does not exceed the second preset threshold, any one of the feature fusion target pairs includes those satisfying the preset association condition One radar target and one image target, where M is a positive integer greater than or equal to 1;
    针对任一所述特征融合目标对,将所述特征融合目标对中的所述雷达目标的回波能量特征和所述图像目标的图像特征分别进行卷积计算后进行拼接,得到所述特征融合目标对所对应的融合特征图;For any of the feature fusion target pairs, the echo energy feature of the radar target and the image feature of the image target in the feature fusion target pair are respectively convoluted and spliced to obtain the feature fusion The fusion feature map corresponding to the target pair;
    将所述融合特征图进行卷积和全连接计算后输入到分类网络进行目标分类,得到所述融合特征图对应的目标类别。After performing convolution and full connection calculation on the fusion feature map, input it to a classification network for target classification to obtain a target category corresponding to the fusion feature map.
  16. 如权利要求15所述的装置,其特征在于,所述第三检测模块,用于:The device according to claim 15, wherein the third detection module is configured to:
    通过所述第一透视矩阵,将第二图像目标从所述图像坐标系映射至所述道路坐标系中,得到所述第二图像目标在所述道路坐标系中对应的位置信息,以及,通过预先存储的第三透视矩阵,将第二雷达目标从所述雷达坐标系映射至所述道路坐标系中,得到所述第二雷达目标在所述道路坐标系中对应的位置信息,所述第二图像目标为置信度未超过第一预设阈值的图像目标,所述第二雷达目标为置信度未超过第二预设阈值的雷达目标;Mapping the second image object from the image coordinate system to the road coordinate system through the first perspective matrix to obtain the corresponding position information of the second image object in the road coordinate system, and, by A pre-stored third perspective matrix, mapping the second radar target from the radar coordinate system to the road coordinate system, to obtain the corresponding position information of the second radar target in the road coordinate system, the first The second image target is an image target whose confidence level does not exceed the first preset threshold, and the second radar target is a radar target whose confidence level does not exceed the second preset threshold;
    根据各个所述第二图像目标在所述道路坐标系中对应的位置信息和各个所 述第二雷达目标在所述道路坐标系中对应的位置信息,对各个所述第二图像目标和各个所述第二雷达目标进行位置关联,得到所述M个特征融合目标对。According to the corresponding position information of each second image target in the road coordinate system and the corresponding position information of each second radar target in the road coordinate system, for each second image target and each position The second radar target performs position correlation to obtain the M feature fusion target pairs.
  17. 如权利要求10所述的装置,其特征在于,所述第一检测模块,用于:The device of claim 10, wherein the first detection module is configured to:
    根据所述车载摄像头提供的当前帧视频图像,以及与所述当前帧视频图像接近的多帧历史帧视频图像,获取所述任意一个所述图像目标的分类置信度、跟踪帧数置信度和位置置信度;Obtain the classification confidence, tracking frame number confidence and position of any one of the image objects according to the current frame video image provided by the vehicle camera and the multi-frame historical frame video image close to the current frame video image Confidence;
    根据所述分类置信度、位置置信度和跟踪帧数置信度中的一个或多个,确定所述图像目标的置信度;Determine the confidence of the image target according to one or more of the classification confidence, position confidence, and tracking frame number confidence;
    所述第二检测模块,用于:The second detection module is used to:
    根据所述当前帧速度距离图像中任一个所述雷达目标的回波能量强度、距离所述车辆的距离,以及所述雷达目标在多帧历史帧速度距离图像中的持续时间,确定所述雷达目标的置信度。Determine the radar according to the intensity of the echo energy of any of the radar targets in the current frame speed distance image, the distance from the vehicle, and the duration of the radar target in the multi-frame historical frame speed distance image The confidence of the goal.
  18. 一种目标检测系统,包括设置在车辆上的雷达、设置在所述车辆上的车载摄像头,以及与所述雷达和所述车载摄像头通信的检测装置,其特征在于,A target detection system includes a radar installed on a vehicle, an on-board camera installed on the vehicle, and a detection device that communicates with the radar and the on-board camera.
    所述车载摄像头,用于对所述车辆周围进行拍摄,得到当前帧视频图像,并向所述检测装置提供拍摄的当前帧视频图像;The on-board camera is used to photograph the surroundings of the vehicle to obtain a video image of the current frame, and provide the video image of the current frame to the detection device;
    所述雷达,用于根据发射的雷达信号和接收的回波信号生成当前帧速度距离图像,并向所述检测装置提供所述当前帧速度距离图像;The radar is used to generate a current frame speed distance image based on the transmitted radar signal and the received echo signal, and provide the current frame speed distance image to the detection device;
    所述检测装置,用于从所述车载摄像头提供的视频图像中检测出车辆周围的各个图像目标,以及各个所述图像目标的置信度,以及各个所述图像目标在图像坐标系中的位置信息;从所述雷达提供的速度距离图像中检测出所述车辆周围的各个雷达目标,以及各个所述雷达目标在雷达坐标系中的位置信息,所述雷达目标的置信度,所述置信度用于表示所述图像目标或所述雷达目标所对应的真实目标的目标类别为指定类别的概率;根据置信度超过第一预设阈值的图像目标的位置信息和置信度超过第二预设阈值的雷达目标的位置信息,获取第一透视矩阵,所述第一透视矩阵用于表示所述图像像坐标系与预设道路坐标系之间的转换关系;通过所述第一透视矩阵,从置信度未超过所述第一预设阈值的图像目标和置信度未超过所述第二预设阈值的雷达目标中检测出目标类 别。The detection device is configured to detect each image target around the vehicle from the video image provided by the on-board camera, the confidence of each image target, and the position information of each image target in the image coordinate system Detecting each radar target around the vehicle from the speed and distance images provided by the radar, and the position information of each radar target in the radar coordinate system, the confidence level of the radar target, the confidence level is The probability that the target category of the real target corresponding to the image target or the radar target is a specified category; according to the location information and confidence level of the image target whose confidence exceeds the first preset threshold Obtain the first perspective matrix of the position information of the radar target, and the first perspective matrix is used to represent the conversion relationship between the image coordinate system and the preset road coordinate system; through the first perspective matrix, from the confidence level The target category is detected from the image target that does not exceed the first preset threshold and the radar target that does not exceed the second preset threshold.
  19. 如权利要求18所述的系统,其特征在于,所述车载摄像头设置在所述车辆的前后方和/或左右两侧,所述雷达设置在所述车辆的前后方。The system according to claim 18, wherein the on-vehicle camera is disposed in front and rear and/or on both sides of the vehicle, and the radar is disposed in front and rear of the vehicle.
  20. 如权利要求18所述的系统,其特征在于,所述雷达为毫米波雷达。The system of claim 18, wherein the radar is a millimeter wave radar.
  21. 一种电子设备,其特征在于,包括:An electronic device, characterized in that it includes:
    至少一个处理器;和At least one processor; and
    至少一个存储器;At least one memory;
    所述至少一个存储器存储有一个或多个指令,所述一个或多个指令被配置成由所述至少一个处理器执行,以执行以下指令:The at least one memory stores one or more instructions, and the one or more instructions are configured to be executed by the at least one processor to execute the following instructions:
    从车载摄像头提供的视频图像中检测出车辆周围的各个图像目标,各个所述图像目标的置信度,以及各个所述图像目标在图像坐标系中的位置信息;Detecting each image object around the vehicle from the video image provided by the onboard camera, the confidence of each image object, and the position information of each image object in the image coordinate system;
    从所述雷达提供的速度距离图像中检测出所述车辆周围的各个雷达目标,各个所述雷达目标在雷达坐标系中的位置信息,以及各个所述雷达目标的置信度,所述置信度用于表示所述图像目标或所述雷达目标所对应的真实目标的目标类别为指定类别的概率;Each radar target around the vehicle is detected from the speed and distance images provided by the radar, the position information of each radar target in the radar coordinate system, and the confidence of each radar target, the confidence is The probability that the target category of the real target corresponding to the image target or the radar target is a specified category;
    根据置信度超过第一预设阈值的图像目标的位置信息和置信度超过第二预设阈值的雷达目标的位置信息,获取第一透视矩阵,所述第一透视矩阵用于表示所述图像像坐标系与预设道路坐标系之间的转换关系;Obtain a first perspective matrix according to the position information of the image target whose confidence exceeds the first preset threshold and the position information of the radar target whose confidence exceeds the second preset threshold, the first perspective matrix is used to represent the image image The conversion relationship between the coordinate system and the preset road coordinate system;
    通过所述第一透视矩阵,从置信度未超过所述第一预设阈值的图像目标和置信度未超过所述第二预设阈值的雷达目标中检测出目标类别。Through the first perspective matrix, target categories are detected from image targets whose confidence does not exceed the first preset threshold and radar targets whose confidence does not exceed the second preset threshold.
PCT/CN2019/122171 2018-11-30 2019-11-29 Target detection method, apparatus and system based on linkage between vehicle-mounted camera and vehicle-mounted radar WO2020108647A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201811459743.X 2018-11-30
CN201811459743.XA CN111257866B (en) 2018-11-30 2018-11-30 Target detection method, device and system for linkage of vehicle-mounted camera and vehicle-mounted radar

Publications (1)

Publication Number Publication Date
WO2020108647A1 true WO2020108647A1 (en) 2020-06-04

Family

ID=70854695

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2019/122171 WO2020108647A1 (en) 2018-11-30 2019-11-29 Target detection method, apparatus and system based on linkage between vehicle-mounted camera and vehicle-mounted radar

Country Status (2)

Country Link
CN (1) CN111257866B (en)
WO (1) WO2020108647A1 (en)

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112099042A (en) * 2020-08-07 2020-12-18 武汉万集信息技术有限公司 Vehicle tracking method and system
CN112116031A (en) * 2020-10-29 2020-12-22 重庆长安汽车股份有限公司 Target fusion method and system based on road side equipment, vehicle and storage medium
CN112183636A (en) * 2020-09-29 2021-01-05 奇瑞汽车股份有限公司 Method and apparatus for detecting living body
CN112349102A (en) * 2020-10-29 2021-02-09 深圳大学 Vehicle type classification method and device, computer equipment and storage medium
CN112433228A (en) * 2021-01-05 2021-03-02 中国人民解放军国防科技大学 Multi-laser radar decision-level fusion method and device for pedestrian detection
CN112528773A (en) * 2020-11-27 2021-03-19 深兰科技(上海)有限公司 Obstacle information fusion method and device, electronic equipment and storage medium
CN112541953A (en) * 2020-12-29 2021-03-23 江苏航天大为科技股份有限公司 Vehicle detection method based on radar signal and video synchronous coordinate mapping
CN114283392A (en) * 2021-12-10 2022-04-05 广州海格星航信息科技有限公司 Acquisition system of difficult sample of road target detection
CN114879177A (en) * 2022-07-11 2022-08-09 浙江大华技术股份有限公司 Target analysis method and device based on radar information
CN114999087A (en) * 2022-05-24 2022-09-02 深圳康佳电子科技有限公司 Monitoring method, device, medium and terminal for protecting privacy
US11473927B2 (en) * 2020-02-05 2022-10-18 Electronic Arts Inc. Generating positions of map items for placement on a virtual map
CN115534801A (en) * 2022-08-29 2022-12-30 深圳市欧冶半导体有限公司 Vehicle lamp self-adaptive dimming method and device, intelligent terminal and storage medium
CN117197182A (en) * 2023-11-07 2023-12-08 华诺星空技术股份有限公司 Lei Shibiao method, apparatus and storage medium
CN117470254A (en) * 2023-12-28 2024-01-30 哈尔滨工业大学(深圳)(哈尔滨工业大学深圳科技创新研究院) Vehicle navigation system and method based on radar service

Families Citing this family (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113870347A (en) * 2020-06-30 2021-12-31 北京市商汤科技开发有限公司 Target vehicle control method and device, electronic equipment and storage medium
CN111932884A (en) * 2020-08-13 2020-11-13 中国第一汽车股份有限公司 Traffic jam recognition method and device, vehicle and storage medium
CN112528763A (en) * 2020-11-24 2021-03-19 浙江大华汽车技术有限公司 Target detection method, electronic device and computer storage medium
CN112572430A (en) * 2020-12-14 2021-03-30 深兰人工智能(深圳)有限公司 Collision risk determination method and device
CN114972935A (en) * 2021-02-27 2022-08-30 上海华为技术有限公司 Information processing method and related equipment
CN113034586B (en) * 2021-04-27 2022-09-23 北京邮电大学 Road inclination angle detection method and detection system
CN115222791B (en) * 2022-07-15 2023-08-15 小米汽车科技有限公司 Target association method, device, readable storage medium and chip
CN115542312B (en) * 2022-11-30 2023-03-21 苏州挚途科技有限公司 Multi-sensor association method and device

Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2012154731A (en) * 2011-01-25 2012-08-16 Panasonic Corp Positioning information formation device, detector, and positioning information formation method
CN103983951A (en) * 2014-05-28 2014-08-13 北京海兰盈华科技有限公司 Display method, device and system of target detected signals
CN104417552A (en) * 2013-08-29 2015-03-18 株式会社电装 Collision probability determination apparatus and program
CN104637059A (en) * 2015-02-09 2015-05-20 吉林大学 Night preceding vehicle detection method based on millimeter-wave radar and machine vision
CN106767852A (en) * 2016-12-30 2017-05-31 东软集团股份有限公司 A kind of method for generating detection target information, device and equipment
CN106908783A (en) * 2017-02-23 2017-06-30 苏州大学 Obstacle detection method based on multi-sensor information fusion
CN107576960A (en) * 2017-09-04 2018-01-12 苏州驾驶宝智能科技有限公司 The object detection method and system of vision radar Spatial-temporal Information Fusion
CN107966700A (en) * 2017-11-20 2018-04-27 天津大学 A kind of front obstacle detecting system and method for pilotless automobile
CN108169743A (en) * 2017-03-10 2018-06-15 南京沃杨机械科技有限公司 Agricultural machinery is unmanned to use farm environment cognitive method
KR20180097004A (en) * 2017-02-22 2018-08-30 한국전자통신연구원 Method of position calculation between radar target lists and vision image ROI
KR20180119344A (en) * 2017-04-25 2018-11-02 재단법인대구경북과학기술원 Region monitoring apparatus and method for monitoring region thereby

Family Cites Families (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104036279B (en) * 2014-06-12 2017-04-05 北京联合大学 A kind of intelligent vehicle traveling control method and system
JP6058072B2 (en) * 2015-05-29 2017-01-11 三菱電機株式会社 Object identification device
US9829888B2 (en) * 2015-11-17 2017-11-28 Ford Global Technologies, Llc Distinguishing lane markings for a vehicle to follow
JP6462630B2 (en) * 2016-05-24 2019-01-30 株式会社デンソー Target detection device
CN106101590B (en) * 2016-06-23 2019-07-19 上海无线电设备研究所 The detection of radar video complex data and processing system and detection and processing method
CN106710240B (en) * 2017-03-02 2019-09-27 公安部交通管理科学研究所 The passing vehicle for merging multiple target radar and video information tracks speed-measuring method
CN107389084B (en) * 2017-06-09 2020-06-05 深圳市速腾聚创科技有限公司 Driving path planning method and storage medium
CN107609522B (en) * 2017-09-19 2021-04-13 东华大学 Information fusion vehicle detection system based on laser radar and machine vision

Patent Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2012154731A (en) * 2011-01-25 2012-08-16 Panasonic Corp Positioning information formation device, detector, and positioning information formation method
CN104417552A (en) * 2013-08-29 2015-03-18 株式会社电装 Collision probability determination apparatus and program
CN103983951A (en) * 2014-05-28 2014-08-13 北京海兰盈华科技有限公司 Display method, device and system of target detected signals
CN104637059A (en) * 2015-02-09 2015-05-20 吉林大学 Night preceding vehicle detection method based on millimeter-wave radar and machine vision
CN106767852A (en) * 2016-12-30 2017-05-31 东软集团股份有限公司 A kind of method for generating detection target information, device and equipment
KR20180097004A (en) * 2017-02-22 2018-08-30 한국전자통신연구원 Method of position calculation between radar target lists and vision image ROI
CN106908783A (en) * 2017-02-23 2017-06-30 苏州大学 Obstacle detection method based on multi-sensor information fusion
CN108169743A (en) * 2017-03-10 2018-06-15 南京沃杨机械科技有限公司 Agricultural machinery is unmanned to use farm environment cognitive method
KR20180119344A (en) * 2017-04-25 2018-11-02 재단법인대구경북과학기술원 Region monitoring apparatus and method for monitoring region thereby
CN107576960A (en) * 2017-09-04 2018-01-12 苏州驾驶宝智能科技有限公司 The object detection method and system of vision radar Spatial-temporal Information Fusion
CN107966700A (en) * 2017-11-20 2018-04-27 天津大学 A kind of front obstacle detecting system and method for pilotless automobile

Cited By (27)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20220412765A1 (en) * 2020-02-05 2022-12-29 Electronic Arts Inc. Generating Positions of Map Items for Placement on a Virtual Map
US11473927B2 (en) * 2020-02-05 2022-10-18 Electronic Arts Inc. Generating positions of map items for placement on a virtual map
US11668581B2 (en) 2020-02-05 2023-06-06 Electronic Arts Inc. Generating positions of map items for placement on a virtual map
CN112099042A (en) * 2020-08-07 2020-12-18 武汉万集信息技术有限公司 Vehicle tracking method and system
CN112099042B (en) * 2020-08-07 2024-04-12 武汉万集信息技术有限公司 Vehicle tracking method and system
CN112183636A (en) * 2020-09-29 2021-01-05 奇瑞汽车股份有限公司 Method and apparatus for detecting living body
CN112183636B (en) * 2020-09-29 2023-04-18 奇瑞汽车股份有限公司 Method and apparatus for detecting living body
CN112349102A (en) * 2020-10-29 2021-02-09 深圳大学 Vehicle type classification method and device, computer equipment and storage medium
CN112116031B (en) * 2020-10-29 2024-02-09 重庆长安汽车股份有限公司 Target fusion method, system, vehicle and storage medium based on road side equipment
CN112349102B (en) * 2020-10-29 2022-05-20 深圳大学 Vehicle type classification method and device, computer equipment and storage medium
CN112116031A (en) * 2020-10-29 2020-12-22 重庆长安汽车股份有限公司 Target fusion method and system based on road side equipment, vehicle and storage medium
CN112528773A (en) * 2020-11-27 2021-03-19 深兰科技(上海)有限公司 Obstacle information fusion method and device, electronic equipment and storage medium
CN112528773B (en) * 2020-11-27 2023-04-07 深兰科技(上海)有限公司 Obstacle information fusion method and device, electronic equipment and storage medium
CN112541953A (en) * 2020-12-29 2021-03-23 江苏航天大为科技股份有限公司 Vehicle detection method based on radar signal and video synchronous coordinate mapping
CN112541953B (en) * 2020-12-29 2023-04-14 江苏航天大为科技股份有限公司 Vehicle detection method based on radar signal and video synchronous coordinate mapping
CN112433228A (en) * 2021-01-05 2021-03-02 中国人民解放军国防科技大学 Multi-laser radar decision-level fusion method and device for pedestrian detection
CN114283392B (en) * 2021-12-10 2024-04-05 广州海格星航信息科技有限公司 Road target detection's difficult sample's acquisition system
CN114283392A (en) * 2021-12-10 2022-04-05 广州海格星航信息科技有限公司 Acquisition system of difficult sample of road target detection
CN114999087A (en) * 2022-05-24 2022-09-02 深圳康佳电子科技有限公司 Monitoring method, device, medium and terminal for protecting privacy
CN114879177B (en) * 2022-07-11 2022-10-28 浙江大华技术股份有限公司 Target analysis method and device based on radar information
CN114879177A (en) * 2022-07-11 2022-08-09 浙江大华技术股份有限公司 Target analysis method and device based on radar information
CN115534801B (en) * 2022-08-29 2023-07-21 深圳市欧冶半导体有限公司 Car lamp self-adaptive dimming method and device, intelligent terminal and storage medium
CN115534801A (en) * 2022-08-29 2022-12-30 深圳市欧冶半导体有限公司 Vehicle lamp self-adaptive dimming method and device, intelligent terminal and storage medium
CN117197182A (en) * 2023-11-07 2023-12-08 华诺星空技术股份有限公司 Lei Shibiao method, apparatus and storage medium
CN117197182B (en) * 2023-11-07 2024-02-27 华诺星空技术股份有限公司 Lei Shibiao method, apparatus and storage medium
CN117470254B (en) * 2023-12-28 2024-03-08 哈尔滨工业大学(深圳)(哈尔滨工业大学深圳科技创新研究院) Vehicle navigation system and method based on radar service
CN117470254A (en) * 2023-12-28 2024-01-30 哈尔滨工业大学(深圳)(哈尔滨工业大学深圳科技创新研究院) Vehicle navigation system and method based on radar service

Also Published As

Publication number Publication date
CN111257866A (en) 2020-06-09
CN111257866B (en) 2022-02-11

Similar Documents

Publication Publication Date Title
WO2020108647A1 (en) Target detection method, apparatus and system based on linkage between vehicle-mounted camera and vehicle-mounted radar
WO2021128777A1 (en) Method, apparatus, device, and storage medium for detecting travelable region
CN110967011B (en) Positioning method, device, equipment and storage medium
CN111126182B (en) Lane line detection method, lane line detection device, electronic device, and storage medium
CN108055402B (en) Shooting method and mobile terminal
CN110986930B (en) Equipment positioning method and device, electronic equipment and storage medium
WO2021103841A1 (en) Control vehicle
CN110044638B (en) Method and device for testing lane keeping function and storage medium
CN111126276B (en) Lane line detection method, lane line detection device, computer equipment and storage medium
CN111104893B (en) Target detection method, target detection device, computer equipment and storage medium
WO2021000956A1 (en) Method and apparatus for upgrading intelligent model
CN110570465B (en) Real-time positioning and map construction method and device and computer readable storage medium
CN113590070A (en) Navigation interface display method, navigation interface display device, terminal and storage medium
CN112406707A (en) Vehicle early warning method, vehicle, device, terminal and storage medium
CN111325701B (en) Image processing method, device and storage medium
CN110942023A (en) Indication method, device and equipment for vehicle vision blind area and storage medium
CN111127541A (en) Vehicle size determination method and device and storage medium
CN112991439B (en) Method, device, electronic equipment and medium for positioning target object
CN111538009B (en) Radar point marking method and device
CN111383243B (en) Method, device, equipment and storage medium for tracking target object
CN111898535A (en) Target identification method, device and storage medium
CN110633336B (en) Method and device for determining laser data search range and storage medium
CN113470116B (en) Verification method, device, equipment and storage medium for calibration data of camera device
CN111179628B (en) Positioning method and device for automatic driving vehicle, electronic equipment and storage medium
CN113255906A (en) Method, device, terminal and storage medium for returning obstacle 3D angle information in automatic driving

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19890325

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 19890325

Country of ref document: EP

Kind code of ref document: A1