WO2023273041A1 - Target detection method and apparatus in vehicle-road coordination, and roadside device - Google Patents

Target detection method and apparatus in vehicle-road coordination, and roadside device Download PDF

Info

Publication number
WO2023273041A1
WO2023273041A1 PCT/CN2021/126163 CN2021126163W WO2023273041A1 WO 2023273041 A1 WO2023273041 A1 WO 2023273041A1 CN 2021126163 W CN2021126163 W CN 2021126163W WO 2023273041 A1 WO2023273041 A1 WO 2023273041A1
Authority
WO
WIPO (PCT)
Prior art keywords
confidence
candidate target
degree
regions
area
Prior art date
Application number
PCT/CN2021/126163
Other languages
French (fr)
Chinese (zh)
Inventor
夏春龙
Original Assignee
阿波罗智联(北京)科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 阿波罗智联(北京)科技有限公司 filed Critical 阿波罗智联(北京)科技有限公司
Priority to JP2022535786A priority Critical patent/JP7436670B2/en
Priority to KR1020227019941A priority patent/KR20220091607A/en
Publication of WO2023273041A1 publication Critical patent/WO2023273041A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/56Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
    • G06V20/58Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/26Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/52Surveillance or monitoring of activities, e.g. for recognising suspicious objects
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W4/00Services specially adapted for wireless communication networks; Facilities therefor
    • H04W4/30Services specially adapted for particular environments, situations or purposes
    • H04W4/40Services specially adapted for particular environments, situations or purposes for vehicles, e.g. vehicle-to-pedestrians [V2P]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/07Target detection
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Definitions

  • the present disclosure relates to the technical field of intelligent transportation, in particular to the technical field of image detection.
  • the present disclosure provides a target detection method, device and roadside equipment in vehicle-road coordination.
  • a method for object detection in vehicle-road coordination comprising:
  • Objects in the image are detected from candidate object regions according to the updated confidence.
  • a device for detecting objects in vehicle-road coordination comprising:
  • An information obtaining module configured to perform target detection on an image, and obtain a candidate target area in the image, a confidence degree of the candidate target area, and an occlusion degree of the candidate target area;
  • Confidence update module for updating the confidence of the candidate target area based on the intersection ratio between the candidate target areas and the degree of occlusion of the candidate target area
  • the object detection module is used to detect the object in the image from the candidate object area according to the updated confidence.
  • an electronic device including:
  • the memory stores instructions that can be executed by the at least one processor, and the instructions are executed by the at least one processor, so that the at least one processor can implement the method for object detection in vehicle-road coordination.
  • a non-transitory computer-readable storage medium storing computer instructions, wherein the computer instructions are used to cause the computer to execute a method for detecting objects in vehicle-road coordination.
  • a computer program product including a computer program, and when the computer program is executed by a processor, a method for detecting objects in vehicle-road coordination is implemented.
  • a roadside device including the above-mentioned electronic device.
  • a cloud control platform including the above-mentioned electronic device.
  • the confidence of the candidate object area is updated according to the intersection ratio between the candidate object areas and the degree of occlusion of the candidate object area, and then based on the updated Confidence, to detect an object in an image from candidate object regions.
  • intersection ratio between candidate object regions can reflect the degree of overlap between candidate object regions
  • occlusion degree of candidate object regions can reflect the degree of occlusion of candidate object regions, therefore, according to the above intersection ratio and occlusion degree, update the candidate
  • the confidence of the target area can refer to the overlap between the target areas, so that the updated confidence of the candidate target area is closer to the actual situation, so that the image can be detected according to the updated confidence, which can improve the accuracy of target detection .
  • FIG. 1 is a schematic flowchart of a method for detecting objects in vehicle-road coordination according to an embodiment of the present disclosure
  • FIG. 2 is a schematic diagram of an image provided according to an embodiment of the present disclosure
  • Fig. 3a is a schematic structural diagram of a network model provided according to an embodiment of the present disclosure.
  • Fig. 3b is a schematic structural diagram of another network model provided according to an embodiment of the present disclosure.
  • Fig. 4 is a schematic structural diagram of an object detection device in vehicle-road coordination provided according to an embodiment of the present disclosure
  • Fig. 5 is a schematic structural diagram of an electronic device provided according to an embodiment of the present disclosure.
  • Embodiments of the present disclosure provide a method, device, and roadside equipment for object detection in vehicle-road coordination.
  • a method for object detection in vehicle-road coordination includes:
  • Objects in the image are detected from candidate object regions based on the updated confidence.
  • intersection ratio between candidate object regions can reflect the degree of overlap between candidate object regions
  • occlusion degree of candidate object regions can reflect the degree of occlusion of candidate object regions, therefore, according to the above intersection ratio and occlusion degree, update the candidate
  • the confidence of the target area can refer to the overlap between the target areas, so that the updated confidence of the candidate target area is closer to the actual situation, so that the image can be detected according to the updated confidence, which can improve the accuracy of target detection .
  • the execution subject of the embodiment of the present disclosure may be an electronic device integrated with a target detection function, wherein the above-mentioned electronic device may be: a desktop computer, a notebook computer, a server, an image acquisition device, and the like.
  • the image acquisition device may include: a video camera, a camera, a driving recorder, and the like.
  • the solutions provided by the embodiments of the present disclosure can be applied to target detection of images collected in application scenarios such as vehicle-road coordination V2X road monitoring and vehicle path planning.
  • the solutions provided by the embodiments of the present disclosure may also be used to perform target detection on images collected in other scenarios.
  • the above-mentioned other scenes may be scenes with a high density of people, such as subway stations, shopping malls, concerts, etc. If images are collected for such scenes, the people contained in the collected images are often dense, and it is easy for some people’s faces to be blocked by others. Situations where a person's face is occluded.
  • the above-mentioned scene can also be a scene with relatively dense personnel such as the entrance of a museum, a bank lobby, etc. For image collection of such a scene, the face of the person may be blocked by other people or buildings in the collected image.
  • the aforementioned objects may be human faces, animals, vehicles, and so on.
  • FIG. 1 is a schematic flowchart of a method for object detection in vehicle-road coordination provided by an embodiment of the present disclosure.
  • the above method includes the following steps S101 - S103.
  • Step S101 Perform target detection on the image to obtain the candidate target area in the image, the confidence level of the candidate target area, and the occlusion degree of the candidate target area.
  • the foregoing images may be images acquired through image acquisition for a specific scene.
  • the above-mentioned scenes can include vehicle driving scenes, parking lot scenes, etc.
  • the above-mentioned objects can be vehicles; the above-mentioned scenes can also include public space scenes such as subway stations and high-speed rail stations.
  • the above-mentioned objects can be people.
  • a preset target detection algorithm may be used to perform target detection on an image to obtain a candidate target area in the image, a confidence degree of the candidate target area, and an occlusion degree of the candidate target area.
  • the aforementioned preset target detection algorithm may be a detection algorithm adopted for different types of targets. For example, when the target is a person, a face detection algorithm, a human body detection algorithm, etc. can be used; when the target is a vehicle, a vehicle detection algorithm, a license plate detection algorithm, etc. can be used.
  • the candidate target area refers to the area where the target may exist after target detection. Taking FIG. 2 as an example, the area surrounded by each rectangular frame in FIG. 2 is a candidate target area obtained by performing animal detection on the image.
  • the confidence of the candidate target area reflects: the possibility of the existence of the target in the candidate target area.
  • the confidence level above can be expressed in decimals, percentages, and the like. The larger the value of the confidence degree, the higher the probability that the target exists in the candidate target area.
  • the target is a person
  • the confidence of the candidate target area A is greater than the confidence of the candidate target area B, it means that the possibility of a person in the candidate target area A is higher than the possibility of a person in the candidate target area B.
  • the degree of occlusion of the candidate target area reflects: the degree of occlusion of the candidate target area.
  • the above occlusion degree can be represented by decimals, percentages, etc., and can also be expressed by the occlusion level serial number, for example: the occlusion level serial number includes 1, 2, 3, wherein, the serial number 1 can indicate that the occlusion level is severe occlusion, and the serial number is 2 can indicate that the occlusion level is moderate occlusion, and the sequence number 3 can indicate that the occlusion level is light occlusion.
  • Step S102 Based on the intersection-over-union ratio between the candidate target areas and the degree of occlusion of the candidate target areas, update the confidence of the candidate target areas.
  • intersection-over-union ratio between candidate object regions is used to describe the coincidence degree between two candidate object regions.
  • intersection ratio is higher, it means that the overlap degree between the two candidate target regions is higher; when the intersection ratio is lower, it means that the overlap degree between the two candidate target regions is lower.
  • the overlapping area between two candidate target regions can be calculated to obtain the first area
  • the sum of the areas of the two candidate target regions can be calculated to obtain the second area
  • the difference between the second area and the first area can be calculated to obtain the second area
  • the ratio between the first area and the third area is determined as the intersection ratio between the candidate target areas.
  • the area of the candidate target area A is 48, and the area of the candidate target area B is 32, wherein, the overlapping area of the candidate target area A and the candidate target area B is 16, that is, the first area is 16, and the candidate target area A and the candidate target area B are 16.
  • a reference area may be selected from each candidate target area, and for every other candidate target area except the reference area in each candidate target area, the intersection ratio between the other candidate target area and the reference area is calculated, The calculated intersection and union ratio is determined as the intersection and union ratio for updating the confidence of the candidate target region.
  • the aforementioned reference region may be the region with the highest confidence among the candidate target regions.
  • an intersection and union ratio may be selected from the intersection and union ratios between the candidate target area and other candidate target areas, and the selected intersection and union ratio may be determined as the The intersection and union ratio of the confidence of the candidate target region is updated.
  • the maximum intersection and union ratio, the average intersection and union ratio, the median intersection and union ratio, or the minimum intersection and union ratio may be selected from the above-mentioned multiple intersection and union ratios.
  • the adjustment coefficient can be calculated according to the preset first weight and second weight according to the intersection ratio between the candidate target areas and the degree of occlusion of the candidate target area , according to the calculated adjustment coefficient, update the confidence of the candidate target region.
  • the product of the intersection ratio between candidate target areas and the first weight can be calculated, and the product of the occlusion degree of the candidate target area and the second weight can be calculated, and the sum of the two calculated products can be used as Adjustment coefficient.
  • the intersection ratio between candidate target areas is 80%
  • the occlusion degree of candidate target areas is 50%
  • the preset first weight is 0.8
  • the preset second weight is 0.2.
  • the product of the adjustment coefficient and the confidence of the candidate target area may be calculated as the updated confidence of the candidate target area.
  • Step S103 Detect the target in the image from the candidate target regions according to the updated confidence.
  • a candidate target area whose updated confidence is greater than a preset confidence threshold may be selected, and the target in the selected candidate target area is determined as the target in the image.
  • the aforementioned preset reliability threshold may be set by staff based on experience, for example: when the confidence is represented by a percentage, the preset reliability threshold may be 90%, 95% and so on.
  • the probability that these candidate target regions contain a target is higher than the probability that other candidate target regions contain a target. Therefore, the target in the candidate target area whose confidence is greater than the preset confidence threshold is determined as the target in the image, and the accuracy of the obtained target is higher.
  • a preset number of candidate target areas with the highest updated confidence may also be selected, and the target in the selected candidate target areas is determined as the target in the image.
  • the above-mentioned preset number can be set by the staff based on experience, for example: the above-mentioned preset number can be 1, 3, 5, etc.
  • the target in the preset number of candidate target areas with the highest confidence is determined as the target in the image, and the accuracy of the obtained target is relatively high.
  • the confidence of the candidate object area is updated according to the intersection ratio between the candidate object areas and the degree of occlusion of the candidate object area, and then based on the updated Confidence, to detect an object in an image from candidate object regions.
  • intersection ratio between candidate object regions can reflect the degree of overlap between candidate object regions
  • occlusion degree of candidate object regions can reflect the degree of occlusion of candidate object regions, therefore, according to the above intersection ratio and occlusion degree, update
  • the confidence of the candidate target area can refer to the overlap between the candidate target areas, so that the updated confidence of the candidate target area is closer to the actual situation, so that the image can be detected according to the updated confidence, which can improve the accuracy of target detection. Accuracy.
  • the occlusion of objects is especially serious.
  • the occlusion degree of each candidate target area is relatively high, so that the target in the candidate target area is incomplete, and the error of the confidence degree of the obtained candidate target area is relatively large. Update the confidence of the candidate target area by the occlusion degree of the candidate target area, which can effectively eliminate the influence of the error on the confidence of each candidate target area when it is occluded, so that the accuracy of the updated confidence is high, and then the detection is accurate. . Therefore, the solutions provided by the embodiments of the present disclosure can be better adapted to situations where occlusion exists in dense scenes, and improve the accuracy of object detection.
  • the first region with the highest confidence degree can be cyclically selected from the region set, and according to the intersection ratio between other regions in the region set and the first region and other The degree of occlusion of the region is used to update the confidence of other regions.
  • a confidence update operation is completed, and the above operations are performed cyclically until one region is included in the region set.
  • Each confidence update operation can be called a cycle.
  • the above region set includes: regions that have not been selected in the candidate target regions. Specifically, at the beginning of the first cycle, the region set includes each candidate target region obtained in step S101; in each cycle, after the first region is selected from the region set, the region set no longer includes the selected first region. area.
  • the first area is: the area with the highest confidence among the candidate target areas obtained in step S101; in each subsequent cycle, the first area is: the updated each target area obtained after the last cycle.
  • the region with the highest confidence in the region is: the region with the highest confidence in the region
  • the aforementioned other areas refer to areas in the area concentration except the first area.
  • the area set includes: area 1, area 2, and area 3, where area 1 is the first area, and areas other than the first area are area 2 and area 3, then area 2 and area 3 are other areas.
  • each region in the region set may be traversed, the confidence of each region is sorted from high to low, and the region with the highest confidence is determined as the first region.
  • the first region may also be stored in the prediction set, and as the number of cycles increases, the number of first regions stored in the prediction set also increases.
  • each candidate target area obtained in step S101 is b1, b2, b3, ... bn.
  • the set of regions B ⁇ b1, b2, b3, . . . , bn ⁇ .
  • the area with the highest confidence in each candidate target area is the area b1, so the area b1 is taken as the first area.
  • the areas other than the area b1 are b2, b3, ..., bn, so ⁇ b2, b3, ..., bn ⁇ are other areas.
  • the area with the highest confidence in the updated ⁇ b2, b3, ..., bn ⁇ is the area b2, so the area b2 is taken as the first area.
  • the areas other than the area b2 are b3, ..., bn, so ⁇ b3, ..., bn ⁇ are other areas.
  • the intersection ratio between other regions ⁇ b3,...,bn ⁇ and region b2 and the occlusion degree of other regions ⁇ b3,...,bn ⁇ update the confidence of other regions ⁇ b3,...,bn ⁇ .
  • the region set B ⁇ b3,...,bn ⁇ .
  • the area with the highest confidence in the updated ⁇ b3, ..., bn ⁇ is the area b3, so the area b3 is taken as the first area.
  • the areas other than the area b3 are b4, ..., bn, so ⁇ b4, ..., bn ⁇ are other areas.
  • the intersection ratio between other regions ⁇ b4,...,bn ⁇ and region b3 and the occlusion degree of other regions ⁇ b4,...,bn ⁇ update the confidence of other regions ⁇ b4,...,bn ⁇ .
  • the confidence of the region in the concentrated region is updated according to the intersection ratio between other regions in the concentrated region and the first region and the degree of occlusion of other regions.
  • the degree of occlusion of other regions reflects the degree of occlusion of other regions.
  • the candidate target region after update can be The accuracy of the confidence degree is high; and the intersection ratio between other regions and the first region reflects the coincidence degree between other regions and the first region, and the first region is the region with the highest confidence, and the region with the highest confidence.
  • the degree of coincidence between them can also effectively adjust the confidence of other regions. Therefore, the confidence levels of other regions can be effectively updated according to the intersection ratio and the degree of occlusion in each cycle. Moreover, the accuracy of the updated confidence level can be further improved by looping and iterating the updating process.
  • Step A1 Calculate the intersection and union ratios between other areas in the area set and the first area.
  • first calculate the overlapping area between other areas and the first area calculate the total area of other areas and the first area; then calculate the difference between the above total area and the overlapping area to obtain the target area, and finally calculate the difference between the overlapping area and the target area
  • the ratio between them is determined as the intersection and union ratio between the candidate target regions.
  • Step A2 Determine a first confidence adjustment value according to the intersection and union ratio and the preset intersection and union ratio threshold.
  • intersection and union ratio thresholds may be set by staff based on experience, for example, the intersection and union ratio thresholds may be 90%, 95%, and so on.
  • intersection ratio is less than a preset intersection ratio threshold, if yes, determine the first confidence adjustment value as the first preset value, and if no, determine the first confidence adjustment value is the second preset value.
  • Both the above-mentioned first preset value and the second preset value are set by the staff based on experience.
  • intersection ratio is less than the preset intersection ratio threshold; if yes, determine the first confidence adjustment value is 1; if no, determine the first confidence adjustment value : The difference between 1 and the intersection and union ratio.
  • the preset intersection ratio threshold is 90%
  • the intersection ratio 95% is greater than the preset intersection ratio threshold 90%
  • determine the first Confidence adjustment value: 1-95% 5%
  • determine the intersection and union ratio between other areas and the first area is 50%
  • determine the intersection and union ratio 50% is less than the preset intersection ratio threshold 90%
  • intersection and union ratio when the intersection and union ratio is less than the preset intersection and union ratio threshold, it means that the overlap between other areas and the first area is small, indicating that a small part of the image content in other areas is blocked, and the confidence of the other areas detected is The accuracy is high, in which case no adjustments to the confidences of the other regions may be made.
  • Setting the first confidence adjustment value to 1 can realize that the confidence of the region is not adjusted.
  • the intersection ratio is not less than the preset intersection ratio threshold, it means that the overlap between other regions and the first region is relatively large, indicating that most of the image content in other regions is blocked, and the confidence of the other regions detected is The accuracy is low. In this case, the confidence of other regions needs to be adjusted. Setting the first confidence adjustment value to 1 and the difference between the intersection and union ratio can make the adjusted confidence approach the actual situation.
  • Step A3 Determine a second confidence adjustment value according to the degree of occlusion of other areas.
  • the product of the degree of occlusion of other regions and a preset adjustment coefficient may be calculated as the second confidence adjustment value.
  • the aforementioned preset adjustment coefficient can be set by the staff based on experience, for example: the preset adjustment coefficient can be 1.2, 1.5, etc.
  • the second confidence adjustment value g(occ_pred) may also be determined according to the following expression:
  • occ_pred is the occlusion degree of other areas
  • is a preset constant, ⁇ >1.
  • the second confidence adjustment value g(occ_pred) increases as the occlusion degree of other regions increases.
  • the accuracy of the confidence of the region is low when the occlusion degree of the region is high, it is necessary to make a large adjustment to the confidence of the region so that the adjusted confidence is close to the actual situation.
  • the second confidence degree adjustment value g(occ_pred) increases as the degree of occlusion of other regions increases, that is, the higher the degree of occlusion of other regions, the larger the second degree of confidence adjustment value.
  • the confidence of other regions is greatly adjusted, so that the adjusted confidence of other regions is close to the actual situation.
  • Step A4 Using the first confidence adjustment value and the second confidence adjustment value, adjust the confidence of other regions.
  • the confidence of other regions can be adjusted according to the following expression:
  • S' represents the confidence degree of other regions after adjustment
  • S represents the confidence degree of other regions before adjustment
  • T1 represents the first confidence degree adjustment value
  • T2 represents the second confidence degree adjustment value
  • the adjusted confidence level is the product of the first confidence level adjustment value, the second confidence level adjustment value and the confidence levels of other regions, and because the first confidence level adjustment value and the second confidence level adjustment value are different from The angle reflects the occlusion situation of other regions. Therefore, the above adjusted confidence refers to the occlusion situation of other regions, making the adjusted confidence closer to the actual situation.
  • the product of the first confidence adjustment value, the second confidence adjustment value, and the confidence of other regions can also be calculated as the reference confidence, and the above reference can be adjusted by the preset confidence error value. Confidence, the adjusted reference confidence is obtained as the adjusted confidence of other regions.
  • the product of the preset confidence error value and the parameter confidence may be calculated, and the calculated product may be determined as the adjusted confidence of other regions.
  • the intersection and union ratio reflects the coincidence degree of other regions and the first region
  • the second confidence adjustment value is based on the intersection and union ratio of other regions
  • the degree of occlusion is determined, and the degree of occlusion reflects the degree of occlusion of other regions, and the first and second confidence adjustment values can reflect the occlusion of other regions from different angles.
  • the first confidence adjustment value and the second confidence adjustment value when using the first confidence adjustment value and the second confidence adjustment value to adjust the confidence of other areas, since the first confidence adjustment value and the second confidence adjustment value reflect the occlusion of other areas from different angles, use When adjusting the first confidence level adjustment value and the second confidence level adjustment value, the confidence level is adjusted based on more accurate occlusion conditions of other regions, so that the adjusted confidence level is closer to the actual situation.
  • each candidate target area is b1, b2, b3, and the preset intersection-over-union ratio threshold Nt is 90%, the confidence and occlusion degree of each candidate target area are shown in Table 1 below.
  • the region set B ⁇ b1, b2, b3 ⁇ , wherein the confidence Cv1 of the region b1 is the highest, the region b1 is the first region, and the regions b2 and b3 are other regions.
  • the region set B ⁇ b2, b3 ⁇ , wherein the confidence Cv21 of the region b2 is the highest, the region b2 is the first region, and the region b3 is other regions.
  • target detection can be performed on the image for different target scales, and the candidate target areas of different scales in the image, the confidence of the candidate target areas, and the occlusion degree of the candidate target areas can be obtained .
  • the target scale refers to: the size of the target.
  • the target scale may be a preset scale value, for example, the target scale may be 16x16, 32x32, or 64x64.
  • multi-layer feature extraction can be performed on the image, and then feature fusion is performed on different features to obtain features of different scales.
  • the features of different scales are used to detect the target of the image, and the candidate target areas of different scales are obtained, and the confidence and occlusion degree of the candidate target areas of different scales are obtained.
  • the feature information of the candidate object regions at different scales is enriched.
  • an image may be input into a pre-trained target detection model, and the candidate target areas in the image output by the target detection model, the confidence of the candidate target areas, and the occlusion degree of the candidate target areas may be obtained.
  • the above object detection model includes: an object detection layer for detecting candidate object areas in an image, and an occlusion degree prediction layer for predicting the degree of occlusion of the candidate object areas.
  • the target detection layer in addition to detecting the candidate target area in the image, can also calculate the confidence of the candidate target area.
  • the network structure of the target detection model can be shown in Figure 3a.
  • the target detection model includes Object detection layer and occlusion prediction layer.
  • the target detection layer in the model detects the candidate target area in the image, calculates the confidence of the candidate target area, and transmits the detection result to the occlusion degree prediction layer; the occlusion degree prediction layer predicts Each candidate target area is occluded; the target detection model outputs the candidate target area, the confidence of the candidate target area, and the occlusion degree.
  • FPN Feature Pyramid Networks, Feature Pyramid Network
  • FPN is used to obtain candidate target areas of various scales, confidence and occlusion of candidate target areas.
  • the network structure of the network model after adding the FPN may be shown in FIG. 3b, and the network structure shown in FIG. 3b includes a backbone network (Backbone) and an FPN.
  • Backbone backbone network
  • FPN FPN
  • the backbone network is used to extract the features of the image, obtain the image features of different levels in the image, and input the image features of different levels into the FPN.
  • each convolutional layer of the convolutional neural network can perform convolution operations on images to obtain image features at different levels.
  • FPN is used to perform feature fusion of image features at different levels to obtain image features of different scales, perform target detection based on image features of different scales, obtain candidate target areas of different scales, and obtain the confidence and occlusion degree of candidate target areas , realizing the divide-and-conquer processing of image features at different levels.
  • the sample image is used as the training sample, and the real candidate target area and the real occlusion degree in the sample image are used as the training labels to train the preset neural network model until the training end condition is met, and the obtained The trained object detection model.
  • the aforementioned preset neural network model may be a CNN (Conv Neural Network, convolutional neural network) model, an RNN (Recurrent Neural Network, recursive neural network) model, a DNN (Deep Neural Network, deep neural network) model, etc.
  • CNN Conv Neural Network, convolutional neural network
  • RNN Recurrent Neural Network, recursive neural network
  • DNN Deep Neural Network, deep neural network
  • the above preset neural network model performs target detection on the sample image, obtains the candidate target area and the degree of occlusion of the sample image, and calculates the candidate target area and the real target area and the difference between the occlusion degree of the candidate target area and the real occlusion degree, adjust the parameters of the neural network model according to the calculated difference, and iteratively adjust the parameters until the preset training end conditions are met.
  • the aforementioned training end conditions may be that the number of training times reaches a preset number of times, the model parameters meet the preset model parameter convergence conditions, and the like.
  • the target detection model Since the target detection model is trained through a large number of training samples, during the training process, the target detection model learns the features of the target area and the occluded features in the sample image, therefore, the target detection model has strong robustness, so the target detection model is adopted.
  • the detection model performs target detection on an image, it can output accurate candidate target areas, confidence levels of candidate target areas, and occlusion degrees.
  • the image in addition to using the target detection model to perform target detection on the image, the image can also be divided into multiple regions, and for each region, the image features in the region are extracted, and the candidate targets in the region are determined according to the image features area.
  • the aforementioned image features include: texture features, color features, edge features, and the like.
  • the confidence of each candidate target is predicted according to the image features of each candidate target area.
  • the degree of occlusion of each candidate target area may also be calculated according to the layer to which each candidate target area belongs and the location information.
  • the layer to which the candidate target areas belong and the relative relationship between positions it can be determined whether occlusion occurs between the candidate target areas, and the ratio between the occluded area and the area of the occluded area can be calculated as the candidate target area degree of occlusion.
  • the candidate target area A when the candidate target area A is located in the foreground layer and the candidate target area B is located in the background layer, and the position information of the candidate target area A and the candidate target area B overlaps, it can be determined that the candidate target area B is blocked, and the candidate target area B is calculated.
  • the ratio of the shaded area of the target area B to the area of the candidate target area B is used as the shaded degree of the candidate target area B.
  • an embodiment of the present disclosure further provides a device for detecting objects in vehicle-road coordination.
  • FIG. 4 is a schematic structural diagram of an object detection device in vehicle-road coordination provided by an embodiment of the present disclosure.
  • the above-mentioned device includes the following modules 401-403.
  • An information obtaining module 401 configured to perform target detection on an image, and obtain a candidate target area in the image, a confidence degree of the candidate target area, and an occlusion degree of the candidate target area;
  • a confidence update module 402 configured to update the confidence of the candidate target region based on the intersection-over-union ratio between the candidate target regions and the degree of occlusion of the candidate target region;
  • the target detection module 403 is configured to detect the target in the image from the candidate target regions according to the updated confidence.
  • the confidence of the candidate object area is updated according to the intersection ratio between the candidate object areas and the degree of occlusion of the candidate object area, and then based on the updated Confidence, to detect an object in an image from candidate object regions. Since the intersection ratio between candidate object regions can reflect the degree of overlap between candidate object regions, and the occlusion degree of candidate object regions can reflect the degree of occlusion of candidate object regions, therefore, according to the above intersection ratio and occlusion degree, update
  • the confidence of the candidate target area can refer to the overlap between the candidate target areas, so that the updated confidence of the candidate target area is more inclined to the actual situation. Therefore, the target detection is performed on the image according to the updated confidence level, which can improve the accuracy of target detection.
  • the confidence update module 402 is specifically configured to cyclically select the first region with the highest confidence from the region set, and according to the intersection ratio between other regions in the region set and the first region and other regions The degree of occlusion of other areas is updated until the set of areas includes an area, wherein the set of areas includes: an unselected area in the candidate target area.
  • the confidence of the region in the concentrated region is updated according to the intersection ratio between other regions in the concentrated region and the first region and the degree of occlusion of other regions.
  • the degree of occlusion of other regions reflects the degree of occlusion of other regions.
  • the candidate target region after update can be The accuracy of the confidence degree is high; and the intersection ratio between other regions and the first region reflects the coincidence degree between other regions and the first region, and the first region is the region with the highest confidence, and the region with the highest confidence.
  • the degree of coincidence between them can also effectively adjust the confidence of other regions. Therefore, the confidence levels of other regions can be effectively updated according to the intersection ratio and the degree of occlusion in each cycle. Moreover, the accuracy of the updated confidence level can be further improved by looping and iterating the updating process.
  • the confidence update module 402 includes:
  • intersection ratio calculation unit configured to calculate intersection ratios between other regions in the region set and the first region
  • a first adjustment value determination unit configured to determine a first confidence adjustment value according to the intersection ratio and a preset intersection ratio threshold
  • a second adjustment value determination unit configured to determine a second confidence adjustment value according to the degree of occlusion of other areas
  • a confidence adjustment unit configured to adjust the confidence of other regions by using the first confidence adjustment value and the second confidence adjustment value.
  • the intersection and union ratio reflects the coincidence degree of other regions and the first region
  • the second confidence adjustment value is based on the intersection and union ratio of other regions
  • the degree of occlusion is determined, and the degree of occlusion reflects the degree of occlusion of other regions, and the first and second confidence adjustment values can reflect the occlusion of other regions from different angles.
  • the first confidence adjustment value and the second confidence adjustment value when using the first confidence adjustment value and the second confidence adjustment value to adjust the confidence of other areas, since the first confidence adjustment value and the second confidence adjustment value reflect the occlusion of other areas from different angles, use When adjusting the first confidence level adjustment value and the second confidence level adjustment value, the confidence level is adjusted based on more accurate occlusion conditions of other regions, so that the adjusted confidence level is closer to the actual situation.
  • the first adjustment value determination unit is specifically configured to judge whether the intersection-over-union ratio is smaller than a preset intersection-over-union ratio threshold; if yes, determine a first confidence adjustment value of 1; If not, determine that the adjusted value of the first confidence level is the difference between 1 and the intersection-over-union ratio.
  • intersection and union ratio when the intersection and union ratio is less than the preset intersection and union ratio threshold, it means that the overlap between other areas and the first area is small, indicating that a small part of the image content in other areas is blocked, and the confidence of the other areas detected is The accuracy is high, in which case no adjustments to the confidences of the other regions may be made.
  • Setting the first confidence adjustment value to 1 can realize that the confidence of the region is not adjusted.
  • the intersection ratio is not less than the preset intersection ratio threshold, it means that the overlap between other regions and the first region is relatively large, indicating that most of the image content in other regions is blocked, and the confidence of the other regions detected is The accuracy is low. In this case, the confidence of other regions needs to be adjusted. Setting the first confidence adjustment value to 1 and the difference between the intersection and union ratio can make the adjusted confidence approach the actual situation.
  • the second adjustment value determination unit is specifically configured to determine the second confidence adjustment value g(occ_pred) according to the following expression:
  • occ_pred is the occlusion degree of other areas
  • is a preset constant, ⁇ >1.
  • the accuracy of the confidence of the region is low when the occlusion degree of the region is high, it is necessary to make a large adjustment to the confidence of the region so that the adjusted confidence is close to the actual situation.
  • the second confidence degree adjustment value g(occ_pred) increases as the degree of occlusion of other regions increases, that is, the higher the degree of occlusion of other regions, the larger the second degree of confidence adjustment value.
  • the confidence of other regions is greatly adjusted, so that the adjusted confidence of other regions is close to the actual situation.
  • the confidence adjustment unit is specifically configured to adjust the confidence of other regions according to the following expression:
  • S' represents the confidence degree of other regions after adjustment
  • S represents the confidence degree of other regions before adjustment
  • T1 represents the first confidence adjustment value
  • T2 represents the second confidence adjustment value
  • the adjusted confidence level is the product of the first confidence level adjustment value, the second confidence level adjustment value and the confidence levels of other regions, and because the first confidence level adjustment value and the second confidence level adjustment value are different from The angle reflects the occlusion situation of other regions. Therefore, the above adjusted confidence refers to the occlusion situation of other regions, making the adjusted confidence closer to the actual situation.
  • the target detection module 403 is specifically configured to select a candidate target area whose updated confidence is greater than a preset confidence threshold, and determine the target in the selected candidate target area as the target in the image or select a preset number of candidate target areas with the highest confidence after updating, and determine the targets in the selected candidate target areas as the targets in the image.
  • the probability that these candidate target regions contain a target is higher than the probability that other candidate target regions contain a target. Therefore, if the target in the candidate target area whose confidence is greater than the preset confidence threshold is determined as the target in the image, the accuracy of the obtained target is relatively high; for the preset number of candidate target areas with the largest confidence, these The probability of an object being contained in a candidate object region is higher than that of other candidate regions. Therefore, the target in the preset number of candidate target areas with the highest confidence is determined as the target in the image, and the accuracy of the obtained target is relatively high.
  • the information obtaining module 401 is specifically configured to perform target detection on an image for different target scales, and obtain candidate target areas of different scales in the image, confidence levels of candidate target areas, and candidate target areas. The degree of occlusion of the target area.
  • the feature information of the candidate object regions at different scales is enriched.
  • the information obtaining module 401 is specifically configured to input an image into a pre-trained target detection model, and obtain the candidate target area and the candidate target area in the image output by the target detection model.
  • the confidence level and the occlusion degree of the candidate target area wherein the target detection model includes: an object detection layer for detecting the candidate target area in the image and an occlusion degree prediction layer for predicting the occlusion degree of the candidate target area.
  • the target detection model since the target detection model is trained through a large number of training samples, during the training process, the target detection model has learned the features of the target area and the occluded features in the sample image, so the target detection model has strong robustness, thus When the target detection model is used to detect the target in the image, it can output the accurate candidate target area, the confidence degree of the candidate target area and the degree of occlusion.
  • the present disclosure also provides an electronic device, a readable storage medium, and a computer program product.
  • an electronic device including:
  • the memory stores instructions that can be executed by the at least one processor, and the instructions are executed by the at least one processor, so that the at least one processor can perform any vehicle-road coordination in the foregoing method embodiments.
  • Object detection method
  • a non-transitory computer-readable storage medium storing computer instructions, wherein the computer instructions are used to enable the computer to execute any vehicle-road coordination target in the foregoing method embodiments. Detection method.
  • a computer program product including a computer program.
  • the computer program is executed by a processor, any method for object detection in vehicle-road coordination in the foregoing method embodiments is implemented.
  • a roadside device including the above-mentioned electronic device.
  • a cloud control platform including the above-mentioned electronic device.
  • FIG. 5 shows a schematic block diagram of an example electronic device 500 that may be used to implement embodiments of the present disclosure.
  • Electronic device is intended to represent various forms of digital computers, such as laptops, desktops, workstations, personal digital assistants, servers, blade servers, mainframes, and other suitable computers.
  • Electronic devices may also represent various forms of mobile devices, such as personal digital processing, cellular telephones, smart phones, wearable devices, and other similar computing devices.
  • the components shown herein, their connections and relationships, and their functions, are by way of example only, and are not intended to limit implementations of the disclosure described and/or claimed herein.
  • the device 500 includes a computing unit 501 that can execute according to a computer program stored in a read-only memory (ROM) 502 or loaded from a storage unit 508 into a random-access memory (RAM) 503. Various appropriate actions and treatments. In the RAM 503, various programs and data necessary for the operation of the device 500 can also be stored.
  • the computing unit 501, ROM 502, and RAM 503 are connected to each other through a bus 504.
  • An input/output (I/O) interface 505 is also connected to the bus 504 .
  • the I/O interface 505 includes: an input unit 506, such as a keyboard, a mouse, etc.; an output unit 507, such as various types of displays, speakers, etc.; a storage unit 508, such as a magnetic disk, an optical disk, etc. ; and a communication unit 509, such as a network card, a modem, a wireless communication transceiver, and the like.
  • the communication unit 509 allows the device 500 to exchange information/data with other devices over a computer network such as the Internet and/or various telecommunication networks.
  • the computing unit 501 may be various general-purpose and/or special-purpose processing components having processing and computing capabilities. Some examples of computing units 501 include, but are not limited to, central processing units (CPUs), graphics processing units (GPUs), various dedicated artificial intelligence (AI) computing chips, various computing units that run machine learning model algorithms, digital signal processing processor (DSP), and any suitable processor, controller, microcontroller, etc.
  • the calculation unit 501 executes various methods and processes described above, for example, the object detection method in vehicle-road coordination.
  • the object detection method in vehicle-road coordination can be implemented as a computer software program, which is tangibly contained in a machine-readable medium, such as the storage unit 508 .
  • part or all of the computer program may be loaded and/or installed on the device 500 via the ROM 502 and/or the communication unit 509.
  • the computer program When the computer program is loaded into the RAM 503 and executed by the computing unit 501, one or more steps of the object detection method in the vehicle-road coordination described above can be executed.
  • the calculation unit 501 may be configured in any other appropriate way (for example, by means of firmware) to execute the object detection method in vehicle-road coordination.
  • Various implementations of the systems and techniques described above herein can be implemented in digital electronic circuit systems, integrated circuit systems, field programmable gate arrays (FPGAs), application specific integrated circuits (ASICs), application specific standard products (ASSPs), systems on chips Implemented in a system of systems (SOC), load programmable logic device (CPLD), computer hardware, firmware, software, and/or combinations thereof.
  • FPGAs field programmable gate arrays
  • ASICs application specific integrated circuits
  • ASSPs application specific standard products
  • SOC system of systems
  • CPLD load programmable logic device
  • computer hardware firmware, software, and/or combinations thereof.
  • programmable processor can be special-purpose or general-purpose programmable processor, can receive data and instruction from storage system, at least one input device, and at least one output device, and transmit data and instruction to this storage system, this at least one input device, and this at least one output device an output device.
  • the roadside equipment may also include communication components, etc., and the electronic equipment and communication components may be integrally integrated or separately provided.
  • Electronic devices can obtain data from sensing devices (such as roadside cameras), such as pictures and videos, for image and video processing and data calculation.
  • the electronic device itself may also have the function of acquiring sensory data and communication functions, such as an AI camera, and the electronic device may directly perform image and video processing and data calculation based on the acquired sensory data.
  • the cloud control platform performs processing on the cloud
  • the electronic devices included in the cloud control platform can obtain data from sensing devices (such as roadside cameras), such as pictures and videos, to perform image and video processing and data calculation; the cloud control platform It can also be called vehicle-road collaborative management platform, edge computing platform, cloud computing platform, central system, cloud server, etc.
  • Program code for implementing the methods of the present disclosure may be written in any combination of one or more programming languages. These program codes may be provided to a processor or controller of a general-purpose computer, a special purpose computer, or other programmable data processing devices, so that the program codes, when executed by the processor or controller, make the functions/functions specified in the flow diagrams and/or block diagrams Action is implemented.
  • the program code may execute entirely on the machine, partly on the machine, as a stand-alone software package partly on the machine and partly on a remote machine or entirely on the remote machine or server.
  • a machine-readable medium may be a tangible medium that may contain or store a program for use by or in conjunction with an instruction execution system, apparatus, or device.
  • a machine-readable medium may be a machine-readable signal medium or a machine-readable storage medium.
  • a machine-readable medium may include, but is not limited to, electronic, magnetic, optical, electromagnetic, infrared, or semiconductor systems, apparatus, or devices, or any suitable combination of the foregoing.
  • machine-readable storage media would include one or more wire-based electrical connections, portable computer discs, hard drives, random access memory (RAM), read only memory (ROM), erasable programmable read only memory (EPROM or flash memory), optical fiber, compact disk read only memory (CD-ROM), optical storage, magnetic storage, or any suitable combination of the foregoing.
  • RAM random access memory
  • ROM read only memory
  • EPROM or flash memory erasable programmable read only memory
  • CD-ROM compact disk read only memory
  • magnetic storage or any suitable combination of the foregoing.
  • the systems and techniques described herein can be implemented on a computer having a display device (e.g., a CRT (cathode ray tube) or LCD (liquid crystal display) monitor) for displaying information to the user. ); and a keyboard and pointing device (eg, a mouse or a trackball) through which a user can provide input to the computer.
  • a display device e.g., a CRT (cathode ray tube) or LCD (liquid crystal display) monitor
  • a keyboard and pointing device eg, a mouse or a trackball
  • Other kinds of devices can also be used to provide interaction with the user; for example, the feedback provided to the user can be any form of sensory feedback (e.g., visual feedback, auditory feedback, or tactile feedback); and can be in any form (including Acoustic input, speech input or, tactile input) to receive input from the user.
  • the systems and techniques described herein can be implemented in a computing system that includes back-end components (e.g., as a data server), or a computing system that includes middleware components (e.g., an application server), or a computing system that includes front-end components (e.g., as a a user computer having a graphical user interface or web browser through which a user can interact with embodiments of the systems and techniques described herein), or including such backend components, middleware components, Or any combination of front-end components in a computing system.
  • the components of the system can be interconnected by any form or medium of digital data communication, eg, a communication network. Examples of communication networks include: Local Area Network (LAN), Wide Area Network (WAN) and the Internet.
  • a computer system may include clients and servers.
  • Clients and servers are generally remote from each other and typically interact through a communication network.
  • the relationship of client and server arises by computer programs running on the respective computers and having a client-server relationship to each other.
  • the server can be a cloud server, a server of a distributed system, or a server combined with a blockchain.
  • steps may be reordered, added or deleted using the various forms of flow shown above.
  • each step described in the present disclosure may be executed in parallel, sequentially, or in a different order, as long as the desired result of the technical solution disclosed in the present disclosure can be achieved, no limitation is imposed herein.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Biomedical Technology (AREA)
  • General Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Software Systems (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Signal Processing (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Image Analysis (AREA)
  • Traffic Control Systems (AREA)

Abstract

The present disclosure relates to the field of intelligent transportation, and in particular, to the technical field of image detection. Disclosed are a target detection method and apparatus in vehicle-road coordination, and a roadside device. A specific implementation solution comprises: performing target detection on an image, and obtaining candidate target regions in the image, confidence levels of the candidate target regions, and the degrees of blockage of the candidate target regions; updating the confidence levels of the candidate target regions on the basis of the intersection over union between the candidate target regions and the degrees of blockage of the candidate target regions; and detecting a target in the image from the candidate target regions according to the updated confidence levels. When the solution provided by embodiments of the present disclosure is used for target detection, the accuracy of target detection is improved.

Description

车路协同中目标检测方法、装置和路侧设备Target detection method, device and roadside equipment in vehicle-road coordination
本申请要求于2021年06月28日提交中国专利局、申请号为202110721853.4发明名称为“车路协同中目标检测方法、装置和路侧设备”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。This application claims the priority of the Chinese patent application with the application number 202110721853.4 submitted to the China Patent Office on June 28, 2021, and the invention title is "Target Detection Method, Device and Roadside Equipment in Vehicle-Infrastructure Coordination", the entire content of which is incorporated by reference incorporated in this application.
技术领域technical field
本公开涉及智能交通技术领域,尤其涉及图像检测技术领域。The present disclosure relates to the technical field of intelligent transportation, in particular to the technical field of image detection.
背景技术Background technique
在车路协同V2X(Vehicle to everything,车用无线通信技术)的道路监控、车辆路径规划等应用场景中,获得图像采集设备采集的图像之后,需要对图像中的人、动物、车辆等目标进行检测,以定位出图像中的目标,进而触发针对上述目标的处理操作,或者结合上述目标进行车辆路径规划等。因此,需要一种车路协同中目标检测方法,以对图像中的目标进行检测。In the application scenarios such as road monitoring and vehicle route planning of V2X (Vehicle to everything, wireless communication technology for vehicles), after obtaining the image collected by the image acquisition device, it is necessary to carry out an Detection, to locate the target in the image, and then trigger the processing operation for the above target, or combine the above target for vehicle path planning, etc. Therefore, a method for object detection in vehicle-road coordination is needed to detect objects in images.
发明内容Contents of the invention
本公开提供了一种车路协同中目标检测方法、装置和路侧设备。The present disclosure provides a target detection method, device and roadside equipment in vehicle-road coordination.
根据本公开的一方面,提供了一种车路协同中目标检测方法,所述方法包括:According to an aspect of the present disclosure, there is provided a method for object detection in vehicle-road coordination, the method comprising:
对图像进行目标检测,得到所述图像中的候选目标区域、候选目标区域的置信度以及候选目标区域的被遮挡度;Carrying out target detection on the image to obtain a candidate target area in the image, a confidence degree of the candidate target area, and an occlusion degree of the candidate target area;
基于候选目标区域间的交并比以及候选目标区域的被遮挡度,更新候选目标区域的置信度;Update the confidence of the candidate target area based on the intersection ratio between the candidate target areas and the occlusion degree of the candidate target area;
根据更新后的置信度,从候选目标区域中检测所述图像中的目标。Objects in the image are detected from candidate object regions according to the updated confidence.
根据本公开的一方面,提供了一种车路协同中目标检测装置,所述装置包括:According to an aspect of the present disclosure, a device for detecting objects in vehicle-road coordination is provided, the device comprising:
信息获得模块,用于对图像进行目标检测,得到所述图像中的候选目标区域、候选目标区域的置信度以及候选目标区域的被遮挡度;An information obtaining module, configured to perform target detection on an image, and obtain a candidate target area in the image, a confidence degree of the candidate target area, and an occlusion degree of the candidate target area;
置信度更新模块,用于基于候选目标区域间的交并比以及候选目标区域的被遮挡度,更新候选目标区域的置信度;Confidence update module, for updating the confidence of the candidate target area based on the intersection ratio between the candidate target areas and the degree of occlusion of the candidate target area;
目标检测模块,用于根据更新后的置信度,从候选目标区域中检测所述图像中的目标。The object detection module is used to detect the object in the image from the candidate object area according to the updated confidence.
根据本公开的另一方面,提供了一种电子设备,包括:According to another aspect of the present disclosure, an electronic device is provided, including:
至少一个处理器;以及at least one processor; and
与所述至少一个处理器通信连接的存储器;其中,a memory communicatively coupled to the at least one processor; wherein,
所述存储器存储有可被所述至少一个处理器执行的指令,所述指令被所述至少一个处理器执行,以使所述至少一个处理器能够实现车路协同中目标检测方法。The memory stores instructions that can be executed by the at least one processor, and the instructions are executed by the at least one processor, so that the at least one processor can implement the method for object detection in vehicle-road coordination.
根据本公开的另一方面,提供了一种存储有计算机指令的非瞬时计算机可读存储介质,其中,所述计算机指令用于使所述计算机执行车路协同中目标检测方法。According to another aspect of the present disclosure, there is provided a non-transitory computer-readable storage medium storing computer instructions, wherein the computer instructions are used to cause the computer to execute a method for detecting objects in vehicle-road coordination.
根据本公开的另一方面,提供了一种计算机程序产品,包括计算机程序,所述计算机程序在被处理器执行时实现车路协同中目标检测方法。According to another aspect of the present disclosure, a computer program product is provided, including a computer program, and when the computer program is executed by a processor, a method for detecting objects in vehicle-road coordination is implemented.
根据本公开的另一方面,提供了一种路侧设备,包括上述电子设备。According to another aspect of the present disclosure, there is provided a roadside device, including the above-mentioned electronic device.
根据本公开的另一方面,提供了一种云控平台,包括上述电子设备。According to another aspect of the present disclosure, a cloud control platform is provided, including the above-mentioned electronic device.
由以上可见,应用本公开的实施例提供的方案进行目标检测时,首先根据候选目标区域间的交并比以及候选目标区域的被遮挡度,更新候选目标区域的置信度,然后基于更新后的置信度,从候选目标区域中检测图像中的目标。由于候选目标区域间的交并比能够反映各候选目标区域间的重合度,候选目标区域的被遮挡度能够反映候选目标区域被遮挡的程度,因此,根据上述交并比和遮挡度,更新候选目标区域的置信度时能够参考目标区域间的重叠情况,使得候选目标区域更新后的置信度更加趋向于实际情况,从而根据更新后的置信度对图像进行目标检测,能够提高目标检测的准确率。It can be seen from the above that when applying the scheme provided by the embodiments of the present disclosure for object detection, firstly, the confidence of the candidate object area is updated according to the intersection ratio between the candidate object areas and the degree of occlusion of the candidate object area, and then based on the updated Confidence, to detect an object in an image from candidate object regions. Since the intersection ratio between candidate object regions can reflect the degree of overlap between candidate object regions, and the occlusion degree of candidate object regions can reflect the degree of occlusion of candidate object regions, therefore, according to the above intersection ratio and occlusion degree, update the candidate The confidence of the target area can refer to the overlap between the target areas, so that the updated confidence of the candidate target area is closer to the actual situation, so that the image can be detected according to the updated confidence, which can improve the accuracy of target detection .
应当理解,本部分所描述的内容并非旨在标识本公开的实施例的关键或重要特征,也不用于限制本公开的范围。本公开的其它特征将通过以下的说明书而变得容易理解。It should be understood that what is described in this section is not intended to identify key or important features of the embodiments of the present disclosure, nor is it intended to limit the scope of the present disclosure. Other features of the present disclosure will be readily understood through the following description.
附图说明Description of drawings
附图用于更好地理解本方案,不构成对本公开的限定。其中:The accompanying drawings are used to better understand the present solution, and do not constitute a limitation to the present disclosure. in:
图1是根据本公开的实施例提供的一种车路协同中目标检测方法的流程示意图;FIG. 1 is a schematic flowchart of a method for detecting objects in vehicle-road coordination according to an embodiment of the present disclosure;
图2是根据本公开的实施例提供的一种图像示意图;FIG. 2 is a schematic diagram of an image provided according to an embodiment of the present disclosure;
图3a是根据本公开的实施例提供的一种网络模型的结构示意图;Fig. 3a is a schematic structural diagram of a network model provided according to an embodiment of the present disclosure;
图3b是根据本公开的实施例提供的另一种网络模型的结构示意图;Fig. 3b is a schematic structural diagram of another network model provided according to an embodiment of the present disclosure;
图4是根据本公开的实施例提供的一种车路协同中目标检测装置的结构示意图;Fig. 4 is a schematic structural diagram of an object detection device in vehicle-road coordination provided according to an embodiment of the present disclosure;
图5是根据本公开的实施例提供的一种电子设备的结构示意图。Fig. 5 is a schematic structural diagram of an electronic device provided according to an embodiment of the present disclosure.
具体实施方式detailed description
以下结合附图对本公开的示范性实施例做出说明,其中包括本公开实施例的各种细节以助于理解,应当将它们认为仅仅是示范性的。因此,本领域普通技术人员应当认识到,可以对这里描述的实施例做出各种改变和修改,而不会背离本公开的范围和精神。同样,为了清楚和简明,以下的描述中省略了对公知功能和结构的描述。Exemplary embodiments of the present disclosure are described below in conjunction with the accompanying drawings, which include various details of the embodiments of the present disclosure to facilitate understanding, and they should be regarded as exemplary only. Accordingly, those of ordinary skill in the art will recognize that various changes and modifications of the embodiments described herein can be made without departing from the scope and spirit of the disclosure. Also, descriptions of well-known functions and constructions are omitted in the following description for clarity and conciseness.
本公开实施例提供了一种车路协同中目标检测方法、装置和路侧设备。Embodiments of the present disclosure provide a method, device, and roadside equipment for object detection in vehicle-road coordination.
本公开的一个实施例中,提供了一种车路协同中目标检测方法,该方法包括:In one embodiment of the present disclosure, a method for object detection in vehicle-road coordination is provided, the method includes:
对图像进行目标检测,得到图像中的候选目标区域、候选目标区域的置信度以及候选目标区域的被遮挡度;Perform target detection on the image to obtain the candidate target area in the image, the confidence of the candidate target area, and the occlusion degree of the candidate target area;
基于候选目标区域间的交并比以及候选目标区域的被遮挡度,更新候选目标区域的置信度;Update the confidence of the candidate target area based on the intersection ratio between the candidate target areas and the occlusion degree of the candidate target area;
根据更新后的置信度,从候选目标区域中检测图像中的目标。Objects in the image are detected from candidate object regions based on the updated confidence.
由于候选目标区域间的交并比能够反映各候选目标区域间的重合度,候选目标区域的被遮挡度能够反映候选目标区域被遮挡的程度,因此,根据上述交并比和遮挡度,更新候选目标区域的置信度时能够参考目标区域间的重叠情况,使得候选目标区域更新后的置信度更加趋向于实际情况,从而根据更新后的置信度对图像进行目标检测,能够提高目标检测的准确率。Since the intersection ratio between candidate object regions can reflect the degree of overlap between candidate object regions, and the occlusion degree of candidate object regions can reflect the degree of occlusion of candidate object regions, therefore, according to the above intersection ratio and occlusion degree, update the candidate The confidence of the target area can refer to the overlap between the target areas, so that the updated confidence of the candidate target area is closer to the actual situation, so that the image can be detected according to the updated confidence, which can improve the accuracy of target detection .
以下对本公开实施例的执行主体进行说明。The subject of execution of the embodiments of the present disclosure will be described below.
本公开实施例的执行主体可以为集成有目标检测功能的电子设备,其中,上述电子设备可以是:台式机、笔记本电脑、服务器、图像采集设备等。其中,图像采集设备可以包括:摄像机、照相机、行车记录仪等。The execution subject of the embodiment of the present disclosure may be an electronic device integrated with a target detection function, wherein the above-mentioned electronic device may be: a desktop computer, a notebook computer, a server, an image acquisition device, and the like. Wherein, the image acquisition device may include: a video camera, a camera, a driving recorder, and the like.
本公开实施例提供的方案可以应用于对车路协同V2X的道路监控、车辆路径规划等应用场景下采集的图像进行目标检测。The solutions provided by the embodiments of the present disclosure can be applied to target detection of images collected in application scenarios such as vehicle-road coordination V2X road monitoring and vehicle path planning.
另外,本公开实施例提供的方案还可以用于对其他场景下采集的图像进行目标检测。例如,上述其他场景可以是地铁站、商场、演唱会等人员高度密集的场景,针对这种场景进行图像采集,所采集的图像中包含的人员往往也会密集,容易出现一些人员的面部被其他人员的面部遮挡的情况。上述场景还可以是博物馆入口处、银行大厅等人员较为密集的场景,针对这种场景进行图像采集,所采集的图像中可能会出现人员的面部被其他人员或者建筑物等遮挡的情况。In addition, the solutions provided by the embodiments of the present disclosure may also be used to perform target detection on images collected in other scenarios. For example, the above-mentioned other scenes may be scenes with a high density of people, such as subway stations, shopping malls, concerts, etc. If images are collected for such scenes, the people contained in the collected images are often dense, and it is easy for some people’s faces to be blocked by others. Situations where a person's face is occluded. The above-mentioned scene can also be a scene with relatively dense personnel such as the entrance of a museum, a bank lobby, etc. For image collection of such a scene, the face of the person may be blocked by other people or buildings in the collected image.
以上仅为本公开实施例的应用场景举例,并不对本公开构成限定。The above are only examples of application scenarios of the embodiments of the present disclosure, and do not limit the present disclosure.
上述目标可以是人脸、动物、车辆等等。The aforementioned objects may be human faces, animals, vehicles, and so on.
以下对本公开的实施例提供的车路协同中目标检测方法进行具体说明。The object detection method in the vehicle-road coordination provided by the embodiments of the present disclosure will be described in detail below.
参见图1,图1为本公开的实施例提供的一种车路协同中目标检测方法的流程示意图,上述方法包括以下步骤S101-步骤S103。Referring to FIG. 1 , FIG. 1 is a schematic flowchart of a method for object detection in vehicle-road coordination provided by an embodiment of the present disclosure. The above method includes the following steps S101 - S103.
步骤S101:对图像进行目标检测,得到图像中的候选目标区域、候选目标区域的置信度以及候选目标区域的被遮挡度。Step S101: Perform target detection on the image to obtain the candidate target area in the image, the confidence level of the candidate target area, and the occlusion degree of the candidate target area.
上述图像可以是针对具体场景进行图像采集得到的图像。上述场景可以包括车辆行驶场景、停车场场景等,这种情况下,上述目标可以为车辆;上述场景还可以包括 地铁站、高铁站等公共空间场景,这种情况下,上述目标可以为人。The foregoing images may be images acquired through image acquisition for a specific scene. The above-mentioned scenes can include vehicle driving scenes, parking lot scenes, etc. In this case, the above-mentioned objects can be vehicles; the above-mentioned scenes can also include public space scenes such as subway stations and high-speed rail stations. In this case, the above-mentioned objects can be people.
在进行目标检测时,一种实施方式中,可以采用预设的目标检测算法对图像进行目标检测,得到图像中的候选目标区域、候选目标区域的置信度以及候选目标区域的被遮挡度。When performing target detection, in one embodiment, a preset target detection algorithm may be used to perform target detection on an image to obtain a candidate target area in the image, a confidence degree of the candidate target area, and an occlusion degree of the candidate target area.
上述预设的目标检测算法可以是针对不同类型的目标所采用的检测算法。例如,当目标为人,可以采用人脸检测算法、人体检测算法等;当目标为车辆,可以采用车辆检测算法、车牌检测算法等。The aforementioned preset target detection algorithm may be a detection algorithm adopted for different types of targets. For example, when the target is a person, a face detection algorithm, a human body detection algorithm, etc. can be used; when the target is a vehicle, a vehicle detection algorithm, a license plate detection algorithm, etc. can be used.
进行目标检测的其他实施方式可以参见后续实施例,在此不进行详述。For other implementation manners of target detection, reference may be made to subsequent embodiments, and details are not described here.
候选目标区域是指:经目标检测认为可能存在目标的区域。以图2为例,图2中各矩形框围成的区域是对图像进行动物检测得到的候选目标区域。The candidate target area refers to the area where the target may exist after target detection. Taking FIG. 2 as an example, the area surrounded by each rectangular frame in FIG. 2 is a candidate target area obtained by performing animal detection on the image.
候选目标区域的置信度反映:候选目标区域中存在目标的可能性大小。上述置信度可以以小数、百分数等表示。置信度的取值越大,表示候选目标区域中存在目标的可能性越高。The confidence of the candidate target area reflects: the possibility of the existence of the target in the candidate target area. The confidence level above can be expressed in decimals, percentages, and the like. The larger the value of the confidence degree, the higher the probability that the target exists in the candidate target area.
例如:在目标为人的情况下,当候选目标区域A的置信度大于候选目标区域B的置信度,表示候选目标区域A中存在人的可能性高于候选目标区域B中存在人的可能性。For example, when the target is a person, when the confidence of the candidate target area A is greater than the confidence of the candidate target area B, it means that the possibility of a person in the candidate target area A is higher than the possibility of a person in the candidate target area B.
候选目标区域的被遮挡度反映:候选目标区域被遮挡的程度。上述被遮挡度可以以小数、百分数等表示,还可以以被遮挡等级序号表示,例如:被遮挡等级序号包括1、2、3,其中,序号为1可以表示被遮挡等级为严重遮挡,序号为2可以表示被遮挡等级为中度遮挡,序号为3可以表示被遮挡等级为轻度遮挡。The degree of occlusion of the candidate target area reflects: the degree of occlusion of the candidate target area. The above occlusion degree can be represented by decimals, percentages, etc., and can also be expressed by the occlusion level serial number, for example: the occlusion level serial number includes 1, 2, 3, wherein, the serial number 1 can indicate that the occlusion level is severe occlusion, and the serial number is 2 can indicate that the occlusion level is moderate occlusion, and the sequence number 3 can indicate that the occlusion level is light occlusion.
步骤S102:基于候选目标区域间的交并比以及候选目标区域的被遮挡度,更新候选目标区域的置信度。Step S102: Based on the intersection-over-union ratio between the candidate target areas and the degree of occlusion of the candidate target areas, update the confidence of the candidate target areas.
候选目标区域间的交并比用于描述两个候选目标区域之间的重合度。当交并比越高,表示两个候选目标区域之间的重合度越高;当交并比越低,表示两个候选目标区域之间的重合度越低。The intersection-over-union ratio between candidate object regions is used to describe the coincidence degree between two candidate object regions. When the intersection ratio is higher, it means that the overlap degree between the two candidate target regions is higher; when the intersection ratio is lower, it means that the overlap degree between the two candidate target regions is lower.
具体的,可以计算两个候选目标区域间的重叠面积,得到第一面积,计算两个候选目标区域的面积之和,得到第二面积,然后计算第二面积与第一面积之差,得到第三面积,将第一面积与第三面积之间的比值确定为候选目标区域间的交并比。Specifically, the overlapping area between two candidate target regions can be calculated to obtain the first area, the sum of the areas of the two candidate target regions can be calculated to obtain the second area, and then the difference between the second area and the first area can be calculated to obtain the second area Three areas, the ratio between the first area and the third area is determined as the intersection ratio between the candidate target areas.
例如:候选目标区域A的面积为48,候选目标区域B的面积为32,其中,候选目标区域A与候选目标区域B的重叠面积为16,也就是第一面积为16,候选目标区域A与候选目标区域B的总面积为(46+32)=80,也就是第二面积为80,计算第二面积与第一面积之差(80-16)=64,也就是第三面积为64,计算第一面积与第三面积之比得到16/64=0.25,0.25为候选目标区域间的交并比。For example: the area of the candidate target area A is 48, and the area of the candidate target area B is 32, wherein, the overlapping area of the candidate target area A and the candidate target area B is 16, that is, the first area is 16, and the candidate target area A and the candidate target area B are 16. The total area of the candidate target area B is (46+32)=80, that is, the second area is 80, and the difference between the second area and the first area is calculated (80-16)=64, that is, the third area is 64, The ratio of the first area to the third area is calculated to obtain 16/64=0.25, and 0.25 is the intersection and union ratio between the candidate target areas.
一种实现方式中,可以从各候选目标区域中选择一基准区域,针对各候选目标区 域中除基准区域外的每一其他候选目标区域,计算该其他候选目标区域与基准区域的交并比,将计算得到的交并比确定为用于更新该候选目标区域的置信度的交并比。In an implementation manner, a reference area may be selected from each candidate target area, and for every other candidate target area except the reference area in each candidate target area, the intersection ratio between the other candidate target area and the reference area is calculated, The calculated intersection and union ratio is determined as the intersection and union ratio for updating the confidence of the candidate target region.
上述基准区域可以是各候选目标区域中置信度最大的区域。The aforementioned reference region may be the region with the highest confidence among the candidate target regions.
另一种实现方式中,还可以针对每一候选目标区域,从该候选目标区域与其他各候选目标区域间的交并比中选择一交并比,将所选择的交并比确定为用于更新该候选目标区域的置信度的交并比。In another implementation manner, for each candidate target area, an intersection and union ratio may be selected from the intersection and union ratios between the candidate target area and other candidate target areas, and the selected intersection and union ratio may be determined as the The intersection and union ratio of the confidence of the candidate target region is updated.
例如:可以从上述多个交并比中选择最大交并比、平均交并比、中值交并比或最小交并比等。For example: the maximum intersection and union ratio, the average intersection and union ratio, the median intersection and union ratio, or the minimum intersection and union ratio may be selected from the above-mentioned multiple intersection and union ratios.
在更新候选目标区域的置信度时,一种实施方式中,可以根据候选目标区域间的交并比以及候选目标区域的被遮挡度,按照预设的第一权重和第二权重,计算调整系数,根据计算得到的调整系数,更新候选目标区域的置信度。When updating the confidence of the candidate target area, in one embodiment, the adjustment coefficient can be calculated according to the preset first weight and second weight according to the intersection ratio between the candidate target areas and the degree of occlusion of the candidate target area , according to the calculated adjustment coefficient, update the confidence of the candidate target region.
具体的,可以计算候选目标区域间的交并比与第一权重之间的乘积,并计算候选目标区域的被遮挡度与第二权重之间的乘积,将计算得到的两个乘积之和作为调整系数。Specifically, the product of the intersection ratio between candidate target areas and the first weight can be calculated, and the product of the occlusion degree of the candidate target area and the second weight can be calculated, and the sum of the two calculated products can be used as Adjustment coefficient.
例如:候选目标区域间的交并比为80%,候选目标区域的被遮挡度为50%,预设的第一权重为0.8,预设的第二权重为0.2,计算候选目标区域间的交并比与第一权重之间的乘积为:0.8*80%=64%,计算候选目标区域的被遮挡度与第二权重之间的乘积为:0.2*50%=10%,计算得到的两个乘积之和为:64%+10%=74%,从而得到调整系数为74%。For example: the intersection ratio between candidate target areas is 80%, the occlusion degree of candidate target areas is 50%, the preset first weight is 0.8, and the preset second weight is 0.2. Calculate the intersection of candidate target areas. The product between the combination ratio and the first weight is: 0.8*80%=64%, the product between the occlusion degree of the calculated candidate target area and the second weight is: 0.2*50%=10%, the calculated two The sum of the products is: 64%+10%=74%, thus the adjustment factor is 74%.
在计算得到调整系数后,可以计算调整系数与候选目标区域的置信度之间的乘积,作为候选目标区域更新后的置信度。After the adjustment coefficient is calculated, the product of the adjustment coefficient and the confidence of the candidate target area may be calculated as the updated confidence of the candidate target area.
更新候选目标区域的置信度的其他实施方式可以参见后续实施例,在此不进行详述。For other implementation manners of updating the confidence of the candidate target region, reference may be made to subsequent embodiments, and details are not described here.
步骤S103:根据更新后的置信度,从候选目标区域中检测图像中的目标。Step S103: Detect the target in the image from the candidate target regions according to the updated confidence.
本公开的一个实施例中,可以选择更新后的置信度大于预设置信度阈值的候选目标区域,将所选择候选目标区域中的目标确定为图像中的目标。In an embodiment of the present disclosure, a candidate target area whose updated confidence is greater than a preset confidence threshold may be selected, and the target in the selected candidate target area is determined as the target in the image.
上述预设置信度阈值可以由工作人员根据经验设定,例如:在置信度以百分数表示时,预设置信度阈值可以为90%,95%等。The aforementioned preset reliability threshold may be set by staff based on experience, for example: when the confidence is represented by a percentage, the preset reliability threshold may be 90%, 95% and so on.
以一个例子说明上述目标确定过程,假设更新后的各候选目标区域的置信度分别为:80%、70%、90%、95%,预设置信度阈值为85%,大于85%的更新后的置信度为90%、95%,其中,区域1更新后的置信度为90%,区域2更新后的置信度为95%,所以,区域1中的目标和区域2中的目标为图像中的目标。An example is used to illustrate the above target determination process, assuming that the confidence levels of the updated candidate target areas are: 80%, 70%, 90%, and 95%, respectively, and the preset confidence threshold is 85%. Confidence levels are 90% and 95%, among which, the confidence level after the update of area 1 is 90%, and the confidence level after update of area 2 is 95%, so the target in area 1 and the target in area 2 are The goal.
这样,对于置信度大于预设置信度阈值的候选目标区域来说,这些候选目标区域中包含目标的可能性高于其他候选目标区域包含目标的可能性。所以,将置信度大于 预设置信度阈值的候选目标区域中的目标确定为图像中的目标,所得到的目标的准确度较高。In this way, for candidate target regions whose confidence is greater than the preset confidence threshold, the probability that these candidate target regions contain a target is higher than the probability that other candidate target regions contain a target. Therefore, the target in the candidate target area whose confidence is greater than the preset confidence threshold is determined as the target in the image, and the accuracy of the obtained target is higher.
本公开的一个实施例中,还可以选择更新后的置信度最大的预设数量个候选目标区域,将所选择候选目标区域中的目标确定为图像中的目标。In an embodiment of the present disclosure, a preset number of candidate target areas with the highest updated confidence may also be selected, and the target in the selected candidate target areas is determined as the target in the image.
上述预设数量可以由工作人员根据经验设定,例如:上述预设数量可以为1个、3个、5个等。The above-mentioned preset number can be set by the staff based on experience, for example: the above-mentioned preset number can be 1, 3, 5, etc.
以一个例子说明上述目标确定过程。假设目标区域的置信度分别为:80%、70%、90%、95%,预设数量为3个,其中更新后的置信度最大的3个分别为95%、90%、80%;将更新后的置信度为95%、90%、80%的候选目标区域中的目标、确定为图像中的目标。An example is used to illustrate the above target determination process. Assume that the confidence levels of the target area are: 80%, 70%, 90%, and 95%, respectively, and the preset number is 3, among which the three with the highest updated confidence levels are 95%, 90%, and 80% respectively; Objects in the candidate object regions with updated confidence levels of 95%, 90%, and 80% are determined as objects in the image.
这样,对于置信度最大的预设数量个候选目标区域来说,这些候选目标区域中包含目标的可能性高于其他候选区域中包含目标的可能性。所以,将置信度最大的预设数量个候选目标区域中的目标确定为图像中的目标,所得到的目标的准确度较高。In this way, for the preset number of candidate target regions with the highest confidence, the possibility of containing the target in these candidate target regions is higher than the possibility of containing the target in other candidate regions. Therefore, the target in the preset number of candidate target areas with the highest confidence is determined as the target in the image, and the accuracy of the obtained target is relatively high.
由以上可见,应用本公开的实施例提供的方案进行目标检测时,首先根据候选目标区域间的交并比以及候选目标区域的被遮挡度,更新候选目标区域的置信度,然后基于更新后的置信度,从候选目标区域中检测图像中的目标。由于候选目标区域间的交并比能够反映各候选目标区域间的重合度,候选目标区域的被遮挡度能够反映候选目标区域被遮挡的程度,因此,根据上述交并比和被遮挡度,更新候选目标区域的置信度时能够参考候选目标区域间的重叠情况,使得候选目标区域更新后的置信度更加趋向于实际情况,从而根据更新后的置信度对图像进行目标检测,能够提高目标检测的准确率。It can be seen from the above that when applying the scheme provided by the embodiments of the present disclosure for object detection, firstly, the confidence of the candidate object area is updated according to the intersection ratio between the candidate object areas and the degree of occlusion of the candidate object area, and then based on the updated Confidence, to detect an object in an image from candidate object regions. Since the intersection ratio between candidate object regions can reflect the degree of overlap between candidate object regions, and the occlusion degree of candidate object regions can reflect the degree of occlusion of candidate object regions, therefore, according to the above intersection ratio and occlusion degree, update The confidence of the candidate target area can refer to the overlap between the candidate target areas, so that the updated confidence of the candidate target area is closer to the actual situation, so that the image can be detected according to the updated confidence, which can improve the accuracy of target detection. Accuracy.
另外,由于在密集场景中,如人流量密集、车流量密集场景,目标发生遮挡的情况尤其严重。对于这些密集场景的图像来说,各候选目标区域的被遮挡度较高,使得候选目标区域内目标不完整,得到的候选目标区域的置信度的误差较大。通过候选目标区域的被遮挡度去更新候选目标区域的置信度,可以有效消除各候选目标区域被遮挡时对置信度的误差影响,使得更新后的置信度准确度高,进而检测得到准确的目标。因此,本公开实施例提供的方案能够更好的适应于密集场景中存在遮挡的情况,提升了目标检测的准确度。In addition, in dense scenes, such as scenes with dense traffic of people and traffic, the occlusion of objects is especially serious. For the images of these dense scenes, the occlusion degree of each candidate target area is relatively high, so that the target in the candidate target area is incomplete, and the error of the confidence degree of the obtained candidate target area is relatively large. Update the confidence of the candidate target area by the occlusion degree of the candidate target area, which can effectively eliminate the influence of the error on the confidence of each candidate target area when it is occluded, so that the accuracy of the updated confidence is high, and then the detection is accurate. . Therefore, the solutions provided by the embodiments of the present disclosure can be better adapted to situations where occlusion exists in dense scenes, and improve the accuracy of object detection.
为准确对候选目标区域的置信度进行更新,本公开的一个实施例中,可以循环从区域集中选择置信度最高的第一区域,根据区域集中其他区域与第一区域间的交并比以及其他区域的被遮挡度,更新其他区域的置信度,至此,完成一次置信度更新操作,循环执行上述操作,直至区域集中包括一个区域。可以将每一次置信度更新操作称为一次循环。In order to accurately update the confidence degree of the candidate target region, in one embodiment of the present disclosure, the first region with the highest confidence degree can be cyclically selected from the region set, and according to the intersection ratio between other regions in the region set and the first region and other The degree of occlusion of the region is used to update the confidence of other regions. At this point, a confidence update operation is completed, and the above operations are performed cyclically until one region is included in the region set. Each confidence update operation can be called a cycle.
上述区域集包括:候选目标区域中未被选择过的区域。具体的,在第一次循环开 始时,区域集包括步骤S101中得到的各个候选目标区域;在每一次循环中,从区域集中选择第一区域后,区域集中不再包括被选择出的第一区域。The above region set includes: regions that have not been selected in the candidate target regions. Specifically, at the beginning of the first cycle, the region set includes each candidate target region obtained in step S101; in each cycle, after the first region is selected from the region set, the region set no longer includes the selected first region. area.
在第一次循环开始时,第一区域为:步骤S101中得到的各个候选目标区域中置信度最高的区域;在之后每一次循环时,第一区域为:上一次循环后得到的更新后各个区域中置信度最高的区域At the beginning of the first cycle, the first area is: the area with the highest confidence among the candidate target areas obtained in step S101; in each subsequent cycle, the first area is: the updated each target area obtained after the last cycle. The region with the highest confidence in the region
上述其他区域是指:区域集中除了第一区域外的区域。如区域集包括:区域1、区域2、区域3,其中,区域1为第一区域,除了第一区域外的区域为区域2、区域3,那么区域2、区域3为其他区域。The aforementioned other areas refer to areas in the area concentration except the first area. For example, the area set includes: area 1, area 2, and area 3, where area 1 is the first area, and areas other than the first area are area 2 and area 3, then area 2 and area 3 are other areas.
在每次循环时,可以遍历区域集中各个区域,对各个区域的置信度从高到低进行排序,将置信度最高的区域确定为第一区域。另外,还可以将第一区域保存至预测集中,随着循环次数的增多,预测集中保存的第一区域的数量也增多。In each cycle, each region in the region set may be traversed, the confidence of each region is sorted from high to low, and the region with the highest confidence is determined as the first region. In addition, the first region may also be stored in the prediction set, and as the number of cycles increases, the number of first regions stored in the prediction set also increases.
以下结合具体示例对上述循环过程进行说明。The above cycle process will be described below with reference to specific examples.
假设步骤S101中得到的各个候选目标区域为b1、b2、b3、……bn。Assume that each candidate target area obtained in step S101 is b1, b2, b3, ... bn.
在第一次循环开始时,区域集B={b1、b2、b3、……、bn}。其中,各候选目标区域中置信度最高的区域为区域b1,所以将区域b1作为第一区域。区域集B中除了区域b1之外的区域为b2、b3、……、bn,所以{b2、b3、……、bn}为其他区域。At the beginning of the first cycle, the set of regions B = {b1, b2, b3, . . . , bn}. Wherein, the area with the highest confidence in each candidate target area is the area b1, so the area b1 is taken as the first area. In the area set B, the areas other than the area b1 are b2, b3, ..., bn, so {b2, b3, ..., bn} are other areas.
根据其他区域{b2、b3、……、bn}与区域b1间的交并比以及其他区域{b2、b3、……、bn}的被遮挡度,更新其他区域{b2、b3、……、bn}的置信度。并且可以将第一区域b1添加至预测集D中,添加后的预测集D={b1};Update other areas {b2, b3, ..., bn} confidence. And the first region b1 can be added to the prediction set D, and the added prediction set D={b1};
第二次循环开始时,由于区域b1已被选择过作为第一区域,那么区域集B中不包括区域b1,区域集B={b2、b3、……、bn}。其中,更新后的{b2、b3、……、bn}中置信度最高的区域为区域b2,所以将区域b2作为第一区域。区域集B中除了区域b2之外的区域为b3、……、bn,所以{b3、……、bn}为其他区域。At the beginning of the second cycle, since the area b1 has been selected as the first area, the area b1 is not included in the area set B, and the area set B={b2, b3, . . . , bn}. Among them, the area with the highest confidence in the updated {b2, b3, ..., bn} is the area b2, so the area b2 is taken as the first area. In the area set B, the areas other than the area b2 are b3, ..., bn, so {b3, ..., bn} are other areas.
根据其他区域{b3、……、bn}与区域b2间的交并比以及其他区域{b3、……、bn}的被遮挡度,更新其他区域{b3、……、bn}的置信度。并且可以将第一区域b2添加至预测集D中,添加后的预测集D={b1,b2}。According to the intersection ratio between other regions {b3,...,bn} and region b2 and the occlusion degree of other regions {b3,...,bn}, update the confidence of other regions {b3,...,bn}. And the first region b2 can be added to the prediction set D, and the added prediction set D={b1, b2}.
第三次循环开始时,由于区域b1、区域b2已被选择过作为第一区域,那么区域集B={b3、……、bn}。其中,更新后的{b3、……、bn}中置信度最高的区域为区域b3,所以将区域b3作为第一区域。区域集B中除了区域b3之外的区域为b4、……、bn,所以{b4、……、bn}为其他区域。At the beginning of the third cycle, since the regions b1 and b2 have been selected as the first regions, then the region set B={b3,...,bn}. Among them, the area with the highest confidence in the updated {b3, ..., bn} is the area b3, so the area b3 is taken as the first area. In the area set B, the areas other than the area b3 are b4, ..., bn, so {b4, ..., bn} are other areas.
根据其他区域{b4、……、bn}与区域b3间的交并比以及其他区域{b4、……、bn}的被遮挡度,更新其他区域{b4、……、bn}的置信度。并且可以将第一区域b3添加至预测集D中,添加后的预测集D={b1,b2,b3}。According to the intersection ratio between other regions {b4,...,bn} and region b3 and the occlusion degree of other regions {b4,...,bn}, update the confidence of other regions {b4,...,bn}. And the first region b3 can be added to the prediction set D, and the added prediction set D={b1, b2, b3}.
按照类似的方式,直至区域集B中区域的数量为1,将区域集B中的唯一一个区 域直接添加至预测集D中,循环结束,得到更新后的各区域的置信度。In a similar manner, until the number of regions in region set B is 1, the only region in region set B is directly added to prediction set D, and the cycle ends to obtain the updated confidence of each region.
这样,在每次循环时根据区域集中其他区域与第一区域间的交并比以及其他区域的被遮挡度,更新区域集中区域的置信度。其中,其他区域的被遮挡度反映其他区域的被遮挡的程度,在区域被遮挡时检测得到的区域的置信度的准确度低,通过引入其他区域的被遮挡度,能够使得更新后候选目标区域的置信度的准确度高;并且其他区域与第一区域间的交并比反映其他区域与第一区域间的重合度,且第一区域是置信度最高的区域,通过与置信度最高的区域之间的重合度,也能够有效调整其他区域的置信度。所以,在每次循环时根据上述交并比以及被遮挡度,能够有效更新其他区域的置信度。并且,通过循环迭代更新过程,能够进一步提高更新后的置信度的准确度。In this way, in each cycle, the confidence of the region in the concentrated region is updated according to the intersection ratio between other regions in the concentrated region and the first region and the degree of occlusion of other regions. Among them, the degree of occlusion of other regions reflects the degree of occlusion of other regions. When the region is occluded, the accuracy of the confidence of the detected region is low. By introducing the degree of occlusion of other regions, the candidate target region after update can be The accuracy of the confidence degree is high; and the intersection ratio between other regions and the first region reflects the coincidence degree between other regions and the first region, and the first region is the region with the highest confidence, and the region with the highest confidence The degree of coincidence between them can also effectively adjust the confidence of other regions. Therefore, the confidence levels of other regions can be effectively updated according to the intersection ratio and the degree of occlusion in each cycle. Moreover, the accuracy of the updated confidence level can be further improved by looping and iterating the updating process.
本公开的一个实施例中,在每一次循环更新其他区域的置信度时,可以按照以下步骤A1-步骤A4实现。In an embodiment of the present disclosure, when updating the confidence levels of other regions in each cycle, it may be implemented according to the following steps A1-A4.
步骤A1:计算区域集中其他区域与第一区域间的交并比。Step A1: Calculate the intersection and union ratios between other areas in the area set and the first area.
具体的,首先计算其他区域与第一区域间的重叠面积,计算其他区域与第一区域的总面积;然后计算上述总面积与重叠面积之差,得到目标面积,最后将重叠面积与目标面积之间的比值确定为候选目标区域间的交并比。Specifically, first calculate the overlapping area between other areas and the first area, calculate the total area of other areas and the first area; then calculate the difference between the above total area and the overlapping area to obtain the target area, and finally calculate the difference between the overlapping area and the target area The ratio between them is determined as the intersection and union ratio between the candidate target regions.
步骤A2:根据交并比和预设的交并比阈值,确定第一置信度调节值。Step A2: Determine a first confidence adjustment value according to the intersection and union ratio and the preset intersection and union ratio threshold.
上述预设的交并比阈值可以由工作人员根据经验设定,例如:交并比阈值可以为90%、95%等。The above-mentioned preset intersection and union ratio thresholds may be set by staff based on experience, for example, the intersection and union ratio thresholds may be 90%, 95%, and so on.
一种实施方式中,可以判断交并比是否小于预设的交并比阈值,若为是,确定第一置信度调节值为第一预设值,若为否,确定第一置信度调节值为第二预设值。In one embodiment, it may be judged whether the intersection ratio is less than a preset intersection ratio threshold, if yes, determine the first confidence adjustment value as the first preset value, and if no, determine the first confidence adjustment value is the second preset value.
上述第一预设值和第二预设值均为工作人员根据经验设定的。Both the above-mentioned first preset value and the second preset value are set by the staff based on experience.
本公开的一个实施例中,还可以判断交并比是否小于预设的交并比阈值;若为是,确定第一置信度调节值为1;若为否,确定第一置信度调节值为:1与交并比之差。In an embodiment of the present disclosure, it may also be judged whether the intersection ratio is less than the preset intersection ratio threshold; if yes, determine the first confidence adjustment value is 1; if no, determine the first confidence adjustment value : The difference between 1 and the intersection and union ratio.
例如:假设预设的交并比阈值为90%,当其他区域与第一区域间的交并比为95%时,交并比95%大于预设的交并比阈值90%,确定第一置信度调节值:1-95%=5%;当其他区域与第一区域间的交并比为50%时,交并比50%小于预设的交并比阈值90%,确定第一置信度调节值:1。For example: Assuming that the preset intersection ratio threshold is 90%, when the intersection ratio between other regions and the first region is 95%, and the intersection ratio 95% is greater than the preset intersection ratio threshold 90%, determine the first Confidence adjustment value: 1-95% = 5%; when the intersection and union ratio between other areas and the first area is 50%, and the intersection and union ratio 50% is less than the preset intersection ratio threshold 90%, determine the first confidence Degree adjustment value: 1.
这样,当交并比小于预设的交并比阈值时,表示其他区域与第一区域间的重合度较小,说明其他区域中小部分图像内容被遮挡,检测得到的该其他区域的置信度的准确度高,在这种情况下,可以不对该其他区域的置信度进行调整。将第一置信度调节值设置为1,能够实现对区域的置信度不进行调整。当交并比不小于预设的交并比阈值时,表示其他区域与第一区域间的重合度较大,说明其他区域中大部分图像内容被遮挡,检测得到的该其他区域的置信度的准确度低,在这种情况下,需要对其他区域的置信度进行调整,将第一置信度调节值设置为1与交并比之差,可以使得调整后的 置信度趋近于实际情况。In this way, when the intersection and union ratio is less than the preset intersection and union ratio threshold, it means that the overlap between other areas and the first area is small, indicating that a small part of the image content in other areas is blocked, and the confidence of the other areas detected is The accuracy is high, in which case no adjustments to the confidences of the other regions may be made. Setting the first confidence adjustment value to 1 can realize that the confidence of the region is not adjusted. When the intersection ratio is not less than the preset intersection ratio threshold, it means that the overlap between other regions and the first region is relatively large, indicating that most of the image content in other regions is blocked, and the confidence of the other regions detected is The accuracy is low. In this case, the confidence of other regions needs to be adjusted. Setting the first confidence adjustment value to 1 and the difference between the intersection and union ratio can make the adjusted confidence approach the actual situation.
步骤A3:根据其他区域的被遮挡度,确定第二置信度调节值。Step A3: Determine a second confidence adjustment value according to the degree of occlusion of other areas.
一种实施方式中,可以计算其他区域的被遮挡度以及预设的调节系数之间的乘积,作为第二置信度调节值。In an implementation manner, the product of the degree of occlusion of other regions and a preset adjustment coefficient may be calculated as the second confidence adjustment value.
上述预设的调节系数可以由工作人员根据经验设定,例如:预设的调节系数可以为1.2、1.5等。The aforementioned preset adjustment coefficient can be set by the staff based on experience, for example: the preset adjustment coefficient can be 1.2, 1.5, etc.
本公开的一个实施例中,还可以按照以下表达式确定第二置信度调节值g(occ_pred):In an embodiment of the present disclosure, the second confidence adjustment value g(occ_pred) may also be determined according to the following expression:
g(occ_pred)=α occ_pred g(occ_pred)=α occ_pred
其中,occ_pred为其他区域的被遮挡度,α为预设的常数,α>1。Among them, occ_pred is the occlusion degree of other areas, α is a preset constant, α>1.
因为α>1,所以第二置信度调节值g(occ_pred)是随着其他区域的被遮挡度增加而增加的。Because α>1, the second confidence adjustment value g(occ_pred) increases as the occlusion degree of other regions increases.
由于区域的被遮挡度较高时,该区域的置信度的准确度低,所以需要对该区域的置信度进行大幅度调整,使得调整后的置信度趋近于实际情况。又由于第二置信度调节值g(occ_pred)是随着其他区域的被遮挡度增加而增加的,也就是其他区域的被遮挡度越高,第二置信度调节值越大,从而能够对该其他区域的置信度进行大幅度调整,使得调整后的其他区域的置信度趋近于实际情况。Since the accuracy of the confidence of the region is low when the occlusion degree of the region is high, it is necessary to make a large adjustment to the confidence of the region so that the adjusted confidence is close to the actual situation. And because the second confidence degree adjustment value g(occ_pred) increases as the degree of occlusion of other regions increases, that is, the higher the degree of occlusion of other regions, the larger the second degree of confidence adjustment value. The confidence of other regions is greatly adjusted, so that the adjusted confidence of other regions is close to the actual situation.
步骤A4:采用第一置信度调节值和第二置信度调节值,调节其他区域的置信度。Step A4: Using the first confidence adjustment value and the second confidence adjustment value, adjust the confidence of other regions.
本公开的一个实施例中,可以按照以下表达式,调节其他区域的置信度:In an embodiment of the present disclosure, the confidence of other regions can be adjusted according to the following expression:
S’=S*T1*T2S'=S*T1*T2
其中,S’表示调节后的其他区域的置信度,S表示调节前的其他区域的置信度,T1表示第一置信度调节值,T2表示第二置信度调节值。Wherein, S' represents the confidence degree of other regions after adjustment, S represents the confidence degree of other regions before adjustment, T1 represents the first confidence degree adjustment value, and T2 represents the second confidence degree adjustment value.
这样,由于调整后的置信度是第一置信度调节值、第二置信度调节值以及其他区域的置信度之间的乘积,又由于第一置信度调节值、第二置信度调节值从不同角度反映其他区域的被遮挡情况,因此,上述调整后的置信度参考了其他区域被遮挡的情况,使得调节后的置信度更加趋近于实际情况。In this way, since the adjusted confidence level is the product of the first confidence level adjustment value, the second confidence level adjustment value and the confidence levels of other regions, and because the first confidence level adjustment value and the second confidence level adjustment value are different from The angle reflects the occlusion situation of other regions. Therefore, the above adjusted confidence refers to the occlusion situation of other regions, making the adjusted confidence closer to the actual situation.
另一种实施方式中,还可以计算第一置信度调节值、第二置信度调节值以及其他区域的置信度之间的乘积,作为参考置信度,通过预设的置信度误差值调整上述参考置信度,得到调整后的参考置信度,作为其他区域调节后的置信度。In another embodiment, the product of the first confidence adjustment value, the second confidence adjustment value, and the confidence of other regions can also be calculated as the reference confidence, and the above reference can be adjusted by the preset confidence error value. Confidence, the adjusted reference confidence is obtained as the adjusted confidence of other regions.
具体的,可以计算预设的置信度误差值与参数置信度之间的乘积,将计算得到的乘积确定为其他区域调节后的置信度。Specifically, the product of the preset confidence error value and the parameter confidence may be calculated, and the calculated product may be determined as the adjusted confidence of other regions.
这样,由于第一置信度调节值是通过其他区域与第一区域间的交并比确定的,交并比反映其他区域与第一区域重合度,且第二置信度调节值是根据其他区域的被遮挡度确定的,被遮挡度反映其他区域的被遮挡程度,第一置信度调节值和第二置信度调 节值从不同角度均能反映其他区域的被遮挡情况。所以,采用第一置信度调节值、第二置信度调节值调节其他区域的置信度时,由于第一置信度调节值、第二置信度调节值从不同角度反映其他区域的被遮挡情况,采用第一置信度调节值和第二置信度调节值调节时,基于较为准确的其他区域的被遮挡情况对置信度进行调节,使得调节后的置信度更加趋近于实际情况。In this way, since the first confidence adjustment value is determined by the intersection and union ratio between other regions and the first region, the intersection and union ratio reflects the coincidence degree of other regions and the first region, and the second confidence adjustment value is based on the intersection and union ratio of other regions The degree of occlusion is determined, and the degree of occlusion reflects the degree of occlusion of other regions, and the first and second confidence adjustment values can reflect the occlusion of other regions from different angles. Therefore, when using the first confidence adjustment value and the second confidence adjustment value to adjust the confidence of other areas, since the first confidence adjustment value and the second confidence adjustment value reflect the occlusion of other areas from different angles, use When adjusting the first confidence level adjustment value and the second confidence level adjustment value, the confidence level is adjusted based on more accurate occlusion conditions of other regions, so that the adjusted confidence level is closer to the actual situation.
以下以一个具体实现过程对上述循环更新置信度的方式进行说明。Hereinafter, a specific implementation process will be used to illustrate the above-mentioned manner of cyclically updating the confidence level.
假设,各候选目标区域为b1、b2、b3,预设的交并比阈值Nt为90%,各候选目标区域的置信度、被遮挡度如下表1所示。Assume that each candidate target area is b1, b2, b3, and the preset intersection-over-union ratio threshold Nt is 90%, the confidence and occlusion degree of each candidate target area are shown in Table 1 below.
表1Table 1
候选目标区域candidate target area 置信度Confidence 被遮挡度degree of occlusion
区域b1area b1 Cv1Cv1 Co1Co1
区域b2area b2 Cv2Cv2 Co2Co2
区域b3area b3 Cv3Cv3 Co3Co3
第一次循环开始时,区域集B={b1、b2、b3},其中,区域b1的置信度Cv1最高,区域b1为第一区域,区域b2、区域b3为其他区域。At the beginning of the first cycle, the region set B={b1, b2, b3}, wherein the confidence Cv1 of the region b1 is the highest, the region b1 is the first region, and the regions b2 and b3 are other regions.
针对区域b2,计算区域b2与区域b1之间的交并比,根据上述交并比与预设的交并比阈值90%,确定第一置信度调节值;并根据区域b2的被遮挡度Co2,确定第二置信度调节值;根据第一置信度调节值和第二置信度调节值,调节区域b2的置信度Cv2,更新后的置信度为Cv21。For the area b2, calculate the intersection ratio between the area b2 and the area b1, and determine the first confidence adjustment value according to the above intersection ratio and the preset intersection ratio threshold of 90%; and according to the occlusion degree Co2 of the area b2 , determine the second confidence adjustment value; according to the first confidence adjustment value and the second confidence adjustment value, adjust the confidence Cv2 of the region b2, and the updated confidence is Cv21.
针对区域b3,计算区域b3与区域b1之间的交并比,根据上述交并比与预设的交并比阈值90%,确定第一置信度调节值;并根据区域b3的被遮挡度Co3,确定第二置信度调节值;根据第一置信度调节值和第二置信度调节值,调节区域b3的置信度Cv3,更新后的置信度为Cv31。For area b3, calculate the intersection ratio between area b3 and area b1, and determine the first confidence adjustment value according to the above intersection ratio and the preset intersection ratio threshold of 90%; and according to the occlusion degree Co3 of area b3 , determine the second confidence adjustment value; according to the first confidence adjustment value and the second confidence adjustment value, adjust the confidence Cv3 of the region b3, and the updated confidence is Cv31.
第一次循环得到的更新后的各候选目标区域的置信度如下表2所示。The confidences of the updated candidate target regions obtained in the first cycle are shown in Table 2 below.
表2Table 2
候选目标区域candidate target area 置信度Confidence
区域b1area b1 Cv1Cv1
区域b2area b2 Cv21Cv21
区域b3area b3 Cv31Cv31
第二次循环开始时,由于区域b1已被选择过,区域集B={b2、b3},其中,区域b2的置信度Cv21最高,区域b2为第一区域,区域b3为其他区域。At the beginning of the second cycle, since the region b1 has been selected, the region set B={b2, b3}, wherein the confidence Cv21 of the region b2 is the highest, the region b2 is the first region, and the region b3 is other regions.
针对区域b3,计算区域b3与区域b2之间的交并比,根据上述交并比与预设的交并比阈值90%,确定第一置信度调节值;并根据区域b3的被遮挡度Co3,确定第二 置信度调节值;根据第一置信度调节值和第二置信度调节值,调节区域b3的置信度Cv31,更新后的置信度为Cv311。For the area b3, calculate the intersection ratio between the area b3 and the area b2, and determine the first confidence adjustment value according to the above intersection ratio and the preset intersection ratio threshold of 90%; and according to the occlusion degree Co3 of the area b3 , determine the second confidence adjustment value; according to the first confidence adjustment value and the second confidence adjustment value, adjust the confidence Cv31 of the region b3, and the updated confidence is Cv311.
由于区域b1、区域b3已被选择过,区域集B={b3},包含一个区域,循环结束。Since the regions b1 and b3 have been selected, the region set B={b3} contains one region, and the loop ends.
最终得到的更新后的各候选目标区域的置信度如下表3所示。The finally updated confidence of each candidate target region is shown in Table 3 below.
表3table 3
候选目标区域candidate target area 置信度Confidence
区域b1area b1 Cv1Cv1
区域b2area b2 Cv21Cv21
区域b3area b3 Cv311Cv311
本公开的一个实施例中,在上述步骤S101中可以针对不同的目标尺度,对图像进行目标检测,得到图像中不同尺度的候选目标区域、候选目标区域的置信度以及候选目标区域的被遮挡度。In one embodiment of the present disclosure, in the above step S101, target detection can be performed on the image for different target scales, and the candidate target areas of different scales in the image, the confidence of the candidate target areas, and the occlusion degree of the candidate target areas can be obtained .
目标尺度是指:目标的尺寸大小。The target scale refers to: the size of the target.
目标尺度可以是预先设定的尺度值,如目标尺度可以为16x16、32x32、64x64。The target scale may be a preset scale value, for example, the target scale may be 16x16, 32x32, or 64x64.
具体的,可以对图像进行多层特征提取,然后对不同的特征进行特征融合,得到不同尺度的特征。采用不同尺度的特征对图像进行目标检测,得到不同尺度的候选目标区域,并得到不同尺度的候选目标区域的置信度以及被遮挡度。Specifically, multi-layer feature extraction can be performed on the image, and then feature fusion is performed on different features to obtain features of different scales. The features of different scales are used to detect the target of the image, and the candidate target areas of different scales are obtained, and the confidence and occlusion degree of the candidate target areas of different scales are obtained.
这样,由于不同尺度的候选目标区域包含的图像特征信息不同,通过得到图像中不同尺度的候选目标区域,丰富了候选目标区域在不同尺度上的特征信息。In this way, since the image feature information contained in the candidate object regions of different scales is different, by obtaining the candidate object regions of different scales in the image, the feature information of the candidate object regions at different scales is enriched.
本公开的一个实施例中,可以将图像输入预先训练得到的目标检测模型,获得目标检测模型输出的图像中的候选目标区域、候选目标区域的置信度以及候选目标区域的被遮挡度。In one embodiment of the present disclosure, an image may be input into a pre-trained target detection model, and the candidate target areas in the image output by the target detection model, the confidence of the candidate target areas, and the occlusion degree of the candidate target areas may be obtained.
上述目标检测模型包括:用于检测图像中候选目标区域的目标检测层和用于预测候选目标区域被遮挡度的遮挡度预测层。The above object detection model includes: an object detection layer for detecting candidate object areas in an image, and an occlusion degree prediction layer for predicting the degree of occlusion of the candidate object areas.
一种实现方式中,目标检测层除了检测图像中候选目标区域,还可以计算候选目标区域的置信度,在这种情况下,目标检测模型的网络结构可以如图3a所示,目标检测模型包括目标检测层和遮挡度预测层。In one implementation, in addition to detecting the candidate target area in the image, the target detection layer can also calculate the confidence of the candidate target area. In this case, the network structure of the target detection model can be shown in Figure 3a. The target detection model includes Object detection layer and occlusion prediction layer.
具体的,图像在输入至目标检测模型后,模型中的目标检测层检测图像中的候选目标区域,并计算候选目标区域的置信度,将检测结果传输至遮挡度预测层;遮挡度预测层预测各个候选目标区域被遮挡度;目标检测模型输出候选目标区域、候选目标区域的置信度以及被遮挡度。Specifically, after the image is input to the target detection model, the target detection layer in the model detects the candidate target area in the image, calculates the confidence of the candidate target area, and transmits the detection result to the occlusion degree prediction layer; the occlusion degree prediction layer predicts Each candidate target area is occluded; the target detection model outputs the candidate target area, the confidence of the candidate target area, and the occlusion degree.
由前述实施例得知,在对图像进行目标检测时,可以针对每一目标尺度,得到不同尺度的候选目标区域、候选目标区域的置信度以及候选目标区域的被遮挡度。在这种情况下,可以在上述网络模型的基础上增加FPN(Feature Pyramid Networks,特征 金字塔网络),FPN用于得到各种不同尺度的候选目标区域、候选目标区域的置信度以及被遮挡度。It can be known from the foregoing embodiments that when performing object detection on an image, for each object scale, candidate object regions of different scales, confidence levels of the candidate object regions, and occlusion degrees of the candidate object regions can be obtained. In this case, FPN (Feature Pyramid Networks, Feature Pyramid Network) can be added on the basis of the above network model. FPN is used to obtain candidate target areas of various scales, confidence and occlusion of candidate target areas.
增加FPN后的网络模型的网络结构可以如图3b所示,图3b所示的网络结构包括骨干网络(Backbone)和FPN。The network structure of the network model after adding the FPN may be shown in FIG. 3b, and the network structure shown in FIG. 3b includes a backbone network (Backbone) and an FPN.
其中,骨干网络用于对图像进行特征提取,得到图像中不同层级的图像特征,并将不同层级的图像特征输入至FPN中。Among them, the backbone network is used to extract the features of the image, obtain the image features of different levels in the image, and input the image features of different levels into the FPN.
例如:骨干网络为卷积神经网络时,卷积神经网络的每一层卷积层都可以对图像进行卷积操作,从而得到不同层级的图像特征。For example, when the backbone network is a convolutional neural network, each convolutional layer of the convolutional neural network can perform convolution operations on images to obtain image features at different levels.
FPN用于对不同层级的图像特征进行特征融合,得到不同尺度的图像特征,基于不同尺度的图像特征进行目标检测,得到不同尺度的候选目标区域,并得到候选目标区域的置信度以及被遮挡度,实现了对不同层级的图像特征进行分治处理。FPN is used to perform feature fusion of image features at different levels to obtain image features of different scales, perform target detection based on image features of different scales, obtain candidate target areas of different scales, and obtain the confidence and occlusion degree of candidate target areas , realizing the divide-and-conquer processing of image features at different levels.
在训练目标检测模型时,将样本图像作为训练样本,并以样本图像中的真实候选目标区域以及真实被遮挡度作为训练标注,对预设的神经网络模型进行训练,直至满足训练结束条件,得到训练完毕的目标检测模型。When training the target detection model, the sample image is used as the training sample, and the real candidate target area and the real occlusion degree in the sample image are used as the training labels to train the preset neural network model until the training end condition is met, and the obtained The trained object detection model.
上述预设的神经网络模型可以为CNN(Conv Neural Network,卷积神经网络)模型、RNN(Recurrent Neural Network,递归神经网络)模型、DNN(Deep Neural Network,深度神经网络)模型等。The aforementioned preset neural network model may be a CNN (Conv Neural Network, convolutional neural network) model, an RNN (Recurrent Neural Network, recursive neural network) model, a DNN (Deep Neural Network, deep neural network) model, etc.
具体的,样本图像在输入至预设的神经网络模型后,上述预设的神经网络模型对样本图像进行目标检测,得到样本图像的候选目标区域以及被遮挡度,计算候选目标区域与真实目标区域间的差异、以及候选目标区域的被遮挡度与真实被遮挡度间的差异,根据计算得到的差异调整神经网络模型的参数,不断迭代调整参数,直至满足预设的训练结束条件。Specifically, after the sample image is input to the preset neural network model, the above preset neural network model performs target detection on the sample image, obtains the candidate target area and the degree of occlusion of the sample image, and calculates the candidate target area and the real target area and the difference between the occlusion degree of the candidate target area and the real occlusion degree, adjust the parameters of the neural network model according to the calculated difference, and iteratively adjust the parameters until the preset training end conditions are met.
上述训练结束条件可以为训练次数到达预设次数、模型参数满足预设的模型参数收敛条件等。The aforementioned training end conditions may be that the number of training times reaches a preset number of times, the model parameters meet the preset model parameter convergence conditions, and the like.
由于目标检测模型是通过大量训练样本训练得到的,训练过程中,目标检测模型学习到了样本图像中目标区域的特征以及被遮挡的特征,因此,目标检测模型具有较强鲁棒性,从而采用目标检测模型对图像进行目标检测时,能够输出准确的候选目标区域、候选目标区域的置信度以及被遮挡度。Since the target detection model is trained through a large number of training samples, during the training process, the target detection model learns the features of the target area and the occluded features in the sample image, therefore, the target detection model has strong robustness, so the target detection model is adopted. When the detection model performs target detection on an image, it can output accurate candidate target areas, confidence levels of candidate target areas, and occlusion degrees.
在上述步骤S101中,除了采用目标检测模型对图像进行目标检测之外,还可以将图像划分为多个区域,针对每一区域,提取该区域中的图像特征,根据图像特征确定区域中候选目标区域。In the above step S101, in addition to using the target detection model to perform target detection on the image, the image can also be divided into multiple regions, and for each region, the image features in the region are extracted, and the candidate targets in the region are determined according to the image features area.
上述图像特征包括:纹理特征、颜色特征、边缘特征等。The aforementioned image features include: texture features, color features, edge features, and the like.
在得到各候选目标区域后,根据各候选目标区域的图像特征,预测各候选目标的置信度。After each candidate target area is obtained, the confidence of each candidate target is predicted according to the image features of each candidate target area.
并且,还可以根据各候选目标区域所属图层以及位置信息,计算各候选目标区域的被遮挡度。In addition, the degree of occlusion of each candidate target area may also be calculated according to the layer to which each candidate target area belongs and the location information.
具体的,可以根据候选目标区域所属图层以及位置间的相对关系,确定候选目标区域之间是否发生遮挡,计算被遮挡的面积与被遮挡的区域的面积之间的比值,作为候选目标区域的被遮挡度。Specifically, according to the layer to which the candidate target areas belong and the relative relationship between positions, it can be determined whether occlusion occurs between the candidate target areas, and the ratio between the occluded area and the area of the occluded area can be calculated as the candidate target area degree of occlusion.
例如:当候选目标区域A位于前景图层、候选目标区域B位于背景图层,且候选目标区域A与候选目标区域B的位置信息之间发生重合,可以确定候选目标区域B被遮挡,计算候选目标区域B的被遮挡面积与候选目标区域B的面积之比,作为候选目标区域B的被遮挡度。For example: when the candidate target area A is located in the foreground layer and the candidate target area B is located in the background layer, and the position information of the candidate target area A and the candidate target area B overlaps, it can be determined that the candidate target area B is blocked, and the candidate target area B is calculated. The ratio of the shaded area of the target area B to the area of the candidate target area B is used as the shaded degree of the candidate target area B.
与上述车路协同中目标检测方法相对应,本公开的实施例还提供了一种车路协同中目标检测装置。Corresponding to the above object detection method in vehicle-road coordination, an embodiment of the present disclosure further provides a device for detecting objects in vehicle-road coordination.
参见图4,图4为本公开的实施例提供的一种车路协同中目标检测装置的结构示意图,上述装置包括以下模块401-403。Referring to FIG. 4 , FIG. 4 is a schematic structural diagram of an object detection device in vehicle-road coordination provided by an embodiment of the present disclosure. The above-mentioned device includes the following modules 401-403.
信息获得模块401,用于对图像进行目标检测,得到所述图像中的候选目标区域、候选目标区域的置信度以及候选目标区域的被遮挡度;An information obtaining module 401, configured to perform target detection on an image, and obtain a candidate target area in the image, a confidence degree of the candidate target area, and an occlusion degree of the candidate target area;
置信度更新模块402,用于基于候选目标区域间的交并比以及候选目标区域的被遮挡度,更新候选目标区域的置信度;A confidence update module 402, configured to update the confidence of the candidate target region based on the intersection-over-union ratio between the candidate target regions and the degree of occlusion of the candidate target region;
目标检测模块403,用于根据更新后的置信度,从候选目标区域中检测所述图像中的目标。The target detection module 403 is configured to detect the target in the image from the candidate target regions according to the updated confidence.
由以上可见,应用本公开的实施例提供的方案进行目标检测时,首先根据候选目标区域间的交并比以及候选目标区域的被遮挡度,更新候选目标区域的置信度,然后基于更新后的置信度,从候选目标区域中检测图像中的目标。由于候选目标区域间的交并比能够反映各候选目标区域间的重合度,候选目标区域的被遮挡度能够反映候选目标区域被遮挡的程度,因此,根据上述交并比和被遮挡度,更新候选目标区域的置信度时能够参考候选目标区域间的重叠情况,使得候选目标区域更新后的置信度更加趋向于实际情况。所以,从而根据更新后的置信度对图像进行目标检测,能够提高目标检测的准确率。It can be seen from the above that when applying the scheme provided by the embodiments of the present disclosure for object detection, firstly, the confidence of the candidate object area is updated according to the intersection ratio between the candidate object areas and the degree of occlusion of the candidate object area, and then based on the updated Confidence, to detect an object in an image from candidate object regions. Since the intersection ratio between candidate object regions can reflect the degree of overlap between candidate object regions, and the occlusion degree of candidate object regions can reflect the degree of occlusion of candidate object regions, therefore, according to the above intersection ratio and occlusion degree, update The confidence of the candidate target area can refer to the overlap between the candidate target areas, so that the updated confidence of the candidate target area is more inclined to the actual situation. Therefore, the target detection is performed on the image according to the updated confidence level, which can improve the accuracy of target detection.
本公开的一个实施例中,所述置信度更新模块402具体用于循环从区域集中选择置信度最高的第一区域,根据区域集中其他区域与所述第一区域间的交并比以及其他区域的被遮挡度,更新其他区域的置信度,直至所述区域集中包括一个区域,其中,所述区域集包括:候选目标区域中未被选择过的区域。In an embodiment of the present disclosure, the confidence update module 402 is specifically configured to cyclically select the first region with the highest confidence from the region set, and according to the intersection ratio between other regions in the region set and the first region and other regions The degree of occlusion of other areas is updated until the set of areas includes an area, wherein the set of areas includes: an unselected area in the candidate target area.
这样,在每次循环时根据区域集中其他区域与第一区域间的交并比以及其他区域的被遮挡度,更新区域集中区域的置信度。其中,其他区域的被遮挡度反映其他区域 的被遮挡的程度,在区域被遮挡时检测得到的区域的置信度的准确度低,通过引入其他区域的被遮挡度,能够使得更新后候选目标区域的置信度的准确度高;并且其他区域与第一区域间的交并比反映其他区域与第一区域间的重合度,且第一区域是置信度最高的区域,通过与置信度最高的区域之间的重合度,也能够有效调整其他区域的置信度。所以,在每次循环时根据上述交并比以及被遮挡度,能够有效更新其他区域的置信度。并且,通过循环迭代更新过程,能够进一步提高更新后的置信度的准确度。In this way, in each cycle, the confidence of the region in the concentrated region is updated according to the intersection ratio between other regions in the concentrated region and the first region and the degree of occlusion of other regions. Among them, the degree of occlusion of other regions reflects the degree of occlusion of other regions. When the region is occluded, the accuracy of the confidence of the detected region is low. By introducing the degree of occlusion of other regions, the candidate target region after update can be The accuracy of the confidence degree is high; and the intersection ratio between other regions and the first region reflects the coincidence degree between other regions and the first region, and the first region is the region with the highest confidence, and the region with the highest confidence The degree of coincidence between them can also effectively adjust the confidence of other regions. Therefore, the confidence levels of other regions can be effectively updated according to the intersection ratio and the degree of occlusion in each cycle. Moreover, the accuracy of the updated confidence level can be further improved by looping and iterating the updating process.
本公开的一个实施例中,所述置信度更新模块402,包括:In an embodiment of the present disclosure, the confidence update module 402 includes:
交并比计算单元,用于计算区域集中其他区域与所述第一区域间的交并比;an intersection ratio calculation unit, configured to calculate intersection ratios between other regions in the region set and the first region;
第一调节值确定单元,用于根据所述交并比和预设的交并比阈值,确定第一置信度调节值;A first adjustment value determination unit, configured to determine a first confidence adjustment value according to the intersection ratio and a preset intersection ratio threshold;
第二调节值确定单元,用于根据其他区域的被遮挡度,确定第二置信度调节值;A second adjustment value determination unit, configured to determine a second confidence adjustment value according to the degree of occlusion of other areas;
置信度调节单元,用于采用所述第一置信度调节值和第二置信度调节值,调节其他区域的置信度。A confidence adjustment unit, configured to adjust the confidence of other regions by using the first confidence adjustment value and the second confidence adjustment value.
这样,由于第一置信度调节值是通过其他区域与第一区域间的交并比确定的,交并比反映其他区域与第一区域重合度,且第二置信度调节值是根据其他区域的被遮挡度确定的,被遮挡度反映其他区域的被遮挡程度,第一置信度调节值和第二置信度调节值从不同角度均能反映其他区域的被遮挡情况。所以,采用第一置信度调节值、第二置信度调节值调节其他区域的置信度时,由于第一置信度调节值、第二置信度调节值从不同角度反映其他区域的被遮挡情况,采用第一置信度调节值和第二置信度调节值调节时,基于较为准确的其他区域的被遮挡情况对置信度进行调节,使得调节后的置信度更加趋近于实际情况。In this way, since the first confidence adjustment value is determined by the intersection and union ratio between other regions and the first region, the intersection and union ratio reflects the coincidence degree of other regions and the first region, and the second confidence adjustment value is based on the intersection and union ratio of other regions The degree of occlusion is determined, and the degree of occlusion reflects the degree of occlusion of other regions, and the first and second confidence adjustment values can reflect the occlusion of other regions from different angles. Therefore, when using the first confidence adjustment value and the second confidence adjustment value to adjust the confidence of other areas, since the first confidence adjustment value and the second confidence adjustment value reflect the occlusion of other areas from different angles, use When adjusting the first confidence level adjustment value and the second confidence level adjustment value, the confidence level is adjusted based on more accurate occlusion conditions of other regions, so that the adjusted confidence level is closer to the actual situation.
本公开的一个实施例中,所述第一调节值确定单元,具体用于判断所述交并比是否小于预设的交并比阈值;若为是,确定第一置信度调节值为1;若为否,确定所述第一置信度调节值为:1与所述交并比之差。In an embodiment of the present disclosure, the first adjustment value determination unit is specifically configured to judge whether the intersection-over-union ratio is smaller than a preset intersection-over-union ratio threshold; if yes, determine a first confidence adjustment value of 1; If not, determine that the adjusted value of the first confidence level is the difference between 1 and the intersection-over-union ratio.
这样,当交并比小于预设的交并比阈值时,表示其他区域与第一区域间的重合度较小,说明其他区域中小部分图像内容被遮挡,检测得到的该其他区域的置信度的准确度高,在这种情况下,可以不对该其他区域的置信度进行调整。将第一置信度调节值设置为1,能够实现对区域的置信度不进行调整。当交并比不小于预设的交并比阈值时,表示其他区域与第一区域间的重合度较大,说明其他区域中大部分图像内容被遮挡,检测得到的该其他区域的置信度的准确度低,在这种情况下,需要对其他区域的置信度进行调整,将第一置信度调节值设置为1与交并比之差,可以使得调整后的置信度趋近于实际情况。In this way, when the intersection and union ratio is less than the preset intersection and union ratio threshold, it means that the overlap between other areas and the first area is small, indicating that a small part of the image content in other areas is blocked, and the confidence of the other areas detected is The accuracy is high, in which case no adjustments to the confidences of the other regions may be made. Setting the first confidence adjustment value to 1 can realize that the confidence of the region is not adjusted. When the intersection ratio is not less than the preset intersection ratio threshold, it means that the overlap between other regions and the first region is relatively large, indicating that most of the image content in other regions is blocked, and the confidence of the other regions detected is The accuracy is low. In this case, the confidence of other regions needs to be adjusted. Setting the first confidence adjustment value to 1 and the difference between the intersection and union ratio can make the adjusted confidence approach the actual situation.
本公开的一个实施例中,所述第二调节值确定单元,具体用于按照以下表达式确定第二置信度调节值g(occ_pred):In an embodiment of the present disclosure, the second adjustment value determination unit is specifically configured to determine the second confidence adjustment value g(occ_pred) according to the following expression:
g(occ_pred)=α occ_pred g(occ_pred)=α occ_pred
其中,occ_pred为其他区域的被遮挡度,α为预设的常数,α>1。Among them, occ_pred is the occlusion degree of other areas, α is a preset constant, α>1.
由于区域的被遮挡度较高时,该区域的置信度的准确度低,所以需要对该区域的置信度进行大幅度调整,使得调整后的置信度趋近于实际情况。又由于第二置信度调节值g(occ_pred)是随着其他区域的被遮挡度增加而增加的,也就是其他区域的被遮挡度越高,第二置信度调节值越大,从而能够对该其他区域的置信度进行大幅度调整,使得调整后的其他区域的置信度趋近于实际情况。Since the accuracy of the confidence of the region is low when the occlusion degree of the region is high, it is necessary to make a large adjustment to the confidence of the region so that the adjusted confidence is close to the actual situation. And because the second confidence degree adjustment value g(occ_pred) increases as the degree of occlusion of other regions increases, that is, the higher the degree of occlusion of other regions, the larger the second degree of confidence adjustment value. The confidence of other regions is greatly adjusted, so that the adjusted confidence of other regions is close to the actual situation.
本公开的一个实施例中,所述置信度调节单元,具体用于按照以下表达式,调节其他区域的置信度:In an embodiment of the present disclosure, the confidence adjustment unit is specifically configured to adjust the confidence of other regions according to the following expression:
S’=S*T1*T2S'=S*T1*T2
其中,S’表示调节后的其他区域的置信度,S表示调节前的其他区域的置信度,T1表示所述第一置信度调节值,T2表示所述第二置信度调节值。Wherein, S' represents the confidence degree of other regions after adjustment, S represents the confidence degree of other regions before adjustment, T1 represents the first confidence adjustment value, and T2 represents the second confidence adjustment value.
这样,由于调整后的置信度是第一置信度调节值、第二置信度调节值以及其他区域的置信度之间的乘积,又由于第一置信度调节值、第二置信度调节值从不同角度反映其他区域的被遮挡情况,因此,上述调整后的置信度参考了其他区域被遮挡的情况,使得调节后的置信度更加趋近于实际情况。In this way, since the adjusted confidence level is the product of the first confidence level adjustment value, the second confidence level adjustment value and the confidence levels of other regions, and because the first confidence level adjustment value and the second confidence level adjustment value are different from The angle reflects the occlusion situation of other regions. Therefore, the above adjusted confidence refers to the occlusion situation of other regions, making the adjusted confidence closer to the actual situation.
本公开的一个实施例中,所述目标检测模块403,具体用于选择更新后的置信度大于预设置信度阈值的候选目标区域,将所选择候选目标区域中的目标确定为所述图像中的目标;或选择更新后的置信度最大的预设数量个候选目标区域,将所选择候选目标区域中的目标确定为所述图像中的目标。In an embodiment of the present disclosure, the target detection module 403 is specifically configured to select a candidate target area whose updated confidence is greater than a preset confidence threshold, and determine the target in the selected candidate target area as the target in the image or select a preset number of candidate target areas with the highest confidence after updating, and determine the targets in the selected candidate target areas as the targets in the image.
这样,对于置信度大于预设置信度阈值的候选目标区域来说,这些候选目标区域中包含目标的可能性高于其他候选目标区域包含目标的可能性。所以,将置信度大于预设置信度阈值的候选目标区域中的目标确定为图像中的目标,所得到的目标准确度较高;对于置信度最大的预设数量个候选目标区域来说,这些候选目标区域中包含目标的可能性高于其他候选区域中包含目标的可能性。所以,将置信度最大的预设数量个候选目标区域中的目标确定为图像中的目标,所得到的目标的准确度较高。In this way, for candidate target regions whose confidence is greater than the preset confidence threshold, the probability that these candidate target regions contain a target is higher than the probability that other candidate target regions contain a target. Therefore, if the target in the candidate target area whose confidence is greater than the preset confidence threshold is determined as the target in the image, the accuracy of the obtained target is relatively high; for the preset number of candidate target areas with the largest confidence, these The probability of an object being contained in a candidate object region is higher than that of other candidate regions. Therefore, the target in the preset number of candidate target areas with the highest confidence is determined as the target in the image, and the accuracy of the obtained target is relatively high.
本公开的一个实施例中,所述信息获得模块401,具体用于针对不同的目标尺度,对图像进行目标检测,得到所述图像中不同尺度的候选目标区域、候选目标区域的置信度以及候选目标区域的被遮挡度。In an embodiment of the present disclosure, the information obtaining module 401 is specifically configured to perform target detection on an image for different target scales, and obtain candidate target areas of different scales in the image, confidence levels of candidate target areas, and candidate target areas. The degree of occlusion of the target area.
这样,由于不同尺度的候选目标区域包含的图像特征信息不同,通过得到图像中不同尺度的候选目标区域,丰富了候选目标区域在不同尺度上的特征信息。In this way, since the image feature information contained in the candidate object regions of different scales is different, by obtaining the candidate object regions of different scales in the image, the feature information of the candidate object regions at different scales is enriched.
本公开的一个实施例中,所述信息获得模块401,具体用于将图像输入预先训练得到的目标检测模型,获得所述目标检测模型输出的所述图像中的候选目标区域、候选目标区域的置信度以及候选目标区域的被遮挡度,其中,所述目标检测模型包括: 用于检测图像中候选目标区域的目标检测层和用于预测候选目标区域被遮挡度的遮挡度预测层。In an embodiment of the present disclosure, the information obtaining module 401 is specifically configured to input an image into a pre-trained target detection model, and obtain the candidate target area and the candidate target area in the image output by the target detection model. The confidence level and the occlusion degree of the candidate target area, wherein the target detection model includes: an object detection layer for detecting the candidate target area in the image and an occlusion degree prediction layer for predicting the occlusion degree of the candidate target area.
这样,由于目标检测模型是通过大量训练样本训练得到的,训练过程中,目标检测模型学习到了样本图像中目标区域的特征以及被遮挡的特征,因此,目标检测模型具有较强鲁棒性,从而采用目标检测模型对图像进行目标检测时,能够输出准确的候选目标区域、候选目标区域的置信度以及被遮挡度。In this way, since the target detection model is trained through a large number of training samples, during the training process, the target detection model has learned the features of the target area and the occluded features in the sample image, so the target detection model has strong robustness, thus When the target detection model is used to detect the target in the image, it can output the accurate candidate target area, the confidence degree of the candidate target area and the degree of occlusion.
本公开的技术方案中,所涉及的用户个人信息的收集、存储、使用、加工、传输、提供和公开等处理,均符合相关法律法规的规定,且不违背公序良俗。In the technical solution of this disclosure, the collection, storage, use, processing, transmission, provision, and disclosure of user personal information involved are all in compliance with relevant laws and regulations, and do not violate public order and good customs.
根据本公开的实施例,本公开还提供了一种电子设备、一种可读存储介质和一种计算机程序产品。According to the embodiments of the present disclosure, the present disclosure also provides an electronic device, a readable storage medium, and a computer program product.
本公开的一个实施例中,提供了一种电子设备,包括:In one embodiment of the present disclosure, an electronic device is provided, including:
至少一个处理器;以及at least one processor; and
与所述至少一个处理器通信连接的存储器;其中,a memory communicatively coupled to the at least one processor; wherein,
所述存储器存储有可被所述至少一个处理器执行的指令,所述指令被所述至少一个处理器执行,以使所述至少一个处理器能够执行前述方法实施例中任一车路协同中目标检测方法。The memory stores instructions that can be executed by the at least one processor, and the instructions are executed by the at least one processor, so that the at least one processor can perform any vehicle-road coordination in the foregoing method embodiments. Object detection method.
本公开的一个实施例中,提供了一种存储有计算机指令的非瞬时计算机可读存储介质,其中,所述计算机指令用于使所述计算机执行前述方法实施例中任一车路协同中目标检测方法。In one embodiment of the present disclosure, a non-transitory computer-readable storage medium storing computer instructions is provided, wherein the computer instructions are used to enable the computer to execute any vehicle-road coordination target in the foregoing method embodiments. Detection method.
本公开的一个实施例中,提供了一种计算机程序产品,包括计算机程序,所述计算机程序在被处理器执行时实现前述方法实施例中任一车路协同中目标检测方法。In one embodiment of the present disclosure, a computer program product is provided, including a computer program. When the computer program is executed by a processor, any method for object detection in vehicle-road coordination in the foregoing method embodiments is implemented.
本公开的一个实施例中,提供了一种路侧设备,包括上述电子设备。In one embodiment of the present disclosure, a roadside device is provided, including the above-mentioned electronic device.
本公开的一个实施例中,提供了一种云控平台,包括上述电子设备。In one embodiment of the present disclosure, a cloud control platform is provided, including the above-mentioned electronic device.
图5示出了可以用来实施本公开的实施例的示例电子设备500的示意性框图。电子设备旨在表示各种形式的数字计算机,诸如,膝上型计算机、台式计算机、工作台、个人数字助理、服务器、刀片式服务器、大型计算机、和其它适合的计算机。电子设备还可以表示各种形式的移动装置,诸如,个人数字处理、蜂窝电话、智能电话、可穿戴设备和其它类似的计算装置。本文所示的部件、它们的连接和关系、以及它们的功能仅仅作为示例,并且不意在限制本文中描述的和/或者要求的本公开的实现。FIG. 5 shows a schematic block diagram of an example electronic device 500 that may be used to implement embodiments of the present disclosure. Electronic device is intended to represent various forms of digital computers, such as laptops, desktops, workstations, personal digital assistants, servers, blade servers, mainframes, and other suitable computers. Electronic devices may also represent various forms of mobile devices, such as personal digital processing, cellular telephones, smart phones, wearable devices, and other similar computing devices. The components shown herein, their connections and relationships, and their functions, are by way of example only, and are not intended to limit implementations of the disclosure described and/or claimed herein.
如图5所示,设备500包括计算单元501,其可以根据存储在只读存储器(ROM)502中的计算机程序或者从存储单元508加载到随机访问存储器(RAM)503中的计算机程序,来执行各种适当的动作和处理。在RAM 503中,还可存储设备500操作所需的各种程序和数据。计算单元501、ROM 502以及RAM 503通过总线504彼此相连。输入/输出(I/O)接口505也连接至总线504。As shown in FIG. 5 , the device 500 includes a computing unit 501 that can execute according to a computer program stored in a read-only memory (ROM) 502 or loaded from a storage unit 508 into a random-access memory (RAM) 503. Various appropriate actions and treatments. In the RAM 503, various programs and data necessary for the operation of the device 500 can also be stored. The computing unit 501, ROM 502, and RAM 503 are connected to each other through a bus 504. An input/output (I/O) interface 505 is also connected to the bus 504 .
设备500中的多个部件连接至I/O接口505,包括:输入单元506,例如键盘、鼠标等;输出单元507,例如各种类型的显示器、扬声器等;存储单元508,例如磁盘、光盘等;以及通信单元509,例如网卡、调制解调器、无线通信收发机等。通信单元509允许设备500通过诸如因特网的计算机网络和/或各种电信网络与其他设备交换信息/数据。Multiple components in the device 500 are connected to the I/O interface 505, including: an input unit 506, such as a keyboard, a mouse, etc.; an output unit 507, such as various types of displays, speakers, etc.; a storage unit 508, such as a magnetic disk, an optical disk, etc. ; and a communication unit 509, such as a network card, a modem, a wireless communication transceiver, and the like. The communication unit 509 allows the device 500 to exchange information/data with other devices over a computer network such as the Internet and/or various telecommunication networks.
计算单元501可以是各种具有处理和计算能力的通用和/或专用处理组件。计算单元501的一些示例包括但不限于中央处理单元(CPU)、图形处理单元(GPU)、各种专用的人工智能(AI)计算芯片、各种运行机器学习模型算法的计算单元、数字信号处理器(DSP)、以及任何适当的处理器、控制器、微控制器等。计算单元501执行上文所描述的各个方法和处理,例如车路协同中目标检测方法。例如,在一些实施例中,车路协同中目标检测方法可被实现为计算机软件程序,其被有形地包含于机器可读介质,例如存储单元508。在一些实施例中,计算机程序的部分或者全部可以经由ROM 502和/或通信单元509而被载入和/或安装到设备500上。当计算机程序加载到RAM 503并由计算单元501执行时,可以执行上文描述的车路协同中目标检测方法的一个或多个步骤。备选地,在其他实施例中,计算单元501可以通过其他任何适当的方式(例如,借助于固件)而被配置为执行车路协同中目标检测方法。The computing unit 501 may be various general-purpose and/or special-purpose processing components having processing and computing capabilities. Some examples of computing units 501 include, but are not limited to, central processing units (CPUs), graphics processing units (GPUs), various dedicated artificial intelligence (AI) computing chips, various computing units that run machine learning model algorithms, digital signal processing processor (DSP), and any suitable processor, controller, microcontroller, etc. The calculation unit 501 executes various methods and processes described above, for example, the object detection method in vehicle-road coordination. For example, in some embodiments, the object detection method in vehicle-road coordination can be implemented as a computer software program, which is tangibly contained in a machine-readable medium, such as the storage unit 508 . In some embodiments, part or all of the computer program may be loaded and/or installed on the device 500 via the ROM 502 and/or the communication unit 509. When the computer program is loaded into the RAM 503 and executed by the computing unit 501, one or more steps of the object detection method in the vehicle-road coordination described above can be executed. Alternatively, in other embodiments, the calculation unit 501 may be configured in any other appropriate way (for example, by means of firmware) to execute the object detection method in vehicle-road coordination.
本文中以上描述的系统和技术的各种实施方式可以在数字电子电路系统、集成电路系统、场可编程门阵列(FPGA)、专用集成电路(ASIC)、专用标准产品(ASSP)、芯片上系统的系统(SOC)、负载可编程逻辑设备(CPLD)、计算机硬件、固件、软件、和/或它们的组合中实现。这些各种实施方式可以包括:实施在一个或者多个计算机程序中,该一个或者多个计算机程序可在包括至少一个可编程处理器的可编程系统上执行和/或解释,该可编程处理器可以是专用或者通用可编程处理器,可以从存储系统、至少一个输入装置、和至少一个输出装置接收数据和指令,并且将数据和指令传输至该存储系统、该至少一个输入装置、和该至少一个输出装置。Various implementations of the systems and techniques described above herein can be implemented in digital electronic circuit systems, integrated circuit systems, field programmable gate arrays (FPGAs), application specific integrated circuits (ASICs), application specific standard products (ASSPs), systems on chips Implemented in a system of systems (SOC), load programmable logic device (CPLD), computer hardware, firmware, software, and/or combinations thereof. These various embodiments may include being implemented in one or more computer programs executable and/or interpreted on a programmable system including at least one programmable processor, the programmable processor Can be special-purpose or general-purpose programmable processor, can receive data and instruction from storage system, at least one input device, and at least one output device, and transmit data and instruction to this storage system, this at least one input device, and this at least one output device an output device.
可选的,路侧设备除了包括电子设备,还可以包括通信部件等,电子设备可以和通信部件一体集成,也可以分体设置。电子设备可以获取感知设备(如路侧相机)的数据,例如图片和视频等,从而进行图像视频处理和数据计算。可选的,电子设备自身也可以具备感知数据获取功能和通信功能,例如是AI相机,电子设备可以直接基于获取的感知数据进行图像视频处理和数据计算。Optionally, in addition to electronic equipment, the roadside equipment may also include communication components, etc., and the electronic equipment and communication components may be integrally integrated or separately provided. Electronic devices can obtain data from sensing devices (such as roadside cameras), such as pictures and videos, for image and video processing and data calculation. Optionally, the electronic device itself may also have the function of acquiring sensory data and communication functions, such as an AI camera, and the electronic device may directly perform image and video processing and data calculation based on the acquired sensory data.
可选的,云控平台在云端执行处理,云控平台包括的电子设备可以获取感知设备(如路侧相机)的数据,例如图片和视频等,从而进行图像视频处理和数据计算;云控平台也可以称为车路协同管理平台、边缘计算平台、云计算平台、中心系统、云端服务器等。Optionally, the cloud control platform performs processing on the cloud, and the electronic devices included in the cloud control platform can obtain data from sensing devices (such as roadside cameras), such as pictures and videos, to perform image and video processing and data calculation; the cloud control platform It can also be called vehicle-road collaborative management platform, edge computing platform, cloud computing platform, central system, cloud server, etc.
用于实施本公开的方法的程序代码可以采用一个或多个编程语言的任何组合来 编写。这些程序代码可以提供给通用计算机、专用计算机或其他可编程数据处理装置的处理器或控制器,使得程序代码当由处理器或控制器执行时使流程图和/或框图中所规定的功能/操作被实施。程序代码可以完全在机器上执行、部分地在机器上执行,作为独立软件包部分地在机器上执行且部分地在远程机器上执行或完全在远程机器或服务器上执行。Program code for implementing the methods of the present disclosure may be written in any combination of one or more programming languages. These program codes may be provided to a processor or controller of a general-purpose computer, a special purpose computer, or other programmable data processing devices, so that the program codes, when executed by the processor or controller, make the functions/functions specified in the flow diagrams and/or block diagrams Action is implemented. The program code may execute entirely on the machine, partly on the machine, as a stand-alone software package partly on the machine and partly on a remote machine or entirely on the remote machine or server.
在本公开的上下文中,机器可读介质可以是有形的介质,其可以包含或存储以供指令执行系统、装置或设备使用或与指令执行系统、装置或设备结合地使用的程序。机器可读介质可以是机器可读信号介质或机器可读储存介质。机器可读介质可以包括但不限于电子的、磁性的、光学的、电磁的、红外的、或半导体系统、装置或设备,或者上述内容的任何合适组合。机器可读存储介质的更具体示例会包括基于一个或多个线的电气连接、便携式计算机盘、硬盘、随机存取存储器(RAM)、只读存储器(ROM)、可擦除可编程只读存储器(EPROM或快闪存储器)、光纤、便捷式紧凑盘只读存储器(CD-ROM)、光学储存设备、磁储存设备、或上述内容的任何合适组合。In the context of the present disclosure, a machine-readable medium may be a tangible medium that may contain or store a program for use by or in conjunction with an instruction execution system, apparatus, or device. A machine-readable medium may be a machine-readable signal medium or a machine-readable storage medium. A machine-readable medium may include, but is not limited to, electronic, magnetic, optical, electromagnetic, infrared, or semiconductor systems, apparatus, or devices, or any suitable combination of the foregoing. More specific examples of machine-readable storage media would include one or more wire-based electrical connections, portable computer discs, hard drives, random access memory (RAM), read only memory (ROM), erasable programmable read only memory (EPROM or flash memory), optical fiber, compact disk read only memory (CD-ROM), optical storage, magnetic storage, or any suitable combination of the foregoing.
为了提供与用户的交互,可以在计算机上实施此处描述的系统和技术,该计算机具有:用于向用户显示信息的显示装置(例如,CRT(阴极射线管)或者LCD(液晶显示器)监视器);以及键盘和指向装置(例如,鼠标或者轨迹球),用户可以通过该键盘和该指向装置来将输入提供给计算机。其它种类的装置还可以用于提供与用户的交互;例如,提供给用户的反馈可以是任何形式的传感反馈(例如,视觉反馈、听觉反馈、或者触觉反馈);并且可以用任何形式(包括声输入、语音输入或者、触觉输入)来接收来自用户的输入。To provide for interaction with the user, the systems and techniques described herein can be implemented on a computer having a display device (e.g., a CRT (cathode ray tube) or LCD (liquid crystal display) monitor) for displaying information to the user. ); and a keyboard and pointing device (eg, a mouse or a trackball) through which a user can provide input to the computer. Other kinds of devices can also be used to provide interaction with the user; for example, the feedback provided to the user can be any form of sensory feedback (e.g., visual feedback, auditory feedback, or tactile feedback); and can be in any form (including Acoustic input, speech input or, tactile input) to receive input from the user.
可以将此处描述的系统和技术实施在包括后台部件的计算系统(例如,作为数据服务器)、或者包括中间件部件的计算系统(例如,应用服务器)、或者包括前端部件的计算系统(例如,具有图形用户界面或者网络浏览器的用户计算机,用户可以通过该图形用户界面或者该网络浏览器来与此处描述的系统和技术的实施方式交互)、或者包括这种后台部件、中间件部件、或者前端部件的任何组合的计算系统中。可以通过任何形式或者介质的数字数据通信(例如,通信网络)来将系统的部件相互连接。通信网络的示例包括:局域网(LAN)、广域网(WAN)和互联网。The systems and techniques described herein can be implemented in a computing system that includes back-end components (e.g., as a data server), or a computing system that includes middleware components (e.g., an application server), or a computing system that includes front-end components (e.g., as a a user computer having a graphical user interface or web browser through which a user can interact with embodiments of the systems and techniques described herein), or including such backend components, middleware components, Or any combination of front-end components in a computing system. The components of the system can be interconnected by any form or medium of digital data communication, eg, a communication network. Examples of communication networks include: Local Area Network (LAN), Wide Area Network (WAN) and the Internet.
计算机系统可以包括客户端和服务器。客户端和服务器一般远离彼此并且通常通过通信网络进行交互。通过在相应的计算机上运行并且彼此具有客户端-服务器关系的计算机程序来产生客户端和服务器的关系。服务器可以是云服务器,也可以为分布式系统的服务器,或者是结合了区块链的服务器。A computer system may include clients and servers. Clients and servers are generally remote from each other and typically interact through a communication network. The relationship of client and server arises by computer programs running on the respective computers and having a client-server relationship to each other. The server can be a cloud server, a server of a distributed system, or a server combined with a blockchain.
应该理解,可以使用上面所示的各种形式的流程,重新排序、增加或删除步骤。例如,本发公开中记载的各步骤可以并行地执行也可以顺序地执行也可以不同的次序执行,只要能够实现本公开公开的技术方案所期望的结果,本文在此不进行限制。It should be understood that steps may be reordered, added or deleted using the various forms of flow shown above. For example, each step described in the present disclosure may be executed in parallel, sequentially, or in a different order, as long as the desired result of the technical solution disclosed in the present disclosure can be achieved, no limitation is imposed herein.
上述具体实施方式,并不构成对本公开保护范围的限制。本领域技术人员应该明白的是,根据设计要求和其他因素,可以进行各种修改、组合、子组合和替代。任何在本公开的精神和原则之内所作的修改、等同替换和改进等,均应包含在本公开保护范围之内。The above specific implementation manners are not intended to limit the protection scope of the present disclosure. It should be apparent to those skilled in the art that various modifications, combinations, sub-combinations and substitutions may be made depending on design requirements and other factors. Any modifications, equivalent replacements and improvements made within the spirit and principles of the present disclosure shall be included within the protection scope of the present disclosure.

Claims (23)

  1. 一种车路协同中目标检测方法,所述方法包括:A method for target detection in vehicle-road coordination, the method comprising:
    对图像进行目标检测,得到所述图像中的候选目标区域、候选目标区域的置信度以及候选目标区域的被遮挡度;Carrying out target detection on the image to obtain a candidate target area in the image, a confidence degree of the candidate target area, and an occlusion degree of the candidate target area;
    基于候选目标区域间的交并比以及候选目标区域的被遮挡度,更新候选目标区域的置信度;Update the confidence of the candidate target area based on the intersection ratio between the candidate target areas and the occlusion degree of the candidate target area;
    根据更新后的置信度,从候选目标区域中检测所述图像中的目标。Objects in the image are detected from candidate object regions according to the updated confidence.
  2. 根据权利要求1所述的方法,其中,所述基于候选目标区域间的交并比以及候选目标区域的被遮挡度,更新候选目标区域的置信度,包括:The method according to claim 1, wherein said updating the confidence of the candidate target area based on the intersection-over-union ratio between the candidate target areas and the degree of occlusion of the candidate target area includes:
    循环从区域集中选择置信度最高的第一区域,根据区域集中其他区域与所述第一区域间的交并比以及其他区域的被遮挡度,更新其他区域的置信度,直至所述区域集中包括一个区域,其中,所述区域集包括:候选目标区域中未被选择过的区域。Loop selects the first region with the highest confidence from the region set, and updates the confidence of other regions according to the intersection ratio between other regions in the region set and the first region and the degree of occlusion of other regions until the region set includes An area, wherein the area set includes: unselected areas in candidate target areas.
  3. 根据权利要求2所述的方法,其中,所述根据区域集中其他区域与所述第一区域间的交并比以及其他区域的被遮挡度,更新其他区域的置信度,包括:The method according to claim 2, wherein the updating the confidence of other regions according to the intersection ratio between other regions in the region set and the first region and the degree of occlusion of other regions includes:
    计算区域集中其他区域与所述第一区域间的交并比;calculating intersection ratios between other areas in the area set and the first area;
    根据所述交并比和预设的交并比阈值,确定第一置信度调节值;Determine a first confidence adjustment value according to the intersection-over-union ratio and a preset intersection-over-union ratio threshold;
    根据其他区域的被遮挡度,确定第二置信度调节值;Determine a second confidence adjustment value according to the degree of occlusion of other areas;
    采用所述第一置信度调节值和第二置信度调节值,调节其他区域的置信度。The confidence of other regions is adjusted by using the first confidence adjustment value and the second confidence adjustment value.
  4. 根据权利要求3所述的方法,其中,所述根据所述交并比和预设的交并比阈值,确定第一置信度调节值,包括:The method according to claim 3, wherein said determining the first confidence adjustment value according to the intersection-over-union ratio and the preset intersection-over-union ratio threshold comprises:
    判断所述交并比是否小于预设的交并比阈值;judging whether the intersection and union ratio is less than a preset intersection and union ratio threshold;
    若为是,确定第一置信度调节值为1;If yes, determine that the first confidence adjustment value is 1;
    若为否,确定所述第一置信度调节值为:1与所述交并比之差。If not, determine that the adjusted value of the first confidence level is the difference between 1 and the intersection-over-union ratio.
  5. 根据权利要求3所述的方法,其中,所述根据其他区域的被遮挡度,确定第二置信度调节值,包括:The method according to claim 3, wherein said determining the second confidence adjustment value according to the degree of occlusion of other regions comprises:
    按照以下表达式确定第二置信度调节值g(occ_pred):Determine the second confidence adjustment value g(occ_pred) according to the following expression:
    g(occ_pred)=α occ_pred g(occ_pred)=α occ_pred
    其中,occ_pred为其他区域的被遮挡度,α为预设的常数,α>1。Among them, occ_pred is the occlusion degree of other areas, α is a preset constant, α>1.
  6. 根据权利要求3-5中任一项所述的方法,其中,所述采用所述第一置信度调节值和第二置信度调节值,调节其他区域的置信度,包括:The method according to any one of claims 3-5, wherein the adjusting the confidence of other regions by using the first confidence adjustment value and the second confidence adjustment value comprises:
    按照以下表达式,调节其他区域的置信度:Adjust the confidence of other regions according to the following expression:
    S’=S*T1*T2S'=S*T1*T2
    其中,S’表示调节后的其他区域的置信度,S表示调节前的其他区域的置信度,T1表示所述第一置信度调节值,T2表示所述第二置信度调节值。Wherein, S' represents the confidence degree of other regions after adjustment, S represents the confidence degree of other regions before adjustment, T1 represents the first confidence adjustment value, and T2 represents the second confidence adjustment value.
  7. 根据权利要求1-5中任一项所述的方法,其中,所述根据更新后的置信度,从候选目标区域中检测所述图像中的目标,包括:The method according to any one of claims 1-5, wherein the detecting the target in the image from the candidate target area according to the updated confidence level comprises:
    选择更新后的置信度大于预设置信度阈值的候选目标区域,将所选择候选目标区域中的目标确定为所述图像中的目标;Select a candidate target area whose updated confidence is greater than a preset confidence threshold, and determine the target in the selected candidate target area as the target in the image;
    or
    选择更新后的置信度最大的预设数量个候选目标区域,将所选择候选目标区域中的目标确定为所述图像中的目标。Selecting a preset number of candidate target areas with the highest confidence after updating, and determining the targets in the selected candidate target areas as targets in the image.
  8. 根据权利要求1-5中任一项所述的方法,其中,所述对图像进行目标检测,得到所述图像中的候选目标区域、候选目标区域的置信度以及候选目标区域的被遮挡度,包括:The method according to any one of claims 1-5, wherein the target detection is performed on the image to obtain the candidate target area in the image, the confidence of the candidate target area, and the occlusion degree of the candidate target area, include:
    针对不同的目标尺度,对图像进行目标检测,得到所述图像中不同尺度的候选目标区域、候选目标区域的置信度以及候选目标区域的被遮挡度。Target detection is performed on the image for different target scales, and candidate target areas of different scales in the image, confidence levels of the candidate target areas, and occlusion degrees of the candidate target areas are obtained.
  9. 根据权利要求1-5中任一项所述的方法,其中,所述对图像进行目标检测,得到所述图像中的候选目标区域、候选目标区域的置信度以及候选目标区域的被遮挡度,包括:The method according to any one of claims 1-5, wherein the target detection is performed on the image to obtain the candidate target area in the image, the confidence of the candidate target area, and the occlusion degree of the candidate target area, include:
    将图像输入预先训练得到的目标检测模型,获得所述目标检测模型输出的所述图像中的候选目标区域、候选目标区域的置信度以及候选目标区域的被遮挡度,其中,所述目标检测模型包括:用于检测图像中候选目标区域的目标检测层和用于预测候选目标区域被遮挡度的遮挡度预测层。Input the image into the pre-trained target detection model, and obtain the candidate target area in the image output by the target detection model, the confidence of the candidate target area, and the occlusion degree of the candidate target area, wherein the target detection model It includes: an object detection layer for detecting candidate object areas in an image, and an occlusion degree prediction layer for predicting the degree of occlusion of the candidate object areas.
  10. 一种车路协同中目标检测装置,所述装置包括:A target detection device in vehicle-road coordination, the device comprising:
    信息获得模块,用于对图像进行目标检测,得到所述图像中的候选目标区域、候选目标区域的置信度以及候选目标区域的被遮挡度;An information obtaining module, configured to perform target detection on an image, and obtain a candidate target area in the image, a confidence degree of the candidate target area, and an occlusion degree of the candidate target area;
    置信度更新模块,用于基于候选目标区域间的交并比以及候选目标区域的被遮挡度,更新候选目标区域的置信度;Confidence update module, for updating the confidence of the candidate target area based on the intersection ratio between the candidate target areas and the degree of occlusion of the candidate target area;
    目标检测模块,用于根据更新后的置信度,从候选目标区域中检测所述图像中的目标。The object detection module is used to detect the object in the image from the candidate object area according to the updated confidence.
  11. 根据权利要求10所述的装置,其中,The apparatus of claim 10, wherein,
    所述置信度更新模块,具体用于循环从区域集中选择置信度最高的第一区域,根据区域集中其他区域与所述第一区域间的交并比以及其他区域的被遮挡度,更新其他区域的置信度,直至所述区域集中包括一个区域,其中,所述区域集包括:候选目标区域中未被选择过的区域。The confidence update module is specifically used to circularly select the first area with the highest confidence from the area set, and update other areas according to the intersection ratio between other areas in the area set and the first area and the occlusion degree of other areas until the set of regions includes a region, wherein the set of regions includes: a region that has not been selected in the candidate target region.
  12. 根据权利要求11所述的装置,其中,所述置信度更新模块,包括:The device according to claim 11, wherein the confidence update module comprises:
    交并比计算单元,用于计算区域集中其他区域与所述第一区域间的交并比;an intersection ratio calculation unit, configured to calculate intersection ratios between other regions in the region set and the first region;
    第一调节值确定单元,用于根据所述交并比和预设的交并比阈值,确定第一置信度调节值;A first adjustment value determination unit, configured to determine a first confidence adjustment value according to the intersection ratio and a preset intersection ratio threshold;
    第二调节值确定单元,用于根据其他区域的被遮挡度,确定第二置信度调节值;A second adjustment value determination unit, configured to determine a second confidence adjustment value according to the degree of occlusion of other areas;
    置信度调节单元,用于采用所述第一置信度调节值和第二置信度调节值,调节其他区 域的置信度。A confidence level adjustment unit, configured to use the first confidence level adjustment value and the second confidence level adjustment value to adjust the confidence level of other regions.
  13. 根据权利要求12所述的装置,其中,The apparatus of claim 12, wherein,
    所述第一调节值确定单元,具体用于判断所述交并比是否小于预设的交并比阈值;若为是,确定第一置信度调节值为1;若为否,确定所述第一置信度调节值为:1与所述交并比之差。The first adjustment value determination unit is specifically configured to judge whether the intersection ratio is less than a preset intersection ratio threshold; if yes, determine the first confidence adjustment value is 1; if no, determine the first adjustment value A confidence adjustment value is the difference between 1 and the cross-over-union ratio.
  14. 根据权利要求12所述的装置,其中,所述第二调节值确定单元,具体用于按照以下表达式确定第二置信度调节值g(occ_pred):The device according to claim 12, wherein the second adjustment value determination unit is specifically configured to determine the second confidence adjustment value g(occ_pred) according to the following expression:
    g(occ_pred)=α occ_pred g(occ_pred)=α occ_pred
    其中,occ_pred为其他区域的被遮挡度,α为预设的常数,α>1。Among them, occ_pred is the occlusion degree of other areas, α is a preset constant, α>1.
  15. 根据权利要求12-14中任一项所述的装置,其中,Apparatus according to any one of claims 12-14, wherein,
    所述置信度调节单元,具体用于按照以下表达式,调节其他区域的置信度:The confidence adjustment unit is specifically configured to adjust the confidence of other regions according to the following expression:
    S’=S*T1*T2S'=S*T1*T2
    其中,S’表示调节后的其他区域的置信度,S表示调节前的其他区域的置信度,T1表示所述第一置信度调节值,T2表示所述第二置信度调节值。Wherein, S' represents the confidence degree of other regions after adjustment, S represents the confidence degree of other regions before adjustment, T1 represents the first confidence adjustment value, and T2 represents the second confidence adjustment value.
  16. 根据权利要求10-14中任一项所述的装置,其中,Apparatus according to any one of claims 10-14, wherein,
    所述目标检测模块,具体用于选择更新后的置信度大于预设置信度阈值的候选目标区域,将所选择候选目标区域中的目标确定为所述图像中的目标;或选择更新后的置信度最大的预设数量个候选目标区域,将所选择候选目标区域中的目标确定为所述图像中的目标。The target detection module is specifically configured to select a candidate target area whose confidence after updating is greater than a preset reliability threshold, and determine the target in the selected candidate target area as the target in the image; or select the updated confidence A preset number of candidate target areas with the highest degrees are selected, and the target in the selected candidate target area is determined as the target in the image.
  17. 根据权利要求10-14中任一项所述的装置,其中,Apparatus according to any one of claims 10-14, wherein,
    所述信息获得模块,具体用于针对不同的目标尺度,对图像进行目标检测,得到所述图像中不同尺度的候选目标区域、候选目标区域的置信度以及候选目标区域的被遮挡度。The information obtaining module is specifically configured to perform target detection on images for different target scales, and obtain candidate target areas of different scales in the image, confidence levels of candidate target areas, and occlusion degrees of candidate target areas.
  18. 根据权利要求10-14中任一项所述的装置,其中,Apparatus according to any one of claims 10-14, wherein,
    所述信息获得模块,具体用于将图像输入预先训练得到的目标检测模型,获得所述目标检测模型输出的所述图像中的候选目标区域、候选目标区域的置信度以及候选目标区域的被遮挡度,其中,所述目标检测模型包括:用于检测图像中候选目标区域的目标检测层和用于预测候选目标区域被遮挡度的遮挡度预测层。The information obtaining module is specifically configured to input the image into the pre-trained target detection model, and obtain the candidate target area in the image output by the target detection model, the confidence of the candidate target area, and the occlusion of the candidate target area degree, wherein the target detection model includes: a target detection layer for detecting candidate target areas in an image and an occlusion degree prediction layer for predicting the degree of occlusion of the candidate target areas.
  19. 一种电子设备,包括:An electronic device comprising:
    至少一个处理器;以及at least one processor; and
    与所述至少一个处理器通信连接的存储器;其中,a memory communicatively coupled to the at least one processor; wherein,
    所述存储器存储有可被所述至少一个处理器执行的指令,所述指令被所述至少一个处理器执行,以使所述至少一个处理器能够执行权利要求1-9中任一项所述的方法。The memory stores instructions executable by the at least one processor, and the instructions are executed by the at least one processor, so that the at least one processor can perform any one of claims 1-9. Methods.
  20. 一种存储有计算机指令的非瞬时计算机可读存储介质,其中,所述计算机指令用于使所述计算机执行根据权利要求1-9中任一项所述的方法。A non-transitory computer-readable storage medium storing computer instructions, wherein the computer instructions are used to cause the computer to execute the method according to any one of claims 1-9.
  21. 一种计算机程序产品,包括计算机程序,所述计算机程序在被处理器执行时实现 根据权利要求1-9中任一项所述的方法。A computer program product comprising a computer program which, when executed by a processor, implements the method according to any one of claims 1-9.
  22. 一种路侧设备,包括如权利要求19所述的电子设备。A roadside device comprising the electronic device as claimed in claim 19.
  23. 一种云控平台,包括如权利要求19所述的电子设备。A cloud control platform, comprising the electronic device according to claim 19.
PCT/CN2021/126163 2021-06-28 2021-10-25 Target detection method and apparatus in vehicle-road coordination, and roadside device WO2023273041A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
JP2022535786A JP7436670B2 (en) 2021-06-28 2021-10-25 Target detection method, device, and roadside equipment in road-vehicle coordination
KR1020227019941A KR20220091607A (en) 2021-06-28 2021-10-25 Target detection method, apparatus and roadside device during vehicle-road collaboration

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202110721853.4 2021-06-28
CN202110721853.4A CN113420682B (en) 2021-06-28 2021-06-28 Target detection method and device in vehicle-road cooperation and road side equipment

Publications (1)

Publication Number Publication Date
WO2023273041A1 true WO2023273041A1 (en) 2023-01-05

Family

ID=77717817

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/126163 WO2023273041A1 (en) 2021-06-28 2021-10-25 Target detection method and apparatus in vehicle-road coordination, and roadside device

Country Status (4)

Country Link
JP (1) JP7436670B2 (en)
KR (1) KR20220091607A (en)
CN (1) CN113420682B (en)
WO (1) WO2023273041A1 (en)

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113420682B (en) * 2021-06-28 2023-08-15 阿波罗智联(北京)科技有限公司 Target detection method and device in vehicle-road cooperation and road side equipment
CN114063079B (en) * 2021-10-12 2022-06-21 福瑞泰克智能系统有限公司 Target confidence coefficient acquisition method and device, radar system and electronic device
WO2023065312A1 (en) * 2021-10-22 2023-04-27 深圳市速腾聚创科技有限公司 Obstacle recognition method and apparatus, storage medium, and electronic device
CN114005095B (en) * 2021-10-29 2023-06-30 北京百度网讯科技有限公司 Vehicle attribute identification method, device, electronic equipment and medium
CN116071724B (en) * 2023-03-03 2023-08-04 安徽蔚来智驾科技有限公司 Vehicle-mounted camera shielding scene recognition method, electronic equipment, storage medium and vehicle
CN116363631B (en) * 2023-05-19 2023-09-05 小米汽车科技有限公司 Three-dimensional target detection method and device and vehicle

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109101859A (en) * 2017-06-21 2018-12-28 北京大学深圳研究生院 The method for punishing pedestrian in detection image using Gauss
CN110222764A (en) * 2019-06-10 2019-09-10 中南民族大学 Shelter target detection method, system, equipment and storage medium
US20200097756A1 (en) * 2018-09-26 2020-03-26 Toyota Jidosha Kabushiki Kaisha Object detection device and object detection method
CN111986252A (en) * 2020-07-16 2020-11-24 浙江工业大学 Method for accurately positioning candidate bounding box in target segmentation network
US20210042592A1 (en) * 2019-08-08 2021-02-11 Toyota Jidosha Kabushiki Kaisha Object detection device, object detection method, and computer program for object detection
CN113011417A (en) * 2021-01-08 2021-06-22 湖南大学 Target matching method based on intersection ratio coverage rate loss and repositioning strategy
CN113420682A (en) * 2021-06-28 2021-09-21 阿波罗智联(北京)科技有限公司 Target detection method and device in vehicle-road cooperation and road side equipment

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
TWI393074B (en) * 2009-12-10 2013-04-11 Ind Tech Res Inst Apparatus and method for moving object detection
CN102044151B (en) * 2010-10-14 2012-10-17 吉林大学 Night vehicle video detection method based on illumination visibility identification
CN110009060B (en) * 2019-04-17 2021-07-23 东北大学 Robustness long-term tracking method based on correlation filtering and target detection
CN110428007B (en) * 2019-08-01 2020-11-24 科大讯飞(苏州)科技有限公司 X-ray image target detection method, device and equipment
CN110659600B (en) * 2019-09-19 2022-04-29 北京百度网讯科技有限公司 Object detection method, device and equipment
CN111182491B (en) * 2019-12-31 2022-03-22 淮安中科晶上智能网联研究院有限公司 Radio frequency tomography-based equipment-free target positioning method and device
CN113033456B (en) * 2021-04-08 2023-12-19 阿波罗智联(北京)科技有限公司 Method and device for determining grounding point of vehicle wheel, road side equipment and cloud control platform

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109101859A (en) * 2017-06-21 2018-12-28 北京大学深圳研究生院 The method for punishing pedestrian in detection image using Gauss
US20200097756A1 (en) * 2018-09-26 2020-03-26 Toyota Jidosha Kabushiki Kaisha Object detection device and object detection method
CN110222764A (en) * 2019-06-10 2019-09-10 中南民族大学 Shelter target detection method, system, equipment and storage medium
US20210042592A1 (en) * 2019-08-08 2021-02-11 Toyota Jidosha Kabushiki Kaisha Object detection device, object detection method, and computer program for object detection
CN111986252A (en) * 2020-07-16 2020-11-24 浙江工业大学 Method for accurately positioning candidate bounding box in target segmentation network
CN113011417A (en) * 2021-01-08 2021-06-22 湖南大学 Target matching method based on intersection ratio coverage rate loss and repositioning strategy
CN113420682A (en) * 2021-06-28 2021-09-21 阿波罗智联(北京)科技有限公司 Target detection method and device in vehicle-road cooperation and road side equipment

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
TIE JUN;SONG WEI;YIN FAN;ZHENG LU;YANG XIN: "Object Detection Algorithm Based on Occlusional Labels", JOURNAL OF SOUTH-CENTRAL UNIVERSITY FOR NATIONALITIES(NATURAL SCIENCE EDITION), vol. 39, no. 3, 15 June 2020 (2020-06-15), pages 302 - 308, XP093019220, ISSN: 1672-4321, DOI: 10.12130/znmdzk.20200314 *

Also Published As

Publication number Publication date
CN113420682A (en) 2021-09-21
CN113420682B (en) 2023-08-15
JP2023536025A (en) 2023-08-23
JP7436670B2 (en) 2024-02-21
KR20220091607A (en) 2022-06-30

Similar Documents

Publication Publication Date Title
WO2023273041A1 (en) Target detection method and apparatus in vehicle-road coordination, and roadside device
WO2022134344A1 (en) Target detection method, system and device, and storage medium
CN107886048B (en) Target tracking method and system, storage medium and electronic terminal
WO2023015941A1 (en) Text detection model training method and apparatus, text detection method, and device
CN110472599B (en) Object quantity determination method and device, storage medium and electronic equipment
CN114549840B (en) Training method of semantic segmentation model and semantic segmentation method and device
US20220351398A1 (en) Depth detection method, method for training depth estimation branch network, electronic device, and storage medium
CN115797736B (en) Training method, device, equipment and medium for target detection model and target detection method, device, equipment and medium
CN113222942A (en) Training method of multi-label classification model and method for predicting labels
WO2019109142A1 (en) Monitoring systems, and computer implemented methods for processing data in monitoring systems, programmed to enable identification and tracking of human targets in crowded environments
WO2023138376A1 (en) Action recognition method and apparatus, model training method and apparatus, and electronic device
US20220172376A1 (en) Target Tracking Method and Device, and Electronic Apparatus
WO2023273344A1 (en) Vehicle line crossing recognition method and apparatus, electronic device, and storage medium
CN114449343A (en) Video processing method, device, equipment and storage medium
WO2022252843A1 (en) Method and apparatus for training spatio-temporal data processing model, and device and storage medium
CN114170627A (en) Pedestrian detection method based on improved Faster RCNN
CN112989987A (en) Method, apparatus, device and storage medium for identifying crowd behavior
CN113989721A (en) Target detection method and training method and device of target detection model
CN113569657A (en) Pedestrian re-identification method, device, equipment and storage medium
CN117372928A (en) Video target detection method and device and related equipment
CN116758280A (en) Target detection method, device, equipment and storage medium
CN117011772A (en) Risk prompting method, device and storage medium for power transmission line
WO2023173633A1 (en) Road condition prediction method and apparatus, and corresponding model training method and apparatus, and device and medium
KR102604426B1 (en) Method and apparatus for determining target detection confidence level, electronic device, storage medium, roadside device, cloud control platform, and computer program product
CN114842498A (en) Smoking behavior detection method and device

Legal Events

Date Code Title Description
ENP Entry into the national phase

Ref document number: 2022535786

Country of ref document: JP

Kind code of ref document: A

ENP Entry into the national phase

Ref document number: 20227019941

Country of ref document: KR

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE