WO2022183484A1 - Method and apparatus for determining object detection model - Google Patents

Method and apparatus for determining object detection model Download PDF

Info

Publication number
WO2022183484A1
WO2022183484A1 PCT/CN2021/079304 CN2021079304W WO2022183484A1 WO 2022183484 A1 WO2022183484 A1 WO 2022183484A1 CN 2021079304 W CN2021079304 W CN 2021079304W WO 2022183484 A1 WO2022183484 A1 WO 2022183484A1
Authority
WO
WIPO (PCT)
Prior art keywords
target
target detection
frame
truth
parameter information
Prior art date
Application number
PCT/CN2021/079304
Other languages
French (fr)
Chinese (zh)
Inventor
高鲁涛
罗达新
Original Assignee
华为技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 华为技术有限公司 filed Critical 华为技术有限公司
Priority to PCT/CN2021/079304 priority Critical patent/WO2022183484A1/en
Priority to CN202180079541.6A priority patent/CN116508073A/en
Publication of WO2022183484A1 publication Critical patent/WO2022183484A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition

Definitions

  • the present application relates to the field of target detection, and more particularly, to a method and apparatus for determining a target detection model.
  • object detection can be achieved.
  • target detection may include obstacle recognition, traffic light recognition, and sign recognition.
  • the task of object detection is an important task in the field of machine vision.
  • the purpose of object detection is to find out the object of interest in an image or video and detect the location of the object.
  • the target detection is implemented based on the target detection model, and the target detection model can be constructed based on a series of training samples.
  • the loss function can measure the gap between the target position detected by the target detection model and the real position of the target. According to the difference between the two, the parameters of the target detection model are modified in the direction of decreasing the loss function, so that the target detection model detects that the target position is getting closer and closer to the real position.
  • the task after target detection may be target ranging, that is, determining the distance between the target and itself according to the position of the target obtained by target detection.
  • target detection and target ranging are two separate tasks, often in the training phase of the target detection model, the design of the loss function only considers the needs of target detection and does not consider the needs of subsequent target ranging, which is likely to cause subsequent The error of target ranging is large.
  • the present application provides a method and device for determining a target detection model, which can improve the accuracy of target ranging.
  • a first aspect provides a method for determining a target detection model, where the target detection model is used for target detection, the method comprising: acquiring a first target detection frame according to a first target detection model, the first target detection frame is the boundary contour obtained by performing target detection on the image to be detected; determine the parameter information of the first target detection frame, and the parameter information of the first target detection frame includes parameters related to target ranging;
  • the parameter information of the value box, the target ground-truth box is the actual boundary contour of the target in the image to be detected, and the parameter information of the target ground-truth box includes the target range related to the target object.
  • a second target detection model is determined according to the loss term, and the second target detection model is used to determine a second target detection frame.
  • the first target detection frame can be understood as the position of the target in the image to be detected obtained through target detection.
  • obtaining the first target detection frame according to the first target detection model may include: inputting the image to be detected into the first target detection model, and the first target detection model detects the target object based on a corresponding algorithm The coordinate value of the area occupied in the image to be detected, and according to the coordinate value of the area occupied by the target in the image to be detected, the boundary contour of the target object, that is, the first target detection frame is obtained.
  • the coordinate value of the area occupied by the target object in the image to be detected may include the coordinate value of the diagonal corner of the area occupied by the target object in the image to be detected.
  • the coordinate value of the area occupied by the target object in the image to be detected may include the width, height and coordinate value of the center point of the area occupied by the target object in the image to be detected.
  • the target ground-truth frame can be understood as the actual position of the target in the image to be detected.
  • the actual boundary contour of the target is manually marked.
  • the coordinate value of the area actually occupied by the target object in the image to be detected can be manually marked, so as to obtain the actual boundary contour of the target object according to the coordinate value of the area actually occupied by the target object in the image to be detected, that is, Target ground truth box.
  • the human can label the coordinate values of the diagonal corners of the area occupied by the target in the image to be detected.
  • the human can annotate the width, height and coordinate value of the center point of the area occupied by the target in the image to be detected.
  • the second target detection model determined according to the loss term is the target detection model.
  • the second target detection model is the first target detection model, and if target detection is performed again on the image to be detected based on the second target detection model, the obtained second target detection frame of the target object is the first target detection frame.
  • the first target detection model needs to be corrected N (N ⁇ 1, N is a positive integer) times until the Nth time
  • the loss term obtained by the revised first target detection model is smaller than the preset value, that is, it is considered that the position of the target detected by the Nth revised first target detection model (for example, the second target detection model) is very close to the real position of the target .
  • the second target detection model is the target detection model.
  • the second target detection model is not the first target detection model.
  • target detection is performed again on the image to be detected based on the second target detection model, the obtained second target detection frame of the target object is different from the first target detection frame. And compared with the first target detection frame, the second target detection frame is closer to the target ground-truth frame.
  • the parameters related to the target ranging are used to determine the deviation between the first target detection frame and the target true value frame, and according to the deviation, the first target detection model is modified to obtain The second target detection model, which can improve the accuracy of target ranging.
  • the parameter information of the first target detection frame includes the coordinate value of the center point of the bottom edge of the first target detection frame, the parameter of the target ground-truth frame The information includes the coordinate value of the center point of the bottom edge of the target ground truth box.
  • the target ranging is ranging by using a landing point ranging method.
  • the parameter information of the first target detection frame includes the area of the first target detection frame
  • the parameter information of the target ground-truth frame includes the target The area of the ground truth box
  • the target ranging is ranging by using a proportional ranging method.
  • determining the loss term according to the parameter information of the first target detection frame and the parameter information of the target ground-truth frame includes: according to the first target The parameter information of the detection frame, the parameter information of the target ground-truth frame, and the parameters of the minimum enclosing rectangular frame between the first target detection frame and the target ground-truth frame are used to determine the loss term.
  • the parameter of the minimum enclosing rectangle includes the length of the diagonal of the minimum enclosing rectangle.
  • the parameter information of the first target detection frame, the parameter information of the target ground-truth frame, and the first target detection frame and the The parameters of the minimum enclosing rectangular box of the target ground-truth box include: determining the loss term according to the following formula:
  • point c is the center point of the bottom edge of the target detection frame
  • point c gt is the center point of the bottom edge of the target ground-truth frame
  • the ⁇ (c,c gt ) is the point c and the point c gt is the distance between points
  • the s is the length of the diagonal of the minimum enclosing rectangle.
  • the parameter of the minimum enclosing rectangle includes an area of the minimum enclosing rectangle.
  • the parameter information of the first target detection frame, the parameter information of the target ground-truth frame, and the first target detection frame and the The parameters of the minimum enclosing rectangular box of the target ground-truth box include: determining the loss term according to the following formula:
  • a1 is the area of the target detection frame
  • a2 is the area of the target ground-truth frame
  • a3 is the area of the minimum circumscribed rectangular frame.
  • an apparatus for determining a target detection model where the target detection model is used for target detection, and the apparatus includes: an obtaining unit configured to obtain a first target detection frame according to a first target detection model, the The first target detection frame is the boundary contour obtained by performing target detection on the image to be detected; the processing unit is used to determine the parameter information of the first target detection frame, and the parameter information of the first target detection frame includes the same as the target detection frame.
  • the processing unit is further configured to determine the parameter information of the target ground truth frame, the target ground truth frame is the actual boundary contour of the target object in the image to be detected, and the target ground truth frame is The parameter information of the value box includes parameters related to target ranging for the target; the processing unit is further configured to, according to the parameter information of the first target detection frame and the parameter information of the target true value frame, determining a loss term, where the loss term is used to indicate a deviation between the first target detection frame and the target ground-truth frame; the processing unit is further configured to determine a second target detection model according to the loss term, The second target detection model is used to determine a second target detection frame.
  • the acquisition unit performs target detection on the image to be detected based on the first target detection model to obtain a first target detection frame; the processing unit determines the parameter information of the first target detection frame and the parameter information of the actual boundary contour of the target in the to-be-detected image , that is, the parameter information of the target ground-truth frame, wherein the parameter information of the first target detection frame and the parameter information of the target ground-truth frame both include parameters related to target ranging;
  • the parameter information of the detection frame and the parameter information of the target detection frame determine the loss term between the first target detection frame and the target ground truth frame; finally, the processing unit also determines the second target detection model according to the loss term.
  • the parameters related to the target ranging are used to determine the deviation between the first target detection frame and the target true value frame, and according to the deviation, the first target detection model is modified to obtain The second target detection model, which can improve the accuracy of target ranging.
  • the parameter information of the first target detection frame includes the coordinate value of the center point of the bottom edge of the first target detection frame, the parameter of the target ground-truth frame The information includes the coordinate value of the center point of the bottom edge of the target ground truth box.
  • the target ranging is ranging by using a landing point ranging method.
  • the parameter information of the first target detection frame includes the area of the first target detection frame
  • the parameter information of the target ground-truth frame includes the target The area of the ground truth box
  • the target ranging is ranging by using a proportional ranging method.
  • the processing unit is further specifically configured to: according to the parameter information of the first target detection frame, the parameter information of the target ground-truth frame, and the The parameters of the minimum enclosing rectangle between the first target detection frame and the target ground-truth frame determine the loss term.
  • the parameter of the minimum enclosing rectangle includes the length of the diagonal of the minimum enclosing rectangle.
  • the processing unit is further specifically configured to: determine the loss term according to the following formula:
  • point c is the center point of the bottom edge of the target detection frame
  • point c gt is the center point of the bottom edge of the target ground-truth frame
  • the ⁇ (c,c gt ) is the point c and the point c gt is the distance between points
  • the s is the length of the diagonal of the minimum enclosing rectangle.
  • the parameter of the minimum enclosing rectangle includes an area of the minimum enclosing rectangle.
  • the processing unit is further specifically configured to: determine the loss term according to the following formula:
  • a1 is the area of the target detection frame
  • a2 is the area of the target ground-truth frame
  • a3 is the area of the minimum circumscribed rectangular frame.
  • the actual boundary contour of the target object is manually marked.
  • a third aspect provides an apparatus for determining a target detection model, comprising at least one memory and at least one processor, wherein the at least one memory is used for storing a program, and the at least one processor is used for running the program to realize the first aspect the method described.
  • program may also be referred to as program code, computer instructions, computer programs, program instructions, or the like.
  • a chip comprising at least one processor and an interface circuit, the interface circuit is configured to provide program instructions or data for the at least one processor, and the at least one processor is configured to execute the program instructions, to implement the method described in the first aspect.
  • the chip system may further include a memory, in which a program is stored, the processor is configured to execute the program stored in the memory, and when the program is executed, the The processor is configured to perform the method in the first aspect.
  • a computer-readable storage medium stores a program code for execution by a device, and when the program code is executed by the device, the method described in the first aspect is implemented.
  • a computer program product comprising a computer program, when the computer program product is executed by a computer, the computer performs the method in the aforementioned first aspect.
  • the method of the first aspect may specifically refer to the method in the first aspect and any one of the various implementation manners of the first aspect.
  • a terminal including the apparatus for determining a target detection model according to the second aspect or the third aspect.
  • the terminal may be an intelligent transportation device (vehicle or drone), a smart home device, an intelligent manufacturing device, or a robot, and the like.
  • the intelligent transportation device may be, for example, an automated guided vehicle (AGV), or an unmanned transportation vehicle.
  • AGV automated guided vehicle
  • FIG. 1 is a schematic diagram of an example of an application scenario of the technical solution provided by the embodiment of the present application.
  • Figure 2 is a schematic diagram of a target detection frame and a target ground-truth frame.
  • FIG. 3 is a schematic diagram of another target detection frame and target ground-truth frame.
  • Figure 4 is a schematic diagram of a set of object detection boxes and object ground-truth boxes.
  • FIG. 5 is a schematic diagram of another target detection frame and target ground-truth frame.
  • Figure 6 is a schematic diagram of another set of target detection boxes and target ground-truth boxes.
  • FIG. 7 is a schematic diagram of distance measurement based on a similar triangle method of landing points.
  • Figure 8 is a schematic diagram of another set of target detection boxes and target ground-truth boxes.
  • FIG. 9 is a schematic flowchart of a method for determining a target detection model provided by the present application.
  • FIG. 10 is a schematic diagram of a set of first target detection frames and target ground-truth frames provided by the present application.
  • FIG. 11 is a schematic diagram of another set of first target detection frames and target ground-truth frames provided by this application.
  • FIG. 12 is a schematic diagram of a set of target detection frames and target ground-truth frames provided by this application.
  • FIG. 13 is a schematic diagram of another set of target detection frames and target ground-truth frames provided by this application.
  • FIG. 14 is a schematic diagram of comparison of target ranging based on different target detection methods.
  • FIG. 15 is a schematic structural diagram of an apparatus for determining a target detection model provided by an embodiment of the present application.
  • FIG. 16 is a schematic structural diagram of another apparatus for determining a target detection model provided by an embodiment of the present application.
  • references to "one embodiment” or “some embodiments” or the like described in the embodiments of the present application mean that a particular feature, structure or characteristic described in connection with the embodiment is included in one or more embodiments of the present application.
  • appearances of the phrases “in one embodiment,” “in some embodiments,” “in other embodiments,” “in other embodiments,” etc. in various places in this specification are not necessarily All refer to the same embodiment, but mean “one or more but not all embodiments” unless specifically emphasized otherwise.
  • the terms “including”, “including”, “having” and their variants mean “including but not limited to” unless specifically emphasized otherwise.
  • FIG. 1 shows a schematic diagram of an example application scenario of the target detection model provided by the embodiment of the present application.
  • the application scenario 100 may include a vehicle 111 and a vehicle 112 driving in front of the vehicle 111 .
  • the vehicle 111 is provided with the target detection model determined by the method 200 , then the vehicle 111 can perform target detection on an object in front of the vehicle 111 (eg, the vehicle 112 ) based on the target detection model.
  • the vehicle 111 may also perform target ranging on the vehicle 112, so as to allow the driver of the vehicle 111 to perceive the condition of the road ahead and make a driving strategy (eg, path planning) in advance.
  • a driving strategy eg, path planning
  • vehicle 111 and the vehicle 112 are schematically shown in FIG. 1 only for ease of understanding, but this should not constitute any limitation to the present application.
  • the scene shown in FIG. 1 may further include more objects (including devices), which is not limited in this application.
  • ADAS advanced driving assistant system
  • vision-based sensors eg, on-board cameras
  • radar-type sensors eg, on-board millimeter-wave radar, on-board lidar, on-board ultrasonic radar
  • ADAS implements functions such as adaptive cruise control (ACC), automatic emergency braking (AEB), lane change assist (LCA), blind spot monitoring (BSD), etc. Can't do without the camera.
  • ACC adaptive cruise control
  • AEB automatic emergency braking
  • LCDA lane change assist
  • BSD blind spot monitoring
  • ADAS functions can be realized, such as lane line detection, freespace detection, obstacle recognition, traffic light recognition, sign recognition and other functions.
  • functions such as lane line detection and freespace detection generally use the semantic segmentation model in the machine learning algorithm to give classification information to pixels belonging to lane lines or pixels belonging to freespace in the image.
  • the functions of obstacle recognition, traffic light recognition and sign recognition generally use the target detection model in the machine learning algorithm to achieve target detection.
  • the task of object detection is an important task in the field of machine vision.
  • the purpose of object detection is to find out the object of interest in the image or video (the object can be called the object), and simultaneously realize the position of the output object and the object classification of the object.
  • the method of outputting the object category of the target and the minimum bounding box of the target on the image (the image involved in this application refers to the image of the target object detected by the terminal device) is called 2D target detection;
  • the target detection of the object category and the length, width, height and rotation angle of the target in the three-dimensional space is called 3D target detection.
  • Target detection also known as target extraction
  • target extraction is a kind of target positioning based on target geometric and statistical characteristics, which combines target positioning and recognition into one, and its accuracy and real-time performance are the most important for the entire system that needs to achieve target positioning. an important capability.
  • target extraction is a kind of target positioning based on target geometric and statistical characteristics, which combines target positioning and recognition into one, and its accuracy and real-time performance are the most important for the entire system that needs to achieve target positioning. an important capability.
  • target extraction and recognition of targets is particularly important.
  • the object detection model corresponding to the obstacle recognition will detect the position of the obstacle in the image, and give the category of the obstacle (for example, vehicle, pedestrian, etc.).
  • the target detection model corresponding to obstacle recognition can be based on the YOLO algorithm to obtain the position of the obstacle from the image collected by the camera, and the obtained position of the obstacle is generally represented by a rectangular frame.
  • the category information of the obstacle and the confidence information corresponding to the category can also be obtained from the image collected by the camera. For example, if the obstacle category is a vehicle, the confidence level corresponding to the obstacle category is a vehicle is 90%.
  • the bounding box is usually used to describe the position of the detected object.
  • the detection frame can be a rectangular frame, which can be determined by the coordinates of the rectangular frame. For example, it can be determined by the coordinates of the opposite corners of the rectangular box.
  • FIG. 2 is a schematic diagram of a 2D detection frame provided by an embodiment of the present application.
  • the dashed box (which may be referred to as a target detection frame) as shown in FIG. 2 is the position of the target obtained through target detection.
  • the solid line box shown in Figure 2 (which can be called the target ground truth box) is the actual position of the target.
  • the embodiments of the present application do not limit how to implement target detection, and target detection can be performed by using an existing method or a method after technical development.
  • the bounding box is predicted by the target detection model.
  • the target detection model will predict the position and size of a detection frame relative to the reference point, the type of objects in the detection frame and the confidence of whether there is an object. , and the object class confidence.
  • the target detection is implemented based on the target detection model, and the target detection model can be constructed based on a series of training samples.
  • the target detection model can be obtained by training through multiple images and the position of the target in each of the multiple images.
  • the target detection model can detect the position of the target in the image through a corresponding algorithm to obtain the position of the detected target.
  • the constructed target detection model may not be able to detect the position of the target well, resulting in the target position detected by the target detection model being far away from the real position of the target, that is, the target
  • the detection accuracy of the detection model is not high. Therefore, it is necessary to revise the target detection model to obtain a target detection model with higher detection accuracy.
  • the loss function Liss Function
  • the loss function can be used to measure the gap between the position of the target detected by the target detection model and the real position of the target.
  • the process of revising the target detection model can be understood as a process of reducing the loss function.
  • the value output by the loss function is smaller than the preset value, it is considered that the position of the target detected by the target detection model is very close to the real position of the target.
  • Mode 1 the calculation method of the loss function considers the distance between the center point of the target detection frame and the center point of the target ground-truth frame.
  • the loss term output by the loss function of the target detection frame 101 and the target ground-truth frame 102 can be:
  • ⁇ (A1, A2) is the distance between point A1 and point A2
  • point A1 is the center point of the target detection frame 101
  • point A2 is the target ground truth frame 102
  • s1 is the target detection frame 101 and the target ground truth frame.
  • the length of the diagonal of the smallest enclosing rectangle 103 of 102 eg, the distance between point A3 and point A4.
  • the loss function of the target detection frame 101 and the target ground-truth frame 102 can be gradually reduced, that is, the center point of the target detection frame 101 is continuously approached to the center point of the target ground-truth frame 102, to Improve the detection accuracy of the target detection model corresponding to Figure 3.
  • the loss term output by the loss function of the target detection frame 101 and the target ground-truth frame 102 shown in (a) of FIG. 4 the target detection frame 101 and the target ground-truth frame 102 shown in (b) of FIG. 4
  • the loss term output by the loss function, the loss term output by the loss function of the target detection frame 101 and the target ground-truth frame 102 shown in (c) in FIG. 4 decrease sequentially.
  • the calculation method of the loss function considers the distance between the aspect ratio of the target detection frame and the aspect ratio of the target ground-truth frame.
  • the loss term output by the loss function of the target detection frame 103 and the target ground-truth frame 104 can be:
  • is a weight parameter
  • w2 is the width of the target ground truth frame 104
  • h2 is the height of the target ground truth frame 104
  • w1 is the width of the target detection frame 103
  • h1 is the height of the target detection frame 103 .
  • the loss function of the target detection frame 103 and the target ground-truth frame 104 can be gradually reduced, that is, the aspect ratio of the target detection frame 101 is constantly approaching the aspect ratio of the target ground-truth frame 102 , to improve the detection accuracy of the target detection model corresponding to Figure 5.
  • the loss term output by the loss function of the target detection frame 103 and the target ground-truth frame 104 shown in (a) of FIG. 6 the target detection frame 103 and the target ground-truth frame 104 shown in (b) of FIG. 6
  • the loss term output by the loss function of , the loss term output by the loss function of the target detection frame 103 and the target ground-truth frame 104 shown in (c) in FIG. 6 decrease sequentially.
  • target detection can be performed based on the revised target detection model to obtain the position or size of the target. Then, the distance to the target, ie, target ranging, can be measured based on the target's location or size, in order to make a driving strategy (eg, path planning) in advance.
  • a driving strategy eg, path planning
  • the following taking the landing point ranging method and the proportional ranging method as an example, introduces how to determine the distance between the target and itself, that is, the target ranging, according to the position or size of the target detection frame obtained by the target detection in the image.
  • the landing point ranging method takes the landing point similar triangle ranging method and the landing point coordinate transformation ranging method as examples, and introduces how to determine the distance between the target and itself according to the position or size of the target detection frame obtained by target detection in the image. .
  • the landing point similarity triangle ranging method uses the triangle similarity relationship to determine the distance between the target and itself.
  • the camera on the vehicle 111 is located at point p, the direction of the optical axis of the camera is parallel to the ground, and I is the imaging plane of the camera. According to the triangle similarity relationship, we can get:
  • y is the distance between the projection point of the target landing point in the image and the optical center of the image
  • the unit is pixel
  • f is the focal length
  • the unit is pixel
  • H is the height of the camera from the ground
  • the unit is m
  • Z is the target landing point distance The horizontal distance of the camera, in m.
  • Z can be considered as the distance between the vehicle 112 and the vehicle 111 .
  • the projection point of the target landing point in the image can be equivalent to the midpoint of the bottom edge of the target detection frame obtained based on the target detection.
  • the midpoint of the bottom edge of the target detection frame as shown in FIG. 2 is O point. That is, the above y is the distance from the midpoint of the bottom edge of the target detection frame obtained by target detection to the optical center of the image.
  • the vehicle 111 can determine the distance of the vehicle 112 from itself based on the target detection and formula (2).
  • the landing point coordinate transformation ranging method is based on the image coordinates of the projection point of the target landing point in the image, combined with the camera internal parameter matrix and the camera external parameter matrix, the distance between the target and itself can be obtained.
  • the projection point of the target landing point in the image can also be equivalent to the bottom edge of the target detection frame obtained based on target detection. point.
  • the proportional ranging method uses the proportional relationship between the real object size and the image target size to measure the distance.
  • the distance between the target and itself is determined according to the proportional relationship between the area of the real object under a certain viewing angle and the imaging area of the target in the image.
  • the size of the target detection frame obtained by target detection is 2m ⁇ 2m. As the target distance itself goes from near to far, the size of the target detection frame in the image will be scaled down. If the real size of the target is known and The proportional relationship can be used to obtain the distance between the target and itself.
  • the target ranging task after target detection that is, detecting the distance of the target, can be completed.
  • target detection and target ranging are two separate processes, often target detection only considers the needs of target detection and does not consider the needs of subsequent target ranging, which is likely to cause large errors in subsequent target ranging.
  • the distance between the center point of the target detection frame 106 and the center point of the target ground-truth frame 105 as shown in (a) of FIG. 8 is equal to that shown in (b) of FIG. 8
  • the distance between the target detection frame 107 and the target ground-truth frame 105 shown in FIG. The distance is equal to the distance between the aspect ratio of the target detection frame 107 and the aspect ratio of the target ground-truth frame 105 as shown in (b) of FIG.
  • the loss items of the target detection frame 106 and the target ground-truth frame 105 shown in (a) of FIG. 8 and the loss items of the target detection frame 106 and the target ground-truth frame 107 shown in (b) of FIG. 8 all the same.
  • the projection point of the target landing point in the image is equivalent to the midpoint of the bottom edge of the target detection frame obtained based on target detection.
  • the target obtained based on the target detection shown in (b) in FIG. the error of the target distance obtained by the measurement is larger than the target distance obtained by the measurement in the target detection frame 106 based on the target detection shown in (a) in FIG. 8 .
  • the embodiment of the present application provides a method for determining a target detection model, and the target detection model is used for target detection, so that the target detection model determined by this method can improve the accuracy of target detection, and can also improve the target detection after target detection. distance accuracy.
  • the upper left corner of the image to be detected is taken as the coordinate origin
  • the horizontal direction of the image to be detected from left to right is the positive direction of the x-axis
  • the vertical direction of the image to be detected from top to bottom is the y-axis
  • the size of the image to be detected in the x-axis direction is the width of the image to be detected
  • the size of the image to be detected in the y-axis direction is the height of the image to be detected.
  • the target detection frame for example, the first target detection frame and the second target detection frame
  • the target ground-truth frame are taken as an example for description.
  • FIG. 9 is a schematic flowchart of a method 200 for determining a target detection model provided by an embodiment of the present application. As shown in Figure 9, the method 200 includes:
  • the first target detection frame is a boundary contour obtained by performing target detection on the image to be detected.
  • the image to be detected may be input into a first target detection model, and the first target detection model detects the coordinate value of the area occupied by the target in the image to be detected based on a corresponding algorithm, so that according to the The coordinate value of the area occupied by the target in the image to be detected is used to obtain the boundary contour of the target, that is, the first target detection frame. That is, the first target detection frame can be understood as the position of the target object in the image to be detected obtained through target detection.
  • the targets all refer to the same target in the image to be detected.
  • the coordinate value of the area occupied by the target object in the image to be detected may include the coordinate value of the diagonal corner of the area occupied by the target object in the image to be detected.
  • the coordinate values of the diagonal corners of the area occupied by the target in the image to be detected include the coordinate values of the upper left corner (x1, y1) and the coordinate value of the lower right corner (x2, y2) of the area occupied by the target in the image to be detected ).
  • the coordinate values of the area occupied by the target in the image to be detected may include the width W 1 , the height H 1 of the area occupied by the target in the image to be detected, and the coordinate values of the center point (x o1 , y o1 ) .
  • the center point can be understood as the center symmetry point of the area occupied by the target in the image to be detected.
  • the first device (the device performing S220 ) includes a camera, and the image to be detected is captured by the camera. In other embodiments, the first device (the device performing S220 ) does not include a camera, and the first device may be obtained from other devices capable of obtaining images to be detected.
  • a plurality of first target detection frames may be acquired according to the first target detection model, and S220 to S250 are performed for each first target detection frame.
  • S220 Determine parameter information of the first target detection frame.
  • the parameter information of the first target detection frame includes parameters related to target ranging for the target.
  • the target ranging needs to obtain the implementation of the bottom edge midpoint of the first target detection frame.
  • the parameter information of the first target detection frame may include the coordinate value of the center point of the bottom edge of the first target detection frame.
  • the coordinate value of the area occupied by the target in the image to be detected includes the coordinate value (x1, y1) of the upper left corner and the coordinate value (x2, y2) of the lower right corner of the area occupied by the target in the image to be detected. Then the coordinate value of the center point of the bottom edge of the target detection frame is ( y2).
  • the coordinate values of the area occupied by the target in the image to be detected include the width W 1 and height H 1 of the area occupied by the target in the image to be detected and the coordinate values of the center point (x o1 , y o1 ), Then the coordinate value of the center point of the bottom edge of the target detection frame is (x o1 , ).
  • the parameter information of the first target detection frame is Including the area of the first target detection frame.
  • the area of the target detection frame is
  • the coordinate values of the area occupied by the target in the image to be detected include the width W 1 and height H 1 of the area occupied by the target in the image to be detected and the coordinate values of the center point (x o1 , y o1 ), Then the area of the target detection frame is W 1 ⁇ H 1 .
  • the target ground-truth frame is the actual boundary contour of the target in the image to be detected, and the parameter information of the target ground-truth frame includes parameters related to target ranging for the target.
  • the coordinate value of the area actually occupied by the target object in the image to be detected may be manually marked, so as to obtain the actual boundary of the target object according to the coordinate value of the area actually occupied by the target object in the image to be detected Contour, that is, the target ground-truth box. That is, the target ground-truth frame can be understood as the actual position of the target in the image to be detected.
  • the coordinate values of the diagonal corners of the area actually occupied by the target in the image to be detected may be manually marked.
  • the coordinates of the upper left corner (x3, y3) and the coordinates of the lower right corner (x4, y4) of the area actually occupied by the target in the image to be detected may be manually marked.
  • the width W 2 , the height H 2 and the coordinate values (x o2 , y o2 ) of the area actually occupied by the target in the image to be detected may be manually marked.
  • the target true value frame includes the coordinate value of the center point of the bottom edge of the target ground truth box.
  • the coordinate value of the area actually occupied by the target in the image to be detected includes the coordinate value of the upper left corner (x3, y3) and the coordinate value of the lower right corner (x4, y4) of the area actually occupied by the target in the image to be detected ), then the coordinate value of the center point of the bottom edge of the target ground-truth box is ( y3).
  • the coordinate values of the diagonal corners of the area actually occupied by the target in the image to be detected include the width L 2 and height w 2 of the area actually occupied by the target in the image to be detected, and the coordinate value of the center point (x o2 ) , y o2 ), then the coordinate value of the center point of the bottom edge of the target truth box is (x o2 , ).
  • the target detection needs to obtain the area of the target detection frame (which can also be understood as the size), and the parameter information of the target ground-truth frame includes the target ground-truth frame. area.
  • the accuracy of target ranging can be improved.
  • the area of the target ground-truth box is
  • the coordinate values of the area actually occupied by the target in the image to be detected include the width L 2 and height W 2 of the area actually occupied by the target in the image to be detected, and the coordinates of the center point (x o2 , y o2 ), then the coordinate value of the center point of the bottom edge of the target ground-truth box is W 2 ⁇ H 2 .
  • S220 and S230 is not limited in this embodiment of the present application.
  • S240 Determine a loss item according to the parameter information of the first target detection frame and the parameter information of the target ground-truth frame.
  • the loss term is used to indicate the deviation between the first target detection frame and the target ground-truth frame.
  • the loss term is determined according to the coordinate value of the center point of the bottom edge of the first target detection frame and the coordinate value of the center point of the bottom edge of the target true value frame.
  • the loss term is determined according to the following formula:
  • point c is the center point of the bottom edge of the first target detection frame
  • point c gt is the center point of the bottom edge of the target ground-truth frame
  • ⁇ (c, c gt ) is the distance between point c and point c gt
  • H 2 is the height of the target ground truth box.
  • the loss term is determined according to the coordinate value of the bottom center point of the first target detection frame, the coordinate value of the bottom center point of the target true value frame, and the length of the diagonal line of the minimum circumscribed rectangular frame.
  • the mode B specifically includes S11 and S12.
  • S11 Determine the length of the diagonal of the smallest circumscribed rectangular frame between the first target detection frame and the target ground truth frame.
  • the coordinate value of the first target detection frame and the coordinate value of the target truth value frame determine the coordinate value of the minimum bounding rectangle frame of the first target detection frame and the target truth value frame, and according to the minimum bounding rectangle frame Coordinate value to determine the length of the diagonal of the smallest bounding rectangle.
  • the coordinate value of the first target detection frame includes the coordinate value of the upper left corner (x1, y1), and the coordinate value of the lower right corner (x2, y2).
  • the coordinate value of the target ground-truth box includes the coordinate value of the upper left corner (x3, y3) and the coordinate value of the lower right corner (x4, y4).
  • the coordinate value of the upper left corner of the minimum bounding rectangle of the first target detection frame and the target ground truth frame is (min[x1, x3], min[y1, y3])
  • the coordinate value of the lower right corner is (max[x2, x4], max[y2, y4]).
  • the length of the diagonal of the smallest bounding rectangle is:
  • the first target detection frame can be determined according to the width, height and coordinate value of the center point.
  • the coordinate value of the target truth value box includes the width and height of the target truth value box and the coordinate value of the center point
  • the upper left corner of the target truth value box can be determined according to the width, height and coordinate value of the center point of the target truth value box.
  • the coordinate value and the coordinate value of the lower right corner Therefore, according to the above example, the length of the diagonal of the minimum circumscribed rectangular frame of the first target detection frame and the target ground-truth frame can be determined.
  • point c is the center point of the bottom edge of the first target detection frame
  • point c gt is the center point of the bottom edge of the target ground-truth frame
  • ⁇ (c, c gt ) is the distance between point c and point c gt
  • s is the length of the diagonal of the smallest bounding rectangle.
  • FIG. 10 is a schematic diagram of a set of first target detection frames and target ground-truth frames provided by this application.
  • the coordinate value of the upper left corner B1 of the target truth value box 107 is (x b1 , y b1 ) and the coordinate value of the lower right corner B2 point is (x b2 , y b2 )
  • the target truth value is The coordinate value of the center point c gt of the bottom edge of the frame 107 is ( y b2 ).
  • the coordinate value of the upper left corner B3 of the first target detection frame 108 is (x b3 , y b3 ) and the coordinate value of the lower right corner B4 point is (x b4 , y b4 ), then the center of the bottom edge of the first target detection frame 108
  • the coordinate value of point c is ( y b4 ).
  • the minimum bounding rectangle of the target ground-truth frame 107 and the first target detection frame 108 is the target ground-truth frame 107, then the length of the diagonal of the minimum bounding rectangle of the target ground-truth frame 107 and the first target detection frame 108 Then according to method B, it can be obtained: the loss term between the target ground-truth frame 107 and the first target detection frame 108 is
  • the loss term is determined according to the area of the first target detection frame, the area of the target ground-truth frame, and the area of the minimum circumscribed rectangular frame.
  • S21 Determine the area of the smallest circumscribed rectangular frame between the first target detection frame and the target ground truth frame.
  • the coordinate value of the first target detection frame and the coordinate value of the target truth value frame determine the coordinate value of the minimum enclosing rectangular frame of the first target detection frame and the target truth value frame, and according to the minimum enclosing rectangle frame Coordinate value to determine the area of the smallest bounding rectangle.
  • the coordinate value of the first target detection frame includes the coordinate value of the upper left corner (x1, y1), and the coordinate value of the lower right corner (x2, y2).
  • the coordinate value of the target ground-truth box includes the coordinate value of the upper left corner (x3, y3) and the coordinate value of the lower right corner (x4, y4).
  • the coordinate value of the upper left corner of the minimum bounding rectangle of the first target detection frame and the target ground truth frame is (min[x1, x3], min[y1, y3])
  • the coordinate value of the lower right corner is (max[x2, x4], max[y2, y4]).
  • the area of the smallest bounding rectangle is:
  • the first target detection frame can be determined according to the width, height and coordinate value of the center point.
  • the coordinate value of the target truth value box includes the width and height of the target truth value box and the coordinate value of the center point
  • the upper left corner of the target truth value box can be determined according to the width, height and coordinate value of the center point of the target truth value box.
  • the coordinate value and the coordinate value of the lower right corner Therefore, according to the above example, the area of the minimum enclosing rectangular frame of the first target detection frame and the target ground-truth frame can be determined.
  • a1 is the area of the first target detection frame
  • a2 is the area of the target ground-truth frame
  • a3 is the area of the minimum bounding rectangle.
  • FIG. 11 is a schematic diagram of a set of first target detection frames and target ground-truth frames provided by this application.
  • the coordinate values of the upper left corner C1 of the target ground truth frame 109 are (x c1 , y c1 ) and the coordinate values of the lower right corner C2 point are (x c2 , y c2 ).
  • the area a2 of the target ground truth box 109 is
  • the coordinate value of point C3 in the upper left corner of the first target detection frame 110 is (x c3 , y c3 ) and the coordinate value of point C4 in the lower right corner is (x c4 , y c4 ), then the area a1 of the first target detection frame 110 is
  • the minimum circumscribed rectangle frame of the target ground truth frame 109 and the first target detection frame 110 is frame 111, and the area a3 of the minimum circumscribed rectangular frame 111 is
  • the loss term is determined according to the area of the first target detection frame and the area of the target ground-truth frame.
  • the loss term is determined according to the following formula:
  • a1 is the area of the first target detection frame
  • a2 is the area of the target ground-truth frame.
  • At least two ways of determining the loss item in the prior art may also be used, Determine the loss term.
  • S250 Determine a second target detection model according to the loss term, where the second target detection model is used to determine a second target detection frame.
  • the determined loss item is smaller than the preset value, it is considered that the position of the target detected by the target detection model corresponding to the current loss item is very close to the real position of the target.
  • the determined second target detection model is the target detection model.
  • the second target detection model is the first target detection model. If target detection is performed again on the image to be detected in S210 based on the second target detection model, the obtained second target detection frame of the target is the first target detection frame.
  • S210 to S240 need to be repeatedly performed to perform N (N ⁇ 1, N is a positive integer) corrections on the first target detection model until the first target detection model is corrected based on
  • N N ⁇ 1, N is a positive integer
  • the loss term obtained by the first target detection model after the Nth revision is smaller than the preset value, that is, it is considered that the position of the target detected by the first target detection model (for example, the second target detection model) after the Nth revision is very close to the target real location.
  • the second target detection model is the target detection model.
  • the second target detection model is not the first target detection model.
  • the obtained second target detection frame of the target object is different from the first target detection frame. And compared with the first target detection frame, the second target detection frame is closer to the target ground-truth frame.
  • the first target detection model needs to be modified for the first time according to the loss item to obtain the first modified first target detection model (for example, the third target detection model). detection model), and replace the first target detection model in S210 to S240 with the third target detection model, and repeat S210 to S240 to obtain the loss term again.
  • the first modified first target detection model for example, the third target detection model. detection model
  • the first target detection model needs to be revised for the second time to obtain the second revised first target detection model (for example, the second target detection model) , and replace the first target detection model in S210 to S240 with the second target detection model, and repeat S210 to S240 to obtain the loss item again, if the loss item determined at this time is less than the preset value.
  • the determined second target detection model is the target detection model.
  • the second target detection model is not the first target detection model.
  • target detection is performed again on the image to be detected in S210 based on the second target detection model
  • the obtained second target detection frame of the target object is different from the first target detection frame.
  • the second target detection frame is closer to the target ground-truth frame.
  • FIG. 12 is a schematic diagram of a set of target detection boxes and target ground-truth boxes provided by this application.
  • the target detection frame and the target ground-truth frame shown in (a) of FIG. 12 are obtained based on the target detection of the prior art.
  • the target detection frame shown in (b) of FIG. 12 is obtained when the loss term determined based on the above-mentioned method B is smaller than the preset value.
  • FIG. 13 is a schematic diagram of another set of target detection frames and target ground-truth frames provided by this application.
  • the target detection frame and the target ground-truth frame shown in (a) of FIG. 13 are obtained based on the target detection of the prior art.
  • the target detection frame shown in (b) of FIG. 13 is obtained when the loss term determined based on the above-mentioned method C is smaller than the preset value.
  • the process of repeatedly performing the above S210 to S240 until the loss term is smaller than the preset value may be referred to as the training process of the first target detection model.
  • target detection can be performed on any graphics to be detected.
  • the trained first target detection model may be applied to the terminal.
  • the terminal may be an intelligent transportation device (vehicle or drone), a smart home device, an intelligent manufacturing device, or a robot, and the like.
  • the intelligent transportation device can be, for example, an AGV or an unmanned transportation vehicle.
  • the target ranging task can be performed.
  • the target ranging task can be performed based on the landing point ranging related algorithm.
  • the landing point ranging may be the landing point similarity triangle ranging method or the landing point coordinate transformation ranging method.
  • the target ranging task in order to improve the accuracy of target ranging, in the target detection model (eg, the second target detection model) determined based on the above method C or method D, the target ranging task can be performed based on a proportional ranging method related algorithm.
  • the target detection model eg, the second target detection model
  • the loss term is determined based on the parameters related to target ranging (the coordinates of the bottom center point of the target detection frame or the area of the target detection frame), and the correction is made.
  • the target detection model completes the training of the target detection model. Therefore, the method 200 can improve the accuracy of target detection, and can also improve the accuracy of target ranging.
  • FIG. 14 is a schematic diagram of a comparison of target ranging based on different target detection models.
  • the target detection model is a loss item determined based on parameters related to target ranging, that is, the coordinate value of the center point of the bottom edge of the target detection frame.
  • the target detection model is also based on the coordinate value of the center point of the bottom edge of the target detection frame, so that the position of the target can be detected more accurately, so that the target can be measured.
  • the method for determining the target detection model of the embodiment of the present application is described above with reference to FIGS. 1 to 14 , and the apparatus for determining the target detection model of the embodiment of the present application is described below with reference to FIGS. 15 and 16 . It should be understood that the description of the apparatus for determining the target detection model corresponds to the description of the method for determining the target detection model. Therefore, for parts not described in detail, reference may be made to the foregoing description of the method for determining the target detection model.
  • FIG. 15 is a schematic structural diagram of an apparatus for determining a target detection model provided by an embodiment of the present application.
  • the apparatus 300 for determining a target detection model includes an acquisition module 310 and a processing module 320, wherein,
  • Obtaining unit 310 is used to obtain the first target detection frame according to the first target detection model, and the first target detection frame is the boundary contour that the target detection is carried out to the image to be detected;
  • a processing unit 320 configured to determine parameter information of the first target detection frame, where the parameter information of the first target detection frame includes parameters related to target ranging;
  • the processing unit 320 is further configured to determine parameter information of the target ground-truth frame, where the target ground-truth frame is the actual boundary contour of the target in the to-be-detected image, and the parameter information of the target ground-truth frame includes: parameters related to target ranging for the target;
  • the processing unit 320 is further configured to determine a loss item according to the parameter information of the first target detection frame and the parameter information of the target ground-truth frame, where the loss item is used to indicate the first target detection frame and the deviation between the target ground-truth boxes;
  • the processing unit 320 is further configured to determine a second target detection model according to the loss term, where the second target detection model is used to determine a second target detection frame.
  • the parameter information of the first target detection frame includes the coordinate value of the bottom center point of the first target detection frame
  • the parameter information of the target truth frame includes the bottom center of the target truth frame. Point coordinate value.
  • the target ranging is a ranging method using a landing point ranging method.
  • the parameter information of the first target detection frame includes the area of the first target detection frame
  • the parameter information of the target ground truth frame includes the area of the target ground truth frame
  • the target ranging is ranging by using a proportional ranging method.
  • the processing unit 320 is further specifically configured to: according to the parameter information of the first target detection frame, the parameter information of the target ground-truth frame, and the first target detection frame and the target ground-truth The parameter of the minimum enclosing rectangular box of the box, which determines the loss term.
  • the parameter of the minimum enclosing rectangle includes the length of the diagonal of the minimum enclosing rectangle.
  • processing unit 320 is further specifically configured to: determine the loss item according to the following formula:
  • point c is the center point of the bottom edge of the target detection frame
  • point c gt is the center point of the bottom edge of the target ground-truth frame
  • the ⁇ (c,c gt ) is the point c and the point c gt is the distance between points
  • the s is the length of the diagonal of the minimum enclosing rectangle.
  • the parameter of the minimum enclosing rectangle includes the area of the minimum enclosing rectangle.
  • processing unit 320 is further specifically configured to: determine the loss item according to the following formula:
  • a1 is the area of the target detection frame
  • a2 is the area of the target ground-truth frame
  • a3 is the area of the minimum circumscribed rectangular frame.
  • the actual boundary contour of the target is manually marked.
  • FIG. 16 is a schematic structural diagram of another apparatus for determining a target detection model provided by an embodiment of the present application.
  • the apparatus 400 for determining a target detection model includes at least one memory 410 and at least one processor 420, the at least one memory 410 is used for storing a program, and the at least one processor 420 is used for running the program to realize the aforementioned method 200.
  • processor in the embodiments of the present application may be a central processing unit (central processing unit, CPU), and the processor may also be other general-purpose processors, digital signal processors (digital signal processors, DSP), application-specific integrated circuits (application specific integrated circuit, ASIC), off-the-shelf programmable gate array (field programmable gate array, FPGA) or other programmable logic devices, discrete gate or transistor logic devices, discrete hardware components, etc.
  • a general purpose processor may be a microprocessor or the processor may be any conventional processor or the like.
  • the memory in the embodiments of the present application may be volatile memory or non-volatile memory, or may include both volatile and non-volatile memory.
  • the non-volatile memory may be read-only memory (ROM), programmable read-only memory (PROM), erasable programmable read-only memory (EPROM), electrically programmable Erase programmable read-only memory (electrically EPROM, EEPROM) or flash memory.
  • Volatile memory may be random access memory (RAM), which acts as an external cache.
  • RAM random access memory
  • SRAM static random access memory
  • DRAM dynamic random access memory
  • DRAM synchronous dynamic random access memory
  • SDRAM synchronous dynamic random access memory
  • DDR SDRAM double data rate synchronous dynamic random access memory
  • enhanced SDRAM enhanced synchronous dynamic random access memory
  • SLDRAM synchronous connection dynamic random access memory Fetch memory
  • direct memory bus random access memory direct rambus RAM, DR RAM
  • Embodiments of the present application further provide a computer-readable storage medium, where the computer-readable storage medium has program instructions, and when the program instructions are directly or indirectly executed, the foregoing method can be implemented.
  • Embodiments of the present application also provide a computer program product containing instructions, which, when running on a computing device, enables the computing device to execute the foregoing method, or enables the computing device to implement the foregoing apparatus for determining a target detection model function.
  • An embodiment of the present application further provides a chip, including at least one processor and an interface circuit, where the interface circuit is configured to provide program instructions or data for the at least one processor, and the at least one processor is configured to execute the program instructions , so that the above method can be realized.
  • An embodiment of the present application further provides a terminal, including the aforementioned apparatus for determining a target detection model.
  • the terminal may be an intelligent transportation device (vehicle or drone), a smart home device, an intelligent manufacturing device, or a robot, and the like.
  • the intelligent transportation device can be, for example, an AGV or an unmanned transportation vehicle.
  • the above embodiments may be implemented in whole or in part by software, hardware, firmware or any other combination.
  • the above-described embodiments may be implemented in whole or in part in the form of a computer program product.
  • the computer program product includes one or more computer instructions or computer programs. When the computer instructions or computer programs are loaded or executed on a computer, all or part of the processes or functions described in the embodiments of the present application are generated.
  • the computer may be a general purpose computer, special purpose computer, computer network, or other programmable device.
  • the computer instructions may be stored in or transmitted from one computer readable storage medium to another computer readable storage medium, for example, the computer instructions may be downloaded from a website site, computer, server or data center Transmission to another website site, computer, server or data center by wire (eg, infrared, wireless, microwave, etc.).
  • the computer-readable storage medium can be any available medium that can be accessed by a computer or a data storage device such as a server, a data center, or the like that contains a set of one or more available media.
  • the usable media may be magnetic media (eg, floppy disks, hard disks, magnetic tapes), optical media (eg, DVDs), or semiconductor media.
  • the semiconductor medium may be a solid state drive.
  • each functional unit in each embodiment of the present application may be integrated into one processing unit, or each unit may exist physically alone, or two or more units may be integrated into one unit.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Artificial Intelligence (AREA)
  • Radar Systems Or Details Thereof (AREA)
  • Image Analysis (AREA)

Abstract

The present application provides a method and apparatus for determining an object detection model, applicable in the field of intelligent driving, autonomous driving or unmanned driving of intelligent vehicles. The method comprises: obtaining, on the basis of a first object detection model, a first object detection box of an image to be detected, determining parameter information of the first object detection box, and determining parameter information of a target ground truth box, wherein the target ground truth box is an actual boundary contour of an object in said image, and the parameter information of the first object detection box and the parameter information of the target ground truth box both comprise parameters related to object ranging of the object; determining, according to the parameter information of the first object detection box and the parameter information of the target ground truth box, a deviation between the first object detection box and the target ground truth box; and determining a second object detection model according to the deviation. The parameters related to object ranging are used to modify the first object detection model, thereby improving the accuracy of object ranging.

Description

确定目标检测模型的方法及其装置Method and device for determining target detection model 技术领域technical field
本申请涉及目标检测领域,更具体地,涉及一种确定目标检测模型的方法及其装置。The present application relates to the field of target detection, and more particularly, to a method and apparatus for determining a target detection model.
背景技术Background technique
随着社会的发展,现代生活中越来越多的机器向自动化、智能化发展,移动出行用的车辆也不例外,智能车辆正在逐步进入人们的日常生活中。例如,在包括无人驾驶系统的车辆中,可以实现目标检测。例如,目标检测可以包括障碍物识别、红绿灯识别及标志牌识别等。With the development of society, more and more machines in modern life are developing towards automation and intelligence, and vehicles for mobile travel are no exception. Intelligent vehicles are gradually entering people's daily life. For example, in vehicles that include unmanned systems, object detection can be achieved. For example, target detection may include obstacle recognition, traffic light recognition, and sign recognition.
目标检测(object detection)的任务是机器视觉领域中的重要任务。目标检测的目的是找出图像或视频中感兴趣的物体,并检测出目标的位置。目标检测基于目标检测模型来实现,目标检测模型可以是基于一系列的训练样本构建的。在目标检测模型的训练阶段,损失函数(Loss Function)可以衡量目标检测模型检测的目标位置和目标真实位置之间的差距。根据两者的差距,朝着损失函数减小的方向修正目标检测模型参数,从而使得目标检测模型检测到目标位置越来越接近真实位置。The task of object detection is an important task in the field of machine vision. The purpose of object detection is to find out the object of interest in an image or video and detect the location of the object. The target detection is implemented based on the target detection model, and the target detection model can be constructed based on a series of training samples. In the training phase of the target detection model, the loss function can measure the gap between the target position detected by the target detection model and the real position of the target. According to the difference between the two, the parameters of the target detection model are modified in the direction of decreasing the loss function, so that the target detection model detects that the target position is getting closer and closer to the real position.
目标检测后的任务可以是目标测距,即根据目标检测得到的目标的位置确定目标距离自身的距离。但是,由于目标检测和目标测距是两个分开的任务,往往在目标检测模型的训练阶段中,损失函数的设计仅考虑了目标检测的需求,未考虑后续目标测距的需求,容易造成后续目标测距的误差较大。The task after target detection may be target ranging, that is, determining the distance between the target and itself according to the position of the target obtained by target detection. However, since target detection and target ranging are two separate tasks, often in the training phase of the target detection model, the design of the loss function only considers the needs of target detection and does not consider the needs of subsequent target ranging, which is likely to cause subsequent The error of target ranging is large.
发明内容SUMMARY OF THE INVENTION
本申请提供一种确定目标检测模型的方法及其装置,可以提高目标测距的精度。The present application provides a method and device for determining a target detection model, which can improve the accuracy of target ranging.
第一方面,提供了一种确定目标检测模型的方法,所述目标检测模型用于目标检测,所述方法包括:根据第一目标检测模型获取第一目标检测框,所述第一目标检测框是对待检测图像进行目标检测得到的边界轮廓;确定所述第一目标检测框的参数信息,所述第一目标检测框的参数信息包括与对目标物进行目标测距相关的参数;确定目标真值框的参数信息,所述目标真值框是所述待检测图像中所述目标物实际的边界轮廓,所述目标真值框的参数信息包括与对所述目标物进行目标测距相关的参数;根据所述第一目标检测框的参数信息和所述目标真值框的参数信息,确定损失项,所述损失项用于指示所述第一目标检测框和所述目标真值框之间的偏差;根据所述损失项确定第二目标检测模型,所述第二目标检测模型用于确定第二目标检测框。A first aspect provides a method for determining a target detection model, where the target detection model is used for target detection, the method comprising: acquiring a first target detection frame according to a first target detection model, the first target detection frame is the boundary contour obtained by performing target detection on the image to be detected; determine the parameter information of the first target detection frame, and the parameter information of the first target detection frame includes parameters related to target ranging; The parameter information of the value box, the target ground-truth box is the actual boundary contour of the target in the image to be detected, and the parameter information of the target ground-truth box includes the target range related to the target object. parameter; according to the parameter information of the first target detection frame and the parameter information of the target ground-truth frame, determine a loss item, and the loss item is used to indicate the difference between the first target detection frame and the target ground-truth frame A second target detection model is determined according to the loss term, and the second target detection model is used to determine a second target detection frame.
其中,第一目标检测框可以理解为通过目标检测得到的目标物在待检测图像中的位置。The first target detection frame can be understood as the position of the target in the image to be detected obtained through target detection.
在一种可实现的方式中,根据第一目标检测模型获取第一目标检测框可以包括:将待检测图像输入第一目标检测模型中,第一目标检测模型基于相应的算法,检测出目标物在 待检测图像中所占区域的坐标值,并根据该目标物在待检测图像中所占区域的坐标值,得到目标物的边界轮廓,即第一目标检测框。In an achievable manner, obtaining the first target detection frame according to the first target detection model may include: inputting the image to be detected into the first target detection model, and the first target detection model detects the target object based on a corresponding algorithm The coordinate value of the area occupied in the image to be detected, and according to the coordinate value of the area occupied by the target in the image to be detected, the boundary contour of the target object, that is, the first target detection frame is obtained.
示例性地,目标物在待检测图像中所占区域的坐标值可以包括目标物在待检测图像中所占区域的对角的坐标值。Exemplarily, the coordinate value of the area occupied by the target object in the image to be detected may include the coordinate value of the diagonal corner of the area occupied by the target object in the image to be detected.
示例性地,目标物在待检测图像中所占区域的坐标值可以包括目标物在待检测图像中所占区域的宽度、高度以及中心点的坐标值。Exemplarily, the coordinate value of the area occupied by the target object in the image to be detected may include the width, height and coordinate value of the center point of the area occupied by the target object in the image to be detected.
其中,目标真值框可以理解为目标物在待检测图像中的实际位置。Among them, the target ground-truth frame can be understood as the actual position of the target in the image to be detected.
在一种可实现的方式中,所述目标物实际的边界轮廓是由人工标注的。In an achievable manner, the actual boundary contour of the target is manually marked.
具体地,可以人工将待检测图像中目标物实际所占区域的坐标值标注出来,从而根据该目标物在待检测图像中实际所占区域的坐标值,得到目标物的实际的边界轮廓,即目标真值框。Specifically, the coordinate value of the area actually occupied by the target object in the image to be detected can be manually marked, so as to obtain the actual boundary contour of the target object according to the coordinate value of the area actually occupied by the target object in the image to be detected, that is, Target ground truth box.
示例性地,人工可以标注目标物在待检测图像中所占区域的对角的坐标值。Exemplarily, the human can label the coordinate values of the diagonal corners of the area occupied by the target in the image to be detected.
示例性地,人工可以标注目标物在待检测图像中所占区域的宽度、高度以及中心点的坐标值。Exemplarily, the human can annotate the width, height and coordinate value of the center point of the area occupied by the target in the image to be detected.
在一种可实现的方式中,若确定的损失项小于预设值,即认为当前损失项对应的目标检测模型检测的目标的位置十分接近目标的真实位置。此时,根据该损失项确定的第二目标检测模型即为目标检测模型。其中,第二目标检测模型即为第一目标检测模型,若基于第二目标检测模型对待检测图像再次进行目标检测,得到的目标物的第二目标检测框即为第一目标检测框。In an achievable manner, if the determined loss item is smaller than the preset value, it is considered that the position of the target detected by the target detection model corresponding to the current loss item is very close to the real position of the target. At this time, the second target detection model determined according to the loss term is the target detection model. Wherein, the second target detection model is the first target detection model, and if target detection is performed again on the image to be detected based on the second target detection model, the obtained second target detection frame of the target object is the first target detection frame.
在另一种可实现的方式中,若确定的损失项大于或等于预设值,则需要对第一目标检测模型进行N(N≥1,N为正整数)次修正,直到基于第N次修正后的第一目标检测模型得到的损失项小于预设值,即认为第N次修正后的第一目标检测模型(例如,第二目标检测模型)检测的目标的位置十分接近目标的真实位置。此时,第二目标检测模型即为目标检测模型。其中,第二目标检测模型不是第一目标检测模型。且若基于第二目标检测模型对待检测图像再次进行目标检测,得到的目标物的第二目标检测框和第一目标检测框不相同。且相比于第一目标检测框,第二目标检测框更贴近目标真值框。In another achievable manner, if the determined loss term is greater than or equal to the preset value, the first target detection model needs to be corrected N (N≥1, N is a positive integer) times until the Nth time The loss term obtained by the revised first target detection model is smaller than the preset value, that is, it is considered that the position of the target detected by the Nth revised first target detection model (for example, the second target detection model) is very close to the real position of the target . At this time, the second target detection model is the target detection model. Wherein, the second target detection model is not the first target detection model. And if target detection is performed again on the image to be detected based on the second target detection model, the obtained second target detection frame of the target object is different from the first target detection frame. And compared with the first target detection frame, the second target detection frame is closer to the target ground-truth frame.
首先,基于第一目标检测模型对待检测图像进行目标检测,得到第一目标检测框,确定第一目标检测框的参数信息,并获取待检测图像中目标物的实际的边界轮廓的参数信息,即目标真值框的参数信息,其中,第一目标检测框的参数信息和目标真值框的参数信息均包括与对目标物进行目标测距相关的参数;其次,根据第一目标检测框的参数信息和目标检测框的参数信息,确定第一目标检测框和目标真值框之间的损失项;最后,根据该损失项确定第二目标检测模型。从而在目标检测的过程中,采用与目标测距相关的参数,来确定第一目标检测框和目标真值框之间的偏差,并根据该偏差,对第一目标检侧模型进行修正,得到第二目标检测模型,这样可以提高目标测距的精度。First, perform target detection on the image to be detected based on the first target detection model, obtain a first target detection frame, determine the parameter information of the first target detection frame, and obtain the parameter information of the actual boundary contour of the target in the image to be detected, that is, The parameter information of the target ground-truth frame, wherein the parameter information of the first target detection frame and the parameter information of the target ground-truth frame both include parameters related to the target ranging for the target; secondly, according to the parameters of the first target detection frame The information and the parameter information of the target detection frame are used to determine the loss item between the first target detection frame and the target ground-truth frame; finally, the second target detection model is determined according to the loss item. Therefore, in the process of target detection, the parameters related to the target ranging are used to determine the deviation between the first target detection frame and the target true value frame, and according to the deviation, the first target detection model is modified to obtain The second target detection model, which can improve the accuracy of target ranging.
结合第一方面,在第一方面的某些实现方式中,所述第一目标检测框的参数信息包括所述第一目标检测框的底边中心点坐标值,所述目标真值框的参数信息包括所述目标真值框的底边中心点坐标值。With reference to the first aspect, in some implementations of the first aspect, the parameter information of the first target detection frame includes the coordinate value of the center point of the bottom edge of the first target detection frame, the parameter of the target ground-truth frame The information includes the coordinate value of the center point of the bottom edge of the target ground truth box.
结合第一方面,在第一方面的某些实现方式中,所述目标测距为采用落地点测距法进行的测距。With reference to the first aspect, in some implementations of the first aspect, the target ranging is ranging by using a landing point ranging method.
结合第一方面,在第一方面的某些实现方式中,所述第一目标检测框的参数信息包括所述第一目标检测框的面积,所述目标真值框的参数信息包括所述目标真值框的面积。With reference to the first aspect, in some implementations of the first aspect, the parameter information of the first target detection frame includes the area of the first target detection frame, and the parameter information of the target ground-truth frame includes the target The area of the ground truth box.
结合第一方面,在第一方面的某些实现方式中,所述目标测距为采用比例测距法进行的测距。With reference to the first aspect, in some implementations of the first aspect, the target ranging is ranging by using a proportional ranging method.
结合第一方面,在第一方面的某些实现方式中,根据所述第一目标检测框的参数信息和所述目标真值框的参数信息,确定损失项,包括:根据所述第一目标检测框的参数信息、所述目标真值框的参数信息以及所述第一目标检测框与所述目标真值框的最小外接矩形框的参数,确定所述损失项。With reference to the first aspect, in some implementations of the first aspect, determining the loss term according to the parameter information of the first target detection frame and the parameter information of the target ground-truth frame includes: according to the first target The parameter information of the detection frame, the parameter information of the target ground-truth frame, and the parameters of the minimum enclosing rectangular frame between the first target detection frame and the target ground-truth frame are used to determine the loss term.
结合第一方面,在第一方面的某些实现方式中,所述最小外接矩形框的参数包括所述最小外接矩形框的对角线的长度。With reference to the first aspect, in some implementations of the first aspect, the parameter of the minimum enclosing rectangle includes the length of the diagonal of the minimum enclosing rectangle.
结合第一方面,在第一方面的某些实现方式中,所述根据所述第一目标检测框的参数信息、所述目标真值框的参数信息以及所述第一目标检测框与所述目标真值框的最小外接矩形框的参数,包括:根据以下公式,确定所述损失项:With reference to the first aspect, in some implementations of the first aspect, the parameter information of the first target detection frame, the parameter information of the target ground-truth frame, and the first target detection frame and the The parameters of the minimum enclosing rectangular box of the target ground-truth box include: determining the loss term according to the following formula:
Figure PCTCN2021079304-appb-000001
Figure PCTCN2021079304-appb-000001
其中,c点为所述目标检测框的底边中心点,c gt点为所述目标真值框的底边中心点,所述ρ(c,c gt)为所述c点和所述c gt点之间的距离,所述s为所述最小外接矩形框的对角线的长度。 Wherein, point c is the center point of the bottom edge of the target detection frame, point c gt is the center point of the bottom edge of the target ground-truth frame, and the ρ(c,c gt ) is the point c and the point c gt is the distance between points, and the s is the length of the diagonal of the minimum enclosing rectangle.
结合第一方面,在第一方面的某些实现方式中,所述最小外接矩形框的参数包括所述最小外接矩形框的面积。With reference to the first aspect, in some implementations of the first aspect, the parameter of the minimum enclosing rectangle includes an area of the minimum enclosing rectangle.
结合第一方面,在第一方面的某些实现方式中,所述根据所述第一目标检测框的参数信息、所述目标真值框的参数信息以及所述第一目标检测框与所述目标真值框的最小外接矩形框的参数,包括:根据以下公式,确定所述损失项:With reference to the first aspect, in some implementations of the first aspect, the parameter information of the first target detection frame, the parameter information of the target ground-truth frame, and the first target detection frame and the The parameters of the minimum enclosing rectangular box of the target ground-truth box include: determining the loss term according to the following formula:
Figure PCTCN2021079304-appb-000002
Figure PCTCN2021079304-appb-000002
其中,a1为所述目标检测框的面积,a2为所述目标真值框的面积,所述a3为所述最小外接矩形框的面积。Wherein, a1 is the area of the target detection frame, a2 is the area of the target ground-truth frame, and a3 is the area of the minimum circumscribed rectangular frame.
第二方面,提供了一种确定目标检测模型的装置,所述目标检测模型用于目标检测,所述装置包括:获取单元,用于根据第一目标检测模型获取第一目标检测框,所述第一目标检测框是对待检测图像进行目标检测得到的边界轮廓;处理单元,用于确定所述第一目标检测框的参数信息,所述第一目标检测框的参数信息包括与对目标物进行目标测距相关的参数;所述处理单元,还用于确定目标真值框的参数信息,所述目标真值框是所述待检测图像中所述目标物实际的边界轮廓,所述目标真值框的参数信息包括与对所述目标物进行目标测距相关的参数;所述处理单元,还用于根据所述第一目标检测框的参数信息和所述目标真值框的参数信息,确定损失项,所述损失项用于指示所述第一目标检测框和所述目标真值框之间的偏差;所述处理单元,还用于根据所述损失项确定第二目标检测模型,所述第二目标检测模型用于确定第二目标检测框。In a second aspect, an apparatus for determining a target detection model is provided, where the target detection model is used for target detection, and the apparatus includes: an obtaining unit configured to obtain a first target detection frame according to a first target detection model, the The first target detection frame is the boundary contour obtained by performing target detection on the image to be detected; the processing unit is used to determine the parameter information of the first target detection frame, and the parameter information of the first target detection frame includes the same as the target detection frame. parameters related to target ranging; the processing unit is further configured to determine the parameter information of the target ground truth frame, the target ground truth frame is the actual boundary contour of the target object in the image to be detected, and the target ground truth frame is The parameter information of the value box includes parameters related to target ranging for the target; the processing unit is further configured to, according to the parameter information of the first target detection frame and the parameter information of the target true value frame, determining a loss term, where the loss term is used to indicate a deviation between the first target detection frame and the target ground-truth frame; the processing unit is further configured to determine a second target detection model according to the loss term, The second target detection model is used to determine a second target detection frame.
首先,获取单元基于第一目标检测模型对待检测图像进行目标检测,得到第一目标检测框;处理单元确定第一目标检测框的参数信息以及待检测图像中目标物的实际的边界轮 廓的参数信息,即目标真值框的参数信息,其中,第一目标检测框的参数信息和目标真值框的参数信息均包括与对目标物进行目标测距相关的参数;其次,处理单元根据第一目标检测框的参数信息和目标检测框的参数信息,确定第一目标检测框和目标真值框之间的损失项;最后,处理单元还根据该损失项确定第二目标检测模型。从而在目标检测的过程中,采用与目标测距相关的参数,来确定第一目标检测框和目标真值框之间的偏差,并根据该偏差,对第一目标检侧模型进行修正,得到第二目标检测模型,这样可以提高目标测距的精度。First, the acquisition unit performs target detection on the image to be detected based on the first target detection model to obtain a first target detection frame; the processing unit determines the parameter information of the first target detection frame and the parameter information of the actual boundary contour of the target in the to-be-detected image , that is, the parameter information of the target ground-truth frame, wherein the parameter information of the first target detection frame and the parameter information of the target ground-truth frame both include parameters related to target ranging; The parameter information of the detection frame and the parameter information of the target detection frame determine the loss term between the first target detection frame and the target ground truth frame; finally, the processing unit also determines the second target detection model according to the loss term. Therefore, in the process of target detection, the parameters related to the target ranging are used to determine the deviation between the first target detection frame and the target true value frame, and according to the deviation, the first target detection model is modified to obtain The second target detection model, which can improve the accuracy of target ranging.
结合第二方面,在第二方面的某些实现方式中,所述第一目标检测框的参数信息包括所述第一目标检测框的底边中心点坐标值,所述目标真值框的参数信息包括所述目标真值框的底边中心点坐标值。With reference to the second aspect, in some implementations of the second aspect, the parameter information of the first target detection frame includes the coordinate value of the center point of the bottom edge of the first target detection frame, the parameter of the target ground-truth frame The information includes the coordinate value of the center point of the bottom edge of the target ground truth box.
结合第二方面,在第二方面的某些实现方式中,所述目标测距为采用落地点测距法进行的测距。With reference to the second aspect, in some implementations of the second aspect, the target ranging is ranging by using a landing point ranging method.
结合第二方面,在第二方面的某些实现方式中,所述第一目标检测框的参数信息包括所述第一目标检测框的面积,所述目标真值框的参数信息包括所述目标真值框的面积。With reference to the second aspect, in some implementations of the second aspect, the parameter information of the first target detection frame includes the area of the first target detection frame, and the parameter information of the target ground-truth frame includes the target The area of the ground truth box.
结合第二方面,在第二方面的某些实现方式中,所述目标测距为采用比例测距法进行的测距。With reference to the second aspect, in some implementations of the second aspect, the target ranging is ranging by using a proportional ranging method.
结合第二方面,在第二方面的某些实现方式中,所述处理单元,还具体用于:根据所述第一目标检测框的参数信息、所述目标真值框的参数信息以及所述第一目标检测框与所述目标真值框的最小外接矩形框的参数,确定所述损失项。With reference to the second aspect, in some implementations of the second aspect, the processing unit is further specifically configured to: according to the parameter information of the first target detection frame, the parameter information of the target ground-truth frame, and the The parameters of the minimum enclosing rectangle between the first target detection frame and the target ground-truth frame determine the loss term.
结合第二方面,在第二方面的某些实现方式中,所述最小外接矩形框的参数包括所述最小外接矩形框的对角线的长度。With reference to the second aspect, in some implementations of the second aspect, the parameter of the minimum enclosing rectangle includes the length of the diagonal of the minimum enclosing rectangle.
结合第二方面,在第二方面的某些实现方式中,所述处理单元,还具体用于:根据以下公式,确定所述损失项:With reference to the second aspect, in some implementations of the second aspect, the processing unit is further specifically configured to: determine the loss term according to the following formula:
Figure PCTCN2021079304-appb-000003
Figure PCTCN2021079304-appb-000003
其中,c点为所述目标检测框的底边中心点,c gt点为所述目标真值框的底边中心点,所述ρ(c,c gt)为所述c点和所述c gt点之间的距离,所述s为所述最小外接矩形框的对角线的长度。 Wherein, point c is the center point of the bottom edge of the target detection frame, point c gt is the center point of the bottom edge of the target ground-truth frame, and the ρ(c,c gt ) is the point c and the point c gt is the distance between points, and the s is the length of the diagonal of the minimum enclosing rectangle.
结合第二方面,在第二方面的某些实现方式中,所述最小外接矩形框的参数包括所述最小外接矩形框的面积。With reference to the second aspect, in some implementations of the second aspect, the parameter of the minimum enclosing rectangle includes an area of the minimum enclosing rectangle.
结合第二方面,在第二方面的某些实现方式中,所述处理单元,还具体用于:根据以下公式,确定所述损失项:With reference to the second aspect, in some implementations of the second aspect, the processing unit is further specifically configured to: determine the loss term according to the following formula:
Figure PCTCN2021079304-appb-000004
Figure PCTCN2021079304-appb-000004
其中,a1为所述目标检测框的面积,a2为所述目标真值框的面积,所述a3为所述最小外接矩形框的面积。Wherein, a1 is the area of the target detection frame, a2 is the area of the target ground-truth frame, and a3 is the area of the minimum circumscribed rectangular frame.
结合第二方面,在第二方面的某些实现方式中,所述目标物实际的边界轮廓是由人工标注的。With reference to the second aspect, in some implementations of the second aspect, the actual boundary contour of the target object is manually marked.
第三方面,提供一种确定目标检测模型的装置,包括至少一个存储器和至少一个处理 器,所述至少一个存储器用于存储程序,所述至少一个处理器用于运行所述程序以实现第一方面所述的方法。A third aspect provides an apparatus for determining a target detection model, comprising at least one memory and at least one processor, wherein the at least one memory is used for storing a program, and the at least one processor is used for running the program to realize the first aspect the method described.
应当理解,程序也可以称为程序代码、计算机指令、计算机程序、程序指令等。It should be understood that a program may also be referred to as program code, computer instructions, computer programs, program instructions, or the like.
第四方面,提供一种芯片,包括至少一个处理器和接口电路,所述接口电路用于为所述至少一个处理器提供程序指令或者数据,所述至少一个处理器用于执行所述程序指令,以实现第一方面所述的方法。In a fourth aspect, a chip is provided, comprising at least one processor and an interface circuit, the interface circuit is configured to provide program instructions or data for the at least one processor, and the at least one processor is configured to execute the program instructions, to implement the method described in the first aspect.
可选地,作为一种实现方式,所述芯片系统还可以包括存储器,所述存储器中存储有程序,所述处理器用于执行所述存储器上存储的程序,当所述程序被执行时,所述处理器用于执行第一方面中的方法。Optionally, as an implementation manner, the chip system may further include a memory, in which a program is stored, the processor is configured to execute the program stored in the memory, and when the program is executed, the The processor is configured to perform the method in the first aspect.
第五方面,提供一种计算机可读存储介质,所述计算机可读介质存储用于设备执行的程序代码,该程序代码被所述设备执行时,实现第一方面所述的方法。In a fifth aspect, a computer-readable storage medium is provided, where the computer-readable medium stores a program code for execution by a device, and when the program code is executed by the device, the method described in the first aspect is implemented.
第六方面,提供一种计算机程序产品,所述计算机程序产品包括计算机程序,当所述计算机程序产品被计算机执行时,该计算机执行前述第一方面中的方法。In a sixth aspect, a computer program product is provided, the computer program product comprising a computer program, when the computer program product is executed by a computer, the computer performs the method in the aforementioned first aspect.
应理解,本申请中,第一方面的方法具体可以是指第一方面以及第一方面中各种实现方式中的任意一种实现方式中的方法。It should be understood that, in this application, the method of the first aspect may specifically refer to the method in the first aspect and any one of the various implementation manners of the first aspect.
第七方面,提供一种终端,包括第二方面或第三方面所述的确定目标检测模型的装置。In a seventh aspect, a terminal is provided, including the apparatus for determining a target detection model according to the second aspect or the third aspect.
进一步,该终端可以为智能运输设备(车辆或者无人机)、智能家居设备、智能制造设备或者机器人等。该智能运输设备例如可以是自动导引运输车(automated guided vehicle,AGV)、或无人运输车。Further, the terminal may be an intelligent transportation device (vehicle or drone), a smart home device, an intelligent manufacturing device, or a robot, and the like. The intelligent transportation device may be, for example, an automated guided vehicle (AGV), or an unmanned transportation vehicle.
附图说明Description of drawings
图1是本申请实施例提供的技术方案的一例应用场景的示意图。FIG. 1 is a schematic diagram of an example of an application scenario of the technical solution provided by the embodiment of the present application.
图2是一种目标检测框和目标真值框的示意图。Figure 2 is a schematic diagram of a target detection frame and a target ground-truth frame.
图3是另一种目标检测框和目标真值框的示意图。FIG. 3 is a schematic diagram of another target detection frame and target ground-truth frame.
图4是一组目标检测框和目标真值框的示意图。Figure 4 is a schematic diagram of a set of object detection boxes and object ground-truth boxes.
图5是另一种目标检测框和目标真值框的示意图。FIG. 5 is a schematic diagram of another target detection frame and target ground-truth frame.
图6是另一组目标检测框和目标真值框的示意图。Figure 6 is a schematic diagram of another set of target detection boxes and target ground-truth boxes.
图7是一种基于落地点相似三角形法测距的示意图。FIG. 7 is a schematic diagram of distance measurement based on a similar triangle method of landing points.
图8是另一组目标检测框和目标真值框的示意图。Figure 8 is a schematic diagram of another set of target detection boxes and target ground-truth boxes.
图9是本申请提供的一种确定目标检测模型的方法的示意性流程图。FIG. 9 is a schematic flowchart of a method for determining a target detection model provided by the present application.
图10是本申请提供的一组第一目标检测框和目标真值框的示意图。FIG. 10 is a schematic diagram of a set of first target detection frames and target ground-truth frames provided by the present application.
图11是本申请提供的另一组第一目标检测框和目标真值框的示意图。FIG. 11 is a schematic diagram of another set of first target detection frames and target ground-truth frames provided by this application.
图12是本申请提供的一组目标检测框和目标真值框的示意图。FIG. 12 is a schematic diagram of a set of target detection frames and target ground-truth frames provided by this application.
图13是本申请提供的另一组目标检测框和目标真值框的示意图。FIG. 13 is a schematic diagram of another set of target detection frames and target ground-truth frames provided by this application.
图14是基于不同目标检测方法进行目标测距的对比示意图。FIG. 14 is a schematic diagram of comparison of target ranging based on different target detection methods.
图15是本申请实施例提供的一种确定目标检测模型的装置的示意性结构图。FIG. 15 is a schematic structural diagram of an apparatus for determining a target detection model provided by an embodiment of the present application.
图16是本申请实施例提供的另一种确定目标检测模型的装置的示意性结构图。FIG. 16 is a schematic structural diagram of another apparatus for determining a target detection model provided by an embodiment of the present application.
具体实施方式Detailed ways
下面将结合附图,对本申请中的技术方案进行描述。The technical solutions in the present application will be described below with reference to the accompanying drawings.
以下实施例中所使用的术语只是为了描述特定实施例的目的,而并非旨在作为对本申请的限制。如在本申请的说明书和所附权利要求书中所使用的那样,单数表达形式“一个”、“一种”、“所述”、“上述”、“该”和“这一”旨在也包括。例如“一个或多个”这种表达形式,除非其上下文中明确地有相反指示。还应当理解,在本申请以下各实施例中,“至少一个”、“一个或多个”是指一个、两个或两个以上。The terms used in the following embodiments are for the purpose of describing particular embodiments only, and are not intended to be limitations of the present application. As used in the specification of this application and the appended claims, the singular expressions "a," "an," "the," "above," "the," and "the" are intended to also include. Such expressions as "one or more" unless the context clearly dictates otherwise. It should also be understood that, in the following embodiments of the present application, "at least one" and "one or more" refer to one, two or more than two.
在本申请实施例中描述的参考“一个实施例”或“一些实施例”等意味着在本申请的一个或多个实施例中包括结合该实施例描述的特定特征、结构或特点。由此,在本说明书中的不同之处出现的语句“在一个实施例中”、“在一些实施例中”、“在其他一些实施例中”、“在另外一些实施例中”等不是必然都参考相同的实施例,而是意味着“一个或多个但不是所有的实施例”,除非是以其他方式另外特别强调。术语“包括”、“包含”、“具有”及它们的变形都意味着“包括但不限于”,除非是以其他方式另外特别强调。References to "one embodiment" or "some embodiments" or the like described in the embodiments of the present application mean that a particular feature, structure or characteristic described in connection with the embodiment is included in one or more embodiments of the present application. Thus, appearances of the phrases "in one embodiment," "in some embodiments," "in other embodiments," "in other embodiments," etc. in various places in this specification are not necessarily All refer to the same embodiment, but mean "one or more but not all embodiments" unless specifically emphasized otherwise. The terms "including", "including", "having" and their variants mean "including but not limited to" unless specifically emphasized otherwise.
以下,将结合图1对本申请实施例的技术方案可以应用的场景进行介绍。Hereinafter, scenarios to which the technical solutions of the embodiments of the present application may be applied will be introduced with reference to FIG. 1 .
图1示出了本申请实施例提供的目标检测模型的一例应用场景的示意图。如图1所示,该应用场景100中可以包括车辆111和行驶在车辆111前方的车辆112。其中,车辆111中设置有通过方法200确定的目标检测模型,那么车辆111基于该目标检测模型,可以对车辆111前方的物体(例如,车辆112)进行目标检测。在目标检测之后,车辆111还可以对车辆112进行目标测距,以便预先让车辆111的驾驶者察觉到前方道路的状况,提前做出驾驶策略(例如,路径规划)。FIG. 1 shows a schematic diagram of an example application scenario of the target detection model provided by the embodiment of the present application. As shown in FIG. 1 , the application scenario 100 may include a vehicle 111 and a vehicle 112 driving in front of the vehicle 111 . Wherein, the vehicle 111 is provided with the target detection model determined by the method 200 , then the vehicle 111 can perform target detection on an object in front of the vehicle 111 (eg, the vehicle 112 ) based on the target detection model. After the target detection, the vehicle 111 may also perform target ranging on the vehicle 112, so as to allow the driver of the vehicle 111 to perceive the condition of the road ahead and make a driving strategy (eg, path planning) in advance.
应理解,图1中仅为了便于理解,示意性示出了车辆111和车辆112,但这不应对本申请构成任何限定。例如,图1所示的场景还可以包括更多的物体(包括设备),本申请对此不做限定。It should be understood that the vehicle 111 and the vehicle 112 are schematically shown in FIG. 1 only for ease of understanding, but this should not constitute any limitation to the present application. For example, the scene shown in FIG. 1 may further include more objects (including devices), which is not limited in this application.
随着社会的发展,现代生活中越来越多的机器向自动化、智能化发展,移动出行用的车辆也不例外,智能车辆正在逐步进入人们的日常生活中。近些年,高级驾驶辅助系统(advanced driving assistant system,ADAS)在智能车辆中发挥着十分重要的作用,它是利用安装在车上的各式各样传感器,在车辆行驶过程中随时来感应周围的环境,收集数据,进行静止、移动物体的辨识、侦测与追踪,并结合导航仪地图数据,进行系统的运算与分析,从而预先让驾驶者察觉到可能发生的危险,有效增加汽车驾驶的舒适性和安全性。With the development of society, more and more machines in modern life are developing towards automation and intelligence, and vehicles for mobile travel are no exception. Intelligent vehicles are gradually entering people's daily life. In recent years, the advanced driving assistant system (ADAS) has played a very important role in intelligent vehicles. It uses various sensors installed on the vehicle to sense the surroundings at any time during the driving process of the vehicle. environment, collect data, identify, detect and track stationary and moving objects, and combine with the map data of the navigator to carry out systematic calculation and analysis, so as to make drivers aware of possible dangers in advance and effectively increase the driving experience of the car. Comfort and safety.
例如,在包括无人驾驶系统的车辆中,可以安装视觉系传感器(例如车载摄像头)、雷达类传感器(例如,车载毫米波雷达、车载激光雷达、车载超声波雷达)等。由于摄像头成本较低、技术比较成熟,率先成为无人驾驶系统的主力传感器。例如,ADAS实现自适应巡航控制(adaptive cruise control,ACC)、自动紧急制动(autonomous emergency braking,AEB)、变道辅助(lane change assist,LCA)、盲点监测(blind spot monitoring,BSD)等功能都离不开摄像头。For example, in a vehicle including an unmanned system, vision-based sensors (eg, on-board cameras), radar-type sensors (eg, on-board millimeter-wave radar, on-board lidar, on-board ultrasonic radar), and the like may be installed. Due to the low cost and relatively mature technology of the camera, it is the first to become the main sensor of the unmanned system. For example, ADAS implements functions such as adaptive cruise control (ACC), automatic emergency braking (AEB), lane change assist (LCA), blind spot monitoring (BSD), etc. Can't do without the camera.
根据摄像头采集的图像,可以实现众多ADAS功能,例如车道线检测、可行驶区域(freespace)检测、障碍物识别、红绿灯识别、标志牌识别等功能。According to the images collected by the camera, many ADAS functions can be realized, such as lane line detection, freespace detection, obstacle recognition, traffic light recognition, sign recognition and other functions.
其中,车道线检测和freespace检测等功能一般都会利用机器学习算法中的语义分割模型,对图像中属于车道线像素或者属于freespace的像素给出分类信息。Among them, functions such as lane line detection and freespace detection generally use the semantic segmentation model in the machine learning algorithm to give classification information to pixels belonging to lane lines or pixels belonging to freespace in the image.
其中,障碍物识别、红绿灯识别及标志牌识别等功能一般都会利用机器学习算法中的 目标检测模型来实现目标检测。Among them, the functions of obstacle recognition, traffic light recognition and sign recognition generally use the target detection model in the machine learning algorithm to achieve target detection.
目标检测的任务是机器视觉领域中的重要任务。目标检测的目的是找出图像或视频中感兴趣的物体(该物体可以称之为目标),并同时实现输出目标的位置和目标的物体分类。例如,输出目标的物体类别和目标在图像(本申请中涉及的图像指的是终端设备检测到的目标物体的影像)上的最小包围框的方式称为2D目标检测;还例如,输出目标的物体类别及目标的在三维空间中的长、宽、高和旋转角等信息的目标检测称为3D目标检测。The task of object detection is an important task in the field of machine vision. The purpose of object detection is to find out the object of interest in the image or video (the object can be called the object), and simultaneously realize the position of the output object and the object classification of the object. For example, the method of outputting the object category of the target and the minimum bounding box of the target on the image (the image involved in this application refers to the image of the target object detected by the terminal device) is called 2D target detection; The target detection of the object category and the length, width, height and rotation angle of the target in the three-dimensional space is called 3D target detection.
目标检测,也可以称为目标提取,是一种基于目标几何和统计特征的目标定位,它将目标的定位和识别合二为一,其准确性和实时性是整个需要实现目标定位的系统的一项重要能力。尤其是在复杂场景中,需要对多个目标进行实时处理时,目标自动提取和识别就显得特别重要。Target detection, also known as target extraction, is a kind of target positioning based on target geometric and statistical characteristics, which combines target positioning and recognition into one, and its accuracy and real-time performance are the most important for the entire system that needs to achieve target positioning. an important capability. Especially in complex scenes, when multiple targets need to be processed in real time, the automatic extraction and recognition of targets is particularly important.
示例性地,障碍物识别对应的目标检测模型会检测出障碍物在图像中的位置,并给出该障碍物的类别(例如,车辆、行人等)。例如,障碍物识别对应的目标检测模型可以基于YOLO算法,从摄像头采集的图像中获取障碍物的位置,获取的障碍物的位置一般用矩形框来表示。同时,还可以从摄像头采集的图像中获取该障碍物的类别信息及对应该类别的置信度信息,如障碍物类别为车辆,对应该障碍物类别为车辆的置信度为90%。Exemplarily, the object detection model corresponding to the obstacle recognition will detect the position of the obstacle in the image, and give the category of the obstacle (for example, vehicle, pedestrian, etc.). For example, the target detection model corresponding to obstacle recognition can be based on the YOLO algorithm to obtain the position of the obstacle from the image collected by the camera, and the obtained position of the obstacle is generally represented by a rectangular frame. At the same time, the category information of the obstacle and the confidence information corresponding to the category can also be obtained from the image collected by the camera. For example, if the obstacle category is a vehicle, the confidence level corresponding to the obstacle category is a vehicle is 90%.
在2D目标检测过程中,通常使用检测框(bounding box)来描述检测的目标的位置。检测框可以是一个矩形框,可以由矩形框的坐标确定。例如,可以由矩形框的对角的坐标确定。In the 2D object detection process, the bounding box is usually used to describe the position of the detected object. The detection frame can be a rectangular frame, which can be determined by the coordinates of the rectangular frame. For example, it can be determined by the coordinates of the opposite corners of the rectangular box.
例如,图2是本申请实施例提供的一种2D检测框的示意图。如图2中所示的虚线框(可称为目标检测框)为经目标检测得到的目标的位置。如图2所示的实线框(可称为目标真值框)为目标的实际位置。For example, FIG. 2 is a schematic diagram of a 2D detection frame provided by an embodiment of the present application. The dashed box (which may be referred to as a target detection frame) as shown in FIG. 2 is the position of the target obtained through target detection. The solid line box shown in Figure 2 (which can be called the target ground truth box) is the actual position of the target.
本申请实施例中对于如何实现目标检测并不限定,可以通过现有的方法或者技术发展之后的方法进行目标检测。The embodiments of the present application do not limit how to implement target detection, and target detection can be performed by using an existing method or a method after technical development.
应理解,在2D目标检测过程中,bounding box是通过目标检测模型预测出来的,目标检测模型会预测一个检测框相对于参考点的位置和大小,检测框内物体类别和是否有物体的置信度,以及物体类别置信度。It should be understood that in the process of 2D target detection, the bounding box is predicted by the target detection model. The target detection model will predict the position and size of a detection frame relative to the reference point, the type of objects in the detection frame and the confidence of whether there is an object. , and the object class confidence.
目标检测基于目标检测模型来实现,目标检测模型可以是基于一系列的训练样本构建的。示例性地,可以通过多个图像以及多个图像中的每个图像中目标的位置,训练得到目标检测模型。当将任意一个图像输入该目标检测模型,该目标检测模型可以通过相应的算法,对该图像中目标的位置进行检测,得到检测到的目标的位置。The target detection is implemented based on the target detection model, and the target detection model can be constructed based on a series of training samples. Exemplarily, the target detection model can be obtained by training through multiple images and the position of the target in each of the multiple images. When any image is input into the target detection model, the target detection model can detect the position of the target in the image through a corresponding algorithm to obtain the position of the detected target.
在构建目标检测模型前期,可能会存在由于训练样本不充分,构建的目标检测模型不能很好的检测出目标的位置,造成目标检测模型检测出的目标位置与目标真实位置相距甚远,即目标检测模型的检测精度不高。因此,需要对目标检测模型进行修正,获取一个检测精度较高的目标检测模型。其中,可以用损失函数(Loss Function)来衡量目标检测模型检测的目标的位置和目标的真实位置之间的差距。若损失函数输出的值越大,则目标检测模型检测的目标的位置和目标的真实位置之间的差距越大,若损失函数输出的值越小,则目标检测模型检测的目标的位置和目标的真实位置越接近。因此,对目标检测模型进行修正的过程可理解为是减小损失函数的过程。当损失函数输出的值小于预设值时,即认为目标检测模型检测的目标的位置十分接近目标的真实位置。In the early stage of building the target detection model, there may be insufficient training samples, and the constructed target detection model may not be able to detect the position of the target well, resulting in the target position detected by the target detection model being far away from the real position of the target, that is, the target The detection accuracy of the detection model is not high. Therefore, it is necessary to revise the target detection model to obtain a target detection model with higher detection accuracy. Among them, the loss function (Loss Function) can be used to measure the gap between the position of the target detected by the target detection model and the real position of the target. If the value output by the loss function is larger, the gap between the position of the target detected by the target detection model and the real position of the target is larger, and if the value output by the loss function is smaller, the position of the target detected by the target detection model is the same as the target. closer to the true location. Therefore, the process of revising the target detection model can be understood as a process of reducing the loss function. When the value output by the loss function is smaller than the preset value, it is considered that the position of the target detected by the target detection model is very close to the real position of the target.
以下,以方式1和方式2为例,介绍如何计算损失函数中的损失项。Hereinafter, taking Mode 1 and Mode 2 as examples, we will introduce how to calculate the loss term in the loss function.
方式1,损失函数的计算方式中考虑目标检测框的中心点和目标真值框的中心点之间的距离。 Mode 1, the calculation method of the loss function considers the distance between the center point of the target detection frame and the center point of the target ground-truth frame.
例如,如图3所示,目标检测框101和目标真值框102的损失函数输出的损失项可以为:For example, as shown in FIG. 3 , the loss term output by the loss function of the target detection frame 101 and the target ground-truth frame 102 can be:
Figure PCTCN2021079304-appb-000005
Figure PCTCN2021079304-appb-000005
其中,ρ(A1,A2)为点A1与点A2之间的距离,点A1为目标检测框101的中心点,点A2为目标真值框102,s1为目标检测框101和目标真值框102的最小外接矩形框103的对角线的长度(例如,点A3与点A4之间的距离)。Among them, ρ(A1, A2) is the distance between point A1 and point A2, point A1 is the center point of the target detection frame 101, point A2 is the target ground truth frame 102, and s1 is the target detection frame 101 and the target ground truth frame. The length of the diagonal of the smallest enclosing rectangle 103 of 102 (eg, the distance between point A3 and point A4).
在图3对应的目标检测模型的构建中,可以逐步减小目标检测框101和目标真值框102的损失函数,即将目标检测框101的中心点不断靠近目标真值框102的中心点,来提高图3对应的目标检测模型的检测精度。例如,图4中的(a)所示的目标检测框101和目标真值框102的损失函数输出的损失项、图4中的(b)所示的目标检测框101和目标真值框102的损失函数输出的损失项、图4中的(c)所示的目标检测框101和目标真值框102的损失函数输出的损失项依次减小。In the construction of the target detection model corresponding to FIG. 3 , the loss function of the target detection frame 101 and the target ground-truth frame 102 can be gradually reduced, that is, the center point of the target detection frame 101 is continuously approached to the center point of the target ground-truth frame 102, to Improve the detection accuracy of the target detection model corresponding to Figure 3. For example, the loss term output by the loss function of the target detection frame 101 and the target ground-truth frame 102 shown in (a) of FIG. 4 , the target detection frame 101 and the target ground-truth frame 102 shown in (b) of FIG. 4 The loss term output by the loss function, the loss term output by the loss function of the target detection frame 101 and the target ground-truth frame 102 shown in (c) in FIG. 4 decrease sequentially.
方式2,损失函数的计算方式中考虑目标检测框的宽高比和目标真值框的宽高比之间的距离。In method 2, the calculation method of the loss function considers the distance between the aspect ratio of the target detection frame and the aspect ratio of the target ground-truth frame.
例如,如图5所示,目标检测框103和目标真值框104的损失函数输出的损失项可以为:For example, as shown in FIG. 5 , the loss term output by the loss function of the target detection frame 103 and the target ground-truth frame 104 can be:
Figure PCTCN2021079304-appb-000006
Figure PCTCN2021079304-appb-000006
其中,α为权重参数,w2为目标真值框104的宽度,h2为目标真值框104的高度,w1为目标检测框103的宽度,h1为目标检测框103的高度。α is a weight parameter, w2 is the width of the target ground truth frame 104 , h2 is the height of the target ground truth frame 104 , w1 is the width of the target detection frame 103 , and h1 is the height of the target detection frame 103 .
在图6对应的目标检测模型的构建中,可以逐步减小目标检测框103和目标真值框104的损失函数,即将目标检测框101的宽高比不断靠近目标真值框102的宽高比,来提高图5对应的目标检测模型的检测精度。例如,图6中的(a)所示的目标检测框103和目标真值框104的损失函数输出的损失项、图6中的(b)所示的目标检测框103和目标真值框104的损失函数输出的损失项、图6中的(c)所示的目标检测框103和目标真值框104的损失函数输出的损失项依次减小。In the construction of the target detection model corresponding to FIG. 6 , the loss function of the target detection frame 103 and the target ground-truth frame 104 can be gradually reduced, that is, the aspect ratio of the target detection frame 101 is constantly approaching the aspect ratio of the target ground-truth frame 102 , to improve the detection accuracy of the target detection model corresponding to Figure 5. For example, the loss term output by the loss function of the target detection frame 103 and the target ground-truth frame 104 shown in (a) of FIG. 6 , the target detection frame 103 and the target ground-truth frame 104 shown in (b) of FIG. 6 The loss term output by the loss function of , the loss term output by the loss function of the target detection frame 103 and the target ground-truth frame 104 shown in (c) in FIG. 6 decrease sequentially.
在对目标检测模型进行修正之后,可以基于修正后的目标检测模型进行目标检测,得到目标的位置或大小。随后,可以基于目标的位置或大小对目标的距离进行测量,即目标测距,以便提前做出驾驶策略(例如,路径规划)。After the target detection model is revised, target detection can be performed based on the revised target detection model to obtain the position or size of the target. Then, the distance to the target, ie, target ranging, can be measured based on the target's location or size, in order to make a driving strategy (eg, path planning) in advance.
以下,以落地点测距法和比例测距法为例,介绍如何根据目标检测得到的目标检测框在图像中的位置或大小,确定目标距离自身的距离,即目标测距。The following, taking the landing point ranging method and the proportional ranging method as an example, introduces how to determine the distance between the target and itself, that is, the target ranging, according to the position or size of the target detection frame obtained by the target detection in the image.
其中,落地点测距法以落地点相似三角形测距法和落地点坐标变换测距法为例,介绍如何根据目标检测得到的目标检测框在图像中的位置或大小,确定目标距离自身的距离。Among them, the landing point ranging method takes the landing point similar triangle ranging method and the landing point coordinate transformation ranging method as examples, and introduces how to determine the distance between the target and itself according to the position or size of the target detection frame obtained by target detection in the image. .
(1)落地点相似三角形测距法(1) Landing point similar triangle ranging method
落地点相似三角形测距法是利用三角形相似关系,确定目标距离自身的距离。The landing point similarity triangle ranging method uses the triangle similarity relationship to determine the distance between the target and itself.
具体的,例如,如图7所示,车辆111上摄像头位于p点,摄像头光轴方向与地面平行,I为摄像机的成像平面。根据三角形相似关系可得:Specifically, for example, as shown in FIG. 7 , the camera on the vehicle 111 is located at point p, the direction of the optical axis of the camera is parallel to the ground, and I is the imaging plane of the camera. According to the triangle similarity relationship, we can get:
Figure PCTCN2021079304-appb-000007
Figure PCTCN2021079304-appb-000007
其中,y为目标落地点在图像中的投影点距离图像光心的距离,单位为pixel,f为焦距,单位为pixel,H为摄像头距离地面的高度,单位为m,Z为目标落地点距离摄像头的水平距离,单位为m。Among them, y is the distance between the projection point of the target landing point in the image and the optical center of the image, the unit is pixel, f is the focal length, the unit is pixel, H is the height of the camera from the ground, the unit is m, and Z is the target landing point distance The horizontal distance of the camera, in m.
进而可得,目标落地点距离摄像头的水平距离Z为:Then, the horizontal distance Z from the target landing point to the camera is:
Figure PCTCN2021079304-appb-000008
Figure PCTCN2021079304-appb-000008
则Z可以认为是车辆112距离车辆111的距离。Then Z can be considered as the distance between the vehicle 112 and the vehicle 111 .
一般情况下,目标落地点在图像中的投影点可等效为基于目标检测得到的目标检测框的底边中点,例如,如图2中所述的目标检测框的底边中点为O点。即上述y为目标检测得到的目标检测框的底边中点距离图像光心的距离。这样,车辆111基于目标检测和公式(2),即可确定车辆112距离自身的距离。In general, the projection point of the target landing point in the image can be equivalent to the midpoint of the bottom edge of the target detection frame obtained based on the target detection. For example, the midpoint of the bottom edge of the target detection frame as shown in FIG. 2 is O point. That is, the above y is the distance from the midpoint of the bottom edge of the target detection frame obtained by target detection to the optical center of the image. In this way, the vehicle 111 can determine the distance of the vehicle 112 from itself based on the target detection and formula (2).
(2)落地点坐标变换测距法(2) Landing point coordinate transformation ranging method
落地点坐标变换测距法则是根据目标落地点在图像中的投影点的图像坐标,结合相机内参矩阵和相机外参矩阵,即可得到目标距离自身的距离。The landing point coordinate transformation ranging method is based on the image coordinates of the projection point of the target landing point in the image, combined with the camera internal parameter matrix and the camera external parameter matrix, the distance between the target and itself can be obtained.
同落地点相似三角形测距法一样,一般情况下,在落地点坐标变换测距法中,目标落地点在图像中的投影点也可等效为基于目标检测得到的目标检测框的底边中点。Like the landing point similar triangle ranging method, in general, in the landing point coordinate transformation ranging method, the projection point of the target landing point in the image can also be equivalent to the bottom edge of the target detection frame obtained based on target detection. point.
(3)比例测距法(3) Proportional ranging method
比例测距法是利用真实物体尺寸与图像目标尺寸比例关系测距。The proportional ranging method uses the proportional relationship between the real object size and the image target size to measure the distance.
一般情况下,会根据真实物体在某个视角下的面积与目标在图像中成像面积之间的比例关系,确定目标距离自身的距离。例如,经目标检测得到的目标检测框的尺寸为2m×2m,随着目标距离本身由近及远,该目标检测框的尺寸在图像中的尺寸会按比例缩小,若已知目标真实尺寸及该比例关系即可求得目标距离自身的距离。In general, the distance between the target and itself is determined according to the proportional relationship between the area of the real object under a certain viewing angle and the imaging area of the target in the image. For example, the size of the target detection frame obtained by target detection is 2m × 2m. As the target distance itself goes from near to far, the size of the target detection frame in the image will be scaled down. If the real size of the target is known and The proportional relationship can be used to obtain the distance between the target and itself.
基于上述(1)至(3)所述的方法,可以完成目标检测后的目标测距任务,即检测目标的距离。Based on the methods described in (1) to (3) above, the target ranging task after target detection, that is, detecting the distance of the target, can be completed.
由于目标检测和目标测距是两个分开的过程,往往目标检测仅考虑了目标检测的需求,未考虑后续目标测距的需求,容易造成后续目标测距的误差较大。Since target detection and target ranging are two separate processes, often target detection only considers the needs of target detection and does not consider the needs of subsequent target ranging, which is likely to cause large errors in subsequent target ranging.
例如,在目标检测任务中,由于如图8中的(a)所示的目标检测框106的中心点与目标真值框105的中心点之间的距离等于如图8中的(b)所示的目标检测框107与目标真值框105之间的距离,以及如图8中的(a)所示的目标检测框106的宽高比与目标真值框105的宽高比之间的距离等于如图8中的(b)所示的目标检测框107的宽高比与目标真值框105的宽高比之间的距离,因此,无论是按照上述方式1还是方式2计算损失项,如图8中的(a)所示的目标检测框106与目标真值框105的损失项和如图8中的(b)所示的目标检测框106与目标真值框107的损失项都是一样的。For example, in the target detection task, since the distance between the center point of the target detection frame 106 and the center point of the target ground-truth frame 105 as shown in (a) of FIG. 8 is equal to that shown in (b) of FIG. 8 The distance between the target detection frame 107 and the target ground-truth frame 105 shown in FIG. The distance is equal to the distance between the aspect ratio of the target detection frame 107 and the aspect ratio of the target ground-truth frame 105 as shown in (b) of FIG. , the loss items of the target detection frame 106 and the target ground-truth frame 105 shown in (a) of FIG. 8 and the loss items of the target detection frame 106 and the target ground-truth frame 107 shown in (b) of FIG. 8 all the same.
然而,在目标测距任务中,若基于上述(1)和(2)所述的方法,即将目标落地点在图像中的投影点等效为基于目标检测得到的目标检测框的底边中点,来完成目标测距的任 务。由于目标检测框106的底边中点比目标检测框107底边中点更接近目标真值框105底边中点,因此,基于图8中的(b)所示的目标检测后得到的目标检测框107,进行测量得到的目标距离的误差大于基于图8中的(a)所示的目标检测后得到的目标检测框106,进行测量得到的目标距离。However, in the target ranging task, if based on the methods described in (1) and (2) above, the projection point of the target landing point in the image is equivalent to the midpoint of the bottom edge of the target detection frame obtained based on target detection. , to complete the task of target ranging. Since the midpoint of the bottom edge of the target detection frame 106 is closer to the midpoint of the bottom edge of the target ground truth frame 105 than the midpoint of the bottom edge of the target detection frame 107 , the target obtained based on the target detection shown in (b) in FIG. In the detection frame 107, the error of the target distance obtained by the measurement is larger than the target distance obtained by the measurement in the target detection frame 106 based on the target detection shown in (a) in FIG. 8 .
因此,本申请实施例提供了一种确定目标检测模型的方法,该目标检测模型用于目标检测,从而通过该方法确定的目标检测模型可以提高目标检测的精度,也可以提高目标检测后目标测距的精度。Therefore, the embodiment of the present application provides a method for determining a target detection model, and the target detection model is used for target detection, so that the target detection model determined by this method can improve the accuracy of target detection, and can also improve the target detection after target detection. distance accuracy.
在本申请实施例中,均以待检测图像的左上角为坐标原点,待检测图像从左向右的水平方向为x轴的正方向,待检测图像从上往下的竖直方向为y轴的正方向,待检测图像在x轴方向上的尺寸为待检测图像的宽度,以及待检测图像在y轴方向上的尺寸为待检测图像的高度为例进行描述。In the embodiments of the present application, the upper left corner of the image to be detected is taken as the coordinate origin, the horizontal direction of the image to be detected from left to right is the positive direction of the x-axis, and the vertical direction of the image to be detected from top to bottom is the y-axis In the positive direction, the size of the image to be detected in the x-axis direction is the width of the image to be detected, and the size of the image to be detected in the y-axis direction is the height of the image to be detected.
在本申请实施例中,均以目标检测框(例如,第一目标检测框和第二目标检测框)和目标真值框为矩形框为例进行描述。In the embodiments of the present application, the target detection frame (for example, the first target detection frame and the second target detection frame) and the target ground-truth frame are taken as an example for description.
以下,将结合具体的附图,介绍本申请实施例提供的确定目标检测模型的方法。Hereinafter, the method for determining a target detection model provided by the embodiments of the present application will be introduced with reference to specific drawings.
例如,图9为本申请实施例提供的确定目标检测模型的方法200的示意性流程图。如图9所示,该方法200包括:For example, FIG. 9 is a schematic flowchart of a method 200 for determining a target detection model provided by an embodiment of the present application. As shown in Figure 9, the method 200 includes:
S210,根据第一目标检测模型获取第一目标检测框。其中,第一目标检测框是对待检测图像进行目标检测得到的边界轮廓。S210: Acquire a first target detection frame according to the first target detection model. The first target detection frame is a boundary contour obtained by performing target detection on the image to be detected.
在一些实施例中,可以将待检测图像输入进行第一目标检测模型中,该第一目标检测模型基于相应的算法,检测出目标物在待检测图像中所占区域的坐标值,从而根据该目标物在待检测图像中所占区域的坐标值,得到目标物的边界轮廓,即第一目标检测框。即第一目标检测框可以理解为通过目标检测得到的目标物在待检测图像中的位置。In some embodiments, the image to be detected may be input into a first target detection model, and the first target detection model detects the coordinate value of the area occupied by the target in the image to be detected based on a corresponding algorithm, so that according to the The coordinate value of the area occupied by the target in the image to be detected is used to obtain the boundary contour of the target, that is, the first target detection frame. That is, the first target detection frame can be understood as the position of the target object in the image to be detected obtained through target detection.
在本申请实施例中,目标物均指的是待检测图像中的同一个目标物。In the embodiments of the present application, the targets all refer to the same target in the image to be detected.
示例性地,目标物在待检测图像中所占区域的坐标值可以包括目标物在待检测图像中所占区域的对角的坐标值。例如,目标物在待检测图像中所占区域的对角的坐标值包括目标物在待检测图像中所占区域的左上角的坐标值(x1,y1)和右下角的坐标值(x2,y2)。Exemplarily, the coordinate value of the area occupied by the target object in the image to be detected may include the coordinate value of the diagonal corner of the area occupied by the target object in the image to be detected. For example, the coordinate values of the diagonal corners of the area occupied by the target in the image to be detected include the coordinate values of the upper left corner (x1, y1) and the coordinate value of the lower right corner (x2, y2) of the area occupied by the target in the image to be detected ).
示例性地,目标物在待检测图像中所占区域的坐标值可以包括目标物在待检测图像中所占区域的宽度W 1、高度H 1以及中心点的坐标值(x o1,y o1)。其中,中心点可理解为目标在待检测图像中所占区域的中心对称点。 Exemplarily, the coordinate values of the area occupied by the target in the image to be detected may include the width W 1 , the height H 1 of the area occupied by the target in the image to be detected, and the coordinate values of the center point (x o1 , y o1 ) . Among them, the center point can be understood as the center symmetry point of the area occupied by the target in the image to be detected.
在一些实施例中,第一设备(执行S220的设备)包括摄像头,并通过该摄像头采集待检测图像。在另一些实施例中,第一设备(执行S220的设备)不包括摄像头,第一设备可以从能够获取待检测图像的其他设备上获取。In some embodiments, the first device (the device performing S220 ) includes a camera, and the image to be detected is captured by the camera. In other embodiments, the first device (the device performing S220 ) does not include a camera, and the first device may be obtained from other devices capable of obtaining images to be detected.
在一些实施例中,可以根据第一目标检测模型获取多个第一目标检测框,针对每个第一目标检测框都执行S220至S250。In some embodiments, a plurality of first target detection frames may be acquired according to the first target detection model, and S220 to S250 are performed for each first target detection frame.
S220,确定第一目标检测框的参数信息。其中,第一目标检测框的参数信息包括与对目标物进行目标测距相关的参数。S220: Determine parameter information of the first target detection frame. The parameter information of the first target detection frame includes parameters related to target ranging for the target.
示例性地,在目标测距是基于落地点相似三角形测距法或落地点坐标变换测距法来完成的实施例中,即目标测距需要获取第一目标检测框的底边中点的实施例,第一目标检测框的参数信息可以包括第一目标检测框的底边中心点坐标值。从而可以提高目标测距的精 度。Exemplarily, in the embodiment where the target ranging is based on the landing point similar triangle ranging method or the landing point coordinate transformation ranging method, that is, the target ranging needs to obtain the implementation of the bottom edge midpoint of the first target detection frame. For example, the parameter information of the first target detection frame may include the coordinate value of the center point of the bottom edge of the first target detection frame. Thus, the accuracy of target ranging can be improved.
例如,若目标物在待检测图像中所占区域的坐标值包括目标物在待检测图像中所占区域的左上角的坐标值(x1,y1)和右下角的坐标值(x2,y2),则目标检测框的底边中心点坐标值为(
Figure PCTCN2021079304-appb-000009
y2)。
For example, if the coordinate value of the area occupied by the target in the image to be detected includes the coordinate value (x1, y1) of the upper left corner and the coordinate value (x2, y2) of the lower right corner of the area occupied by the target in the image to be detected, Then the coordinate value of the center point of the bottom edge of the target detection frame is (
Figure PCTCN2021079304-appb-000009
y2).
又例如,若目标物在待检测图像中所占区域的坐标值包括目标物在待检测图像中所占区域的宽度W 1和高度H 1以及中心点的坐标值(x o1,y o1),则目标检测框的底边中心点坐标值为(x o1
Figure PCTCN2021079304-appb-000010
)。
For another example, if the coordinate values of the area occupied by the target in the image to be detected include the width W 1 and height H 1 of the area occupied by the target in the image to be detected and the coordinate values of the center point (x o1 , y o1 ), Then the coordinate value of the center point of the bottom edge of the target detection frame is (x o1 ,
Figure PCTCN2021079304-appb-000010
).
示例性地,在目标测距是基于比例测距法来完成的实施例中,即目标检测需要获取目标检测框的面积(也可以理解为尺寸)的实施例,第一目标检测框的参数信息包括第一目标检测框的面积。从而可以提高目标测距的精度。Exemplarily, in the embodiment in which the target ranging is completed based on the proportional ranging method, that is, in the embodiment in which the target detection needs to obtain the area of the target detection frame (which can also be understood as the size), the parameter information of the first target detection frame is Including the area of the first target detection frame. Thus, the accuracy of target ranging can be improved.
例如,若目标物在待检测图像中所占区域的坐标值包括目标物在待检测图像中所占区域的左上角的坐标值(x1,y1)和右下角的坐标值(x2,y2),则目标检测框的面积为|x2-x1|×|y2-y1|。For example, if the coordinate value of the area occupied by the target in the image to be detected includes the coordinate value (x1, y1) of the upper left corner and the coordinate value (x2, y2) of the lower right corner of the area occupied by the target in the image to be detected, Then the area of the target detection frame is |x2-x1|×|y2-y1|.
又例如,若目标物在待检测图像中所占区域的坐标值包括目标物在待检测图像中所占区域的宽度W 1和高度H 1以及中心点的坐标值(x o1,y o1),则目标检测框的面积为W 1×H 1For another example, if the coordinate values of the area occupied by the target in the image to be detected include the width W 1 and height H 1 of the area occupied by the target in the image to be detected and the coordinate values of the center point (x o1 , y o1 ), Then the area of the target detection frame is W 1 ×H 1 .
S230,确定目标真值框的参数信息。其中,目标真值框是待检测图像中目标物实际的边界轮廓,目标真值框的参数信息包括与对目标物进行目标测距相关的参数。S230: Determine parameter information of the target ground truth box. The target ground-truth frame is the actual boundary contour of the target in the image to be detected, and the parameter information of the target ground-truth frame includes parameters related to target ranging for the target.
在一些实施例中,可以人工将待检测图像中目标物实际所占区域的坐标值标注出来,从而根据该目标物在待检测图像中实际所占区域的坐标值,得到目标物的实际的边界轮廓,即目标真值框。即目标真值框可以理解为目标物在待检测图像中的实际位置。In some embodiments, the coordinate value of the area actually occupied by the target object in the image to be detected may be manually marked, so as to obtain the actual boundary of the target object according to the coordinate value of the area actually occupied by the target object in the image to be detected Contour, that is, the target ground-truth box. That is, the target ground-truth frame can be understood as the actual position of the target in the image to be detected.
示例性地,可以人工标注目标物在待检测图像中实际所占区域的对角的坐标值。例如,目标物在待检测图像中实际所占区域的左上角的坐标值(x3,y3)和右下角的坐标值(x4,y4)。Exemplarily, the coordinate values of the diagonal corners of the area actually occupied by the target in the image to be detected may be manually marked. For example, the coordinates of the upper left corner (x3, y3) and the coordinates of the lower right corner (x4, y4) of the area actually occupied by the target in the image to be detected.
示例性地,可以人工标注目标物在待检测图像中实际所占区域的宽度W 2、高度H 2以及中心点的坐标值(x o2,y o2)。 Exemplarily, the width W 2 , the height H 2 and the coordinate values (x o2 , y o2 ) of the area actually occupied by the target in the image to be detected may be manually marked.
在目标测距是基于落地点相似三角形测距法或落地点坐标变换测距法来完成的实施例中,即目标测距需要获取目标检测框的底边中点的实施例,目标真值框的参数信息包括目标真值框的底边中心点坐标值。从而可以提高目标测距的精度。In the embodiment in which the target ranging is completed based on the landing point similar triangle ranging method or the landing point coordinate transformation ranging method, that is, the embodiment in which the target ranging needs to obtain the midpoint of the bottom edge of the target detection frame, the target true value frame The parameter information includes the coordinate value of the center point of the bottom edge of the target ground truth box. Thus, the accuracy of target ranging can be improved.
例如,若目标物在待检测图像中实际所占区域的坐标值包括目标物在待检测图像中实际所占区域的左上角的坐标值(x3,y3)和右下角的坐标值(x4,y4),则目标真值框的底边中心点坐标值为(
Figure PCTCN2021079304-appb-000011
y3)。
For example, if the coordinate value of the area actually occupied by the target in the image to be detected includes the coordinate value of the upper left corner (x3, y3) and the coordinate value of the lower right corner (x4, y4) of the area actually occupied by the target in the image to be detected ), then the coordinate value of the center point of the bottom edge of the target ground-truth box is (
Figure PCTCN2021079304-appb-000011
y3).
又例如,若目标物在待检测图像中实际所占区域的对角的坐标值包括目标物在待检测图像中实际所占区域的宽度L 2和高度w 2以及中心点的坐标值(x o2,y o2),则目标真值框的底边中心点坐标值为(x o2
Figure PCTCN2021079304-appb-000012
)。
For another example, if the coordinate values of the diagonal corners of the area actually occupied by the target in the image to be detected include the width L 2 and height w 2 of the area actually occupied by the target in the image to be detected, and the coordinate value of the center point (x o2 ) , y o2 ), then the coordinate value of the center point of the bottom edge of the target truth box is (x o2 ,
Figure PCTCN2021079304-appb-000012
).
在目标测距是基于比例测距法来完成的实施例中,即目标检测需要获取目标检测框的面积(也可以理解为尺寸)的实施例,目标真值框的参数信息包括目标真值框的面积。从而可以提高目标测距的精度。In the embodiment in which the target ranging is completed based on the proportional ranging method, that is, the target detection needs to obtain the area of the target detection frame (which can also be understood as the size), and the parameter information of the target ground-truth frame includes the target ground-truth frame. area. Thus, the accuracy of target ranging can be improved.
例如,若目标物在待检测图像中实际所占区域的坐标值包括目标物在待检测图像中实际所占区域的左上角的坐标值(x3,y3)和右下角的坐标值(x4,y4),则目标真值框的面积为|x3-x4|×|y3-y4|。For example, if the coordinate value of the area actually occupied by the target in the image to be detected includes the coordinate value of the upper left corner (x3, y3) and the coordinate value of the lower right corner (x4, y4) of the area actually occupied by the target in the image to be detected ), then the area of the target ground-truth box is |x3-x4|×|y3-y4|.
又例如,若目标物在待检测图像中实际所占区域的坐标值包括目标物在待检测图像中实际所占区域的宽度L 2和高度W 2以及中心点的坐标值(x o2,y o2),则目标真值框的底边中心点坐标值为W 2×H 2For another example, if the coordinate values of the area actually occupied by the target in the image to be detected include the width L 2 and height W 2 of the area actually occupied by the target in the image to be detected, and the coordinates of the center point (x o2 , y o2 ), then the coordinate value of the center point of the bottom edge of the target ground-truth box is W 2 ×H 2 .
可选地,本申请实施例对S220和S230的先后顺序不作限定。Optionally, the order of S220 and S230 is not limited in this embodiment of the present application.
S240,根据第一目标检测框的参数信息和目标真值框的参数信息,确定损失项。其中,损失项用于指示第一目标检测框和目标真值框之间的偏差。S240: Determine a loss item according to the parameter information of the first target detection frame and the parameter information of the target ground-truth frame. Among them, the loss term is used to indicate the deviation between the first target detection frame and the target ground-truth frame.
以下,将以方式A、方式B、方式C和方式D为例,具体介绍S240。Hereinafter, S240 will be described in detail by taking Mode A, Mode B, Mode C, and Mode D as examples.
方式A,根据第一目标检测框的底边中心点坐标值和目标真值框的底边中心点坐标值,确定损失项。Mode A, the loss term is determined according to the coordinate value of the center point of the bottom edge of the first target detection frame and the coordinate value of the center point of the bottom edge of the target true value frame.
根据以下公式确定损失项:The loss term is determined according to the following formula:
Figure PCTCN2021079304-appb-000013
Figure PCTCN2021079304-appb-000013
其中,c点为第一目标检测框的底边中心点,c gt点为目标真值框的底边中心点,ρ(c,c gt)为c点和c gt点之间的距离,H 2为目标真值框的的高度。 Among them, point c is the center point of the bottom edge of the first target detection frame, point c gt is the center point of the bottom edge of the target ground-truth frame, ρ(c, c gt ) is the distance between point c and point c gt , H 2 is the height of the target ground truth box.
方式B,根据第一目标检测框的底边中心点坐标值、目标真值框的底边中心点坐标值和最小外接矩形框的对角线的长度,确定损失项。Mode B, the loss term is determined according to the coordinate value of the bottom center point of the first target detection frame, the coordinate value of the bottom center point of the target true value frame, and the length of the diagonal line of the minimum circumscribed rectangular frame.
该方式B具体包括S11和S12。The mode B specifically includes S11 and S12.
S11,确定第一目标检测框与目标真值框的最小外接矩形框的对角线的长度。S11: Determine the length of the diagonal of the smallest circumscribed rectangular frame between the first target detection frame and the target ground truth frame.
具体地,根据上述第一目标检测框的坐标值和上述目标真值框的坐标值,确定第一目标检测框和目标真值框的最小外接矩形框的坐标值,并根据最小外接矩形框的坐标值,确定最小外接矩形框的对角线的长度。Specifically, according to the coordinate value of the first target detection frame and the coordinate value of the target truth value frame, determine the coordinate value of the minimum bounding rectangle frame of the first target detection frame and the target truth value frame, and according to the minimum bounding rectangle frame Coordinate value to determine the length of the diagonal of the smallest bounding rectangle.
示例性地,若第一目标检测框的坐标值包括左上角的坐标值为(x1,y1),右下角的坐标值为(x2,y2)。若目标真值框的坐标值包括左上角的坐标值(x3,y3)和右下角的坐标值(x4,y4)。则第一目标检测框和目标真值框的最小外接矩形框的左上角的坐标值为(min[x1,x3],min[y1,y3]),右下角的坐标值为(max[x2,x4],max[y2,y4])。从而最小外接矩形框的对角线的长度为:Exemplarily, if the coordinate value of the first target detection frame includes the coordinate value of the upper left corner (x1, y1), and the coordinate value of the lower right corner (x2, y2). If the coordinate value of the target ground-truth box includes the coordinate value of the upper left corner (x3, y3) and the coordinate value of the lower right corner (x4, y4). Then the coordinate value of the upper left corner of the minimum bounding rectangle of the first target detection frame and the target ground truth frame is (min[x1, x3], min[y1, y3]), and the coordinate value of the lower right corner is (max[x2, x4], max[y2, y4]). Thus the length of the diagonal of the smallest bounding rectangle is:
Figure PCTCN2021079304-appb-000014
Figure PCTCN2021079304-appb-000014
示例性地,若第一目标检测框的坐标值包括第一目标检测框的宽度、高度以及中心点的坐标值,可以根据第一目标检测框的宽度、高度以及中心点的坐标值确定出第一目标检测框的左上角的坐标值和右下角的坐标值。若目标真值框的坐标值包括目标真值框的宽度、高度以及中心点的坐标值,可以根据目标真值框的宽度、高度以及中心点的坐标值确定出目标真值框的左上角的坐标值和右下角的坐标值。从而可以根据上述示例,确定第一目标检测框和目标真值框的最小外接矩形框的对角线的长度。Exemplarily, if the coordinate value of the first target detection frame includes the width, height and coordinate value of the center point of the first target detection frame, the first target detection frame can be determined according to the width, height and coordinate value of the center point. The coordinate value of the upper left corner and the coordinate value of the lower right corner of a target detection frame. If the coordinate value of the target truth value box includes the width and height of the target truth value box and the coordinate value of the center point, the upper left corner of the target truth value box can be determined according to the width, height and coordinate value of the center point of the target truth value box. The coordinate value and the coordinate value of the lower right corner. Therefore, according to the above example, the length of the diagonal of the minimum circumscribed rectangular frame of the first target detection frame and the target ground-truth frame can be determined.
S12,根据以下公式确定损失项:S12, the loss term is determined according to the following formula:
Figure PCTCN2021079304-appb-000015
Figure PCTCN2021079304-appb-000015
其中,c点为第一目标检测框的底边中心点,c gt点为目标真值框的底边中心点,ρ(c,c gt)为c点和c gt点之间的距离,s为最小外接矩形框的对角线的长度。 Among them, point c is the center point of the bottom edge of the first target detection frame, point c gt is the center point of the bottom edge of the target ground-truth frame, ρ(c, c gt ) is the distance between point c and point c gt , s is the length of the diagonal of the smallest bounding rectangle.
例如,图10为本申请提供的一组第一目标检测框和目标真值框的示意图。如图10所示,其中,目标真值框107的左上角B1点的坐标值为(x b1,y b1)和右下角B2点的坐标值为(x b2,y b2),则目标真值框107的底边中心点c gt点的坐标值为(
Figure PCTCN2021079304-appb-000016
y b2)。第一目标检测框108的左上角B3点的坐标值为(x b3,y b3)和右下角B4点的坐标值为(x b4,y b4),则第一目标检测框108的底边中心点c点的坐标值为(
Figure PCTCN2021079304-appb-000017
y b4)。从而c点和c gt点之间的距离
Figure PCTCN2021079304-appb-000018
此外,目标真值框107和第一目标检测框108的最小外接矩形框为目标真值框107,则目标真值框107和第一目标检测框108的最小外接矩形框的对角线的长度
Figure PCTCN2021079304-appb-000019
则按照方式B可得:目标真值框107和第一目标检测框108之间的损失项为
Figure PCTCN2021079304-appb-000020
For example, FIG. 10 is a schematic diagram of a set of first target detection frames and target ground-truth frames provided by this application. As shown in FIG. 10 , where the coordinate value of the upper left corner B1 of the target truth value box 107 is (x b1 , y b1 ) and the coordinate value of the lower right corner B2 point is (x b2 , y b2 ), then the target truth value is The coordinate value of the center point c gt of the bottom edge of the frame 107 is (
Figure PCTCN2021079304-appb-000016
y b2 ). The coordinate value of the upper left corner B3 of the first target detection frame 108 is (x b3 , y b3 ) and the coordinate value of the lower right corner B4 point is (x b4 , y b4 ), then the center of the bottom edge of the first target detection frame 108 The coordinate value of point c is (
Figure PCTCN2021079304-appb-000017
y b4 ). Thus the distance between point c and point c gt
Figure PCTCN2021079304-appb-000018
In addition, the minimum bounding rectangle of the target ground-truth frame 107 and the first target detection frame 108 is the target ground-truth frame 107, then the length of the diagonal of the minimum bounding rectangle of the target ground-truth frame 107 and the first target detection frame 108
Figure PCTCN2021079304-appb-000019
Then according to method B, it can be obtained: the loss term between the target ground-truth frame 107 and the first target detection frame 108 is
Figure PCTCN2021079304-appb-000020
方式C,根据第一目标检测框的面积、目标真值框的面积和最小外接矩形框的面积,确定损失项。Mode C, the loss term is determined according to the area of the first target detection frame, the area of the target ground-truth frame, and the area of the minimum circumscribed rectangular frame.
S21,确定第一目标检测框与目标真值框的最小外接矩形框的面积。S21: Determine the area of the smallest circumscribed rectangular frame between the first target detection frame and the target ground truth frame.
具体地,根据上述第一目标检测框的坐标值和上述目标真值框的坐标值,确定第一目标检测框与目标真值框的最小外接矩形框的坐标值,并根据最小外接矩形框的坐标值,确定最小外接矩形框的面积。Specifically, according to the coordinate value of the first target detection frame and the coordinate value of the target truth value frame, determine the coordinate value of the minimum enclosing rectangular frame of the first target detection frame and the target truth value frame, and according to the minimum enclosing rectangle frame Coordinate value to determine the area of the smallest bounding rectangle.
示例性地,若第一目标检测框的坐标值包括左上角的坐标值为(x1,y1),右下角的坐标值为(x2,y2)。若目标真值框的坐标值包括左上角的坐标值(x3,y3)和右下角的坐标值(x4,y4)。则第一目标检测框和目标真值框的最小外接矩形框的左上角的坐标值为(min[x1,x3],min[y1,y3]),右下角的坐标值为(max[x2,x4],max[y2,y4])。从而最小外接矩形框的面积为:Exemplarily, if the coordinate value of the first target detection frame includes the coordinate value of the upper left corner (x1, y1), and the coordinate value of the lower right corner (x2, y2). If the coordinate value of the target ground-truth box includes the coordinate value of the upper left corner (x3, y3) and the coordinate value of the lower right corner (x4, y4). Then the coordinate value of the upper left corner of the minimum bounding rectangle of the first target detection frame and the target ground truth frame is (min[x1, x3], min[y1, y3]), and the coordinate value of the lower right corner is (max[x2, x4], max[y2, y4]). Thus the area of the smallest bounding rectangle is:
|min[x1,x3]-max[x2,x4]|×|min[y1,y3]-max[y2,y4]||min[x1,x3]-max[x2,x4]|×|min[y1,y3]-max[y2,y4]|
示例性地,若第一目标检测框的坐标值包括第一目标检测框的宽度、高度以及中心点的坐标值,可以根据第一目标检测框的宽度、高度以及中心点的坐标值确定出第一目标检测框的左上角的坐标值和右下角的坐标值。若目标真值框的坐标值包括目标真值框的宽度、高度以及中心点的坐标值,可以根据目标真值框的宽度、高度以及中心点的坐标值确定出目标真值框的左上角的坐标值和右下角的坐标值。从而可以根据上述示例,确定第一目标检测框和目标真值框的最小外接矩形框的面积。Exemplarily, if the coordinate value of the first target detection frame includes the width, height and coordinate value of the center point of the first target detection frame, the first target detection frame can be determined according to the width, height and coordinate value of the center point. The coordinate value of the upper left corner and the coordinate value of the lower right corner of a target detection frame. If the coordinate value of the target truth value box includes the width and height of the target truth value box and the coordinate value of the center point, the upper left corner of the target truth value box can be determined according to the width, height and coordinate value of the center point of the target truth value box. The coordinate value and the coordinate value of the lower right corner. Therefore, according to the above example, the area of the minimum enclosing rectangular frame of the first target detection frame and the target ground-truth frame can be determined.
S22,根据以下公式确定损失项:S22, the loss term is determined according to the following formula:
Figure PCTCN2021079304-appb-000021
Figure PCTCN2021079304-appb-000021
其中,a1为第一目标检测框的面积,a2为目标真值框的面积,a3为最小外接矩形框的面积。Among them, a1 is the area of the first target detection frame, a2 is the area of the target ground-truth frame, and a3 is the area of the minimum bounding rectangle.
例如,图11为本申请提供的一组第一目标检测框和目标真值框的示意图。如图11所示,其中,目标真值框109的左上角C1点的坐标值为(x c1,y c1)和右下角C2点的坐标值为(x c2,y c2)。则目标真值框109的面积a2为|x c1-x c2|×|y c1-y c2|。第一目标检测框110的左上角C3点的坐标值为(x c3,y c3)和右下角C4点的坐标值为(x c4,y c4),则第一目标检测框110的面积a1为|x c3-x c4|×|y c3-y c4|。此外,目标真值框109和第一目标检测框110的最小外接矩形框为框111,则最小外接矩形框111的面积a3为|x c1-x c4|×|y c1-y c2|。则按照方式c可得:目标真值框109和第一目标检测框110之间损失项为
Figure PCTCN2021079304-appb-000022
For example, FIG. 11 is a schematic diagram of a set of first target detection frames and target ground-truth frames provided by this application. As shown in FIG. 11 , the coordinate values of the upper left corner C1 of the target ground truth frame 109 are (x c1 , y c1 ) and the coordinate values of the lower right corner C2 point are (x c2 , y c2 ). Then the area a2 of the target ground truth box 109 is |x c1 -x c2 |×|y c1 -y c2 |. The coordinate value of point C3 in the upper left corner of the first target detection frame 110 is (x c3 , y c3 ) and the coordinate value of point C4 in the lower right corner is (x c4 , y c4 ), then the area a1 of the first target detection frame 110 is |x c3 -x c4 |×|y c3 -y c4 |. In addition, the minimum circumscribed rectangle frame of the target ground truth frame 109 and the first target detection frame 110 is frame 111, and the area a3 of the minimum circumscribed rectangular frame 111 is |x c1 -x c4 |×|y c1 -y c2 |. Then according to method c, we can obtain: the loss term between the target ground-truth frame 109 and the first target detection frame 110 is:
Figure PCTCN2021079304-appb-000022
方式D,根据第一目标检测框的面积和目标真值框的面积,确定损失项。Mode D, the loss term is determined according to the area of the first target detection frame and the area of the target ground-truth frame.
根据以下公式确定损失项:The loss term is determined according to the following formula:
Figure PCTCN2021079304-appb-000023
Figure PCTCN2021079304-appb-000023
其中,a1为第一目标检测框的面积,a2为目标真值框的面积。Among them, a1 is the area of the first target detection frame, and a2 is the area of the target ground-truth frame.
在一些实施例中,还可以通过现有技术中任一种确定损失项的方式(例如,上述方式1或方式2)、方式A、方式B、方式C、方式D中的至少两种方式,确定损失项。In some embodiments, at least two ways of determining the loss item in the prior art (for example, the above-mentioned way 1 or way 2), way A, way B, way C, and way D may also be used, Determine the loss term.
S250,根据损失项确定第二目标检测模型,第二目标检测模型用于确定第二目标检测框。S250: Determine a second target detection model according to the loss term, where the second target detection model is used to determine a second target detection frame.
在一些实施例中,若确定的损失项小于预设值,即认为当前损失项对应的目标检测模型检测的目标的位置十分接近目标的真实位置。此时,根据S240得到的损失项,确定的第二目标检测模型即为目标检测模型。其中,第二目标检测模型即为第一目标检测模型,若基于第二目标检测模型对S210中的待检测图像再次进行目标检测,得到的目标物的第二目标检测框即为第一目标检测框。In some embodiments, if the determined loss item is smaller than the preset value, it is considered that the position of the target detected by the target detection model corresponding to the current loss item is very close to the real position of the target. At this time, according to the loss term obtained in S240, the determined second target detection model is the target detection model. The second target detection model is the first target detection model. If target detection is performed again on the image to be detected in S210 based on the second target detection model, the obtained second target detection frame of the target is the first target detection frame.
在另一些实施例中,若确定的损失项大于或等于预设值,则需要重复执行S210至S240,对第一目标检测模型进行N(N≥1,N为正整数)次修正,直到基于第N次修正后的第一目标检测模型得到的损失项小于预设值,即认为第N次修正后的第一目标检测模型(例如,第二目标检测模型)检测的目标的位置十分接近目标的真实位置。此时,第二目标检测模型即为目标检测模型。其中,第二目标检测模型不是第一目标检测模型。且若基于第二目标检测模型对S210中的待检测图像再次进行目标检测,得到的目标物的第二目标检测框和第一目标检测框不相同。且相比于第一目标检测框,第二目标检测框更贴近目标真值框。In other embodiments, if the determined loss term is greater than or equal to the preset value, S210 to S240 need to be repeatedly performed to perform N (N≥1, N is a positive integer) corrections on the first target detection model until the first target detection model is corrected based on The loss term obtained by the first target detection model after the Nth revision is smaller than the preset value, that is, it is considered that the position of the target detected by the first target detection model (for example, the second target detection model) after the Nth revision is very close to the target real location. At this time, the second target detection model is the target detection model. Wherein, the second target detection model is not the first target detection model. And if target detection is performed again on the image to be detected in S210 based on the second target detection model, the obtained second target detection frame of the target object is different from the first target detection frame. And compared with the first target detection frame, the second target detection frame is closer to the target ground-truth frame.
例如,若确定的损失项大于或等于预设值,则需要根据该损失项对第一目标检测模型进行第一次修正,得到第1次修正后的第一目标检测模型(例如,第三目标检测模型),并将S210至S240中的第一目标检测模型用第三目标检测模型替换,重复执行S210至S240,再次得到损失项。若此时确定的损失项仍然大于或等于预设值,则需要对第一目标检测模型进行第二次修正,得到第2次修正后的第一目标检测模型(例如,第二目标检测模型),并将S210至S240中的第一目标检测模型用第二目标检测模型替换,重复执行S210至S240,再次得到损失项,若此时确定的损失项小于预设值。此时,根据S240得到的损失项,确定的第二目标检测模型即为目标检测模型。其中,第二目标检测模型不是第 一目标检测模型。且若基于第二目标检测模型对S210中的待检测图像再次进行目标检测,得到的目标物的第二目标检测框和第一目标检测框不相同。且相比于第一目标检测框,第二目标检测框更贴近目标真值框。For example, if the determined loss item is greater than or equal to the preset value, the first target detection model needs to be modified for the first time according to the loss item to obtain the first modified first target detection model (for example, the third target detection model). detection model), and replace the first target detection model in S210 to S240 with the third target detection model, and repeat S210 to S240 to obtain the loss term again. If the loss term determined at this time is still greater than or equal to the preset value, the first target detection model needs to be revised for the second time to obtain the second revised first target detection model (for example, the second target detection model) , and replace the first target detection model in S210 to S240 with the second target detection model, and repeat S210 to S240 to obtain the loss item again, if the loss item determined at this time is less than the preset value. At this time, according to the loss term obtained in S240, the determined second target detection model is the target detection model. Wherein, the second target detection model is not the first target detection model. And if target detection is performed again on the image to be detected in S210 based on the second target detection model, the obtained second target detection frame of the target object is different from the first target detection frame. And compared with the first target detection frame, the second target detection frame is closer to the target ground-truth frame.
例如,图12是本申请提供的一组目标检测框和目标真值框的示意图。其中,如图12中的(a)所示的目标检测框和目标真值框是基于现有技术的目标检测得到的。如图12中的(b)所示的目标检测框是基于上述方式B确定的损失项小于预设值时所得到。For example, FIG. 12 is a schematic diagram of a set of target detection boxes and target ground-truth boxes provided by this application. Among them, the target detection frame and the target ground-truth frame shown in (a) of FIG. 12 are obtained based on the target detection of the prior art. The target detection frame shown in (b) of FIG. 12 is obtained when the loss term determined based on the above-mentioned method B is smaller than the preset value.
又例如,图13是本申请提供的另一组目标检测框和目标真值框的示意图。其中,如图13中的(a)所示的目标检测框和目标真值框是基于现有技术的目标检测得到的。如图13中的(b)所示的目标检测框是基于上述方式C确定的损失项小于预设值时所得到。For another example, FIG. 13 is a schematic diagram of another set of target detection frames and target ground-truth frames provided by this application. Among them, the target detection frame and the target ground-truth frame shown in (a) of FIG. 13 are obtained based on the target detection of the prior art. The target detection frame shown in (b) of FIG. 13 is obtained when the loss term determined based on the above-mentioned method C is smaller than the preset value.
通过重复执行上述S210至S240,直到损失项小于预设值的过程可以称为第一目标检测模型的训练过程。通过训练好的第一目标检测模型(例如,第二目标检测模型),可以对任意待检测的图形进行目标检测。The process of repeatedly performing the above S210 to S240 until the loss term is smaller than the preset value may be referred to as the training process of the first target detection model. Through the trained first target detection model (for example, the second target detection model), target detection can be performed on any graphics to be detected.
示例性地,可以将训练好的第一目标检测模型应用到终端上。Exemplarily, the trained first target detection model may be applied to the terminal.
进一步地,该终端可以为智能运输设备(车辆或者无人机)、智能家居设备、智能制造设备或者机器人等。该智能运输设备例如可以是AGV或无人运输车。Further, the terminal may be an intelligent transportation device (vehicle or drone), a smart home device, an intelligent manufacturing device, or a robot, and the like. The intelligent transportation device can be, for example, an AGV or an unmanned transportation vehicle.
示例性地,将训练好的第一目标检测模型应用到终端之后,可进行目标测距任务。Exemplarily, after the trained first target detection model is applied to the terminal, the target ranging task can be performed.
示例性地,为了提高目标测距的精度,在基于上述方式A或方式B确定的目标检测模型(例如,第二目标检测模型),可基于落地点测距相关算法进行目标测距任务。例如,落地点测距可以是落地点相似三角形测距法或落地点坐标变换测距法。Exemplarily, in order to improve the accuracy of target ranging, in the target detection model (eg, the second target detection model) determined based on the above method A or method B, the target ranging task can be performed based on the landing point ranging related algorithm. For example, the landing point ranging may be the landing point similarity triangle ranging method or the landing point coordinate transformation ranging method.
示例性地,为了提高目标测距的精度,在基于上述方式C或方式D确定的目标检测模型(例如,第二目标检测模型),可基于比例测距法相关算法进行目标测距任务。Exemplarily, in order to improve the accuracy of target ranging, in the target detection model (eg, the second target detection model) determined based on the above method C or method D, the target ranging task can be performed based on a proportional ranging method related algorithm.
由于在本申请实施例提供的确定目标检测模型的方法200中,是基于与目标测距相关的参数(目标检测框的底边中心点坐标或目标检测框的面积),确定损失项,并修正目标检测模型,完成目标检测模型的训练。因此,该方法200可以提高目标检测精度,还可以提高目标测距的精度。Because in the method 200 for determining a target detection model provided by this embodiment of the present application, the loss term is determined based on the parameters related to target ranging (the coordinates of the bottom center point of the target detection frame or the area of the target detection frame), and the correction is made. The target detection model completes the training of the target detection model. Therefore, the method 200 can improve the accuracy of target detection, and can also improve the accuracy of target ranging.
例如,图14是基于不同目标检测模型进行目标测距的对比示意图。其中,如果基于现有技术的目标检测模型进行目标测距,确定的车辆112距离车辆111的距离为Z1。如果基于上述方式B确定的目标检测模型进行目标测距,确定的车辆112距离车辆111的距离为Z。由于在本申请实施例提供的上述方式B中,目标检测模型是基于与目标测距相关的参数,即目标检测框的底边中心点坐标值,确定的损失项。且在后续目标测距中,也是基于目标检测框的底边中心点坐标值,这样,可以更加精确地检测出目标的位置,从而对目标进行测距。For example, FIG. 14 is a schematic diagram of a comparison of target ranging based on different target detection models. Wherein, if the target ranging is performed based on the target detection model of the prior art, the determined distance between the vehicle 112 and the vehicle 111 is Z1. If target ranging is performed based on the target detection model determined in the above manner B, the determined distance between the vehicle 112 and the vehicle 111 is Z. Because in the above-mentioned method B provided by the embodiment of the present application, the target detection model is a loss item determined based on parameters related to target ranging, that is, the coordinate value of the center point of the bottom edge of the target detection frame. And in the subsequent target ranging, it is also based on the coordinate value of the center point of the bottom edge of the target detection frame, so that the position of the target can be detected more accurately, so that the target can be measured.
上文结合图1至图14的描述了本申请实施例的确定目标检测模型的方法,下面结合图15和图16,描述本申请实施例的确定目标检测模型的装置。应理解,确定目标检测模型的装置的描述与确定目标检测模型的方法的描述相互对应,因此,未详细描述的部分可以参见前面对确定目标检测模型的方法的说明。The method for determining the target detection model of the embodiment of the present application is described above with reference to FIGS. 1 to 14 , and the apparatus for determining the target detection model of the embodiment of the present application is described below with reference to FIGS. 15 and 16 . It should be understood that the description of the apparatus for determining the target detection model corresponds to the description of the method for determining the target detection model. Therefore, for parts not described in detail, reference may be made to the foregoing description of the method for determining the target detection model.
图15是本申请实施例提供的一种确定目标检测模型的装置的示意性结构图。如图15所示,该确定目标检测模型的装置300包括获取模块310和处理模块320,其中,FIG. 15 is a schematic structural diagram of an apparatus for determining a target detection model provided by an embodiment of the present application. As shown in FIG. 15, the apparatus 300 for determining a target detection model includes an acquisition module 310 and a processing module 320, wherein,
获取单元310,用于根据第一目标检测模型获取第一目标检测框,所述第一目标检测 框是对待检测图像进行目标检测得到的边界轮廓;Obtaining unit 310 is used to obtain the first target detection frame according to the first target detection model, and the first target detection frame is the boundary contour that the target detection is carried out to the image to be detected;
处理单元320,用于确定所述第一目标检测框的参数信息,所述第一目标检测框的参数信息包括与对目标物进行目标测距相关的参数;a processing unit 320, configured to determine parameter information of the first target detection frame, where the parameter information of the first target detection frame includes parameters related to target ranging;
所述处理单元320,还用于确定目标真值框的参数信息,所述目标真值框是所述待检测图像中所述目标物实际的边界轮廓,所述目标真值框的参数信息包括与对所述目标物进行目标测距相关的参数;The processing unit 320 is further configured to determine parameter information of the target ground-truth frame, where the target ground-truth frame is the actual boundary contour of the target in the to-be-detected image, and the parameter information of the target ground-truth frame includes: parameters related to target ranging for the target;
所述处理单元320,还用于根据所述第一目标检测框的参数信息和所述目标真值框的参数信息,确定损失项,所述损失项用于指示所述第一目标检测框和所述目标真值框之间的偏差;The processing unit 320 is further configured to determine a loss item according to the parameter information of the first target detection frame and the parameter information of the target ground-truth frame, where the loss item is used to indicate the first target detection frame and the deviation between the target ground-truth boxes;
所述处理单元320,还用于根据所述损失项确定第二目标检测模型,所述第二目标检测模型用于确定第二目标检测框。The processing unit 320 is further configured to determine a second target detection model according to the loss term, where the second target detection model is used to determine a second target detection frame.
可选地,所述第一目标检测框的参数信息包括所述第一目标检测框的底边中心点坐标值,所述目标真值框的参数信息包括所述目标真值框的底边中心点坐标值。Optionally, the parameter information of the first target detection frame includes the coordinate value of the bottom center point of the first target detection frame, and the parameter information of the target truth frame includes the bottom center of the target truth frame. Point coordinate value.
可选地,所述目标测距为采用落地点测距法进行的测距。Optionally, the target ranging is a ranging method using a landing point ranging method.
可选地,所述第一目标检测框的参数信息包括所述第一目标检测框的面积,所述目标真值框的参数信息包括所述目标真值框的面积。Optionally, the parameter information of the first target detection frame includes the area of the first target detection frame, and the parameter information of the target ground truth frame includes the area of the target ground truth frame.
可选地,所述目标测距为采用比例测距法进行的测距。Optionally, the target ranging is ranging by using a proportional ranging method.
可选地,所述处理单元320,还具体用于:根据所述第一目标检测框的参数信息、所述目标真值框的参数信息以及所述第一目标检测框与所述目标真值框的最小外接矩形框的参数,确定所述损失项。Optionally, the processing unit 320 is further specifically configured to: according to the parameter information of the first target detection frame, the parameter information of the target ground-truth frame, and the first target detection frame and the target ground-truth The parameter of the minimum enclosing rectangular box of the box, which determines the loss term.
可选地,所述最小外接矩形框的参数包括所述最小外接矩形框的对角线的长度。Optionally, the parameter of the minimum enclosing rectangle includes the length of the diagonal of the minimum enclosing rectangle.
可选地,所述处理单元320,还具体用于:根据以下公式,确定所述损失项:Optionally, the processing unit 320 is further specifically configured to: determine the loss item according to the following formula:
Figure PCTCN2021079304-appb-000024
Figure PCTCN2021079304-appb-000024
其中,c点为所述目标检测框的底边中心点,c gt点为所述目标真值框的底边中心点,所述ρ(c,c gt)为所述c点和所述c gt点之间的距离,所述s为所述最小外接矩形框的对角线的长度。 Wherein, point c is the center point of the bottom edge of the target detection frame, point c gt is the center point of the bottom edge of the target ground-truth frame, and the ρ(c,c gt ) is the point c and the point c gt is the distance between points, and the s is the length of the diagonal of the minimum enclosing rectangle.
可选地,所述最小外接矩形框的参数包括所述最小外接矩形框的面积。Optionally, the parameter of the minimum enclosing rectangle includes the area of the minimum enclosing rectangle.
可选地,所述处理单元320,还具体用于:根据以下公式,确定所述损失项:Optionally, the processing unit 320 is further specifically configured to: determine the loss item according to the following formula:
Figure PCTCN2021079304-appb-000025
Figure PCTCN2021079304-appb-000025
其中,a1为所述目标检测框的面积,a2为所述目标真值框的面积,所述a3为所述最小外接矩形框的面积。Wherein, a1 is the area of the target detection frame, a2 is the area of the target ground-truth frame, and a3 is the area of the minimum circumscribed rectangular frame.
可选地,所述目标物实际的边界轮廓是由人工标注的。Optionally, the actual boundary contour of the target is manually marked.
图16是本申请实施例提供的另一种确定目标检测模型的装置的示意性结构图。FIG. 16 is a schematic structural diagram of another apparatus for determining a target detection model provided by an embodiment of the present application.
确定目标检测模型的装置400包括至少一个存储器410和至少一个处理器420,所述至少一个存储器410用于存储程序,所述至少一个处理器420用于运行所述程序,以实现前文所述的方法200。The apparatus 400 for determining a target detection model includes at least one memory 410 and at least one processor 420, the at least one memory 410 is used for storing a program, and the at least one processor 420 is used for running the program to realize the aforementioned method 200.
应理解,本申请实施例中的处理器可以为中央处理单元(central processing unit,CPU), 该处理器还可以是其他通用处理器、数字信号处理器(digital signal processor,DSP)、专用集成电路(application specific integrated circuit,ASIC)、现成可编程门阵列(field programmable gate array,FPGA)或者其他可编程逻辑器件、分立门或者晶体管逻辑器件、分立硬件组件等。通用处理器可以是微处理器或者该处理器也可以是任何常规的处理器等。It should be understood that the processor in the embodiments of the present application may be a central processing unit (central processing unit, CPU), and the processor may also be other general-purpose processors, digital signal processors (digital signal processors, DSP), application-specific integrated circuits (application specific integrated circuit, ASIC), off-the-shelf programmable gate array (field programmable gate array, FPGA) or other programmable logic devices, discrete gate or transistor logic devices, discrete hardware components, etc. A general purpose processor may be a microprocessor or the processor may be any conventional processor or the like.
还应理解,本申请实施例中的存储器可以是易失性存储器或非易失性存储器,或可包括易失性和非易失性存储器两者。其中,非易失性存储器可以是只读存储器(read-only memory,ROM)、可编程只读存储器(programmable ROM,PROM)、可擦除可编程只读存储器(erasable PROM,EPROM)、电可擦除可编程只读存储器(electrically EPROM,EEPROM)或闪存。易失性存储器可以是随机存取存储器(random access memory,RAM),其用作外部高速缓存。通过示例性但不是限制性说明,许多形式的随机存取存储器(random access memory,RAM)可用,例如静态随机存取存储器(static RAM,SRAM)、动态随机存取存储器(DRAM)、同步动态随机存取存储器(synchronous DRAM,SDRAM)、双倍数据速率同步动态随机存取存储器(double data rate SDRAM,DDR SDRAM)、增强型同步动态随机存取存储器(enhanced SDRAM,ESDRAM)、同步连接动态随机存取存储器(synchlink DRAM,SLDRAM)和直接内存总线随机存取存储器(direct rambus RAM,DR RAM)。It should also be understood that the memory in the embodiments of the present application may be volatile memory or non-volatile memory, or may include both volatile and non-volatile memory. The non-volatile memory may be read-only memory (ROM), programmable read-only memory (PROM), erasable programmable read-only memory (EPROM), electrically programmable Erase programmable read-only memory (electrically EPROM, EEPROM) or flash memory. Volatile memory may be random access memory (RAM), which acts as an external cache. By way of example and not limitation, many forms of random access memory (RAM) are available, such as static random access memory (SRAM), dynamic random access memory (DRAM), synchronous dynamic random access memory (DRAM) Access memory (synchronous DRAM, SDRAM), double data rate synchronous dynamic random access memory (double data rate SDRAM, DDR SDRAM), enhanced synchronous dynamic random access memory (enhanced SDRAM, ESDRAM), synchronous connection dynamic random access memory Fetch memory (synchlink DRAM, SLDRAM) and direct memory bus random access memory (direct rambus RAM, DR RAM).
上述各个附图对应的流程的描述各有侧重,某个流程中没有详述的部分,可以参见其他流程的相关描述。The descriptions of the processes corresponding to the above figures have their own emphasis, and for parts that are not described in detail in a certain process, please refer to the relevant descriptions of other processes.
本申请实施例还提供一种计算机可读存储介质,所述计算机可读存储介质具有程序指令,当所述程序指令被直接或者间接执行时,使得前文中的方法得以实现。Embodiments of the present application further provide a computer-readable storage medium, where the computer-readable storage medium has program instructions, and when the program instructions are directly or indirectly executed, the foregoing method can be implemented.
本申请实施例还提供了一种包含指令的计算机程序产品,当其在计算设备上运行时,使得计算设备执行前文中的方法,或者使得所述计算设备实现前文中的确定目标检测模型的装置的功能。Embodiments of the present application also provide a computer program product containing instructions, which, when running on a computing device, enables the computing device to execute the foregoing method, or enables the computing device to implement the foregoing apparatus for determining a target detection model function.
本申请实施例还提供一种芯片,包括至少一个处理器和接口电路,所述接口电路用于为所述至少一个处理器提供程序指令或者数据,所述至少一个处理器用于执行所述程序指令,使得前文中的方法得以实现。An embodiment of the present application further provides a chip, including at least one processor and an interface circuit, where the interface circuit is configured to provide program instructions or data for the at least one processor, and the at least one processor is configured to execute the program instructions , so that the above method can be realized.
本申请实施例还提供一种终端,包括前文所述的确定目标检测模型的装置。An embodiment of the present application further provides a terminal, including the aforementioned apparatus for determining a target detection model.
进一步,该终端可以为智能运输设备(车辆或者无人机)、智能家居设备、智能制造设备或者机器人等。该智能运输设备例如可以是AGV或无人运输车。Further, the terminal may be an intelligent transportation device (vehicle or drone), a smart home device, an intelligent manufacturing device, or a robot, and the like. The intelligent transportation device can be, for example, an AGV or an unmanned transportation vehicle.
上述实施例,可以全部或部分地通过软件、硬件、固件或其他任意组合来实现。当使用软件实现时,上述实施例可以全部或部分地以计算机程序产品的形式实现。所述计算机程序产品包括一个或多个计算机指令或计算机程序。在计算机上加载或执行所述计算机指令或计算机程序时,全部或部分地产生按照本申请实施例所述的流程或功能。所述计算机可以为通用计算机、专用计算机、计算机网络、或者其他可编程装置。所述计算机指令可以存储在计算机可读存储介质中,或者从一个计算机可读存储介质向另一个计算机可读存储介质传输,例如,所述计算机指令可以从一个网站站点、计算机、服务器或数据中心通过有线(例如红外、无线、微波等)方式向另一个网站站点、计算机、服务器或数据中心进行传输。所述计算机可读存储介质可以是计算机能够存取的任何可用介质或者是包含一 个或多个可用介质集合的服务器、数据中心等数据存储设备。所述可用介质可以是磁性介质(例如,软盘、硬盘、磁带)、光介质(例如,DVD)、或者半导体介质。半导体介质可以是固态硬盘。The above embodiments may be implemented in whole or in part by software, hardware, firmware or any other combination. When implemented in software, the above-described embodiments may be implemented in whole or in part in the form of a computer program product. The computer program product includes one or more computer instructions or computer programs. When the computer instructions or computer programs are loaded or executed on a computer, all or part of the processes or functions described in the embodiments of the present application are generated. The computer may be a general purpose computer, special purpose computer, computer network, or other programmable device. The computer instructions may be stored in or transmitted from one computer readable storage medium to another computer readable storage medium, for example, the computer instructions may be downloaded from a website site, computer, server or data center Transmission to another website site, computer, server or data center by wire (eg, infrared, wireless, microwave, etc.). The computer-readable storage medium can be any available medium that can be accessed by a computer or a data storage device such as a server, a data center, or the like that contains a set of one or more available media. The usable media may be magnetic media (eg, floppy disks, hard disks, magnetic tapes), optical media (eg, DVDs), or semiconductor media. The semiconductor medium may be a solid state drive.
所属领域的技术人员可以清楚地了解到,为描述的方便和简洁,上述描述的系统、装置和单元的具体工作过程,可以参考前述方法实施例中的对应过程,在此不再赘述。Those skilled in the art can clearly understand that, for the convenience and brevity of description, the specific working process of the above-described systems, devices and units may refer to the corresponding processes in the foregoing method embodiments, which will not be repeated here.
另外,在本申请各个实施例中的各功能单元可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中。In addition, each functional unit in each embodiment of the present application may be integrated into one processing unit, or each unit may exist physically alone, or two or more units may be integrated into one unit.
以上所述,仅为本申请的具体实施方式,但本申请的保护范围并不局限于此,任何熟悉本技术领域的技术人员在本申请揭露的技术范围内,可轻易想到变化或替换,都应涵盖在本申请的保护范围之内。因此,本申请的保护范围应以所述权利要求的保护范围为准。The above are only specific embodiments of the present application, but the protection scope of the present application is not limited to this. should be covered within the scope of protection of this application. Therefore, the protection scope of the present application should be subject to the protection scope of the claims.

Claims (25)

  1. 一种确定目标检测模型的方法,所述目标检测模型用于目标检测,其特征在于,所述方法包括:A method for determining a target detection model, wherein the target detection model is used for target detection, wherein the method comprises:
    根据第一目标检测模型获取第一目标检测框,所述第一目标检测框是对待检测图像进行目标检测得到的边界轮廓;Obtain a first target detection frame according to the first target detection model, where the first target detection frame is a boundary contour obtained by performing target detection on the image to be detected;
    确定所述第一目标检测框的参数信息,所述第一目标检测框的参数信息包括与对目标物进行目标测距相关的参数;determining parameter information of the first target detection frame, where the parameter information of the first target detection frame includes parameters related to target ranging;
    确定目标真值框的参数信息,所述目标真值框是所述待检测图像中所述目标物实际的边界轮廓,所述目标真值框的参数信息包括与对所述目标物进行目标测距相关的参数;Determine the parameter information of the target ground-truth frame, the target ground-truth frame is the actual boundary contour of the target in the image to be detected, and the parameter information of the target ground-truth frame includes and is related to the target detection of the target. distance-related parameters;
    根据所述第一目标检测框的参数信息和所述目标真值框的参数信息,确定损失项,所述损失项用于指示所述第一目标检测框和所述目标真值框之间的偏差;According to the parameter information of the first target detection frame and the parameter information of the target ground-truth frame, a loss item is determined, and the loss item is used to indicate the difference between the first target detection frame and the target ground-truth frame. deviation;
    根据所述损失项确定第二目标检测模型,所述第二目标检测模型用于确定第二目标检测框。A second object detection model is determined according to the loss term, and the second object detection model is used to determine a second object detection frame.
  2. 根据权利要求1所述的方法,其特征在于,所述第一目标检测框的参数信息包括所述第一目标检测框的底边中心点坐标值,所述目标真值框的参数信息包括所述目标真值框的底边中心点坐标值。The method according to claim 1, wherein the parameter information of the first target detection frame includes a coordinate value of a bottom center point of the first target detection frame, and the parameter information of the target ground-truth frame includes all the The coordinate value of the center point of the bottom edge of the target ground truth box.
  3. 根据权利要求2所述的方法,其特征在于,所述目标测距为采用落地点测距法进行的测距。The method according to claim 2, wherein the target ranging is a ranging method using a landing point ranging method.
  4. 根据权利要求1所述的方法,其特征在于,所述第一目标检测框的参数信息包括所述第一目标检测框的面积,所述目标真值框的参数信息包括所述目标真值框的面积。The method according to claim 1, wherein the parameter information of the first target detection frame includes an area of the first target detection frame, and the parameter information of the target ground-truth frame includes the target ground-truth frame area.
  5. 根据权利要求4所述的方法,其特征在于,所述目标测距为采用比例测距法进行的测距。The method according to claim 4, characterized in that, the target ranging is ranging by using a proportional ranging method.
  6. 根据权利要求2至5中任一项所述的方法,其特征在于,The method according to any one of claims 2 to 5, wherein,
    根据所述第一目标检测框的参数信息和所述目标真值框的参数信息,确定损失项,包括:According to the parameter information of the first target detection frame and the parameter information of the target ground-truth frame, the loss term is determined, including:
    根据所述第一目标检测框的参数信息、所述目标真值框的参数信息以及所述第一目标检测框与所述目标真值框的最小外接矩形框的参数,确定所述损失项。The loss term is determined according to the parameter information of the first target detection frame, the parameter information of the target ground-truth frame, and the parameters of the minimum enclosing rectangular frame between the first target detection frame and the target ground-truth frame.
  7. 根据权利要求6所述的方法,其特征在于,所述最小外接矩形框的参数包括所述最小外接矩形框的对角线的长度。The method according to claim 6, wherein the parameter of the minimum enclosing rectangle includes the length of the diagonal of the minimum enclosing rectangle.
  8. 根据权利要求7所述的方法,其特征在于,所述根据所述第一目标检测框的参数信息、所述目标真值框的参数信息以及所述第一目标检测框与所述目标真值框的最小外接矩形框的参数,包括:The method according to claim 7, wherein, according to the parameter information of the first target detection frame, the parameter information of the target ground-truth frame, and the first target detection frame and the target ground-truth The parameters of the minimum bounding rectangle of the box, including:
    根据以下公式,确定所述损失项:The loss term is determined according to the following formula:
    Figure PCTCN2021079304-appb-100001
    Figure PCTCN2021079304-appb-100001
    其中,c点为所述目标检测框的底边中心点,c gt点为所述目标真值框的底边中心点,所述ρ(c,c gt)为所述c点和所述c gt点之间的距离,所述s为所述最小外接矩形框的对角线的长度。 Wherein, point c is the center point of the bottom edge of the target detection frame, point c gt is the center point of the bottom edge of the target ground-truth frame, and the ρ(c,c gt ) is the point c and the point c gt is the distance between points, and the s is the length of the diagonal of the minimum enclosing rectangle.
  9. 根据权利要求6所述的方法,其特征在于,所述最小外接矩形框的参数包括所述最小外接矩形框的面积。The method according to claim 6, wherein the parameter of the minimum enclosing rectangle includes the area of the minimum enclosing rectangle.
  10. 根据权利要求9所述的方法,其特征在于,所述根据所述第一目标检测框的参数信息、所述目标真值框的参数信息以及所述第一目标检测框与所述目标真值框的最小外接矩形框的参数,包括:The method according to claim 9, wherein, according to the parameter information of the first target detection frame, the parameter information of the target ground-truth frame, and the first target detection frame and the target ground-truth The parameters of the minimum bounding rectangle of the box, including:
    根据以下公式,确定所述损失项:The loss term is determined according to the following formula:
    Figure PCTCN2021079304-appb-100002
    Figure PCTCN2021079304-appb-100002
    其中,a1为所述目标检测框的面积,a2为所述目标真值框的面积,所述a3为所述最小外接矩形框的面积。Wherein, a1 is the area of the target detection frame, a2 is the area of the target ground-truth frame, and a3 is the area of the minimum circumscribed rectangular frame.
  11. 根据权利要求1至10中任一项所述的方法,其特征在于,所述目标物实际的边界轮廓是由人工标注的。The method according to any one of claims 1 to 10, wherein the actual boundary contour of the target is manually marked.
  12. 一种确定目标检测模型的装置,所述目标检测模型用于目标检测,其特征在于,所述装置包括:A device for determining a target detection model, wherein the target detection model is used for target detection, wherein the device comprises:
    获取单元,用于根据第一目标检测模型获取第一目标检测框,所述第一目标检测框是对待检测图像进行目标检测得到的边界轮廓;an obtaining unit, configured to obtain a first target detection frame according to a first target detection model, where the first target detection frame is a boundary contour obtained by performing target detection on the image to be detected;
    处理单元,用于确定所述第一目标检测框的参数信息,所述第一目标检测框的参数信息包括与对目标物进行目标测距相关的参数;a processing unit, configured to determine parameter information of the first target detection frame, where the parameter information of the first target detection frame includes parameters related to target ranging;
    所述处理单元,还用于确定目标真值框的参数信息,所述目标真值框是所述待检测图像中所述目标物实际的边界轮廓,所述目标真值框的参数信息包括与对所述目标物进行目标测距相关的参数;The processing unit is further configured to determine the parameter information of the target ground truth frame, the target ground truth frame is the actual boundary contour of the target in the image to be detected, and the parameter information of the target ground truth frame includes and Performing target ranging related parameters on the target;
    所述处理单元,还用于根据所述第一目标检测框的参数信息和所述目标真值框的参数信息,确定损失项,所述损失项用于指示所述第一目标检测框和所述目标真值框之间的偏差;The processing unit is further configured to determine a loss item according to the parameter information of the first target detection frame and the parameter information of the target ground-truth frame, where the loss item is used to indicate the first target detection frame and the all target detection frame. describe the deviation between the target ground-truth boxes;
    所述处理单元,还用于根据所述损失项确定第二目标检测模型,所述第二目标检测模型用于确定第二目标检测框。The processing unit is further configured to determine a second target detection model according to the loss term, where the second target detection model is used to determine a second target detection frame.
  13. 根据权利要求12所述的装置,其特征在于,所述第一目标检测框的参数信息包括所述第一目标检测框的底边中心点坐标值,所述目标真值框的参数信息包括所述目标真值框的底边中心点坐标值。The device according to claim 12, wherein the parameter information of the first target detection frame includes a coordinate value of a bottom center point of the first target detection frame, and the parameter information of the target ground-truth frame includes all the The coordinate value of the center point of the bottom edge of the target ground truth box.
  14. 根据权利要求13所述的装置,其特征在于,所述目标测距为采用落地点测距法进行的测距。The device according to claim 13, wherein the target ranging is a ranging method using a landing point ranging method.
  15. 根据权利要求12所述的装置,其特征在于,所述第一目标检测框的参数信息包括所述第一目标检测框的面积,所述目标真值框的参数信息包括所述目标真值框的面积。The device according to claim 12, wherein the parameter information of the first target detection frame comprises an area of the first target detection frame, and the parameter information of the target ground truth frame comprises the target ground truth frame area.
  16. 根据权利要求15所述的装置,其特征在于,所述目标测距为采用比例测距法进行的测距。The device according to claim 15, wherein the target ranging is a ranging method using a proportional ranging method.
  17. 根据权利要求13至16中任一项所述的装置,其特征在于,所述处理单元,还具体用于:The device according to any one of claims 13 to 16, wherein the processing unit is further specifically configured to:
    根据所述第一目标检测框的参数信息、所述目标真值框的参数信息以及所述第一目标检测框与所述目标真值框的最小外接矩形框的参数,确定所述损失项。The loss term is determined according to the parameter information of the first target detection frame, the parameter information of the target ground-truth frame, and the parameters of the minimum enclosing rectangular frame between the first target detection frame and the target ground-truth frame.
  18. 根据权利要求17所述的装置,其特征在于,所述最小外接矩形框的参数包括所述最小外接矩形框的对角线的长度。The apparatus according to claim 17, wherein the parameter of the minimum circumscribed rectangle comprises the length of the diagonal of the minimum circumscribed rectangle.
  19. 根据权利要求18所述的装置,其特征在于,所述处理单元,还具体用于:The device according to claim 18, wherein the processing unit is further specifically configured to:
    根据以下公式,确定所述损失项:The loss term is determined according to the following formula:
    Figure PCTCN2021079304-appb-100003
    Figure PCTCN2021079304-appb-100003
    其中,c点为所述目标检测框的底边中心点,c gt点为所述目标真值框的底边中心点,所述ρ(c,c gt)为所述c点和所述c gt点之间的距离,所述s为所述最小外接矩形框的对角线的长度。 Wherein, point c is the center point of the bottom edge of the target detection frame, point c gt is the center point of the bottom edge of the target ground-truth frame, and the ρ(c,c gt ) is the point c and the point c gt is the distance between points, and the s is the length of the diagonal of the minimum enclosing rectangle.
  20. 根据权利要求17所述的装置,其特征在于,所述最小外接矩形框的参数包括所述最小外接矩形框的面积。The apparatus according to claim 17, wherein the parameter of the minimum enclosing rectangle includes an area of the minimum enclosing rectangle.
  21. 根据权利要求20所述的装置,其特征在于,所述处理单元,还具体用于:The device according to claim 20, wherein the processing unit is further specifically configured to:
    根据以下公式,确定所述损失项:The loss term is determined according to the following formula:
    Figure PCTCN2021079304-appb-100004
    Figure PCTCN2021079304-appb-100004
    其中,a1为所述目标检测框的面积,a2为所述目标真值框的面积,所述a3为所述最小外接矩形框的面积。Wherein, a1 is the area of the target detection frame, a2 is the area of the target ground-truth frame, and a3 is the area of the minimum circumscribed rectangular frame.
  22. 根据权利要求12至21中任一项所述的装置,其特征在于,所述目标物实际的边界轮廓是由人工标注的。The device according to any one of claims 12 to 21, wherein the actual boundary contour of the target is manually marked.
  23. 一种确定目标检测模型的装置,其特征在于,包括至少一个存储器和至少一个处理器,所述至少一个存储器用于存储程序,所述至少一个处理器用于运行所述程序,以实现如权利要求1至11中任一项所述的方法。A device for determining a target detection model, characterized in that it comprises at least one memory and at least one processor, the at least one memory is used for storing a program, and the at least one processor is used for running the program, so as to realize the method as claimed in the claims The method of any one of 1 to 11.
  24. 一种计算机可读存储介质,其特征在于,所述计算机可读存储介质上存储有程序或指令,所述程序或指令被执行时使得计算机执行如权利要求1至11中任一项所述的方法。A computer-readable storage medium, characterized in that, a program or an instruction is stored on the computer-readable storage medium, and when the program or instruction is executed, a computer executes the method according to any one of claims 1 to 11. method.
  25. 一种芯片,其特征在于,包括至少一个处理器和接口电路,所述接口电路用于为所述至少一个处理器提供程序指令或者数据,所述至少一个处理器用于执行所述程序指令,以实现如权利要求1至11中任一项所述的方法。A chip, characterized by comprising at least one processor and an interface circuit, wherein the interface circuit is configured to provide program instructions or data for the at least one processor, and the at least one processor is configured to execute the program instructions to A method as claimed in any one of claims 1 to 11 is implemented.
PCT/CN2021/079304 2021-03-05 2021-03-05 Method and apparatus for determining object detection model WO2022183484A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
PCT/CN2021/079304 WO2022183484A1 (en) 2021-03-05 2021-03-05 Method and apparatus for determining object detection model
CN202180079541.6A CN116508073A (en) 2021-03-05 2021-03-05 Method and device for determining target detection model

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2021/079304 WO2022183484A1 (en) 2021-03-05 2021-03-05 Method and apparatus for determining object detection model

Publications (1)

Publication Number Publication Date
WO2022183484A1 true WO2022183484A1 (en) 2022-09-09

Family

ID=83154872

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/079304 WO2022183484A1 (en) 2021-03-05 2021-03-05 Method and apparatus for determining object detection model

Country Status (2)

Country Link
CN (1) CN116508073A (en)
WO (1) WO2022183484A1 (en)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109029363A (en) * 2018-06-04 2018-12-18 泉州装备制造研究所 A kind of target ranging method based on deep learning
CN109117831A (en) * 2018-09-30 2019-01-01 北京字节跳动网络技术有限公司 The training method and device of object detection network
US20200110970A1 (en) * 2018-10-03 2020-04-09 Idemia Identity & Security France Parameter training method for a convolutional neural network and method for detecting items of interest visible in an image
CN112016605A (en) * 2020-08-19 2020-12-01 浙江大学 Target detection method based on corner alignment and boundary matching of bounding box
CN112329873A (en) * 2020-11-12 2021-02-05 苏州挚途科技有限公司 Training method of target detection model, target detection method and device

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109029363A (en) * 2018-06-04 2018-12-18 泉州装备制造研究所 A kind of target ranging method based on deep learning
CN109117831A (en) * 2018-09-30 2019-01-01 北京字节跳动网络技术有限公司 The training method and device of object detection network
US20200110970A1 (en) * 2018-10-03 2020-04-09 Idemia Identity & Security France Parameter training method for a convolutional neural network and method for detecting items of interest visible in an image
CN112016605A (en) * 2020-08-19 2020-12-01 浙江大学 Target detection method based on corner alignment and boundary matching of bounding box
CN112329873A (en) * 2020-11-12 2021-02-05 苏州挚途科技有限公司 Training method of target detection model, target detection method and device

Also Published As

Publication number Publication date
CN116508073A (en) 2023-07-28

Similar Documents

Publication Publication Date Title
US20210082181A1 (en) Method and apparatus for object detection, intelligent driving method and device, and storage medium
WO2020052540A1 (en) Object labeling method and apparatus, movement control method and apparatus, device, and storage medium
US11138750B2 (en) Distance estimating method and apparatus
WO2022099511A1 (en) Method and apparatus for ground segmentation based on point cloud data, and computer device
WO2022188663A1 (en) Target detection method and apparatus
WO2022104774A1 (en) Target detection method and apparatus
US11475628B2 (en) Monocular 3D vehicle modeling and auto-labeling using semantic keypoints
WO2018120040A1 (en) Obstacle detection method and device
KR20170106963A (en) Object detection using location data and scale space representations of image data
WO2022161140A1 (en) Target detection method and apparatus, and computer device and storage medium
JP7135665B2 (en) VEHICLE CONTROL SYSTEM, VEHICLE CONTROL METHOD AND COMPUTER PROGRAM
CN112753038B (en) Method and device for identifying lane change trend of vehicle
US20200125860A1 (en) Method and apparatus for determining driving information
WO2022165802A1 (en) Road boundary recognition method and apparatus
JP7509040B2 (en) Apparatus and method for tracking objects in roadway surveillance video
CN115147328A (en) Three-dimensional target detection method and device
CN114705121B (en) Vehicle pose measurement method and device, electronic equipment and storage medium
Sakic et al. Camera-LIDAR object detection and distance estimation with application in collision avoidance system
CN114972427A (en) Target tracking method based on monocular vision, terminal equipment and storage medium
US20220171975A1 (en) Method for Determining a Semantic Free Space
Fu et al. Camera-based semantic enhanced vehicle segmentation for planar lidar
WO2022078342A1 (en) Dynamic occupancy grid estimation method and apparatus
WO2022183484A1 (en) Method and apparatus for determining object detection model
CN116189150A (en) Monocular 3D target detection method, device, equipment and medium based on fusion output
Chen et al. Integrated vehicle and lane detection with distance estimation

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21928567

Country of ref document: EP

Kind code of ref document: A1

WWE Wipo information: entry into national phase

Ref document number: 202180079541.6

Country of ref document: CN

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 21928567

Country of ref document: EP

Kind code of ref document: A1