WO2022017140A1 - 目标检测方法及装置、电子设备和存储介质 - Google Patents
目标检测方法及装置、电子设备和存储介质 Download PDFInfo
- Publication number
- WO2022017140A1 WO2022017140A1 PCT/CN2021/103286 CN2021103286W WO2022017140A1 WO 2022017140 A1 WO2022017140 A1 WO 2022017140A1 CN 2021103286 W CN2021103286 W CN 2021103286W WO 2022017140 A1 WO2022017140 A1 WO 2022017140A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- result
- target
- detection
- correction
- target object
- Prior art date
Links
- 238000001514 detection method Methods 0.000 title claims abstract description 299
- 238000003860 storage Methods 0.000 title claims abstract description 25
- 238000012937 correction Methods 0.000 claims abstract description 224
- 238000005457 optimization Methods 0.000 claims abstract description 190
- 238000000034 method Methods 0.000 claims abstract description 47
- 238000004590 computer program Methods 0.000 claims description 13
- 238000012545 processing Methods 0.000 description 24
- 238000010586 diagram Methods 0.000 description 18
- 238000004891 communication Methods 0.000 description 13
- 230000006870 function Effects 0.000 description 10
- 238000005516 engineering process Methods 0.000 description 9
- 230000008569 process Effects 0.000 description 7
- 238000013528 artificial neural network Methods 0.000 description 5
- 230000005540 biological transmission Effects 0.000 description 4
- 230000003287 optical effect Effects 0.000 description 4
- 230000005236 sound signal Effects 0.000 description 4
- 230000009471 action Effects 0.000 description 3
- 230000001133 acceleration Effects 0.000 description 2
- 238000003491 array Methods 0.000 description 2
- 230000003190 augmentative effect Effects 0.000 description 2
- 238000011161 development Methods 0.000 description 2
- 239000000835 fiber Substances 0.000 description 2
- 230000003993 interaction Effects 0.000 description 2
- 230000001902 propagating effect Effects 0.000 description 2
- 230000003068 static effect Effects 0.000 description 2
- RYGMFSIKBFXOCR-UHFFFAOYSA-N Copper Chemical compound [Cu] RYGMFSIKBFXOCR-UHFFFAOYSA-N 0.000 description 1
- 230000002159 abnormal effect Effects 0.000 description 1
- 230000001413 cellular effect Effects 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 229910052802 copper Inorganic materials 0.000 description 1
- 239000010949 copper Substances 0.000 description 1
- 238000013480 data collection Methods 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 238000003384 imaging method Methods 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000002093 peripheral effect Effects 0.000 description 1
- 230000002441 reversible effect Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 239000013589 supplement Substances 0.000 description 1
- 230000002123 temporal effect Effects 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
- 230000001052 transient effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V2201/00—Indexing scheme relating to image or video recognition or understanding
- G06V2201/07—Target detection
Definitions
- Embodiments of the present disclosure relate to the technical field of computer vision, and in particular, to a method and apparatus for target detection, an electronic device, and a storage medium.
- Computer vision technology can simulate biological vision through electronic equipment. With the development of computer vision technology, more and more work can be done using electronic equipment to provide people with convenient conditions.
- Object detection is an important task in computer vision, and its task goal is to estimate the position information of objects within the field of view. Stable object detection techniques can not only estimate the position information of objects, but also help to optimize the pose of the camera or be used in the development of other applications such as augmented reality and indoor navigation.
- the embodiment of the present disclosure proposes a technical solution for target detection.
- a target detection method comprising: acquiring a first detection result obtained by performing target detection on a current data frame of a target scene; A detection result is updated to obtain the first observation result of the target object in the current data frame; the first observation result is corrected according to the point cloud data corresponding to the first observation result, and the first observation result of the target object is obtained. A correction result.
- the updating the first detection result based on the historical optimization result of the target scene to obtain the first observation result of the target object in the current data frame includes: based on the target The historical optimization result of the scene determines the object information of the first detection result, wherein the object information is used to identify the target object; the first detection result is updated according to the object information of the first detection result , to obtain the first observation result of the target object in the current data frame. In this way, by determining the object information of the first detection result, the connection between the historical optimization result and the first detection result can be established, thereby improving the accuracy of object detection.
- the determining the object information of the first detection result based on the historical optimization result of the target scene includes: comparing the historical optimization result of the target scene with the first detection result Matching; if the first detection result matches the historical optimization result, determine the object information of the historical optimization result as the object information of the first detection result. In this way, the first detection result can be further updated to obtain the first observation result with accurate object information.
- the determining the object information of the first detection result based on the historical optimization result of the target scene includes: when the first detection result does not match the historical optimization result Next, new object information is set for the first detection result. In this way, the first detection result can be made to correspond to the newly observed target object.
- the matching the historical optimization result of the target scene with the first detection result includes: determining the overlap between the first detection result and a detection frame of a historical optimization result. the first volume, and determine the total volume occupied by the detection frame of the first detection result and the detection frame of the historical optimization result; determine the first detection result according to the ratio of the first volume to the total volume How well it matches the historical optimization results. In this way, the matching degree between the detection result and the historical optimization result can be determined more accurately.
- the updating the first detection result based on the historical optimization result of the target scene to obtain the first observation result of the target object in the current data frame includes: based on the target The historical optimization result of the scene, when it is determined that the current data frame has a target object that is not detected by the first detection result, the historical optimization result of the undetected target object is determined as the current data frame.
- the first observation of an undetected target object in . In this way, the phenomenon of missed detection can be reduced, and the reliability of target detection can be greatly increased.
- the modifying the first observation result according to the point cloud data corresponding to the first observation result to obtain the first correction result of the target object includes: The point cloud data corresponding to the historical optimization result and the point cloud data corresponding to the first observation result are merged to obtain merged point cloud data; based on the merged point cloud data, a first correction result obtained by modifying the first observation result is obtained .
- the merged point cloud data of the same object is used to obtain a first correction result with more accurate position information, and the historical information of the same target object can be considered in the target detection process, which can improve the accuracy of target detection.
- the merging of the point cloud data corresponding to the historical optimization result of the same target object and the point cloud data corresponding to the first observation result includes: for the same target object, combining the current data frame The point cloud data corresponding to the historical optimization result of the previous data frame is merged with the point cloud data corresponding to the first observation result. In this way, the first observation result of the current data frame is corrected by using the historical optimization result of the previous data frame, so that the obtained first correction result is more accurate.
- the method further includes: acquiring a correction result of the target object, wherein the correction result includes the first correction result and a second correction result, and the second correction result is based on The historical data frame of the target scene is obtained by performing target detection; based on the target result in the correction result, the current optimization result of the target object is determined. In this way, the current optimization result of the target object can be obtained by using multiple correction results, so that the target detection is more accurate.
- the method further includes: determining errors between a first correction result in the correction results and a plurality of second correction results, wherein the first correction result is any one of the correction results As a result, the second correction result is a correction observation frame other than the first correction result; the number of inliers corresponding to the first correction result is counted, wherein the number of inliers is the same as the first correction result The number of second correction results whose error is less than the error threshold; the target result in the correction result is determined according to the number of inliers corresponding to the first correction result. In this way, the current optimization result of the target object is determined according to the relatively accurate target result in the correction result, and the correction result with lower accuracy is removed, so that the accuracy of target detection can be further improved.
- the determining the target result in the correction result according to the number of inliers corresponding to the first correction result includes: determining that the number of inliers in the plurality of first correction results is the largest the first correction result; the first correction result with the largest number of inliers and the second correction result in which the error with the first correction result with the largest number of inliers is less than the error threshold are determined as the correction results target result in .
- the first correction result of the target object can be further optimized, so that the current optimization result obtained after optimization can more accurately indicate the position of the target object.
- the sum of errors between the current optimization result and a plurality of the target results is minimized.
- a target detection apparatus comprising: an acquisition module configured to acquire a first detection result obtained by performing target detection on a current data frame of a target scene; a determination module configured to The historical optimization result of the target scene updates the first detection result, and obtains the first observation result of the target object in the current data frame; the correction module is configured to perform a correction on the first detection result according to the point cloud data corresponding to the first observation result. The first observation result is corrected to obtain the first correction result of the target object.
- an electronic device including:
- memory for storing processor-executable instructions
- the processor is configured to: execute the above target detection method.
- a computer-readable storage medium having computer program instructions stored thereon, the computer program instructions implementing the above target detection method when executed by a processor.
- An embodiment of the present disclosure provides a computer program, including computer-readable code, when the computer-readable code is executed in an electronic device, a processor in the electronic device executes and is configured to achieve any of the above objectives. Detection method.
- the first detection result obtained by performing target detection on the current data frame of the target scene may be obtained, and then the first detection result is updated based on the historical optimization result of the target scene to obtain the target object in the current data frame.
- the first observation result is corrected according to the historical optimization result of the target object and the point cloud data corresponding to the first observation result, so as to obtain the first correction result of the target object.
- the first detection result of the target scene can be combined with the historical optimization result, and the correlation between the first detection result and the historical optimization result can be considered, so that the obtained first correction result can more accurately represent the position of the target object.
- FIG. 1A is a schematic diagram of a system architecture to which a target detection method according to an embodiment of the present disclosure can be applied;
- FIG. 1B shows a flowchart of a target detection method according to an embodiment of the present disclosure.
- FIG. 2 shows a flowchart of an example of a target detection method according to an embodiment of the present disclosure.
- FIG. 3 shows a block diagram of a target detection apparatus according to an embodiment of the present disclosure.
- FIG. 4 shows a block diagram of an example of an electronic device according to an embodiment of the present disclosure.
- FIG. 5 shows a block diagram of an example of an electronic device according to an embodiment of the present disclosure.
- the target detection solution provided by the embodiment of the present disclosure can obtain the first detection result obtained by performing target detection on the current data frame of the target scene, and then update the first detection result based on the historical optimization result of the target scene, and obtain the target scene in the current data frame.
- the first observation result is corrected according to the historical optimization result of the target object and the point cloud data corresponding to the first observation result, and a first correction result obtained by correcting the first observation result is obtained.
- the first observation result obtained by combining the first detection result and the historical optimization result can more accurately indicate the target object in the current data frame, and further through the point cloud data corresponding to the historical optimization result and the first observation result, it is possible to The first observation result is adjusted so that the first correction result can more accurately indicate the position of the target object.
- target detection is usually performed separately for each data frame collected from the target scene.
- this method of target detection has great limitations. For example, the detection results for the same object will shake, or when the target object in the target scene has occlusion or truncation, it is difficult to detect The position of the target object is estimated accurately, so the accuracy of the detection result is poor.
- the embodiment of the present disclosure can combine the first detection result of the current data frame of the target scene with the historical optimization result, so that the temporal continuity of the position of the same target object can be considered, and the accuracy of estimating the position of the target object can be improved.
- the technical solutions provided by the embodiments of the present disclosure can be applied to the expansion of application scenarios such as target detection, target tracking, positioning, and navigation, which are not limited in the embodiments of the present disclosure.
- the augmented reality technology that can be applied to the terminal can realize indoor positioning and/or indoor navigation by obtaining the first correction result of the target object in the indoor scene.
- FIG. 1A is a schematic diagram of a system architecture to which a target detection method according to an embodiment of the present disclosure can be applied; as shown in FIG. 1A , the system architecture includes a data frame collection terminal 131 , a network 132 and a target detection terminal 133 .
- the data frame collection terminal 131 and the target detection terminal 133 may establish a communication connection through the network 132, and the data frame collection terminal 131 sends the collected current data frame to the target detection terminal 133 through the network 132.
- the detection terminal 133 first, obtains the first detection result after the target detection of the current data frame; then, updates the first detection result through the historical optimization result to obtain the first observation result of the target object;
- the point cloud data is used to correct the result, and the final correction result of the target object can be obtained.
- the first detection result of the target scene is combined with the historical optimization result, and the correlation between the first detection result and the historical optimization result is considered, so that the position of the target object can be more accurately represented.
- the data frame acquisition terminal 131 may be an image acquisition device with a camera, and the target detection terminal 133 may include a computer device with certain computing capabilities, such as a terminal device or a server or other processing device.
- the network 132 can be wired or wireless.
- the data frame acquisition terminal 131 can be connected to the server through wired connection, such as data communication through a bus; when the target detection terminal 133 is a terminal device, the data frame acquisition terminal 131 can It is connected to the target detection terminal 133 by means of wireless connection, and then performs data communication.
- the target detection terminal 133 may be a vision processing device with a video capture module, or a host with a camera.
- the target detection method in the embodiment of the present disclosure may be executed by the target detection terminal, and the above-mentioned system architecture may not include the network 132 and the data frame collection terminal 131 .
- FIG. 1B shows a flowchart of a target detection method according to an embodiment of the present disclosure.
- the target detection method can be performed by a terminal device, a server or other types of electronic devices, wherein the terminal device can be a user equipment (User Equipment, UE), a mobile device, a user terminal, a terminal, a cellular phone, a cordless phone, a personal digital processor (Personal Digital Assistant, PDA), handheld devices, computing devices, in-vehicle devices, wearable devices, etc.
- the object detection method may be implemented by the processor calling computer-readable instructions stored in the memory.
- the target detection method of the embodiment of the present disclosure will be described below by taking an electronic device as an execution subject as an example.
- Step S11 acquiring a first detection result obtained by performing target detection on the current data frame of the target scene.
- the electronic device may perform data collection on the target scene to obtain the current data frame of the target scene, or the electronic device may obtain the current data frame of the target scene from other devices.
- the current data frame may be an image frame, for example, the current data frame may be a depth image of the target scene, or the current data frame may also be point cloud data collected for the target scene.
- the depth image may include a common RGB image (a color image with three color channels of red (R), green (G), and blue (B)) and a depth image.
- target detection may be performed on the current data frame to obtain a first detection result.
- any detection method can be used to perform target detection on the current data frame.
- the first detection result may be a detection frame obtained by performing target detection on the current data frame, and the detection frame may indicate the position and size of the target object, so that the first detection result may include position information and size information.
- the detection frame may be a three-dimensional (3-Dimension, 3D) detection frame, and the position and size of the target object indicated by the detection frame may be the position and size of the target object in the target scene.
- the first detection result can be considered as a relatively rough detection result.
- the electronic device also directly obtains the first detection result from other devices.
- the position of the target object indicated by the first detection result may be the position of the target object in the world coordinate system of the target scene.
- the first detection result may be the coordinates of the target object in the world coordinate system.
- the electronic device may directly acquire the first detection result including the position of the target object in the world coordinate system.
- the position of the target object in the coordinate system of the image acquisition device can be obtained first, and then according to the relative position transformation relationship between the coordinate system of the image acquisition device and the world coordinate system, the position of the target object in the coordinate system of the image acquisition device can be obtained.
- the target object may be an object, a person, etc. existing in the target scene.
- the target objects can be pedestrians, tables, chairs, etc.
- the first detection result may further include object information of the indicated target object, so that the target object indicated by the first detection result may be determined according to the object information of the first detection result.
- Step S12 Update the first detection result based on the historical optimization result of the target scene to obtain the first observation result of the target object in the current data frame.
- the historical optimization result of the target scene may be the detection result of the target object obtained by performing optimization based on the second detection result, and the historical optimization result may more accurately indicate the location of the target object.
- the second detection result may be obtained by performing target detection on all or part of the historical data frame of the target scene, the historical data frame may be the data frame collected before the current data frame, and the second detection result may be the historical detection result of the target object.
- the manner of acquiring the second detection result may be similar to the manner of acquiring the above-mentioned first detection result, which will not be repeated here.
- the second detection result may be a detection frame obtained by performing target detection on the historical data frame, and the second detection result may include position information and size information.
- a target object in the target scene may correspond to a historical optimization result, that is, a plurality of second detection results obtained by target detection based on all or part of the historical data frame, a historical optimization of each target object can be obtained.
- the stored historical optimization result can be updated, so that one target object corresponds to one historical optimization result, thereby reducing the stored historical optimization result.
- the optimization result corresponding to each data frame may also be stored, which is not limited in this embodiment of the present disclosure.
- step S12 is performed at this time.
- the historical optimization results mentioned in can be considered as the optimization results corresponding to the previous data frame of the current data frame.
- the first detection result may be updated by using the historical optimization result of the target scene.
- the historical optimization result may be matched with the first detection result, and a known knowledge corresponding to the target object corresponding to the first detection result and the historical optimization result may be established. Associations between target objects.
- the first detection result may be updated according to the association between the target object corresponding to the first detection result and the known target object corresponding to the historical optimization result; for example, the object information of the first detection result may be determined, or the same
- the historical optimization result of the target object and the first detection result are combined, for example, the detection frame corresponding to the historical optimization result and the detection frame corresponding to the first detection result are combined.
- the connection between the target object of the current data frame and the target object of the historical data frame can be established, so that the obtained first observation result has more accurate object information .
- the first observation result may also be a detection frame, and correspondingly, the first observation result may include position information and size information of the target object.
- Step S13 correcting the first observation result according to the point cloud data corresponding to the first observation result to obtain a first correction result of the target object.
- a target object in the target scene may exist in the current data frame or may exist in one or more historical data frames, so that a target object in the current data frame may have the first observation results.
- the first observation result can be corrected according to the point cloud data corresponding to the first observation result of the target object to obtain the first correction of the target object. result.
- the point cloud data corresponding to the first observation result of the target object and the point cloud data corresponding to the historical optimization result can be used to determine The first observation result is corrected to obtain the first correction result of the target object.
- the point cloud data corresponding to the first observation result of the target object and/or the obvious abnormal data in the historical optimization result may be deleted, or the missing data in the point cloud data corresponding to the first observation result may be deleted. Supplement, and then the first correction result of the target object. In this way, the first correction result may more accurately indicate the position of the target object in the current data frame in the target scene.
- the image frame may be converted into point cloud data according to depth information of the image frame. Then, point cloud data corresponding to the historical optimization result and/or the first observation result can be obtained.
- the first detection result may be updated through the historical optimization result of the target scene, so that the association between the current data frame and the historical data frame may be established.
- the process of obtaining the first observation result of the target object in the current data frame will be described below through an implementation manner.
- the object information of the first detection result is determined. Then, the first detection result is updated according to the object information of the first detection result to obtain the first observation result of the target object in the current data frame. The object information is used to identify the target object.
- the historical optimization result of the target object in the target scene may be used to determine the object information of the first detection result.
- the detection frame of the historical optimization result may overlap with the detection frame of the first detection result. Then, it can be considered that the target object indicated by the historical optimization result and the target object indicated by the first detection result are the same target object, so that the object information of the historical optimization result can be used as the object information corresponding to the first detection result.
- the detection frame of any historical optimization result does not overlap with the detection frame of the first detection result, it can be considered that the target object indicated by the first detection result is the newly detected target object in the target scene, so that it can be The new object information is generated to identify the target object indicated by the first detection result.
- the historical optimization result of the target scene may be matched with the first detection result, and if the first detection result matches the historical optimization result, the object information of the historical optimization result may be matched The object information determined as the first detection result.
- the historical optimization result of the target scene may be matched with the first detection result.
- the detection frame of the historical optimization result may be determined to match the detection frame of the first detection result, and the historical optimization result may be determined to match the detection frame of the first detection result.
- the matching degree of the detection result For a first detection result, the historical optimization result with the highest matching degree with the first detection result and greater than the matching degree threshold can be determined as the historical optimization result matching the first detection result, and then the historical optimization result matching the first detection result can be determined.
- the object information of the historical optimization result of the result matching is used as the object information of the first detection result, and the first observation result of the target object is obtained, and the first observation result may be the first detection result after updating the object information.
- the first volume of the overlapping portion of a detection frame of a first detection result and a detection frame of a historical optimization result it is possible to determine the first volume of the detection frame of the first detection result.
- the total volume occupied by the detection frame of the result and the detection frame of the historical optimization result, and then the ratio of the first volume to the total volume can be used as the matching degree of the historical optimization result and the first detection result. That is, the three-dimensional intersection ratio (3-Dimensional Intersection over Union) between the detection frame of a first detection result and the detection frame of a historical optimization result may be used as the matching degree between the detection result and the historical optimization result.
- new object information is set for the first detection result.
- the first detection result does not match any historical optimization result, so it can be considered that the first detection result does not match any historical optimization result. is the detection result of the newly observed target object in the target scene, thereby setting new object information for the first detection result. In the case that the first detection result does not match the historical optimization result in the current scene, by setting new object information for the first detection result, the first detection result can be made to correspond to the newly observed target object.
- the historical The optimization result is determined as the first observation result of the undetected target object in the current data frame.
- each historical optimization result may be obtained by performing target detection based on historical data frames of the target scene, and the same target object detected in multiple historical data frames may correspond to one historical optimization result.
- the result may include the position information and object information of the target object, and the target object existing in the target scene may be determined according to the historical optimization result of the historical data frame.
- a target object can be observed within the field of view of the current data frame determined according to the historical optimization results, but the first detection result of the current data frame indicates that the target object is not detected in the current data frame, and it can be considered that the current data frame has missed detection Therefore, the historical optimization result of the undetected target object can be determined as the first observation result of the target object in the current data frame, thereby reducing the phenomenon of missed detection and greatly increasing the reliability of target detection.
- the first observation result may be corrected to obtain the first correction result.
- the first correction result has more accurate position information, thereby making the target detection more accurate. The process of obtaining the first correction result will be described below through a possible implementation manner.
- the point cloud data of the historical optimization result of the same target object and the point cloud data corresponding to the first observation result may be merged to obtain merged point cloud data. Then, based on the merged point cloud data, a first correction result obtained by correcting the first observation result is obtained.
- the historical optimization result and the first observation result belonging to the same target object may be determined according to the object information of the historical optimization result and the object information of the first observation result. Since the object information can mark the target object, if the object information is the same, it can be considered that the historical optimization result and the first observation result belong to the same target object. For the same target object, the point cloud data in the detection frame of the historical optimization result and the point cloud data in the detection frame of the first observation result can be obtained, and the point cloud data corresponding to the historical optimization result and the point corresponding to the first observation result can be obtained.
- the cloud data is merged, for example, the point cloud data corresponding to the historical optimization result and the point cloud data corresponding to the first observation result are merged to obtain the merged point cloud data of a target object.
- the first observation result can be corrected according to the merged point cloud data to obtain the first correction result of the target object.
- the merged point cloud data of a target object can be input into a neural network, and the position information of the first observation result can be corrected by using the neural network to obtain the first corrected result output by the neural network.
- the merged point cloud data of the same object can be used to obtain the first correction result with more accurate position information, so that the historical information of the same target object (such as the position information of the historical optimization result) can be considered in the target detection process, improving the The accuracy of object detection.
- each data frame may correspond to an optimization result of a target object, so that when correction and optimization are performed for the first observation result of the current data frame , for the same target object, the point cloud data corresponding to the historical optimization result of the previous data frame of the current data frame can be merged with the point cloud data corresponding to the first observation result, and the historical optimization result of the previous data frame of the current data frame can be used.
- the first observation result of the current data frame is corrected. Since the historical optimization result of the previous data frame of the current data frame is the latest stored, it is more accurate than the historical optimization results corresponding to other historical data frames.
- the historical optimization result of the data frame corrects the first observation result of the current data frame, which can make the obtained first correction result more accurate.
- correction and optimization are performed on the first observations of some of the collected data frames, for example, a data frame for which correction and optimization of the first observations are to be performed every certain data frame is selected, then not every data frame is selected for correction and optimization.
- Each frame corresponds to the optimization result of a target object.
- the latest stored historical optimization result of the target object may be selected to correct the first observation result of the current data frame.
- the first correction result may be optimized. The process of further optimizing the first correction result will be described below.
- a correction result of the target object may be obtained, wherein the correction result includes a first correction result and a second correction result, and the second correction result is obtained by performing target detection based on the historical data frame of the target scene. Based on the target result in the correction result, the current optimization result of the target object can be determined.
- the first correction result of the current data frame may be combined with the second correction result of the historical data frame to further optimize the first correction result.
- the second correction result may be obtained based on a second detection result of target detection performed on a historical data frame of the target scene, and the second detection result may be a historical detection result.
- the manner of determining the second correction result may be the same as the manner of determining the first correction result, which will not be repeated here.
- Each historical data frame may correspond to a second correction result of the target object, and the same target object may correspond to a series of second correction results as data frames are continuously collected on the target scene.
- a correction result including the first correction result and the second correction result can be obtained, so that the target detection information (second correction result) of the historical data frame can be combined.
- the current optimization result of the target object can be determined. For example, one or several correction results can be selected from the correction results of a target object as the target result, and the target result can be used as the current optimization result, or, Take the average or median of multiple target results as the current optimization result. Since the position change of the target object may be small, the correction results of the target object obtained from different data frames can be consistent, so that the current optimization result of the target object can be obtained by using multiple correction results, so that the target detection is more accurate.
- errors between the first correction result in the correction results and the plurality of second correction results may be determined, wherein the first correction result is any one correction result, and the second correction result It is a correction result other than the first correction result.
- the target result in the correction result is determined according to the number of interior points corresponding to the first correction result.
- an example of determining the target result in the correction result is provided.
- any one of the correction results may be used as the first correction result, and the correction results other than the first correction result among the multiple correction results may be used as the second correction result.
- the error between the first correction result and the plurality of second correction results can be calculated respectively, and according to the error between the first correction result and the plurality of second correction results, the internal error corresponding to the first correction result can be counted. number of points. For example, the error between the position information of the first correction result and the position information of a second correction result can be calculated.
- the second correction result is an interior point of the first correction result, and the number of interior points of the first correction result can be used as the number of interior points corresponding to the first correction result, that is, the second correction whose error with the first correction result is smaller than the error threshold the number of results.
- the target result in the correction result may be determined according to the number of inliers corresponding to the first correction result. For example, the first correction result with the largest number of inliers is determined as the target result among the correction results. In this way, the current optimization result of the target object can be determined according to the relatively accurate target result in the correction result, and the correction result with lower accuracy can be removed, so that the accuracy of target detection can be further improved.
- a first correction result with the largest number of inliers among the plurality of first correction results is determined. Then, the first correction result with the largest number of inliers and the second correction result with an error smaller than the error threshold from the first correction result with the largest number of inliers are determined as target results in the correction results.
- the second correction result whose error with the first correction result is smaller than the error threshold may be an inner point of the first correction result, and the first correction result with the largest number of inner points may be the first correction result with the largest number of inner points.
- Correction results A first correction result of a target object has the largest number of inliers, which can indicate that in the case where the position of the target object changes less, the first correction result and the inliers of the first correction result are closer to the target object. Therefore, the first correction result and the inner point of the first correction result can be determined as the target result in the correction result of the target object.
- the current optimization result of a target object may be determined based on multiple target results in the correction results of a target object, so that the first correction result of the target object may be further optimized, so that the optimization
- the current optimization result obtained after can more accurately indicate the position of the target object. For example, an optimal value can be estimated according to the position information of the target object in each target result, so that the optimal value reaches a specific condition, and this optimal value can be used as the current optimization result of the target object.
- a current optimization result is estimated according to the position information of the target object in each target result, so that the sum of the distances between the current optimization result and multiple target results can be minimized, for example, the The current optimization result is regarded as an unknown variable, and the equation of the sum of the squares of the errors between the unknown variable and each target result is established, and then the value of the unknown variable under the condition of the minimum sum of distances is solved.
- the solved value of the unknown variable can be used as The current optimization result for this target object.
- the sum of the distances between the obtained current optimization result and the position information of multiple target results can be minimized. In this way, the current optimization result can be used as the final detection result of the target object, thereby improving the accuracy of target detection.
- the current optimization result of a target object may be saved, or the saved historical optimization result of the target object may be updated to the obtained current optimization result .
- FIG. 2 shows a flowchart of an example of a target detection method according to an embodiment of the present disclosure.
- Step S201 obtaining the 3D detection frame (first detection result) of the current data frame of the target scene
- Step S202 matching the historical optimal estimation frame (historical optimization result) of the known object in the target scene with the 3D detection frame of the current data frame to obtain the current observation frame (first observation result) of the target object in the current data frame;
- Step S203 for each target object, use the optimal estimation frame of the target object and the current observation frame of the current data frame to segment the point cloud data of the target scene, and retain the historical optimal estimation frame and/or the current optimal estimation frame of the target object.
- Step S204 input the optimal estimation frame of each target object and/or the point cloud data in the current observation frame and the current observation frame corresponding to the target object into the neural network, and use the neural network to observe the current observation of each target object.
- the frame is corrected to obtain the current correction frame (the first correction result) of each target object in the current data frame;
- step S205 the current correction frame and the historical correction frame of each target object are jointly optimized to obtain the current optimal estimation frame (current optimization result) of each target object.
- the target detection solution provided by the embodiments of the present disclosure can improve the accuracy of target detection. Even if there is occlusion or truncation in the target scene, the obtained detection result has strong robustness and improves the stability of target detection.
- embodiments of the present disclosure also provide apparatuses, electronic devices, computer-readable storage media, and programs, all of which can be used to implement any target detection method provided by the embodiments of the present disclosure. Corresponding records will not be repeated.
- the writing order of each step does not mean a strict execution order but constitutes any limitation on the implementation process, and the specific execution order of each step should be based on its function and possible Internal logic is determined.
- FIG. 3 shows a block diagram of a target detection apparatus according to an embodiment of the present disclosure. As shown in FIG. 3 , the apparatus includes:
- the obtaining module 31 is configured to obtain a first detection result obtained by performing target detection on the current data frame of the target scene;
- a determination module 32 configured to update the first detection result based on the historical optimization result of the target scene to obtain the first observation result of the target object in the current data frame;
- the correction module 33 is configured to correct the first observation result according to the point cloud data corresponding to the first observation result to obtain a first correction result of the target object.
- the determining module 32 is configured to determine object information of the first detection result based on historical optimization results of the target scene, where the object information is used to identify the target object ; Update the first detection result according to the object information of the first detection result to obtain the first observation result of the target object in the current data frame.
- the determining module 32 is configured to match the historical optimization result of the target scene with the first detection result; when the first detection result matches the historical optimization result In this case, the object information of the historical optimization result is determined as the object information of the first detection result.
- the determining module 32 is configured to set new object information for the first detection result in the case that the first detection result does not match the historical optimization result.
- the determining module 32 is configured to determine the first volume of the overlapping portion of the first detection result and a detection frame of a historical optimization result, and to determine the detection of the first detection result
- the total volume occupied by the frame and the detection frame of the historical optimization result; the matching degree between the first detection result and the historical optimization result is determined according to the ratio of the first volume to the total volume.
- the determining module 32 is configured to, based on the historical optimization result of the target scene, in the case that it is determined that the current data frame has a target object that is not detected by the first detection result, The historical optimization result of the undetected target object is determined as the first observation result of the undetected target object in the current data frame.
- the correction module 33 is configured to merge the point cloud data corresponding to the historical optimization result of the same target object and the point cloud data corresponding to the first observation result to obtain the merged point cloud data;
- the merged point cloud data is used to obtain a first correction result obtained by correcting the first observation result.
- the correction module 33 is configured to, for the same target object, compare the point cloud data corresponding to the historical optimization result of the previous data frame of the current data frame with the point cloud data corresponding to the first observation result. Merge point cloud data.
- the apparatus further includes: an optimization module configured to obtain a correction result of the target object, wherein the correction result includes the first correction result and the second correction result, the first correction result
- the second correction result is obtained by performing target detection based on the historical data frame of the target scene; based on the target result in the correction result, the current optimization result of the target object is determined.
- the optimization module is further configured to determine an error between a first correction result in the correction results and a plurality of second correction results, wherein the first correction result is any one the correction result, the second correction result is the correction observation frame other than the first correction result; the number of inliers corresponding to the first correction result is counted, wherein the number of inliers is the same as the first correction result.
- the number of second correction results whose error of the correction result is smaller than the error threshold; the target result in the correction result is determined according to the number of inliers corresponding to the first correction result.
- the optimization module is configured to determine a first correction result with the largest number of inliers among the plurality of first correction results; A second correction result whose error from the first correction result with the largest number of inliers is smaller than the error threshold is determined as the target result in the correction results.
- the sum of errors between the current optimization result and a plurality of the target results is minimized.
- the functions or modules included in the apparatuses provided in the embodiments of the present disclosure may be used to execute the methods described in the above method embodiments.
- FIG. 4 is a block diagram of an electronic device 800 according to an exemplary embodiment.
- electronic device 800 may be a mobile phone, computer, digital broadcast terminal, messaging device, game console, tablet device, medical device, fitness device, personal digital assistant, and the like.
- the electronic device 800 may include one or more of the following components: a processing component 802, a memory 804, a power supply component 806, a multimedia component 808, an audio component 810, an input/output (I/O) interface 812, a sensor component 814 , and the communication component 816 .
- the processing component 802 generally controls the overall operation of the electronic device 800, such as operations associated with display, phone calls, data communications, camera operations, and recording operations.
- the processing component 802 can include one or more processors 820 to execute instructions to perform all or some of the steps of the methods described above.
- processing component 802 may include one or more modules that facilitate interaction between processing component 802 and other components.
- processing component 802 may include a multimedia module to facilitate interaction between multimedia component 808 and processing component 802.
- Memory 804 is configured to store various types of data to support operation at electronic device 800 . Examples of such data include instructions for any application or method operating on electronic device 800, contact data, phonebook data, messages, pictures, videos, and the like. Memory 804 may be implemented by any type of volatile or nonvolatile storage device or combination thereof, such as static random access memory (SRAM), electrically erasable programmable read only memory (EEPROM), erasable Programmable Read Only Memory (EPROM), Programmable Read Only Memory (PROM), Read Only Memory (ROM), Magnetic Memory, Flash Memory, Magnetic or Optical Disk.
- SRAM static random access memory
- EEPROM electrically erasable programmable read only memory
- EPROM erasable Programmable Read Only Memory
- PROM Programmable Read Only Memory
- ROM Read Only Memory
- Magnetic Memory Flash Memory
- Magnetic or Optical Disk Magnetic Disk
- Power supply assembly 806 provides power to various components of electronic device 800 .
- Power supply components 806 may include a power management system, one or more power supplies, and other components associated with generating, managing, and distributing power to electronic device 800 .
- Multimedia component 808 includes a screen that provides an output interface between the electronic device 800 and the user.
- the screen may include a liquid crystal display (LCD) and a touch panel (TP). If the screen includes a touch panel, the screen may be implemented as a touch screen to receive input signals from a user.
- the touch panel includes one or more touch sensors to sense touch, swipe, and gestures on the touch panel. The touch sensor may not only sense the boundaries of a touch or swipe action, but also detect the duration and pressure associated with the touch or swipe action.
- the multimedia component 808 includes a front-facing camera and/or a rear-facing camera. When the electronic device 800 is in an operation mode, such as a shooting mode or a video mode, the front camera and/or the rear camera may receive external multimedia data. Each of the front and rear cameras can be a fixed optical lens system or have focal length and optical zoom capability.
- Audio component 810 is configured to output and/or input audio signals.
- audio component 810 includes a microphone (MIC) that is configured to receive external audio signals when electronic device 800 is in operating modes, such as calling mode, recording mode, and voice recognition mode.
- the received audio signal may be further stored in memory 804 or transmitted via communication component 816 .
- audio component 810 also includes a speaker for outputting audio signals.
- the I/O interface 812 provides an interface between the processing component 802 and a peripheral interface module, which may be a keyboard, a click wheel, a button, or the like. These buttons may include, but are not limited to: home button, volume buttons, start button, and lock button.
- Sensor assembly 814 includes one or more sensors for providing status assessment of various aspects of electronic device 800 .
- the sensor assembly 814 can detect the on/off state of the electronic device 800, the relative positioning of the components, such as the display and the keypad of the electronic device 800, the sensor assembly 814 can also detect the electronic device 800 or one of the electronic device 800 Changes in the position of components, presence or absence of user contact with the electronic device 800 , orientation or acceleration/deceleration of the electronic device 800 and changes in the temperature of the electronic device 800 .
- Sensor assembly 814 may include a proximity sensor configured to detect the presence of nearby objects in the absence of any physical contact.
- Sensor assembly 814 may also include a light sensor, such as a CMOS or CCD image sensor, for use in imaging applications.
- the sensor assembly 814 may also include an acceleration sensor, a gyroscope sensor, a magnetic sensor, a pressure sensor, or a temperature sensor.
- Communication component 816 is configured to facilitate wired or wireless communication between electronic device 800 and other devices.
- Electronic device 800 may access wireless networks based on communication standards, such as WiFi, 2G or 3G, or a combination thereof.
- the communication component 816 receives broadcast signals or broadcast related information from an external broadcast management system via a broadcast channel.
- the communication component 816 further includes a near field communication (NFC) module to facilitate short-range communication.
- NFC near field communication
- the NFC module may be implemented based on radio frequency identification (RFID) technology, infrared data association (IrDA) technology, ultra-wideband (UWB) technology, Bluetooth (BT) technology and other technologies.
- RFID radio frequency identification
- IrDA infrared data association
- UWB ultra-wideband
- Bluetooth Bluetooth
- electronic device 800 may be implemented by one or more application specific integrated circuits (ASICs), digital signal processors (DSPs), digital signal processing devices (DSPDs), programmable logic devices (PLDs), field programmable A programmed gate array (FPGA), controller, microcontroller, microprocessor or other electronic component implementation is used to perform the above method.
- ASICs application specific integrated circuits
- DSPs digital signal processors
- DSPDs digital signal processing devices
- PLDs programmable logic devices
- FPGA field programmable A programmed gate array
- controller microcontroller, microprocessor or other electronic component implementation is used to perform the above method.
- a non-volatile computer-readable storage medium such as a memory 804 comprising computer program instructions executable by the processor 820 of the electronic device 800 to perform the above method is also provided.
- An embodiment of the present disclosure further provides an electronic device, including: a processor; a memory for storing instructions executable by the processor; wherein the processor is configured to perform the above method.
- the electronic device may be provided as a terminal, server or other form of device.
- FIG. 5 is a block diagram of an electronic device 1900 according to an exemplary embodiment.
- the electronic device 1900 may be provided as a server.
- electronic device 1900 includes processing component 1922, which further includes one or more processors, and a memory resource represented by memory 1932 for storing instructions executable by processing component 1922, such as applications.
- An application program stored in memory 1932 may include one or more modules, each corresponding to a set of instructions.
- the processing component 1922 is configured to execute instructions to perform the above-described methods.
- the electronic device 1900 may also include a power supply assembly 1926 configured to perform power management of the electronic device 1900, a wired or wireless network interface 1950 configured to connect the electronic device 1900 to a network, and an input output (I/O) interface 1958 .
- the electronic device 1900 can operate based on an operating system stored in the memory 1932, such as a Microsoft server operating system (Windows ServerTM), a graphical user interface based operating system (Mac OS XTM) introduced by Apple, a multi-user multi-process computer operating system (UnixTM). ), Free and Open Source Unix-like Operating System (LinuxTM), Open Source Unix-like Operating System (FreeBSDTM) or similar.
- a non-volatile computer-readable storage medium such as memory 1932 comprising computer program instructions executable by processing component 1922 of electronic device 1900 to perform the above-described method.
- Embodiments of the present disclosure may be systems, methods and/or computer program products.
- the computer program product may include a computer-readable storage medium having computer-readable program instructions loaded thereon for causing a processor to implement various aspects of the embodiments of the present disclosure.
- a computer-readable storage medium may be a tangible device that can hold and store instructions for use by the instruction execution device.
- the computer-readable storage medium may be, for example, but not limited to, an electrical storage device, a magnetic storage device, an optical storage device, an electromagnetic storage device, a semiconductor storage device, or any suitable combination of the foregoing.
- Non-exhaustive list of computer readable storage media include: portable computer disks, hard disks, random access memory (RAM), read only memory (ROM), erasable programmable read only memory (EPROM) or flash memory), static random access memory (SRAM), portable compact disk read only memory (CD-ROM), digital versatile disk (DVD), memory sticks, floppy disks, mechanically coded devices, such as printers with instructions stored thereon Hole cards or raised structures in grooves, and any suitable combination of the above.
- RAM random access memory
- ROM read only memory
- EPROM erasable programmable read only memory
- flash memory static random access memory
- SRAM static random access memory
- CD-ROM compact disk read only memory
- DVD digital versatile disk
- memory sticks floppy disks
- mechanically coded devices such as printers with instructions stored thereon Hole cards or raised structures in grooves, and any suitable combination of the above.
- Computer-readable storage media are not to be construed as transient signals per se, such as radio waves or other freely propagating electromagnetic waves, electromagnetic waves propagating through waveguides or other transmission media (eg, light pulses through fiber optic cables), or through electrical wires transmitted electrical signals.
- the computer readable program instructions described herein may be downloaded to various computing/processing devices from a computer readable storage medium, or to an external computer or external storage device over a network such as the Internet, a local area network, a wide area network, and/or a wireless network.
- the network may include copper transmission cables, fiber optic transmission, wireless transmission, routers, firewalls, switches, gateway computers, and/or edge servers.
- a network adapter card or network interface in each computing/processing device receives computer-readable program instructions from a network and forwards the computer-readable program instructions for storage in a computer-readable storage medium in each computing/processing device .
- the computer program instructions for carrying out the operations of the disclosed embodiments may be assembly instructions, instruction set architecture (ISA) instructions, machine instructions, machine dependent instructions, microcode, firmware instructions, state setting data, or programmed in one or more Source or object code written in any combination of languages, including object-oriented programming languages, such as Smalltalk, C++, etc., and conventional procedural programming languages, such as the "C" language or similar programming languages.
- the computer readable program instructions may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer, or entirely on the remote computer or server implement.
- the remote computer may be connected to the user's computer through any kind of network, including a local area network (LAN) or a wide area network (WAN), or may be connected to an external computer (eg, using an Internet service provider through the Internet connect).
- LAN local area network
- WAN wide area network
- custom electronic circuits such as programmable logic circuits, field programmable gate arrays (FPGAs), or programmable logic arrays (PLAs)
- FPGAs field programmable gate arrays
- PDAs programmable logic arrays
- Computer readable program instructions are executed to implement various aspects of the embodiments of the present disclosure.
- These computer readable program instructions may be provided to a processor of a general purpose computer, special purpose computer or other programmable data processing apparatus to produce a machine that causes the instructions when executed by the processor of the computer or other programmable data processing apparatus , resulting in means for implementing the functions/acts specified in one or more blocks of the flowchart and/or block diagrams.
- These computer readable program instructions can also be stored in a computer readable storage medium, these instructions cause a computer, programmable data processing apparatus and/or other equipment to operate in a specific manner, so that the computer readable medium on which the instructions are stored includes An article of manufacture comprising instructions for implementing various aspects of the functions/acts specified in one or more blocks of the flowchart and/or block diagrams.
- Computer readable program instructions can also be loaded onto a computer, other programmable data processing apparatus, or other equipment to cause a series of operational steps to be performed on the computer, other programmable data processing apparatus, or other equipment to produce a computer-implemented process , thereby causing instructions executing on a computer, other programmable data processing apparatus, or other device to implement the functions/acts specified in one or more blocks of the flowcharts and/or block diagrams.
- each block in the flowchart or block diagrams may represent a module, segment, or portion of instructions, which comprises one or more functions for implementing the specified logical function(s) executable instructions.
- the functions noted in the blocks may occur out of the order noted in the figures. For example, two blocks in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved.
- each block of the block diagrams and/or flowchart illustrations, and combinations of blocks in the block diagrams and/or flowchart illustrations can be implemented in dedicated hardware-based systems that perform the specified functions or actions , or can be implemented in a combination of dedicated hardware and computer instructions.
- Embodiments of the present disclosure relate to a target detection method and apparatus, electronic device, and storage medium, wherein the method includes: acquiring a first detection result obtained by performing target detection on a current data frame of a target scene; The historical optimization result updates the first detection result to obtain the first observation result of the target object in the current data frame; the first observation result is corrected according to the point cloud data corresponding to the first observation result, Obtain the first correction result of the target object.
Abstract
Description
Claims (16)
- 一种目标检测方法,所述方法由电子设备执行,包括:获取针对目标场景的当前数据帧进行目标检测得到的第一检测结果;基于所述目标场景的历史优化结果对所述第一检测结果进行更新,得到所述当前数据帧中目标对象的第一观测结果;根据所述第一观测结果对应的点云数据对所述第一观测结果进行修正,得到所述目标对象的第一修正结果。
- 根据权利要求1所述的方法,其中,所述基于所述目标场景的历史优化结果对所述第一检测结果进行更新,得到所述当前数据帧中目标对象的第一观测结果,包括:基于所述目标场景的历史优化结果,确定所述第一检测结果的对象信息,其中,所述对象信息用于标识所述目标对象;根据所述第一检测结果的对象信息对所述第一检测结果进行更新,得到所述当前数据帧中目标对象的第一观测结果。
- 根据权利要求2所述的方法,其中,所述基于所述目标场景的历史优化结果,确定所述第一检测结果的对象信息,包括:将所述目标场景的历史优化结果与所述第一检测结果进行匹配;在所述第一检测结果与所述历史优化结果匹配的情况下,将所述历史优化结果的对象信息确定为所述第一检测结果的对象信息。
- 根据权利要求3所述的方法,其中,所述基于所述目标场景的历史优化结果,确定所述第一检测结果的对象信息,包括:在所述第一检测结果与所述历史优化结果不匹配的情况下,为所述第一检测结果设置新的对象信息。
- 根据权利要求3或4所述的方法,其中,所述将所述目标场景的历史优化结果与所述第一检测结果进行匹配,包括:确定所述第一检测结果与一个历史优化结果的检测框交叠部分的第一体积,以及,确定所述第一检测结果的检测框与该历史优化结果的检测框共同占据的总体积;根据所述第一体积与所述总体积的比值确定所述第一检测结果与该的历史优化结果的匹配程度。
- 根据权利要求1至5任意一项所述的方法,其中,所述基于所述目标场景的历史优化结果对所述第一检测结果进行更新,得到所述当前数据帧中目标对象的第一观测结果,包括:基于所述目标场景的历史优化结果,在确定所述当前数据帧存在所述第一检测结果未检测到的目标对象的情况下,将所述未检测到的目标对象的历史优化结果确定为所述当前数据帧中未检测到的目标对象的第一观测结果。
- 根据权利要求1至6任意一项所述的方法,其中,所述根据所述第一观测结果对应的点云数据对所述第一观测结果进行修正,得到所述目标对象的第一修正结果,包括:对同一目标对象的历史优化结果对应的点云数据与第一观测结果对应的点云数据进行合并,得到合并点云数据;基于所述合并点云数据,得到对所述第一观测结果进行修正的第一修正结果。
- 根据权利要求7所述的方法,其中,所述对同一目标对象的历史优化结果对应的点云数据与第一观测结果对应的点云数据进行合并,包括:针对同一目标对象,将所述当前数据帧的前一个数据帧的历史优化结果对应的点云数据与所述第一观测结果对应的点云数据进行合并。
- 根据权利要求1至8任意一项所述的方法,其中,所述方法还包括:获取所述目标对象的修正结果,其中,所述修正结果包括所述第一修正结果和第二修正结果,所述第二修正结果为基于目标场景的历史数据帧进行目标检测得到的;基于所述修正结果中的目标结果,确定所述目标对象的当前优化结果。
- 根据权利要求9所述的方法,其中,所述方法还包括:确定所述修正结果中的第一修正结果分别与多个第二修正结果的误差,其中,所述第一修正结果为任意一个所述 修正结果,所述第二修正结果为所述第一修正结果之外的修正观测框;统计所述第一修正结果对应的内点数量,其中,所述内点数量为与所述第一修正结果的误差小于误差阈值的第二修正结果的数量;根据所述第一修正结果对应的内点数量确定所述修正结果中的目标结果。
- 根据权利要求10所述的方法,其中,所述根据所述第一修正结果对应的内点数量确定所述修正结果中的目标结果,包括:确定多个所述第一修正结果中所述内点数量最大的第一修正结果;将所述内点数量最大的第一修正结果以及与所述内点数量最大的第一修正结果的误差小于所述误差阈值的第二修正结果,确定为所述修正结果中的目标结果。
- 根据权利要求9至11中任意一项所述的方法,其中,所述当前优化结果与多个所述目标结果的误差之和达到最小。
- 一种目标检测装置,其中,包括:获取模块,配置为获取针对目标场景的当前数据帧进行目标检测得到的第一检测结果;确定模块,配置为基于所述目标场景的历史优化结果对所述第一检测结果进行更新,得到所述当前数据帧中目标对象的第一观测结果;修正模块,配置为根据所述第一观测结果对应的点云数据对所述第一观测结果进行修正,得到所述目标对象的第一修正结果。
- 一种电子设备,其中,包括:处理器;用于存储处理器可执行指令的存储器;其中,所述处理器被配置为调用所述存储器存储的指令,以执行权利要求1至12中任意一项所述的目标检测方法。
- 一种计算机可读存储介质,其上存储有计算机程序指令,其中,所述计算机程序指令被处理器执行时实现权利要求1至12中任意一项所述的目标检测方法。
- 一种计算机程序,包括计算机可读代码,当所述计算机可读代码在电子设备中运行时,所述电子设备中的处理器执行用于实现权利要求1至12任一项所述的方法。
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010725039.5 | 2020-07-24 | ||
CN202010725039.5A CN111860373B (zh) | 2020-07-24 | 2020-07-24 | 目标检测方法及装置、电子设备和存储介质 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2022017140A1 true WO2022017140A1 (zh) | 2022-01-27 |
Family
ID=72949590
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2021/103286 WO2022017140A1 (zh) | 2020-07-24 | 2021-06-29 | 目标检测方法及装置、电子设备和存储介质 |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN111860373B (zh) |
WO (1) | WO2022017140A1 (zh) |
Families Citing this family (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111860373B (zh) * | 2020-07-24 | 2022-05-20 | 浙江商汤科技开发有限公司 | 目标检测方法及装置、电子设备和存储介质 |
CN112330717B (zh) * | 2020-11-11 | 2023-03-10 | 北京市商汤科技开发有限公司 | 目标跟踪方法及装置、电子设备和存储介质 |
CN112528889B (zh) * | 2020-12-16 | 2024-02-06 | 中国平安财产保险股份有限公司 | Ocr信息检测修正方法、装置、终端及存储介质 |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108053427A (zh) * | 2017-10-31 | 2018-05-18 | 深圳大学 | 一种基于KCF与Kalman的改进型多目标跟踪方法、系统及装置 |
CN109636829A (zh) * | 2018-11-24 | 2019-04-16 | 华中科技大学 | 一种基于语义信息和场景信息的多目标跟踪方法 |
CN110827202A (zh) * | 2019-11-07 | 2020-02-21 | 上海眼控科技股份有限公司 | 目标检测方法、装置、计算机设备和存储介质 |
WO2020108311A1 (zh) * | 2018-11-29 | 2020-06-04 | 北京市商汤科技开发有限公司 | 目标对象3d检测方法、装置、介质及设备 |
CN111860373A (zh) * | 2020-07-24 | 2020-10-30 | 浙江商汤科技开发有限公司 | 目标检测方法及装置、电子设备和存储介质 |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108256506B (zh) * | 2018-02-14 | 2020-11-24 | 北京市商汤科技开发有限公司 | 一种视频中物体检测方法及装置、计算机存储介质 |
CN110555339A (zh) * | 2018-05-31 | 2019-12-10 | 北京嘀嘀无限科技发展有限公司 | 一种目标检测方法、系统、装置及存储介质 |
CN108734360B (zh) * | 2018-06-19 | 2021-09-07 | 哈尔滨工业大学 | 一种基于修正的elm预测模型多维遥测数据智能判读方法 |
-
2020
- 2020-07-24 CN CN202010725039.5A patent/CN111860373B/zh active Active
-
2021
- 2021-06-29 WO PCT/CN2021/103286 patent/WO2022017140A1/zh active Application Filing
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108053427A (zh) * | 2017-10-31 | 2018-05-18 | 深圳大学 | 一种基于KCF与Kalman的改进型多目标跟踪方法、系统及装置 |
CN109636829A (zh) * | 2018-11-24 | 2019-04-16 | 华中科技大学 | 一种基于语义信息和场景信息的多目标跟踪方法 |
WO2020108311A1 (zh) * | 2018-11-29 | 2020-06-04 | 北京市商汤科技开发有限公司 | 目标对象3d检测方法、装置、介质及设备 |
CN110827202A (zh) * | 2019-11-07 | 2020-02-21 | 上海眼控科技股份有限公司 | 目标检测方法、装置、计算机设备和存储介质 |
CN111860373A (zh) * | 2020-07-24 | 2020-10-30 | 浙江商汤科技开发有限公司 | 目标检测方法及装置、电子设备和存储介质 |
Also Published As
Publication number | Publication date |
---|---|
CN111860373B (zh) | 2022-05-20 |
CN111860373A (zh) | 2020-10-30 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2022017140A1 (zh) | 目标检测方法及装置、电子设备和存储介质 | |
WO2021051857A1 (zh) | 目标对象匹配方法及装置、电子设备和存储介质 | |
US9953506B2 (en) | Alarming method and device | |
CN111105454B (zh) | 一种获取定位信息的方法、装置及介质 | |
CN111551191B (zh) | 传感器外参数标定方法及装置、电子设备和存储介质 | |
CN107692997B (zh) | 心率检测方法及装置 | |
WO2022043741A1 (zh) | 网络训练、行人重识别方法及装置、存储介质、计算机程序 | |
WO2020181728A1 (zh) | 图像处理方法及装置、电子设备和存储介质 | |
JP2016531361A (ja) | 画像分割方法、画像分割装置、画像分割デバイス、プログラム及び記録媒体 | |
TWI718631B (zh) | 人臉圖像的處理方法及裝置、電子設備和儲存介質 | |
CN111323007B (zh) | 定位方法及装置、电子设备和存储介质 | |
WO2022021872A1 (zh) | 目标检测方法及装置、电子设备和存储介质 | |
WO2022134475A1 (zh) | 点云地图构建方法及装置、电子设备、存储介质和程序 | |
EP3147802A1 (en) | Method and apparatus for processing information | |
WO2015196715A1 (zh) | 图像重定位方法、装置及终端 | |
WO2023273498A1 (zh) | 深度检测方法及装置、电子设备和存储介质 | |
CN111563138A (zh) | 定位方法及装置、电子设备和存储介质 | |
WO2023273499A1 (zh) | 深度检测方法及装置、电子设备和存储介质 | |
WO2022222379A1 (zh) | 一种位置确定方法及装置、电子设备和存储介质 | |
WO2022179080A1 (zh) | 定位方法、装置、电子设备、存储介质、程序及产品 | |
CN112837372A (zh) | 数据生成方法及装置、电子设备和存储介质 | |
WO2023155393A1 (zh) | 特征点匹配方法、装置、电子设备、存储介质和计算机程序产品 | |
WO2022110801A1 (zh) | 数据处理方法及装置、电子设备和存储介质 | |
WO2022237071A1 (zh) | 定位方法及装置、电子设备、存储介质和计算机程序 | |
CN112949568A (zh) | 人脸和人体匹配的方法及装置、电子设备和存储介质 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 21847104 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 21847104 Country of ref document: EP Kind code of ref document: A1 |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 21847104 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 31/07/2023) |