WO2022135594A1 - Method and apparatus for detecting target object, fusion processing unit, and medium - Google Patents

Method and apparatus for detecting target object, fusion processing unit, and medium Download PDF

Info

Publication number
WO2022135594A1
WO2022135594A1 PCT/CN2021/141370 CN2021141370W WO2022135594A1 WO 2022135594 A1 WO2022135594 A1 WO 2022135594A1 CN 2021141370 W CN2021141370 W CN 2021141370W WO 2022135594 A1 WO2022135594 A1 WO 2022135594A1
Authority
WO
WIPO (PCT)
Prior art keywords
target object
motion state
data
information
target
Prior art date
Application number
PCT/CN2021/141370
Other languages
French (fr)
Chinese (zh)
Inventor
吴臻志
杨哲宇
马欣
祝夭龙
Original Assignee
北京灵汐科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from CN202011562118.5A external-priority patent/CN112816995B/en
Priority claimed from CN202011560817.6A external-priority patent/CN112666550B/en
Application filed by 北京灵汐科技有限公司 filed Critical 北京灵汐科技有限公司
Publication of WO2022135594A1 publication Critical patent/WO2022135594A1/en

Links

Images

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S13/00Systems using the reflection or reradiation of radio waves, e.g. radar systems; Analogous systems using reflection or reradiation of waves whose nature or wavelength is irrelevant or unspecified
    • G01S13/86Combinations of radar systems with non-radar systems, e.g. sonar, direction finder
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S17/00Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
    • G01S17/02Systems using the reflection of electromagnetic waves other than radio waves
    • G01S17/04Systems determining the presence of a target

Definitions

  • the present disclosure relates to the technical field of target detection, and in particular, to a target object detection method, a target object detection device, a fusion processing unit, and a computer-readable medium.
  • Object detection is an important technology for video image analysis and understanding, and an important preprocessing step for some computational vision tasks, such as object recognition and moving object tracking.
  • Inter-frame difference method and optical flow method are common methods for detecting moving objects in some related technologies.
  • the inter-frame difference method obtains the outline of the moving target by calculating the difference between two adjacent frames of images, specifically, subtracting the two frames of images to obtain the absolute value of the brightness difference between the two frames of images, and by judging whether the calculated absolute value is greater than the threshold value. Analyze the motion characteristics of video or image sequences.
  • the law of optical flow is to describe the motion of the observation target, surface or edge caused by the motion of the observer relative to the motion of the observer.
  • moving objects are detected based on temporal perception, but the spatial perception ability is weak, and multi-dimensional visual perception cannot be performed on moving objects and/or stationary objects at the same time.
  • the detection effect of the moving object detection algorithm is not ideal.
  • the present disclosure provides a method for detecting a target object, a device for detecting a target object, a fusion processing unit, and a computer-readable medium.
  • an embodiment of the present disclosure provides a method for detecting a target object, where the method includes:
  • the event data represents light intensity change information in the target plane, and the light intensity change information is used to determine at least one target object in the target plane;
  • the radar detection data is information describing the motion state of the target object
  • the event data and the radar detection data are fused to generate multi-dimensional motion state information of the target object.
  • an embodiment of the present disclosure provides a detection device for a target object, the detection device includes:
  • a first sensor for detecting light intensity change information in the target plane to generate event data, the light intensity change information being used to determine at least one target object in the target plane;
  • a radar for acquiring radar detection data for the target object, where the radar detection data is information describing the motion state of the target object;
  • the fusion processing unit is configured to perform fusion processing on the event data and the radar detection data to generate multi-dimensional motion state information of the target object.
  • an embodiment of the present disclosure provides a fusion processing unit, which is applied to a target object detection device, and the fusion processing unit includes:
  • processors one or more processors
  • a storage device on which one or more programs are stored, when the one or more programs are executed by the one or more processors, the one or more processors implement the first aspect of the embodiments of the present disclosure; The detection method of the target object described.
  • an embodiment of the present disclosure provides a computer-readable medium on which a computer program is stored, and when the program is executed by a processor, implements the method for detecting a target object described in the first aspect of the embodiment of the present disclosure.
  • the event data representing the light intensity change information in the target plane and the radar detection data for the target object are fused to realize the collection of multi-dimensional motion state information of the target object, so that at least one The stationary or moving target object can be clearly judged by motion, so that the detection device has biological-like vision, and the biological-like visual perception of the stationary or moving target object is realized, and the collected event data is detected by the motion-sensitive sensor.
  • Real-time dynamic response is generated, which effectively reduces the influence of redundant data on the detection effect, realizes the real-time dynamic response of target object detection, and effectively improves the efficiency of moving object detection.
  • FIG. 1 is a schematic flowchart of a detection method in an embodiment of the present disclosure
  • FIG. 2 is a schematic flowchart of some steps in a detection method according to an embodiment of the present disclosure
  • FIG. 3 is a schematic flowchart of some steps in a detection method according to an embodiment of the present disclosure
  • FIG. 4 is a schematic diagram of an implementation manner of determining an offset in an embodiment of the present disclosure
  • FIG. 5 is a schematic flowchart of some steps in a detection method according to an embodiment of the present disclosure.
  • FIG. 6 is a schematic flowchart of some steps in a detection method according to an embodiment of the present disclosure.
  • FIG. 7 is a schematic flowchart of some steps in a detection method according to an embodiment of the present disclosure.
  • FIG. 8 is a schematic flowchart of some steps in a detection method according to an embodiment of the present disclosure.
  • FIG. 9 is a schematic flowchart of some steps in a detection method according to an embodiment of the present disclosure.
  • FIG. 10 is a block diagram of the composition of a detection device in an embodiment of the present disclosure.
  • FIG. 11 is a block diagram of another detection device in an embodiment of the present disclosure.
  • FIG. 12 is a block diagram of another detection device in an embodiment of the present disclosure.
  • FIG. 13 is a block diagram of a fusion processing unit in an embodiment of the present disclosure.
  • an embodiment of the present disclosure provides a method for detecting a target object.
  • the detecting method includes steps S100 to S300.
  • step S100 event data is acquired, where the event data represents light intensity change information in the target plane, and the light intensity change information is used to determine at least one target object in the target plane.
  • step S200 radar detection data for the target object is acquired, where the radar detection data is information describing the motion state of the target object.
  • step S300 the event data and the radar detection data are fused to generate multi-dimensional motion state information of the target object.
  • a device for detecting a target object which is used to detect the motion state of the target object.
  • the detection device of the target object includes a first sensor and a radar.
  • the first sensor in the detection device is sensitive to motion and can respond dynamically to scene changes in real time.
  • the event data is generated by collecting the light intensity change information in the target plane by the first sensor, wherein the first sensor is a sensor imitating the working mechanism of biological vision, A sensor only retains dynamic information when generating event data.
  • the target plane may represent the image captured by the first sensor for the target scene, and the first sensor may generate event data by sensing the light intensity change information of each pixel in the captured image of the target scene, and the event data may be stored in the target plane.
  • the first sensor is a Dynamic Vision Sensor (DVS, Dynamic Vision Sensor).
  • step S200 the motion state of the target object is detected by the radar to generate radar detection data.
  • step S100 and step S200 there is no special restriction on the execution sequence of step S100 and step S200, and step S100 and step S200 may be executed at the same time to obtain event data and radar detection data corresponding to the same time point; Steps S100 and S200 are executed to obtain event data corresponding to multiple time points and radar detection data corresponding to multiple time points.
  • step S300 the fusion processing of event data and radar detection data is performed by aligning and calibrating event data and radar detection data corresponding to the same time point to generate multi-dimensional motion state information of the target object.
  • the multi-dimensional motion state information may include three-dimensional velocity and three-dimensional position coordinates of the target object at various time points, and may also include velocity components of the target object in multiple predetermined directions including the first direction at each time point, and further It may include the movement trajectory of the target object generated according to the three-dimensional velocity and three-dimensional position coordinates of multiple time points, and the like.
  • the multi-dimensional motion state information may include at least one of the distance (distance between the target object and the detection device), orientation, height, speed, attitude, shape and other information of the target object.
  • the event data representing the light intensity change information in the target plane and the radar detection data for the target object are fused to realize the collection of multi-dimensional motion state information of the target object, so that at least one The stationary or moving target object can be clearly judged by motion, so that the detection device has biological-like vision, and the biological-like visual perception of the stationary or moving target object is realized, and the collected event data is detected by the motion-sensitive sensor.
  • Real-time dynamic response is generated, which effectively reduces the influence of redundant data on the detection effect, realizes the real-time dynamic response of target object detection, and effectively improves the efficiency of moving object detection.
  • the above-mentioned radar for collecting radar detection data may be a pulse Doppler radar.
  • the pulse Doppler radar detects the target object by transmitting a pulse signal to the target object and receiving the pulse signal reflected by the target object.
  • the movement state of the radar can be obtained to obtain radar detection data.
  • the radar detection data may include first motion state component information of the target object in a first direction, and the first direction is perpendicular to the target plane.
  • the first motion state component information may include the velocity component of the target object in the first direction, the distance from the target object to the detection device, and the like.
  • the first direction is a direction parallel to the radial direction of the radar
  • the target plane is a plane perpendicular to the radial direction of the radar.
  • the motion state component information of the target object on the target plane can be determined according to the event data, and the multi-dimensional motion of the target object can be determined according to the motion state component information of the target object on the target plane and the motion state component information in the first direction status information.
  • step S300 may further include steps S310A-step S320A.
  • step S310A the second motion state component information of the target object on the target plane is determined according to the event data.
  • step S320A multi-dimensional motion state information of the target object is generated according to the second motion state component information and the first motion state component information.
  • the target object may be detected by radar to generate the first motion state component information; in other embodiments, the target object may also be detected by radar to generate the initial detection signal, for example, The Doppler frequency shift corresponding to the target object measured by the pulse Doppler radar, and then the first motion state component information is determined according to the initial detection information.
  • This embodiment of the present disclosure makes no special limitation on this.
  • the detection method may further include the step of determining the first motion state component information according to the initial detection signal.
  • the first sensor in the detection device may be a Dynamic Vision Sensor (DVS, Dynamic Vision Sensor).
  • DVS Dynamic Vision Sensor
  • DVS is a sensor that imitates the working mechanism of biological vision. It can detect the change of light and output the address and information of the pixel where the light intensity changes. It eliminates redundant data and can dynamically respond to scene changes in real time.
  • the event data collected by the DVS is the two-dimensional data of the target plane. Determine the target object in the target plane, and determine the motion state component information of the target object in the target plane.
  • the DVS does not need to read all the pixels in the picture, but only needs to obtain the address and information of the pixels whose light intensity changes; specifically, when the DVS detects the light of a certain pixel When the intensity change is greater than or equal to the preset threshold value, the event signal of the pixel is sent; if the change of the light intensity is a positive change, that is, the pixel jumps from low brightness to high brightness, it will send out with "+1" If the light intensity change is a negative change, that is, the pixel jumps from high brightness to low brightness, an event signal represented by "-1" is sent and marked as negative.
  • Event if the change of light intensity is less than the preset threshold value, no event signal will be sent, and it will be marked as no event; DVS constitutes event data by marking the event of each pixel where the light intensity changes. Among them, in the event data, both positive events and negative events can be used to represent the light intensity change information of the pixel point.
  • the above-mentioned event data may include the coordinates of the pixels where the light intensity changes in the target plane and the light intensity change information, and may further include time information, and the time information may indicate that the light intensity changes. time; the second motion state component information includes the position coordinates of the target object in the target plane; referring to FIG. 3 , step S310A may further include steps S311A and S312A.
  • step S311A an event frame is generated according to the coordinates and light intensity change information of each pixel in the event data within a preset sampling period.
  • step S312A the position coordinates of the target object in the target plane are determined according to the event frame.
  • the event data sampled in the preset sampling period is framed to generate an event frame, and the event frame can represent the data generated for each pixel in the preset sampling period.
  • the output data of the DVS is event data consisting of a plurality of 4-tuple data, each 4-tuple data corresponds to a pixel whose light intensity changes in the target plane, and the 4-tuple data includes pixels whose light intensity changes.
  • the coordinates of the point in the target plane (abscissa x, ordinate y), light intensity change information, and time information.
  • the 4-tuple data corresponding to the same time point is framed, thereby generating a corresponding event frame.
  • a target detection algorithm may be used to determine the position coordinates of the target object in the target plane according to the event frame. For example, when the target object in the picture moves relatively, the light intensity of the corresponding pixel will change to varying degrees. For example, when the target object appears, the light intensity of the pixel in the area where the target object appears will increase significantly, and when the target object disappears , the brightness of the pixels in the disappearing area of the target object will be significantly reduced.
  • the position coordinates of the pixels where the light intensity changes it can be determined which pixels in the picture may have the target object, and the contour area of the target object can be determined, and then Obtain the location area of the target object, and determine the location coordinates of the target object.
  • the coordinates of any point on the target object in the target plane may be used as the position coordinates of the target object in the target plane.
  • the coordinates of the center point of the target object may be used as the coordinates of the target object in the target plane. location coordinates in . This embodiment of the present disclosure makes no special limitation on this.
  • the position coordinates of the target object in the target plane may be represented by (x, y).
  • x corresponds to one of the second direction and the third direction
  • y corresponds to the other of the second direction and the third direction
  • the second direction and the third direction are both parallel to the target plane
  • the second direction and the third direction are parallel to the target plane.
  • Three directions are vertical.
  • the second motion state component information further includes the second velocity component of the target object in the second direction and the third velocity component in the third direction; referring to FIG. 3 , step S310A further includes It may further include: step S313A-step S315A.
  • step S313A the second offset of the target object in the second direction and the third offset of the target object in the third direction are respectively determined according to the position coordinates of the target object in the target plane.
  • step S314A a second velocity component of the target object in the second direction is determined according to the second offset.
  • step S315A the third velocity component of the target object in the third direction is determined according to the third offset.
  • the second direction and the third direction in the target plane are not particularly limited.
  • the first direction, the second direction, and the third direction constitute a three-dimensional rectangular coordinate system
  • the first direction is the ordinate direction of the three-dimensional rectangular coordinate system
  • the second direction is the abscissa direction of the three-dimensional rectangular coordinate system direction
  • the third direction is the vertical coordinate direction of the three-dimensional Cartesian coordinate system.
  • FIG. 4 shows an optional implementation manner of determining the third offset in the third direction according to the coordinates of the pixel points where the light intensity changes in the target plane at adjacent time points. As shown in Figure 4, the absolute offset of the target object in the third direction can be obtained. Similarly, the absolute offset of the target object in the second direction can be obtained.
  • step S314A and step S315A when determining the second velocity component and the third velocity component, the time difference between adjacent time points is also combined.
  • the first motion state component information includes a first velocity component of the target object in a first direction and a distance parameter, where the distance parameter represents a distance between the target object and the detection device;
  • the multi-dimensional motion state information includes a three-dimensional Speed and three-dimensional coordinates; referring to FIG. 3 , step S320A may further include: step S321A and step S322A.
  • step S321A the position coordinates of the target object in the target plane, the first velocity component, the second velocity component, the third velocity component and the distance parameter of the target object in the first direction corresponding to the same time point are determined.
  • step S322A the target object is determined according to the position coordinates of the target object corresponding to the same time point in the target plane, the first velocity component, the second velocity component, the third velocity component and the distance parameter of the target object in the first direction The 3D velocity and the 3D coordinates.
  • the three-dimensional coordinates represent the position coordinates of the target object in the three-dimensional space
  • the three-dimensional velocity represents the relative movement speed of the target object in the three-dimensional space.
  • the three-dimensional coordinates can be represented by (x, y, z). Wherein, z corresponds to the first direction, x corresponds to one of the second direction and the third direction, and y corresponds to the other of the second direction and the third direction.
  • step S100 event data is acquired through the DVS.
  • step S100 includes: in step S110 , acquiring event data in response to changes in the light intensity of pixels in the target plane, where the event data includes pixels whose light intensity changes in the target plane coordinates and light intensity change information.
  • the event data may further include time information, and the time information may represent the time when the light intensity of the pixel changes.
  • step S200 may further include: in step S210 , acquiring the first velocity component and distance parameter of the target object in the first direction as the first motion state component information.
  • step S300 the event data and the first motion state component information are fused to generate multi-dimensional motion state information of the target object.
  • the event data representing the light intensity change information in the target plane and the first motion state component information of the target object in the first direction detected by the pulse Doppler radar are fused to achieve
  • the multi-dimensional motion state information of the target object is collected, in which the event data is generated by the real-time dynamic response of the scene change by the motion-sensitive sensor, which effectively reduces the influence of redundant data on the detection effect and realizes the detection of the target object.
  • Real-time dynamic response effectively improves the efficiency of moving object detection.
  • the method for detecting a target object may further include: in step S400 , outputting multi-dimensional motion state information.
  • the multi-dimensional motion state information may include the three-dimensional velocity and three-dimensional position coordinates of the target object at various time points, and may also include the velocity components of the target object in the first direction, the second direction, and the third direction at each time point , and may also include a motion trajectory of the target object generated according to the three-dimensional velocity and three-dimensional position coordinates of multiple time points.
  • the multi-dimensional motion state information may include at least one of the distance (distance between the target object and the detection device), orientation, height, speed, attitude, shape and other information of the target object.
  • This embodiment of the present disclosure does not specifically limit how to output the multi-dimensional motion state information.
  • the multi-dimensional motion state information of the target object can be displayed on the display screen.
  • the above-mentioned radar for collecting radar detection data may be a lidar (Lidar), a laser radar, and the collected radar detection data is laser point cloud data; A set of vectors in a three-dimensional coordinate system.
  • the vectors in the set can be represented in the form of X, Y, and Z three-dimensional coordinates.
  • Lidar is a radar system that emits laser beams to detect the position, velocity and other characteristic quantities of target objects.
  • Lidar generates laser point cloud data by scanning, which can characterize the motion state of at least one target object such as distance, azimuth, height, speed, attitude, and shape.
  • the lidar can detect moving objects as well as stationary objects. Therefore, in the embodiment of the present disclosure, the target object may be a moving object or a stationary object. This embodiment of the present disclosure makes no special limitation on this.
  • the multi-dimensional motion state information may include at least one of the distance (distance between the target object and the detection device), azimuth, altitude, speed, attitude, and shape of the target object.
  • the multi-dimensional motion state information of the target object is generated by acquiring laser point cloud data and event data representing light intensity change information in the target plane, and fusing the laser point cloud data and the event data. , so that at least one stationary or moving target object can be clearly judged on the motion, and the biological-like visual perception of the stationary or moving target object is realized.
  • step S300 does not specifically limit how to perform step S300 to perform fusion processing on event data and laser point cloud data.
  • a neural network (such as a convolutional neural network) is used to perform fusion processing on event data and laser point cloud data.
  • the input of the neural network is a three-dimensional image and an event frame, wherein the three-dimensional image is generated according to laser point cloud data, and the event frame is generated by framing according to the acquired event data.
  • step S300 may further include: steps S310B to S330B.
  • step S310B a three-dimensional image is generated according to the laser point cloud data.
  • the point cloud data includes the distance information between the point corresponding to the reflected signal in all the emitted laser signals and the emission source (ie, the laser radar).
  • the corresponding three-dimensional image can be obtained by transforming the spatial position information from spherical coordinates to XYZ three-dimensional coordinates, and the generated three-dimensional image may refer to a three-dimensional point cloud image.
  • step S320B an event frame is generated according to the event data within the preset sampling period.
  • step S320B For the description of this step S320B, reference may be made to the above description of step S311A, and details are not repeated here.
  • step S330B the three-dimensional image and the event frame are input into the neural network for processing to generate multi-dimensional motion state information of the target object.
  • step S330B the three-dimensional image is projected onto the same two-dimensional plane as the event frame, and the two images are stitched together in the channel dimension and then input to a neural network (eg, a convolutional neural network) for Feature extraction, so as to obtain the multi-dimensional motion state information of the target object.
  • a neural network eg, a convolutional neural network
  • the inputs to a neural network are two-dimensional images and event frames.
  • the two-dimensional image is the top view and front view obtained by projecting the three-dimensional laser point cloud data along the front-view direction and the top-view direction respectively, so as to obtain the two-dimensional image representation of the three-dimensional laser point cloud data
  • the event frame is the preset sampling period.
  • the event data is framed and fed into the neural network frame by frame.
  • step S300 may further include steps S310C to S330C.
  • step S310C the laser point cloud data is processed to generate a front view and a top view of the laser point cloud data.
  • the point cloud data is represented by three-dimensional coordinates.
  • the three-dimensional coordinates of the point cloud data can be transformed into two-dimensional coordinates to obtain the corresponding projected view.
  • the data is projected and mapped along the front-view and top-view directions, and the corresponding front-view and top-view views can be obtained respectively.
  • step S320C an event frame is generated according to the event data in the preset sampling period.
  • step S320C For the description of this step S320C, reference may be made to the above description of step S311A, and details are not repeated here.
  • step S330C the front view, the top view, and the event frame are input into the neural network for processing to generate multi-dimensional motion state information of the target object.
  • the front view, the top view and the event frame may be spliced together in the channel dimension and then input to the neural network for feature extraction, so as to obtain the multi-dimensional motion state information of the target object.
  • step S300 may further include steps S310D-S320D.
  • step S310D at least one target area is determined according to the event data, first coordinate information of the at least one target area is obtained, and each target area corresponds to a target object.
  • step S320D the second coordinate information of the target area in the laser point cloud data is determined according to the first coordinate information, and multi-dimensional motion state information of the target object is generated.
  • a target detection algorithm may be used to detect a target object in at least one target area in the event frame generated according to the event data, so as to determine the first coordinate information of each target area in the event frame.
  • step S320D through the coordinate change algorithm in three-dimensional space, the laser point cloud data and the images of different angles in the event frame can be transformed to the same angle, and then the points in the images can be correlated, thereby Determine the multi-dimensional motion state information of the target object.
  • the motion state information of the target object on two different planes can also be obtained through two DVSs installed in different positions, and then the multi-dimensional motion state information of the target object can be uniquely determined through calculation.
  • the target object can also be detected by one or more image sensors, and the signals generated by the one or more image sensors can be fused with event data and laser point cloud data to form a multi-dimensional image of the target object.
  • the image sensor is a Complementary Metal Oxide Semiconductor (CMOS, Complementary Metal Oxide Semiconductor) sensor.
  • CMOS Complementary Metal Oxide Semiconductor
  • the target detection method further includes: in step S500 , acquiring at least one channel of RGB image signals.
  • step S300 may further include: in step S310E, fusing at least one channel of RGB image signal with laser point cloud data and event data to generate multi-dimensional motion information of the target object.
  • the neural network may include a plurality of processing branches, each processing branch correspondingly processes a channel of RGB image signals, and the RGB image signals of different channels are image signals collected by different RGB image sensors.
  • the RGB image signals are input to the neural network through corresponding processing branches.
  • the RGB image signal is input to the neural network frame by frame.
  • RGB image signals, laser point cloud data, and event frames captured from different angles can be registered through key point detection (the images captured at different angles and different fields of view are The operation becomes a picture of the same angle and the same field of view), and the registered three sets of pictures are spliced together in the channel dimension and then input to the network for feature extraction to obtain the multi-dimensional motion information of the target object.
  • RGB image signals, laser point cloud data and event frames can also be input into a neural network with multiple inputs, and the network outputs the motion state information of the target object such as speed, position, etc.
  • the neural network needs to pass the marked
  • the training samples are obtained by network training.
  • an embodiment of the present disclosure provides a detection device for a target object.
  • the detection device can be used to implement the above detection method. Referring to FIG.
  • the first sensor 101 is used to detect light intensity change information in the target plane to generate event data, and the light intensity change information is used to determine at least one target object in the target plane.
  • the radar 102 is configured to acquire radar detection data for the target object, where the radar detection data is information describing the motion state of the target object.
  • the fusion processing unit 103 is configured to perform fusion processing on event data and radar detection data to generate multi-dimensional motion state information of the target object.
  • the detection device provided by the embodiment of the present disclosure can be applied to automatic driving, and is used to detect a target object.
  • the fusion processing unit 103 can execute the method for detecting a target object described in the first aspect of the embodiment of the present disclosure, and fuse the event data with the radar detection data to generate multi-dimensional motion state information of the target object.
  • the first sensor 101 in the detection device may be a dynamic vision sensor (DVS, Dynamic Vision Sensor).
  • DVS Dynamic Vision Sensor
  • DVS is a sensor that imitates the working mechanism of biological vision. It can detect the change of light and output the address (position coordinates) and information of the pixel where the light intensity changes, effectively reduce redundant data, and can dynamically respond to scene changes in real time.
  • the first sensor 101 is a dynamic vision sensor; the dynamic vision sensor is used to detect changes in the light intensity of each pixel in the target plane, and generate event data, where the event data includes the target
  • the coordinates and light intensity change information of the pixel points where the light intensity changes in the plane may further include time information.
  • the radar 102 in the detection device may be a pulse Doppler radar, and the radar detection data includes first motion state component information of the target object in a first direction, and the first direction is perpendicular to the target plane.
  • the pulse Doppler radar is used for sending and receiving pulse signals to determine the first motion state component information of the target object in the first direction; the fusion processing unit 103 can obtain the first motion state component information from the pulse Doppler radar.
  • Doppler radar refers to the radar that uses the Doppler effect to measure the radial velocity component of the target relative to the radar, or to extract the radial direction of the target with a specific radial velocity.
  • a pulsed Doppler radar is a Doppler radar that transmits pulsed signals.
  • the pulse Doppler radar scans the air with a pulse wave at a fixed frequency, if it encounters a moving target, the frequency of the reflected echo and the frequency of the transmitted wave will have a frequency difference, that is, the Doppler frequency shift.
  • the Doppler shift is proportional to the relative radial velocity of the moving target and the radar. From the magnitude of the Doppler shift, the radial velocity of the moving target can be determined. The magnitude of the Doppler shift is calculated from the phase of the signal. Therefore, in the embodiment of the present disclosure, the radar 102 is a coherent radar, so that phase information can be preserved.
  • the Doppler frequency shift is positive; when the target object moves away from the radar, the Doppler frequency shift is negative.
  • the detection apparatus further includes: an output unit 104 for outputting multi-dimensional motion state information of the target object.
  • the output unit is a display screen, and the multi-dimensional motion state information of the target object is displayed on the display screen.
  • the fusion processing unit 103 is configured to: determine the second motion state component information of the target object on the target plane according to the event data; according to the second motion state component information and the first motion state component information, Generate multi-dimensional motion state information of the target object.
  • the radar 102 is a laser radar, and the radar detection data is laser point cloud data; the laser radar is used for emitting a laser beam to detect at least one target object to generate laser point cloud data.
  • the fusion processing unit 103 may include (ISP, Image Signal Processing) 131 and a first neural network 132.
  • the first image signal processor 131 is used for: generating a three-dimensional image according to the laser point cloud data; and generating an event frame according to the event data in a preset sampling period.
  • the first neural network 132 is used to process three-dimensional images and event frames to generate multi-dimensional motion state information of the target object.
  • the fusion processing unit 103 may include a second image signal processor 133 and a second neural network (eg, a convolutional neural network) 134 .
  • the second image signal processor 133 is used for processing the laser point cloud data to generate the front view and top view of the laser point cloud data; according to the event data in the preset sampling period, the event frame is generated; the second neural network 134 uses It is used to process the front view, top view and event frame to generate multi-dimensional motion state information of the target object.
  • the fusion processing unit is configured to: determine at least one target area according to the event data, obtain first coordinate information of the at least one target area, each target area corresponds to a target object; determine the target area according to the first coordinate information
  • the second coordinate information in the laser point cloud data generates multi-dimensional motion state information of the target object.
  • the detection device further includes at least one second sensor 140; the second sensor 140 is used to acquire RGB images and generate RGB image signals; the fusion processing unit 103 is used to combine at least one RGB image
  • the image signal, laser point cloud data and event data are fused to generate multi-dimensional motion state information of the target object.
  • the first neural network 132 may include a plurality of processing branches, each processing branch corresponds to a second sensor 140 , and the RGB image signals output by the second sensor 140 pass through the corresponding processing branch Input to the first neural network 132 .
  • the second neural network 134 includes a plurality of processing branches, each processing branch corresponds to a second sensor 140 , and the RGB image signals output by the second sensor 140 are input through the corresponding processing branch The second neural network 134 .
  • the RGB image signals are input to the first neural network 132 or the second neural network 134 frame by frame.
  • the detection apparatus is used to implement the detection method provided by any of the foregoing embodiments.
  • reference may be made to the specific description of the detection method in the foregoing embodiment, which will not be repeated here.
  • an embodiment of the present disclosure provides a fusion processing unit, which is applied to a target object detection device.
  • the fusion processing unit includes: one or more processors 201 ; and a memory 202 that stores one or more processors thereon.
  • a program when one or more programs are executed by one or more processors, so that one or more processors implement the target object detection method described in the first aspect of the embodiment of the present disclosure; one or more I/O interfaces 203 , connected between the processor 201 and the memory 202 , and configured to implement information interaction between the processor 201 and the memory 202 .
  • the processor 201 is a device with data processing capability, including but not limited to a central processing unit (CPU), etc.; the memory 202 is a device with data storage capability, including but not limited to random access memory (RAM, more specifically Such as SDRAM, DDR, etc.), read only memory (ROM), electrified erasable programmable read only memory (EEPROM), flash memory (FLASH); I/O interface (read and write interface) 203 is connected between the processor 201 and the memory 202 , can realize the information interaction between the processor 201 and the memory 202, which includes but is not limited to a data bus (Bus) and the like.
  • RAM random access memory
  • ROM read only memory
  • EEPROM electrified erasable programmable read only memory
  • FLASH flash memory
  • I/O interface (read and write interface) 203 is connected between the processor 201 and the memory 202 , can realize the information interaction between the processor 201 and the memory 202, which includes but is not limited to a data bus (Bus) and
  • processor 201 memory 202, and I/O interface 203 are interconnected by bus 204, which in turn is connected to other components of the computing device.
  • an embodiment of the present disclosure provides a computer-readable medium on which a computer program is stored, and when the program is executed by a processor, implements the method for detecting a target object described in the first aspect of the embodiment of the present disclosure.
  • Such software may be distributed on computer-readable media, which may include computer storage media (or non-transitory media) and communication media (or transitory media).
  • computer storage media includes both volatile and nonvolatile implemented in any method or technology for storage of information, such as computer readable instructions, data structures, program modules or other data flexible, removable and non-removable media.
  • Computer storage media include, but are not limited to, RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, digital versatile disk (DVD) or other optical disk storage, magnetic cartridges, magnetic tape, magnetic disk storage or other magnetic storage devices, or may Any other medium used to store desired information and which can be accessed by a computer.
  • communication media typically embodies computer readable instructions, data structures, program modules, or other data in a modulated data signal such as a carrier wave or other transport mechanism, and can include any information delivery media, as is well known to those of ordinary skill in the art .

Abstract

A method for detecting a moving object. The method comprises: acquiring event data, wherein the event data represents light intensity change information in a target plane, and the light intensity change information is used for determining at least one target object in the target plane (S100); acquiring radar detection data for the target object, wherein the radar detection data is information for describing the motion state of the target object (S200); and performing fusion processing on the event data and the radar detection data, so as to generate multi-dimensional motion state information of the target object (S300). Provided are an apparatus for detecting a target object, a fusion processing unit, and a computer-readable medium.

Description

目标物体的检测方法及装置、融合处理单元、介质Target object detection method and device, fusion processing unit, medium 技术领域technical field
本公开涉及目标检测技术领域,特别涉及一种目标物体的检测方法、一种目标物体的检测装置、一种融合处理单元、一种计算机可读介质。The present disclosure relates to the technical field of target detection, and in particular, to a target object detection method, a target object detection device, a fusion processing unit, and a computer-readable medium.
背景技术Background technique
目标物体检测是视频图像分析与理解的重要技术,是部分计算视觉任务重要的预处理步骤,如物体识别、运动物体跟踪等。Object detection is an important technology for video image analysis and understanding, and an important preprocessing step for some computational vision tasks, such as object recognition and moving object tracking.
帧间差分法和光流法是一些相关技术中对运动物体进行检测的常用方法。其中帧间差分法通过计算相邻两帧图像的差值来获得运动目标轮廓,具体是将两帧图像相减得到两帧图像亮度差的绝对值,通过判断计算得到的绝对值是否大于阈值来分析视频或图像序列的运动特性。光流法则是用光流描述相对于观察者的运动所造成的观测目标、表面或边缘的运动。Inter-frame difference method and optical flow method are common methods for detecting moving objects in some related technologies. The inter-frame difference method obtains the outline of the moving target by calculating the difference between two adjacent frames of images, specifically, subtracting the two frames of images to obtain the absolute value of the brightness difference between the two frames of images, and by judging whether the calculated absolute value is greater than the threshold value. Analyze the motion characteristics of video or image sequences. The law of optical flow is to describe the motion of the observation target, surface or edge caused by the motion of the observer relative to the motion of the observer.
在一些相关技术中,基于时间感知对运动物体进行检测,但是空间感知能力较弱,无法同时对运动的物体和/或静止的物体进行多维视觉感知。In some related technologies, moving objects are detected based on temporal perception, but the spatial perception ability is weak, and multi-dimensional visual perception cannot be performed on moving objects and/or stationary objects at the same time.
在一些相关运用场景中,运动物体检测算法的检测效果不够理想。In some related application scenarios, the detection effect of the moving object detection algorithm is not ideal.
发明内容SUMMARY OF THE INVENTION
本公开提供一种目标物体的检测方法、一种目标物体的检测装置、一种融合处理单元、一种计算机可读介质。The present disclosure provides a method for detecting a target object, a device for detecting a target object, a fusion processing unit, and a computer-readable medium.
第一方面,本公开实施例提供一种目标物体的检测方法,该检测方法包括:In a first aspect, an embodiment of the present disclosure provides a method for detecting a target object, where the method includes:
所述事件数据表征目标平面中的光强变化信息,所述光强变化信息用于确定所述目标平面中的至少一个目标物体;The event data represents light intensity change information in the target plane, and the light intensity change information is used to determine at least one target object in the target plane;
获取针对所述目标物体的雷达探测数据,所述雷达探测数据为描述所述目标物体的运动状态的信息;acquiring radar detection data for the target object, where the radar detection data is information describing the motion state of the target object;
将所述事件数据与所述雷达探测数据进行融合处理,生成所述目标物体的多维运动状态信息。The event data and the radar detection data are fused to generate multi-dimensional motion state information of the target object.
第二方面,本公开实施例提供一种目标物体的检测装置,该检测装置包括:In a second aspect, an embodiment of the present disclosure provides a detection device for a target object, the detection device includes:
第一传感器,用于检测目标平面中的光强变化信息,以生成事件数据,所述光强变化 信息用于确定所述目标平面中的至少一个目标物体;a first sensor for detecting light intensity change information in the target plane to generate event data, the light intensity change information being used to determine at least one target object in the target plane;
雷达,用于获取针对所述目标物体的雷达探测数据,所述雷达探测数据为描述所述目标物体的运动状态的信息;a radar, for acquiring radar detection data for the target object, where the radar detection data is information describing the motion state of the target object;
融合处理单元,用于将所述事件数据与所述雷达探测数据进行融合处理,生成所述目标物体的多维运动状态信息。The fusion processing unit is configured to perform fusion processing on the event data and the radar detection data to generate multi-dimensional motion state information of the target object.
第三方面,本公开实施例提供一种融合处理单元,应用于目标物体的检测装置,所述融合处理单元包括:In a third aspect, an embodiment of the present disclosure provides a fusion processing unit, which is applied to a target object detection device, and the fusion processing unit includes:
一个或多个处理器;one or more processors;
存储装置,其上存储有一个或多个程序,当所述一个或多个程序被所述一个或多个处理器执行,使得所述一个或多个处理器实现本公开实施例第一方面所述的目标物体的检测方法。A storage device on which one or more programs are stored, when the one or more programs are executed by the one or more processors, the one or more processors implement the first aspect of the embodiments of the present disclosure; The detection method of the target object described.
第四方面,本公开实施例提供一种计算机可读介质,其上存储有计算机程序,所述程序被处理器执行时实现本公开实施例第一方面所述的目标物体的检测方法。In a fourth aspect, an embodiment of the present disclosure provides a computer-readable medium on which a computer program is stored, and when the program is executed by a processor, implements the method for detecting a target object described in the first aspect of the embodiment of the present disclosure.
在本公开实施例中,将表征目标平面中的光强变化信息的事件数据与针对目标物体的雷达探测数据进行融合处理,实现了对目标物体的多维运动状态信息的采集,从而能够对至少一个静止或运动的目标物体进行清晰的运动判断,使得检测装置具有类生物视觉,实现了对静止或运动的目标物体的类生物视觉感知,且采集的事件数据是通过具有运动敏感的传感器对场景变化实时动态响应产生的,从而有效减少了冗余数据对检测效果的影响,实现了目标物体检测的实时动态响应,有效提高了运动物体检测的效率。In the embodiment of the present disclosure, the event data representing the light intensity change information in the target plane and the radar detection data for the target object are fused to realize the collection of multi-dimensional motion state information of the target object, so that at least one The stationary or moving target object can be clearly judged by motion, so that the detection device has biological-like vision, and the biological-like visual perception of the stationary or moving target object is realized, and the collected event data is detected by the motion-sensitive sensor. Real-time dynamic response is generated, which effectively reduces the influence of redundant data on the detection effect, realizes the real-time dynamic response of target object detection, and effectively improves the efficiency of moving object detection.
应当理解,本部分所描述的内容并非旨在标识本公开的实施例的关键或重要特征,也不用于限制本公开的范围。本公开的其它特征将通过以下的说明书而变得容易理解。It should be understood that what is described in this section is not intended to identify key or critical features of embodiments of the disclosure, nor is it intended to limit the scope of the disclosure. Other features of the present disclosure will become readily understood from the following description.
附图说明Description of drawings
附图用来提供对本公开的进一步理解,并且构成说明书的一部分,与本公开的实施例一起用于解释本公开,并不构成对本公开的限制。通过参考附图对详细示例实施例进行描述,以上和其他特征和优点对本领域技术人员将变得更加显而易见,在附图中:The accompanying drawings are used to provide a further understanding of the present disclosure and constitute a part of the specification, and together with the embodiments of the present disclosure, they are used to explain the present disclosure, and are not intended to limit the present disclosure. The above and other features and advantages will become more apparent to those skilled in the art by describing detailed example embodiments with reference to the accompanying drawings, in which:
图1是本公开实施例中一种检测方法的流程示意图;1 is a schematic flowchart of a detection method in an embodiment of the present disclosure;
图2是本公开实施例中一种检测方法中部分步骤的流程示意图;2 is a schematic flowchart of some steps in a detection method according to an embodiment of the present disclosure;
图3是本公开实施例中一种检测方法中部分步骤的流程示意图;3 is a schematic flowchart of some steps in a detection method according to an embodiment of the present disclosure;
图4是本公开实施例中确定偏移量的一种实施方式的示意图;FIG. 4 is a schematic diagram of an implementation manner of determining an offset in an embodiment of the present disclosure;
图5是本公开实施例中一种检测方法中部分步骤的流程示意图;5 is a schematic flowchart of some steps in a detection method according to an embodiment of the present disclosure;
图6是本公开实施例中一种检测方法中部分步骤的流程示意图;6 is a schematic flowchart of some steps in a detection method according to an embodiment of the present disclosure;
图7是本公开实施例中一种检测方法中部分步骤的流程示意图;7 is a schematic flowchart of some steps in a detection method according to an embodiment of the present disclosure;
图8是本公开实施例中一种检测方法中部分步骤的流程示意图;8 is a schematic flowchart of some steps in a detection method according to an embodiment of the present disclosure;
图9是本公开实施例中一种检测方法中部分步骤的流程示意图;9 is a schematic flowchart of some steps in a detection method according to an embodiment of the present disclosure;
图10是本公开实施例中一种检测装置的组成框图;10 is a block diagram of the composition of a detection device in an embodiment of the present disclosure;
图11是本公开实施例中另一种检测装置的组成框图;11 is a block diagram of another detection device in an embodiment of the present disclosure;
图12是本公开实施例中另一种检测装置的组成框图;12 is a block diagram of another detection device in an embodiment of the present disclosure;
图13是本公开实施例中一种融合处理单元的组成框图。FIG. 13 is a block diagram of a fusion processing unit in an embodiment of the present disclosure.
具体实施方式Detailed ways
为使本领域的技术人员更好地理解本公开的技术方案,以下结合附图对本公开的示范性实施例做出说明,其中包括本公开实施例的各种细节以助于理解,应当将它们认为仅仅是示范性的。因此,本领域普通技术人员应当认识到,可以对这里描述的实施例做出各种改变和修改,而不会背离本公开的范围和精神。同样,为了清楚和简明,以下的描述中省略了对公知功能和结构的描述。In order for those skilled in the art to better understand the technical solutions of the present disclosure, the exemplary embodiments of the present disclosure are described below with reference to the accompanying drawings, including various details of the embodiments of the present disclosure to facilitate understanding, and they should be considered to be exemplary only. Accordingly, those of ordinary skill in the art will recognize that various changes and modifications of the embodiments described herein can be made without departing from the scope and spirit of the present disclosure. Also, descriptions of well-known functions and constructions are omitted from the following description for clarity and conciseness.
在不冲突的情况下,本公开各实施例及实施例中的各特征可相互组合。Various embodiments of the present disclosure and various features of the embodiments may be combined with each other without conflict.
如本文所使用的,术语“和/或”包括一个或多个相关列举条目的任何和所有组合。As used herein, the term "and/or" includes any and all combinations of one or more of the associated listed items.
本文所使用的术语仅用于描述特定实施例,且不意欲限制本公开。如本文所使用的,单数形式“一个”和“该”也意欲包括复数形式,除非上下文另外清楚指出。还将理解的是,当本说明书中使用术语“包括”和/或“由……制成”时,指定存在所述特征、整体、步骤、操作、元件和/或组件,但不排除存在或添加一个或多个其它特征、整体、步骤、操作、元件、组件和/或其群组。“连接”或者“相连”等类似的词语并非限定于物理的或者机械的连接,而是可以包括电性的连接,不管是直接的还是间接的。The terminology used herein is used to describe particular embodiments only and is not intended to limit the present disclosure. As used herein, the singular forms "a" and "the" are intended to include the plural forms as well, unless the context clearly dictates otherwise. It will also be understood that when the terms "comprising" and/or "made of" are used in this specification, the stated features, integers, steps, operations, elements and/or components are specified to be present, but not precluded or Add one or more other features, integers, steps, operations, elements, components, and/or groups thereof. Words like "connected" or "connected" are not limited to physical or mechanical connections, but may include electrical connections, whether direct or indirect.
除非另外限定,否则本文所用的所有术语(包括技术和科学术语)的含义与本领域普通技术人员通常理解的含义相同。还将理解,诸如那些在常用字典中限定的那些术语应当被解释为具有与其在相关技术以及本公开的背景下的含义一致的含义,且将不解释为具有理想化或过度形式上的含义,除非本文明确如此限定。Unless otherwise defined, all terms (including technical and scientific terms) used herein have the same meaning as commonly understood by one of ordinary skill in the art. It will also be understood that terms such as those defined in common dictionaries should be construed as having meanings consistent with their meanings in the context of the related art and the present disclosure, and will not be construed as having idealized or over-formal meanings, unless expressly so limited herein.
第一方面,参照图1,本公开实施例提供一种目标物体的检测方法,该检测方法包括:步骤S100-步骤S300。In a first aspect, referring to FIG. 1 , an embodiment of the present disclosure provides a method for detecting a target object. The detecting method includes steps S100 to S300.
在步骤S100中,获取事件数据,事件数据表征目标平面中的光强变化信息,光强变化信息用于确定目标平面中的至少一个目标物体。In step S100, event data is acquired, where the event data represents light intensity change information in the target plane, and the light intensity change information is used to determine at least one target object in the target plane.
在步骤S200中,获取针对目标物体的雷达探测数据,雷达探测数据为描述目标物体的运动状态的信息。In step S200, radar detection data for the target object is acquired, where the radar detection data is information describing the motion state of the target object.
在步骤S300中,将事件数据与雷达探测数据进行融合处理,生成目标物体的多维运动状态信息。In step S300, the event data and the radar detection data are fused to generate multi-dimensional motion state information of the target object.
在本公开实施例中,提供了一种目标物体的检测装置,用于对目标物体的运动状态进行检测。目标物体的检测装置包括第一传感器和雷达。检测装置中的第一传感器具有运动敏感,能够对场景变化进行实时动态响应。In an embodiment of the present disclosure, a device for detecting a target object is provided, which is used to detect the motion state of the target object. The detection device of the target object includes a first sensor and a radar. The first sensor in the detection device is sensitive to motion and can respond dynamically to scene changes in real time.
在本公开实施例中,在上述步骤S100中,事件数据为通过第一传感器对目标平面中的光强变化信息进行采集生成的,其中,第一传感器是模仿生物视觉的工作机理的传感器,第一传感器在生成事件数据时仅保留动态信息。其中,目标平面可以表示第一传感器针对目标场景的采集画面,第一传感器通过对目标场景的采集画面中各像素点的光强变化信息进行感知,从而生成事件数据,事件数据可以由目标平面中光强发生变化的像素点的信息组成。示例性的,第一传感器为动态视觉传感器(DVS,Dynamic Vision Sensor)。In the embodiment of the present disclosure, in the above step S100, the event data is generated by collecting the light intensity change information in the target plane by the first sensor, wherein the first sensor is a sensor imitating the working mechanism of biological vision, A sensor only retains dynamic information when generating event data. The target plane may represent the image captured by the first sensor for the target scene, and the first sensor may generate event data by sensing the light intensity change information of each pixel in the captured image of the target scene, and the event data may be stored in the target plane. The information composition of the pixels where the light intensity changes. Exemplarily, the first sensor is a Dynamic Vision Sensor (DVS, Dynamic Vision Sensor).
在本公开实施例中,在步骤S200中,通过雷达对目标物体的运动状态进行检测,生成雷达探测数据。In the embodiment of the present disclosure, in step S200, the motion state of the target object is detected by the radar to generate radar detection data.
在本公开实施例中,本公开实施例对于步骤S100和步骤S200的执行先后顺序不作特殊限制,可以同时执行步骤S100和步骤S200,获取对应相同时间点的事件数据和雷达探测数据;也可以分别执行步骤S100和步骤S200,获得对应多个时间点的事件数据和对应多个时间点的雷达探测数据。In this embodiment of the present disclosure, there is no special restriction on the execution sequence of step S100 and step S200, and step S100 and step S200 may be executed at the same time to obtain event data and radar detection data corresponding to the same time point; Steps S100 and S200 are executed to obtain event data corresponding to multiple time points and radar detection data corresponding to multiple time points.
在本公开实施例中,在步骤S300中,将事件数据和雷达探测数据进行融合处理,是将对应相同时间点的事件数据与雷达探测数据进行对齐校准,以生成目标物体的多维运动状态信息。In the embodiment of the present disclosure, in step S300, the fusion processing of event data and radar detection data is performed by aligning and calibrating event data and radar detection data corresponding to the same time point to generate multi-dimensional motion state information of the target object.
本公开实施例对经步骤S300生成的目标物体的多维运动状态信息不做特殊限定。在一些实施例中,多维运动状态信息可以包括目标物体在各个时间点的三维速度、三维位置坐标,还可以包括目标物体在各个时间点在包括第一方向的多个预定方向的速度分量,还可以包括根据多个时间点的三维速度、三维位置坐标生成的目标物体的运动轨迹等。在一些实施例中,多维运动状态信息可以包括目标物体的距离(与检测装置之间的距离)、方位、高度、速度、姿态、形状等信息中的至少一者。This embodiment of the present disclosure does not specifically limit the multi-dimensional motion state information of the target object generated in step S300. In some embodiments, the multi-dimensional motion state information may include three-dimensional velocity and three-dimensional position coordinates of the target object at various time points, and may also include velocity components of the target object in multiple predetermined directions including the first direction at each time point, and further It may include the movement trajectory of the target object generated according to the three-dimensional velocity and three-dimensional position coordinates of multiple time points, and the like. In some embodiments, the multi-dimensional motion state information may include at least one of the distance (distance between the target object and the detection device), orientation, height, speed, attitude, shape and other information of the target object.
在本公开实施例中,将表征目标平面中的光强变化信息的事件数据与针对目标物体的雷达探测数据进行融合处理,实现了对目标物体的多维运动状态信息的采集,从而能够对至少一个静止或运动的目标物体进行清晰的运动判断,使得检测装置具有类生物视觉,实 现了对静止或运动的目标物体的类生物视觉感知,且采集的事件数据是通过具有运动敏感的传感器对场景变化实时动态响应产生的,从而有效减少了冗余数据对检测效果的影响,实现了目标物体检测的实时动态响应,有效提高了运动物体检测的效率。In the embodiment of the present disclosure, the event data representing the light intensity change information in the target plane and the radar detection data for the target object are fused to realize the collection of multi-dimensional motion state information of the target object, so that at least one The stationary or moving target object can be clearly judged by motion, so that the detection device has biological-like vision, and the biological-like visual perception of the stationary or moving target object is realized, and the collected event data is detected by the motion-sensitive sensor. Real-time dynamic response is generated, which effectively reduces the influence of redundant data on the detection effect, realizes the real-time dynamic response of target object detection, and effectively improves the efficiency of moving object detection.
作为一种可选的实施方式,上述用于采集雷达探测数据的雷达可以为脉冲多普勒雷达,脉冲多普勒雷达通过向目标物体发射脉冲信号并接收目标物体反射的脉冲信号,检测目标物体的运动状态,从而获取雷达探测数据。此种情况下,雷达探测数据可以包括目标物体在第一方向的第一运动状态分量信息,第一方向垂直于目标平面。As an optional implementation manner, the above-mentioned radar for collecting radar detection data may be a pulse Doppler radar. The pulse Doppler radar detects the target object by transmitting a pulse signal to the target object and receiving the pulse signal reflected by the target object. The movement state of the radar can be obtained to obtain radar detection data. In this case, the radar detection data may include first motion state component information of the target object in a first direction, and the first direction is perpendicular to the target plane.
本公开实施例对第一运动状态分量信息不做特殊限定。例如,第一运动状态分量信息可以包括目标物体在第一方向的速度分量、目标物体到检测装置的距离等。This embodiment of the present disclosure does not specifically limit the first motion state component information. For example, the first motion state component information may include the velocity component of the target object in the first direction, the distance from the target object to the detection device, and the like.
需要说明的是,在一些实施例中,第一方向为与雷达的径向平行的方向,目标平面为与雷达的径向垂直的平面。It should be noted that, in some embodiments, the first direction is a direction parallel to the radial direction of the radar, and the target plane is a plane perpendicular to the radial direction of the radar.
在一些实施例中,根据事件数据能够确定目标物体在目标平面的运动状态分量信息,并根据目标物体在目标平面的运动状态分量信息和在第一方向的运动状态分量信息确定目标物体的多维运动状态信息。In some embodiments, the motion state component information of the target object on the target plane can be determined according to the event data, and the multi-dimensional motion of the target object can be determined according to the motion state component information of the target object on the target plane and the motion state component information in the first direction status information.
相应地,在一些实施例中,雷达探测数据包括目标物体在第一方向的第一运动状态分量信息,参照图2,步骤S300可以进一步包括:步骤S310A-步骤S320A。Correspondingly, in some embodiments, the radar detection data includes first motion state component information of the target object in the first direction. Referring to FIG. 2 , step S300 may further include steps S310A-step S320A.
在步骤S310A中,根据事件数据确定目标物体在目标平面的第二运动状态分量信息。In step S310A, the second motion state component information of the target object on the target plane is determined according to the event data.
在步骤S320A中,根据第二运动状态分量信息与第一运动状态分量信息,生成目标物体的多维运动状态信息。In step S320A, multi-dimensional motion state information of the target object is generated according to the second motion state component information and the first motion state component information.
需要说明的是,在一些实施例中,可以通过雷达对目标物体进行检测生成第一运动状态分量信息;在另一些实施例中,也可以由雷达对目标物体进行检测生成初始检测信号,例如,脉冲多普勒雷达测量的目标物体对应的多普勒频移,然后根据初始检测信息确定第一运动状态分量信息。本公开实施例对此不做特殊限定。It should be noted that, in some embodiments, the target object may be detected by radar to generate the first motion state component information; in other embodiments, the target object may also be detected by radar to generate the initial detection signal, for example, The Doppler frequency shift corresponding to the target object measured by the pulse Doppler radar, and then the first motion state component information is determined according to the initial detection information. This embodiment of the present disclosure makes no special limitation on this.
在通过雷达对目标物体进行检测生成初始检测信号的场景下,检测方法还可以包括根据初始检测信号确定第一运动状态分量信息的步骤。In the scenario where the target object is detected by the radar to generate the initial detection signal, the detection method may further include the step of determining the first motion state component information according to the initial detection signal.
本公开实施例对检测装置中的第一传感器不做特殊限定。作为一种可选的实施方式,检测装置中的第一传感器可以为动态视觉传感器(DVS,Dynamic Vision Sensor)。DVS是一种模仿生物视觉的工作机理的传感器,能够检测光的改变并输出光强发生变化像素的地址和信息,消除了冗余数据,并能够对场景变化实时动态响应。The embodiment of the present disclosure does not make any special limitation on the first sensor in the detection device. As an optional implementation manner, the first sensor in the detection device may be a Dynamic Vision Sensor (DVS, Dynamic Vision Sensor). DVS is a sensor that imitates the working mechanism of biological vision. It can detect the change of light and output the address and information of the pixel where the light intensity changes. It eliminates redundant data and can dynamically respond to scene changes in real time.
在本公开的一些实施例中,DVS采集的事件数据为目标平面的二维数据,根据DVS提供的每个时间点光强发生变化的像素点的地址(位置坐标)和光强变化信息,能够确定 目标平面中的目标物体,并确定目标物体在目标平面中的运动状态分量信息。In some embodiments of the present disclosure, the event data collected by the DVS is the two-dimensional data of the target plane. Determine the target object in the target plane, and determine the motion state component information of the target object in the target plane.
在本公开的一些实施例中,DVS不需要对画面中的所有像素点进行读取,仅需要获取光强度变化的像素点的地址和信息;具体的,当DVS检测到某个像素点的光强度变化大于等于预设门限数值时,则发出该像素点的事件信号;其中,如果该光强度变化为正向变化,即该像素点由低亮度跳变至高亮度,则发出用“+1”表示的事件信号,并标注为正事件;如果该光强度变化为负向变化,即该像素点由高亮度跳变至低亮度,则发出用“-1”表示的事件信号,并标注为负事件;如果光强度变化小于预设门限数值,则不发出事件信号,标注为无事件;DVS通过对各光强发生变化的像素点进行的事件标注,以构成事件数据。其中,事件数据中,正事件和负事件均可以用于表示像素点的光强变化信息。In some embodiments of the present disclosure, the DVS does not need to read all the pixels in the picture, but only needs to obtain the address and information of the pixels whose light intensity changes; specifically, when the DVS detects the light of a certain pixel When the intensity change is greater than or equal to the preset threshold value, the event signal of the pixel is sent; if the change of the light intensity is a positive change, that is, the pixel jumps from low brightness to high brightness, it will send out with "+1" If the light intensity change is a negative change, that is, the pixel jumps from high brightness to low brightness, an event signal represented by "-1" is sent and marked as negative. Event; if the change of light intensity is less than the preset threshold value, no event signal will be sent, and it will be marked as no event; DVS constitutes event data by marking the event of each pixel where the light intensity changes. Among them, in the event data, both positive events and negative events can be used to represent the light intensity change information of the pixel point.
相应地,在一些实施例中,上述事件数据可以包括所述目标平面中光强发生变化的像素点的坐标和光强变化信息,还可以进一步包括时间信息,时间信息可以表示光强发生变化的时间;第二运动状态分量信息包括目标物体在目标平面中的位置坐标;参照图3,步骤S310A可以进一步包括步骤S311A和S312A。Correspondingly, in some embodiments, the above-mentioned event data may include the coordinates of the pixels where the light intensity changes in the target plane and the light intensity change information, and may further include time information, and the time information may indicate that the light intensity changes. time; the second motion state component information includes the position coordinates of the target object in the target plane; referring to FIG. 3 , step S310A may further include steps S311A and S312A.
在步骤S311A中,根据预设采样周期内的事件数据中各像素点的坐标和光强变化信息,生成事件帧。In step S311A, an event frame is generated according to the coordinates and light intensity change information of each pixel in the event data within a preset sampling period.
在步骤S312A中,根据事件帧确定目标物体在目标平面中的位置坐标。In step S312A, the position coordinates of the target object in the target plane are determined according to the event frame.
在本公开实施例中,在步骤S311A中,将预设采样周期内采样得到的事件数据进行组帧,从而生成事件帧,事件帧可以表示在预设采样周期内,对每个像素点产生的所有事件(如正事件或负事件)进行汇总后显示的图像帧。示例性的,DVS的输出数据为由多个4元组数据组成的事件数据,每一个4元组数据对应目标平面中光强发生变化的像素点,4元组数据包括光强发生变化的像素点在目标平面中的坐标(横坐标x、纵坐标y)以及光强变化信息、时间信息。在步骤S311A中,根据4元组数据携带的时间信息,将对应于同一个时间点的4元组数据进行组帧,从而生成相应的事件帧。In the embodiment of the present disclosure, in step S311A, the event data sampled in the preset sampling period is framed to generate an event frame, and the event frame can represent the data generated for each pixel in the preset sampling period. Image frame displayed after summing all events, such as positive or negative events. Exemplarily, the output data of the DVS is event data consisting of a plurality of 4-tuple data, each 4-tuple data corresponds to a pixel whose light intensity changes in the target plane, and the 4-tuple data includes pixels whose light intensity changes. The coordinates of the point in the target plane (abscissa x, ordinate y), light intensity change information, and time information. In step S311A, according to the time information carried by the 4-tuple data, the 4-tuple data corresponding to the same time point is framed, thereby generating a corresponding event frame.
在本公开实施例中,对于如何实施步骤S312A不做特殊限定。作为一种可选的实施方式,可以采用目标检测算法根据事件帧确定目标物体在目标平面中的位置坐标。例如,画面中目标物体相对运动时,其对应的像素点的光亮强度会存在不同程度的变化,例如,目标物体出现时,目标物体出现区域的像素点的光亮强度会显著增加,目标物体消失时,目标物体消失区域的像素点的光亮强度会显著降低,因此根据光强发生变化的像素点的位置坐标,可以确定画面中哪些像素点可能存在目标物体,可以确定出目标物体的轮廓区域,进而获取目标物体的位置区域,确定目标物体的位置坐标。In this embodiment of the present disclosure, there is no special limitation on how to implement step S312A. As an optional implementation manner, a target detection algorithm may be used to determine the position coordinates of the target object in the target plane according to the event frame. For example, when the target object in the picture moves relatively, the light intensity of the corresponding pixel will change to varying degrees. For example, when the target object appears, the light intensity of the pixel in the area where the target object appears will increase significantly, and when the target object disappears , the brightness of the pixels in the disappearing area of the target object will be significantly reduced. Therefore, according to the position coordinates of the pixels where the light intensity changes, it can be determined which pixels in the picture may have the target object, and the contour area of the target object can be determined, and then Obtain the location area of the target object, and determine the location coordinates of the target object.
在本公开的一些实施例中,可以将目标物体上的任意一点在目标平面中的坐标作为目 标物体在目标平面中的位置坐标,例如,将目标物体的中心点的坐标作为目标物体在目标平面中的位置坐标。本公开实施例对此不做特殊限定。In some embodiments of the present disclosure, the coordinates of any point on the target object in the target plane may be used as the position coordinates of the target object in the target plane. For example, the coordinates of the center point of the target object may be used as the coordinates of the target object in the target plane. location coordinates in . This embodiment of the present disclosure makes no special limitation on this.
在本公开的一些实施例中,目标物体在目标平面中的位置坐标可以用(x,y)表示。其中,x对应第二方向和第三方向中的一者,y对应第二方向和第三方向中的另一者,第二方向和第三方向均平行于目标平面,且第二方向与第三方向垂直。In some embodiments of the present disclosure, the position coordinates of the target object in the target plane may be represented by (x, y). Wherein, x corresponds to one of the second direction and the third direction, y corresponds to the other of the second direction and the third direction, the second direction and the third direction are both parallel to the target plane, and the second direction and the third direction are parallel to the target plane. Three directions are vertical.
相应地,在一些实施例中,所述第二运动状态分量信息还包括所述目标物体在第二方向的第二速度分量和在第三方向的第三速度分量;参照图3,步骤S310A还可以进一步包括:步骤S313A-步骤S315A。Correspondingly, in some embodiments, the second motion state component information further includes the second velocity component of the target object in the second direction and the third velocity component in the third direction; referring to FIG. 3 , step S310A further includes It may further include: step S313A-step S315A.
在步骤S313A中,根据目标物体在目标平面中的位置坐标分别确定目标物体在第二方向的第二偏移量和在第三方向的第三偏移量。In step S313A, the second offset of the target object in the second direction and the third offset of the target object in the third direction are respectively determined according to the position coordinates of the target object in the target plane.
在步骤S314A中,根据第二偏移量确定目标物体在第二方向的第二速度分量。In step S314A, a second velocity component of the target object in the second direction is determined according to the second offset.
在步骤S315A中,根据所述第三偏移量确定目标物体在第三方向的第三速度分量。In step S315A, the third velocity component of the target object in the third direction is determined according to the third offset.
在本公开的一些实施例中,对目标平面中的第二方向和第三方向不做特殊限定。作为一种可选的实施方式,第一方向、第二方向、第三方向构成三维直角坐标系,第一方向为三维直角坐标系的纵坐标方向,第二方向为三维直角坐标系的横坐标方向,第三方向为三维直角坐标系的竖坐标方向。In some embodiments of the present disclosure, the second direction and the third direction in the target plane are not particularly limited. As an optional implementation manner, the first direction, the second direction, and the third direction constitute a three-dimensional rectangular coordinate system, the first direction is the ordinate direction of the three-dimensional rectangular coordinate system, and the second direction is the abscissa direction of the three-dimensional rectangular coordinate system direction, the third direction is the vertical coordinate direction of the three-dimensional Cartesian coordinate system.
本公开实施例对如何执行步骤S313A确定第二偏移量和第三偏移量不做特殊限定。图4示出了根据相邻时间点目标平面中光强发生变化的像素点的坐标确定第三方向的第三偏移量的一种可选实施方式。如图4所示,可以得到目标物体在第三方向的绝对偏移量。同理,可以得到目标物体在第二方向的绝对偏移量。This embodiment of the present disclosure does not specifically limit how to perform step S313A to determine the second offset and the third offset. FIG. 4 shows an optional implementation manner of determining the third offset in the third direction according to the coordinates of the pixel points where the light intensity changes in the target plane at adjacent time points. As shown in Figure 4, the absolute offset of the target object in the third direction can be obtained. Similarly, the absolute offset of the target object in the second direction can be obtained.
还需要说明的是,在步骤S314A和步骤S315A中,在确定第二速度分量和第三速度分量时,还结合相邻时间点的时间差。It should also be noted that, in step S314A and step S315A, when determining the second velocity component and the third velocity component, the time difference between adjacent time points is also combined.
在一些实施例中,第一运动状态分量信息包括目标物体在第一方向的第一速度分量和距离参数,距离参数表示目标物体与检测装置之间的距离;多维运动状态信息包括目标物体的三维速度和三维坐标;参照图3,步骤S320A可以进一步包括:步骤S321A和步骤S322A。In some embodiments, the first motion state component information includes a first velocity component of the target object in a first direction and a distance parameter, where the distance parameter represents a distance between the target object and the detection device; the multi-dimensional motion state information includes a three-dimensional Speed and three-dimensional coordinates; referring to FIG. 3 , step S320A may further include: step S321A and step S322A.
在步骤S321A中,确定对应于同一时间点的目标物体在目标平面中的位置坐标、第一速度分量、第二速度分量、第三速度分量和目标物体在第一方向的距离参数。In step S321A, the position coordinates of the target object in the target plane, the first velocity component, the second velocity component, the third velocity component and the distance parameter of the target object in the first direction corresponding to the same time point are determined.
在步骤S322A中,根据对应于同一时间点的目标物体在目标平面中的位置坐标、第一速度分量、第二速度分量、第三速度分量和目标物体在第一方向的距离参数,确定目标物体的三维速度和所述三维坐标。In step S322A, the target object is determined according to the position coordinates of the target object corresponding to the same time point in the target plane, the first velocity component, the second velocity component, the third velocity component and the distance parameter of the target object in the first direction The 3D velocity and the 3D coordinates.
在本公开的一些实施例中,三维坐标表示目标物体在三维空间中的位置坐标,三维速度表示目标物体在三维空间中的相对运动速度。其中三维坐标可以用(x,y,z)表示。其中,z对应第一方向,x对应第二方向和第三方向中的一者,y对应第二方向和第三方向中的另一者。In some embodiments of the present disclosure, the three-dimensional coordinates represent the position coordinates of the target object in the three-dimensional space, and the three-dimensional velocity represents the relative movement speed of the target object in the three-dimensional space. The three-dimensional coordinates can be represented by (x, y, z). Wherein, z corresponds to the first direction, x corresponds to one of the second direction and the third direction, and y corresponds to the other of the second direction and the third direction.
作为一种可选的实施方式,如前所述,在步骤S100中,通过DVS获取事件数据。相应地,在一些实施例中,参照图5,步骤S100包括:在步骤S110中,响应于目标平面中像素点的光强度变化获取事件数据,事件数据包括目标平面中光强发生变化的像素点的坐标和光强变化信息。As an optional implementation manner, as described above, in step S100, event data is acquired through the DVS. Correspondingly, in some embodiments, referring to FIG. 5 , step S100 includes: in step S110 , acquiring event data in response to changes in the light intensity of pixels in the target plane, where the event data includes pixels whose light intensity changes in the target plane coordinates and light intensity change information.
其中事件数据还可以进一步包括时间信息,时间信息可以表示像素点的光强发生变化的时间。The event data may further include time information, and the time information may represent the time when the light intensity of the pixel changes.
作为一种可选的实施方式,通过脉冲多普勒雷达获取第一运动状态分量信息,脉冲多普勒雷达能够探测目标物体的速度和距离等信息。相应地,在一些实施例中,参照图5,步骤S200可以进一步包括:在步骤S210中,获取目标物体在第一方向的第一速度分量和距离参数,以作为第一运动状态分量信息。As an optional implementation manner, the information of the first motion state component is obtained through a pulse Doppler radar, and the pulse Doppler radar can detect information such as the speed and distance of the target object. Correspondingly, in some embodiments, referring to FIG. 5 , step S200 may further include: in step S210 , acquiring the first velocity component and distance parameter of the target object in the first direction as the first motion state component information.
在步骤S300中,将事件数据与第一运动状态分量信息进行融合处理,生成目标物体的多维运动状态信息。In step S300, the event data and the first motion state component information are fused to generate multi-dimensional motion state information of the target object.
在本公开的一些实施例中,将表征目标平面中的光强变化信息的事件数据与通过脉冲多普勒雷达探测得到的目标物体在第一方向的第一运动状态分量信息进行融合处理,实现了对目标物体的多维运动状态信息的采集,其中事件数据是通过具有运动敏感的传感器对场景变化实时动态响应产生的,从而有效减少了冗余数据对检测效果的影响,实现了目标物体检测的实时动态响应,有效提高了运动物体检测的效率。In some embodiments of the present disclosure, the event data representing the light intensity change information in the target plane and the first motion state component information of the target object in the first direction detected by the pulse Doppler radar are fused to achieve The multi-dimensional motion state information of the target object is collected, in which the event data is generated by the real-time dynamic response of the scene change by the motion-sensitive sensor, which effectively reduces the influence of redundant data on the detection effect and realizes the detection of the target object. Real-time dynamic response effectively improves the efficiency of moving object detection.
在一些实施例中,参照图5,目标物体的检测方法还可以进一步包括:在步骤S400中,输出多维运动状态信息。In some embodiments, referring to FIG. 5 , the method for detecting a target object may further include: in step S400 , outputting multi-dimensional motion state information.
本公开实施例对多维运动状态信息不做特殊限定。在一些实施例中,多维运动状态信息可以包括目标物体在各个时间点的三维速度、三维位置坐标,还可以包括目标物体在各个时间点在第一方向、第二方向、第三方向的速度分量,还可以包括根据多个时间点的三维速度、三维位置坐标生成的目标物体的运动轨迹。在一些实施例中,多维运动状态信息可以包括目标物体的距离(与检测装置之间的距离)、方位、高度、速度、姿态、形状等信息中的至少一者。This embodiment of the present disclosure does not specifically limit the multi-dimensional motion state information. In some embodiments, the multi-dimensional motion state information may include the three-dimensional velocity and three-dimensional position coordinates of the target object at various time points, and may also include the velocity components of the target object in the first direction, the second direction, and the third direction at each time point , and may also include a motion trajectory of the target object generated according to the three-dimensional velocity and three-dimensional position coordinates of multiple time points. In some embodiments, the multi-dimensional motion state information may include at least one of the distance (distance between the target object and the detection device), orientation, height, speed, attitude, shape and other information of the target object.
本公开实施例对如何输出多维运动状态信息不做特殊限定。例如,可以在显示屏上显示目标物体的多维运动状态信息。This embodiment of the present disclosure does not specifically limit how to output the multi-dimensional motion state information. For example, the multi-dimensional motion state information of the target object can be displayed on the display screen.
作为一种可选的实施方式,上述用于采集雷达探测数据的雷达可以为激光雷达(Lidar),激光雷达,所采集的雷达探测数据为激光点云数据;其中,点云数据可以是指在一个三维坐标系统中的一组向量的集合,集合中的向量可以以X、Y、Z三维坐标的形式表示。激光雷达是以发射激光束探测目标物体的位置、速度等特征量的雷达系统。激光雷达通过扫描生成激光点云数据,能够表征至少一个目标物体的距离、方位、高度、速度、姿态、形状等目标物体的运动状态。As an optional implementation manner, the above-mentioned radar for collecting radar detection data may be a lidar (Lidar), a laser radar, and the collected radar detection data is laser point cloud data; A set of vectors in a three-dimensional coordinate system. The vectors in the set can be represented in the form of X, Y, and Z three-dimensional coordinates. Lidar is a radar system that emits laser beams to detect the position, velocity and other characteristic quantities of target objects. Lidar generates laser point cloud data by scanning, which can characterize the motion state of at least one target object such as distance, azimuth, height, speed, attitude, and shape.
在一些实施例中,激光雷达能够对运动物体进行探测,也能够对静止物体的进行探测。因此,在本公开实施例中,目标物体可以是运动中的物体,也可以是静止的物体。本公开实施例对此不做特殊限定。In some embodiments, the lidar can detect moving objects as well as stationary objects. Therefore, in the embodiment of the present disclosure, the target object may be a moving object or a stationary object. This embodiment of the present disclosure makes no special limitation on this.
在一些实施例中,在雷达为激光雷达的情况下,多维运动状态信息可以包括目标物体的距离(与检测装置之间的距离)、方位、高度、速度、姿态、形状等信息中的至少一者。In some embodiments, in the case where the radar is a lidar, the multi-dimensional motion state information may include at least one of the distance (distance between the target object and the detection device), azimuth, altitude, speed, attitude, and shape of the target object. By.
在本公开的一些实施例中,通过获取激光点云数据和表征目标平面中的光强变化信息的事件数据,并将激光点云数据和事件数据进行融合处理,生成目标物体的多维运动状态信息,从而能够对至少一个静止或运动的目标物体进行清晰的运动判断,实现了对静止或运动的目标物体的类生物视觉感知。In some embodiments of the present disclosure, the multi-dimensional motion state information of the target object is generated by acquiring laser point cloud data and event data representing light intensity change information in the target plane, and fusing the laser point cloud data and the event data. , so that at least one stationary or moving target object can be clearly judged on the motion, and the biological-like visual perception of the stationary or moving target object is realized.
本公开实施例对于如何执行步骤S300对事件数据和激光点云数据进行融合处理不做特殊限定。作为一种可选的实施方式,利用神经网络(如卷积神经网络)对事件数据和激光点云数据进行融合处理。在一些实施例中,神经网络的输入为三维图像和事件帧,其中,三维图像是根据激光点云数据生成的,事件帧是根据获取的事件数据进行组帧生成的。相应地,参照图6,步骤S300可以进一步包括:步骤S310B至步骤S330B。This embodiment of the present disclosure does not specifically limit how to perform step S300 to perform fusion processing on event data and laser point cloud data. As an optional implementation manner, a neural network (such as a convolutional neural network) is used to perform fusion processing on event data and laser point cloud data. In some embodiments, the input of the neural network is a three-dimensional image and an event frame, wherein the three-dimensional image is generated according to laser point cloud data, and the event frame is generated by framing according to the acquired event data. Correspondingly, referring to FIG. 6 , step S300 may further include: steps S310B to S330B.
在步骤S310B中,根据激光点云数据生成三维图像。In step S310B, a three-dimensional image is generated according to the laser point cloud data.
其中,点云数据包括所有发出的激光信号中对应有反射信号的点与发射源(即激光雷达)的距离信息,根据这些距离信息可以得到这些反射点构成的障碍物即目标物体在三维空间中的位置信息,将这些空间位置信息从球坐标变换到XYZ三维坐标即可得到相应的三维图像,生成的三维图像可以是指三维点云图。Among them, the point cloud data includes the distance information between the point corresponding to the reflected signal in all the emitted laser signals and the emission source (ie, the laser radar). The corresponding three-dimensional image can be obtained by transforming the spatial position information from spherical coordinates to XYZ three-dimensional coordinates, and the generated three-dimensional image may refer to a three-dimensional point cloud image.
在步骤S320B中,根据预设采样周期内的事件数据,生成事件帧。In step S320B, an event frame is generated according to the event data within the preset sampling period.
关于该步骤S320B的描述可参见上述对步骤S311A的描述,此处不再赘述。For the description of this step S320B, reference may be made to the above description of step S311A, and details are not repeated here.
在步骤S330B中,将三维图像和事件帧输入神经网络进行处理,生成目标物体的多维运动状态信息。In step S330B, the three-dimensional image and the event frame are input into the neural network for processing to generate multi-dimensional motion state information of the target object.
在一些实施例中,在步骤S330B中,将三维图像投影到和事件帧相同的二维平面,再将两种图像在通道维度上拼接在一起后一同输入神经网络(如卷积神经网络)进行特征 提取,从而得到目标物体的多维运动状态信息。In some embodiments, in step S330B, the three-dimensional image is projected onto the same two-dimensional plane as the event frame, and the two images are stitched together in the channel dimension and then input to a neural network (eg, a convolutional neural network) for Feature extraction, so as to obtain the multi-dimensional motion state information of the target object.
在一些实施例中,神经网络(如卷积神经网络)的输入为二维图像和事件帧。其中,二维图像是将三维激光点云数据分别沿前视方向和俯视方向投影得到的俯视图和前视图,从而得到三维激光点云数据的二维图像表示;事件帧是将预设采样周期的事件数据进行组帧,按帧输入神经网络。相应地,在一些实施例中,参照图7,步骤S300可以进一步包括:步骤S310C至步骤S330C。In some embodiments, the inputs to a neural network (eg, a convolutional neural network) are two-dimensional images and event frames. Among them, the two-dimensional image is the top view and front view obtained by projecting the three-dimensional laser point cloud data along the front-view direction and the top-view direction respectively, so as to obtain the two-dimensional image representation of the three-dimensional laser point cloud data; the event frame is the preset sampling period. The event data is framed and fed into the neural network frame by frame. Correspondingly, in some embodiments, referring to FIG. 7 , step S300 may further include steps S310C to S330C.
在步骤S310C中,对激光点云数据进行处理,生成激光点云数据的前视图和俯视图。In step S310C, the laser point cloud data is processed to generate a front view and a top view of the laser point cloud data.
可以理解的是,点云数据由三维坐标表示,通过对点云数据进行某一方向的投影映射,可以将点云数据的三维坐标变换为二维坐标,得到对应的投影视图,而对点云数据沿前视和俯视方向进行投影映射,则可以分别得到相应的前视图和俯视图。It can be understood that the point cloud data is represented by three-dimensional coordinates. By performing projection mapping on the point cloud data in a certain direction, the three-dimensional coordinates of the point cloud data can be transformed into two-dimensional coordinates to obtain the corresponding projected view. The data is projected and mapped along the front-view and top-view directions, and the corresponding front-view and top-view views can be obtained respectively.
在步骤S320C中,根据预设采样周期内的事件数据,生成事件帧。In step S320C, an event frame is generated according to the event data in the preset sampling period.
关于该步骤S320C的描述可参见上述对步骤S311A的描述,此处不再赘述。For the description of this step S320C, reference may be made to the above description of step S311A, and details are not repeated here.
在步骤S330C中,将前视图、俯视图、事件帧输入神经网络进行处理,生成目标物体的多维运动状态信息。In step S330C, the front view, the top view, and the event frame are input into the neural network for processing to generate multi-dimensional motion state information of the target object.
在一些实施例中,可以将前视图和俯视图以及事件帧在通道维度上拼接在一起后一同输入神经网络进行特征提取,从而获得目标物体的多维运动状态信息。In some embodiments, the front view, the top view and the event frame may be spliced together in the channel dimension and then input to the neural network for feature extraction, so as to obtain the multi-dimensional motion state information of the target object.
本公开实施例还提供了将激光点云数据与事件数据进行融合处理的非神经网络的处理方式。在一些实施例中,参照图8,步骤S300可以进一步包括步骤S310D-步骤S320D。The embodiment of the present disclosure also provides a non-neural network processing method for fusion processing of laser point cloud data and event data. In some embodiments, referring to FIG. 8 , step S300 may further include steps S310D-S320D.
在步骤S310D中,根据事件数据确定至少一个目标区域,得到至少一个目标区域的第一坐标信息,每一个目标区域对应一个目标物体。In step S310D, at least one target area is determined according to the event data, first coordinate information of the at least one target area is obtained, and each target area corresponds to a target object.
在步骤S320D中,根据第一坐标信息确定目标区域在激光点云数据中的第二坐标信息,生成目标物体的多维运动状态信息。In step S320D, the second coordinate information of the target area in the laser point cloud data is determined according to the first coordinate information, and multi-dimensional motion state information of the target object is generated.
在步骤S310D中,可以利用目标检测算法检测根据事件数据生成的事件帧中至少一个目标区域的目标物体,从而确定各目标区域在事件帧中第一坐标信息。In step S310D, a target detection algorithm may be used to detect a target object in at least one target area in the event frame generated according to the event data, so as to determine the first coordinate information of each target area in the event frame.
在一些实施例中,在步骤S320D中,通过三维空间的坐标变化算法,可以将激光点云数据和事件帧中不同角度的图像变换到同一个角度,进而可以将图像中的点相关联,从而确定目标物体的多维运动状态信息。In some embodiments, in step S320D, through the coordinate change algorithm in three-dimensional space, the laser point cloud data and the images of different angles in the event frame can be transformed to the same angle, and then the points in the images can be correlated, thereby Determine the multi-dimensional motion state information of the target object.
在一些实施例中,还可以通过两个安装在不同位置的DVS可以得到目标物体在两个不同平面上的运动状态信息,进而通过计算可以唯一确定目标物体的多维运动状态信息。In some embodiments, the motion state information of the target object on two different planes can also be obtained through two DVSs installed in different positions, and then the multi-dimensional motion state information of the target object can be uniquely determined through calculation.
在一些实施例中,还可以通过一个或多个图像传感器对目标物体进行探测,并将一个或多个图像传感器产生的信号与事件数据、激光点云数据进行融合处理,形成对目标物体 的多维感知。作为一种可选的实施方式,图像传感器为互补金属氧化物半导体(CMOS,Complementary Metal Oxide Semiconductor)传感器。相应地,参照图9,目标检测方法还包括:在步骤S500中,获取至少一路RGB图像信号。此种情况下,步骤S300还可以进一步包括:在步骤S310E中,将至少一路RGB图像信号和激光点云数据、事件数据进行融合处理,生成目标物体的多维运动信息。In some embodiments, the target object can also be detected by one or more image sensors, and the signals generated by the one or more image sensors can be fused with event data and laser point cloud data to form a multi-dimensional image of the target object. perception. As an optional implementation manner, the image sensor is a Complementary Metal Oxide Semiconductor (CMOS, Complementary Metal Oxide Semiconductor) sensor. Correspondingly, referring to FIG. 9 , the target detection method further includes: in step S500 , acquiring at least one channel of RGB image signals. In this case, step S300 may further include: in step S310E, fusing at least one channel of RGB image signal with laser point cloud data and event data to generate multi-dimensional motion information of the target object.
在本公开的一些实施例中,神经网络可以包括多条处理支路,每一条处理支路对应处理一路RGB图像信号,不同路的RGB图像信号是通过不同的RGB图像传感器采集的图像信号。RGB图像信号通过对应的处理支路输入神经网络。在本公开的一些实施例中,RGB图像信号按帧输入神经网络。In some embodiments of the present disclosure, the neural network may include a plurality of processing branches, each processing branch correspondingly processes a channel of RGB image signals, and the RGB image signals of different channels are image signals collected by different RGB image sensors. The RGB image signals are input to the neural network through corresponding processing branches. In some embodiments of the present disclosure, the RGB image signal is input to the neural network frame by frame.
在本公开的一些实施例中,从不同角度拍摄得到的RGB图像信号、激光点云数据和事件帧可以通过关键点检测等方式进行配准(将不同角度、不同视野范围拍摄得到的图片通过算法操作变为同一个角度同样视野范围的图片),将配准后的三组图片在通道维度拼接在一起后一同输入网络做特征提取,从而得到目标物体的多维运动信息。In some embodiments of the present disclosure, RGB image signals, laser point cloud data, and event frames captured from different angles can be registered through key point detection (the images captured at different angles and different fields of view are The operation becomes a picture of the same angle and the same field of view), and the registered three sets of pictures are spliced together in the channel dimension and then input to the network for feature extraction to obtain the multi-dimensional motion information of the target object.
此外,还可以将上述RGB图像信号、激光点云数据和事件帧输入到一个具有多个输入的神经网络,网络输出目标物体的运动状态信息如速度、位置等,该神经网络需要通过被标记的训练样本进行网络训练得到。In addition, the above RGB image signals, laser point cloud data and event frames can also be input into a neural network with multiple inputs, and the network outputs the motion state information of the target object such as speed, position, etc. The neural network needs to pass the marked The training samples are obtained by network training.
第二方面,本公开实施例提供一种目标物体的检测装置,检测装置可以用于实现上述检测方法,参照图10,该检测装置可以包括:第一传感器101、雷达102和融合处理单元103。In a second aspect, an embodiment of the present disclosure provides a detection device for a target object. The detection device can be used to implement the above detection method. Referring to FIG.
其中,第一传感器101,用于检测目标平面中的光强变化信息,以生成事件数据,光强变化信息用于确定目标平面中的至少一个目标物体。The first sensor 101 is used to detect light intensity change information in the target plane to generate event data, and the light intensity change information is used to determine at least one target object in the target plane.
雷达102,用于获取针对目标物体的雷达探测数据,雷达探测数据为描述目标物体的运动状态的信息。The radar 102 is configured to acquire radar detection data for the target object, where the radar detection data is information describing the motion state of the target object.
融合处理单元103,用于将事件数据与雷达探测数据进行融合处理,生成目标物体的多维运动状态信息。The fusion processing unit 103 is configured to perform fusion processing on event data and radar detection data to generate multi-dimensional motion state information of the target object.
本公开实施例提供的检测装置能够应用于自动驾驶,用于检测目标物体。The detection device provided by the embodiment of the present disclosure can be applied to automatic driving, and is used to detect a target object.
在本公开实施例中,融合处理单元103能够执行本公开实施例第一方面所述的目标物体的检测方法,将所述事件数据与雷达探测数据进行融合,生成目标物体的多维运动状态信息。In the embodiment of the present disclosure, the fusion processing unit 103 can execute the method for detecting a target object described in the first aspect of the embodiment of the present disclosure, and fuse the event data with the radar detection data to generate multi-dimensional motion state information of the target object.
本公开实施例对检测装置中的第一传感器101不做特殊限定。作为一种可选的实施方式,检测装置中的第一传感器101可以为动态视觉传感器(DVS,Dynamic Vision Sensor)。 DVS是一种模仿生物视觉的工作机理的传感器,能够检测光的改变并输出光强发生变化像素的地址(位置坐标)和信息,有效减少可冗余数据,并能够对场景变化实时动态响应。The embodiment of the present disclosure does not specifically limit the first sensor 101 in the detection device. As an optional implementation manner, the first sensor 101 in the detection device may be a dynamic vision sensor (DVS, Dynamic Vision Sensor). DVS is a sensor that imitates the working mechanism of biological vision. It can detect the change of light and output the address (position coordinates) and information of the pixel where the light intensity changes, effectively reduce redundant data, and can dynamically respond to scene changes in real time.
相应地,在一些实施例中,所述第一传感器101为动态视觉传感器;所述动态视觉传感器用于检测所述目标平面中各像素点的光强的变化,生成事件数据,事件数据包括目标平面中光强发生变化的像素点的坐标和光强变化信息,还可以进一步包括时间信息。Correspondingly, in some embodiments, the first sensor 101 is a dynamic vision sensor; the dynamic vision sensor is used to detect changes in the light intensity of each pixel in the target plane, and generate event data, where the event data includes the target The coordinates and light intensity change information of the pixel points where the light intensity changes in the plane may further include time information.
作为一种可选的实施方式,检测装置中的雷达102可以为脉冲多普勒雷达,雷达探测数据包括目标物体在第一方向的第一运动状态分量信息,第一方向垂直于目标平面。其中,脉冲多普勒雷达用于发送并接收脉冲信号,以确定目标物体在第一方向的第一运动状态分量信息;融合处理单元103可以从脉冲多普勒雷达获取第一运动状态分量信息。As an optional implementation manner, the radar 102 in the detection device may be a pulse Doppler radar, and the radar detection data includes first motion state component information of the target object in a first direction, and the first direction is perpendicular to the target plane. The pulse Doppler radar is used for sending and receiving pulse signals to determine the first motion state component information of the target object in the first direction; the fusion processing unit 103 can obtain the first motion state component information from the pulse Doppler radar.
下面对脉冲多普勒雷达做简要介绍。The following is a brief introduction to pulse Doppler radar.
多普勒雷达是指利用多普勒效应,对目标相对于雷达的径向速度分量进行测定、或对具有特定径向速度的目标径向提取的雷达。脉冲多普勒雷达为发送脉冲信号的多普勒雷达。Doppler radar refers to the radar that uses the Doppler effect to measure the radial velocity component of the target relative to the radar, or to extract the radial direction of the target with a specific radial velocity. A pulsed Doppler radar is a Doppler radar that transmits pulsed signals.
脉冲多普勒雷达以固定频率发生脉冲波对空扫描时,如遇到活动目标,反射回波的频率和发射波的频率回出现频率差,即多普勒频移。多普勒频移与活动目标与雷达的相对径向速度成正比。根据多普勒频移的幅度,可以确定活动目标的径向速度。多普勒频移的幅度通过信号的相位进行计算。因此,在本公开实施例中,雷达102为相参雷达,从而可以保留相位信息。When the pulse Doppler radar scans the air with a pulse wave at a fixed frequency, if it encounters a moving target, the frequency of the reflected echo and the frequency of the transmitted wave will have a frequency difference, that is, the Doppler frequency shift. The Doppler shift is proportional to the relative radial velocity of the moving target and the radar. From the magnitude of the Doppler shift, the radial velocity of the moving target can be determined. The magnitude of the Doppler shift is calculated from the phase of the signal. Therefore, in the embodiment of the present disclosure, the radar 102 is a coherent radar, so that phase information can be preserved.
还需要说明的是,对于一个目标物体,当目标物体向着雷达运动时,多普勒频移为正;当目标物体背离雷达运动时,多普勒频移为负。It should also be noted that, for a target object, when the target object moves toward the radar, the Doppler frequency shift is positive; when the target object moves away from the radar, the Doppler frequency shift is negative.
在一些实施例中,参照图10,检测装置还包括:输出单元104,用于输出目标物体的多维运动状态信息。In some embodiments, referring to FIG. 10 , the detection apparatus further includes: an output unit 104 for outputting multi-dimensional motion state information of the target object.
本公开实施例对输出单元不做特殊限制。例如,输出单元为显示屏,在显示屏上显示目标物体的多维运动状态信息。The embodiment of the present disclosure does not impose special restrictions on the output unit. For example, the output unit is a display screen, and the multi-dimensional motion state information of the target object is displayed on the display screen.
在本公开的一些实施例中,融合处理单元103用于:根据事件数据确定目标物体在所述目标平面的第二运动状态分量信息;根据第二运动状态分量信息与第一运动状态分量信息,生成目标物体的多维运动状态信息。In some embodiments of the present disclosure, the fusion processing unit 103 is configured to: determine the second motion state component information of the target object on the target plane according to the event data; according to the second motion state component information and the first motion state component information, Generate multi-dimensional motion state information of the target object.
在本公开的一些实施例中,雷达102为激光雷达,雷达探测数据为激光点云数据;激光雷达,用于发射激光束对至少一个目标物体进行探测,生成激光点云数据。In some embodiments of the present disclosure, the radar 102 is a laser radar, and the radar detection data is laser point cloud data; the laser radar is used for emitting a laser beam to detect at least one target object to generate laser point cloud data.
在一些实施例中,参照图11,融合处理单103可以包括(ISP,Image Signal Processing)131和第一神经网络132。其中,第一图像信号处理器131用于:根据激光点云数据生成 三维图像;根据预设采样周期内的事件数据,生成事件帧。第一神经网络132用于对三维图像和事件帧进行处理,生成目标物体的多维运动状态信息。In some embodiments, referring to FIG. 11 , the fusion processing unit 103 may include (ISP, Image Signal Processing) 131 and a first neural network 132. Wherein, the first image signal processor 131 is used for: generating a three-dimensional image according to the laser point cloud data; and generating an event frame according to the event data in a preset sampling period. The first neural network 132 is used to process three-dimensional images and event frames to generate multi-dimensional motion state information of the target object.
在一些实施例中,参照图12,融合处理单103可以包括第二图像信号处理器133和第二神经网络(如卷积神经网络)134。其中,第二图像信号处理器133用于对激光点云数据进行处理,生成激光点云数据的前视图和俯视图;根据预设采样周期内的事件数据,生成事件帧;第二神经网络134用于对前视图、俯视图、事件帧进行处理,生成目标物体的多维运动状态信息。In some embodiments, referring to FIG. 12 , the fusion processing unit 103 may include a second image signal processor 133 and a second neural network (eg, a convolutional neural network) 134 . Among them, the second image signal processor 133 is used for processing the laser point cloud data to generate the front view and top view of the laser point cloud data; according to the event data in the preset sampling period, the event frame is generated; the second neural network 134 uses It is used to process the front view, top view and event frame to generate multi-dimensional motion state information of the target object.
在一些实施例中,融合处理单元用于:根据事件数据确定至少一个目标区域,得到至少一个目标区域的第一坐标信息,每一个目标区域对应一个目标物体;根据第一坐标信息确定目标区域在激光点云数据中的第二坐标信息,生成目标物体的多维运动状态信息。In some embodiments, the fusion processing unit is configured to: determine at least one target area according to the event data, obtain first coordinate information of the at least one target area, each target area corresponds to a target object; determine the target area according to the first coordinate information The second coordinate information in the laser point cloud data generates multi-dimensional motion state information of the target object.
在一些实施例中,参照图11和图12,检测装置还包括至少一个第二传感器140;第二传感器140用于获取RGB图像,并生成RGB图像信号;融合处理单元103用于将至少一路RGB图像信号和激光点云数据、事件数据进行融合处理,生成目标物体的多维运动状态信息。In some embodiments, referring to FIG. 11 and FIG. 12 , the detection device further includes at least one second sensor 140; the second sensor 140 is used to acquire RGB images and generate RGB image signals; the fusion processing unit 103 is used to combine at least one RGB image The image signal, laser point cloud data and event data are fused to generate multi-dimensional motion state information of the target object.
在一些实施例中,参照图11,第一神经网络132可以包括多条处理支路,每一条处理支路对应一个第二传感器140,第二传感器140输出的RGB图像信号通过对应的处理支路输入第一神经网络132。In some embodiments, referring to FIG. 11 , the first neural network 132 may include a plurality of processing branches, each processing branch corresponds to a second sensor 140 , and the RGB image signals output by the second sensor 140 pass through the corresponding processing branch Input to the first neural network 132 .
在一些实施例中,参照图12,第二神经网络134包括多条处理支路,每一条处理支路对应一个第二传感器140,第二传感器140输出的RGB图像信号通过对应的处理支路输入第二神经网络134。In some embodiments, referring to FIG. 12 , the second neural network 134 includes a plurality of processing branches, each processing branch corresponds to a second sensor 140 , and the RGB image signals output by the second sensor 140 are input through the corresponding processing branch The second neural network 134 .
在一些实施例中,RGB图像信号按帧输入第一神经网络132或第二神经网络134。In some embodiments, the RGB image signals are input to the first neural network 132 or the second neural network 134 frame by frame.
在本公开实施例中,检测装置用于实现上述任一实施例提供的检测方法,其他相关描述可参见上述实施例的检测方法中具体的描述,此处不再赘述。In this embodiment of the present disclosure, the detection apparatus is used to implement the detection method provided by any of the foregoing embodiments. For other related descriptions, reference may be made to the specific description of the detection method in the foregoing embodiment, which will not be repeated here.
第三方面,本公开实施例提供一种融合处理单元,应用于目标物体的检测装置,参照图13,融合处理单元包括:一个或多个处理器201;存储器202,其上存储有一个或多个程序,当一个或多个程序被一个或多个处理器执行,使得一个或多个处理器实现本公开实施例第一方面所述的目标物体的检测方法;一个或多个I/O接口203,连接在处理器201与存储器202之间,配置为实现处理器201与存储器202的信息交互。In a third aspect, an embodiment of the present disclosure provides a fusion processing unit, which is applied to a target object detection device. Referring to FIG. 13 , the fusion processing unit includes: one or more processors 201 ; and a memory 202 that stores one or more processors thereon. a program, when one or more programs are executed by one or more processors, so that one or more processors implement the target object detection method described in the first aspect of the embodiment of the present disclosure; one or more I/O interfaces 203 , connected between the processor 201 and the memory 202 , and configured to implement information interaction between the processor 201 and the memory 202 .
其中,处理器201为具有数据处理能力的器件,其包括但不限于中央处理器(CPU)等;存储器202为具有数据存储能力的器件,其包括但不限于随机存取存储器(RAM,更具体如SDRAM、DDR等)、只读存储器(ROM)、带电可擦可编程只读存储器 (EEPROM)、闪存(FLASH);I/O接口(读写接口)203连接在处理器201与存储器202间,能实现处理器201与存储器202的信息交互,其包括但不限于数据总线(Bus)等。The processor 201 is a device with data processing capability, including but not limited to a central processing unit (CPU), etc.; the memory 202 is a device with data storage capability, including but not limited to random access memory (RAM, more specifically Such as SDRAM, DDR, etc.), read only memory (ROM), electrified erasable programmable read only memory (EEPROM), flash memory (FLASH); I/O interface (read and write interface) 203 is connected between the processor 201 and the memory 202 , can realize the information interaction between the processor 201 and the memory 202, which includes but is not limited to a data bus (Bus) and the like.
在一些实施例中,处理器201、存储器202和I/O接口203通过总线204相互连接,进而与计算设备的其它组件连接。In some embodiments, processor 201, memory 202, and I/O interface 203 are interconnected by bus 204, which in turn is connected to other components of the computing device.
第四方面,本公开实施例提供一种计算机可读介质,其上存储有计算机程序,所述程序被处理器执行时实现本公开实施例第一方面所述的目标物体的检测方法。In a fourth aspect, an embodiment of the present disclosure provides a computer-readable medium on which a computer program is stored, and when the program is executed by a processor, implements the method for detecting a target object described in the first aspect of the embodiment of the present disclosure.
本领域普通技术人员可以理解,上文中所公开方法中的全部或某些步骤、系统、装置中的功能模块/单元可以被实施为软件、固件、硬件及其适当的组合。在硬件实施方式中,在以上描述中提及的功能模块/单元之间的划分不一定对应于物理组件的划分;例如,一个物理组件可以具有多个功能,或者一个功能或步骤可以由若干物理组件合作执行。某些物理组件或所有物理组件可以被实施为由处理器,如中央处理器、数字信号处理器或微处理器执行的软件,或者被实施为硬件,或者被实施为集成电路,如专用集成电路。这样的软件可以分布在计算机可读介质上,计算机可读介质可以包括计算机存储介质(或非暂时性介质)和通信介质(或暂时性介质)。如本领域普通技术人员公知的,术语计算机存储介质包括在用于存储信息(诸如计算机可读指令、数据结构、程序模块或其他数据)的任何方法或技术中实施的易失性和非易失性、可移除和不可移除介质。计算机存储介质包括但不限于RAM、ROM、EEPROM、闪存或其他存储器技术、CD-ROM、数字多功能盘(DVD)或其他光盘存储、磁盒、磁带、磁盘存储或其他磁存储装置、或者可以用于存储期望的信息并且可以被计算机访问的任何其他的介质。此外,本领域普通技术人员公知的是,通信介质通常包含计算机可读指令、数据结构、程序模块或者诸如载波或其他传输机制之类的调制数据信号中的其他数据,并且可包括任何信息递送介质。Those of ordinary skill in the art can understand that all or some of the steps in the methods disclosed above, functional modules/units in the systems, and devices can be implemented as software, firmware, hardware, and appropriate combinations thereof. In a hardware implementation, the division between functional modules/units mentioned in the above description does not necessarily correspond to the division of physical components; for example, one physical component may have multiple functions, or one function or step may be composed of several physical components Components execute cooperatively. Some or all physical components may be implemented as software executed by a processor, such as a central processing unit, digital signal processor or microprocessor, or as hardware, or as an integrated circuit, such as an application specific integrated circuit . Such software may be distributed on computer-readable media, which may include computer storage media (or non-transitory media) and communication media (or transitory media). As known to those of ordinary skill in the art, the term computer storage media includes both volatile and nonvolatile implemented in any method or technology for storage of information, such as computer readable instructions, data structures, program modules or other data flexible, removable and non-removable media. Computer storage media include, but are not limited to, RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, digital versatile disk (DVD) or other optical disk storage, magnetic cartridges, magnetic tape, magnetic disk storage or other magnetic storage devices, or may Any other medium used to store desired information and which can be accessed by a computer. In addition, communication media typically embodies computer readable instructions, data structures, program modules, or other data in a modulated data signal such as a carrier wave or other transport mechanism, and can include any information delivery media, as is well known to those of ordinary skill in the art .
本文已经公开了示例实施例,并且虽然采用了具体术语,但它们仅用于并仅应当被解释为一般说明性含义,并且不用于限制的目的。在一些实例中,对本领域技术人员显而易见的是,除非另外明确指出,否则可单独使用与特定实施例相结合描述的特征、特性和/或元素,或可与其他实施例相结合描述的特征、特性和/或元件组合使用。因此,本领域技术人员将理解,在不脱离由所附的权利要求阐明的本公开的范围的情况下,可进行各种形式和细节上的改变。Example embodiments have been disclosed herein, and although specific terms are employed, they are used and should only be construed in a general descriptive sense and not for purposes of limitation. In some instances, it will be apparent to those skilled in the art that features, characteristics and/or elements described in connection with a particular embodiment may be used alone or in combination with other embodiments, unless expressly stated otherwise. Features and/or elements are used in combination. Accordingly, it will be understood by those of ordinary skill in the art that various changes in form and details may be made therein without departing from the scope of the present disclosure as set forth in the appended claims.

Claims (23)

  1. 一种目标物体的检测方法,包括:A method for detecting a target object, comprising:
    获取事件数据,所述事件数据表征目标平面中的光强变化信息,所述光强变化信息用于确定所述目标平面中的至少一个目标物体;acquiring event data, where the event data represents light intensity change information in the target plane, and the light intensity change information is used to determine at least one target object in the target plane;
    获取针对所述目标物体的雷达探测数据,所述雷达探测数据为描述所述目标物体的运动状态的信息;acquiring radar detection data for the target object, where the radar detection data is information describing the motion state of the target object;
    将所述事件数据与所述雷达探测数据进行融合处理,生成所述目标物体的多维运动状态信息。The event data and the radar detection data are fused to generate multi-dimensional motion state information of the target object.
  2. 根据权利要求1所述的检测方法,其中,所述雷达探测数据包括所述目标物体在第一方向的第一运动状态分量信息,所述第一方向垂直于所述目标平面;The detection method according to claim 1, wherein the radar detection data includes first motion state component information of the target object in a first direction, and the first direction is perpendicular to the target plane;
    所述将所述事件数据与所述雷达探测数据进行融合处理,生成所述目标物体的多维运动状态信息的步骤,包括:The step of performing fusion processing on the event data and the radar detection data to generate the multi-dimensional motion state information of the target object includes:
    根据所述事件数据确定所述目标物体在所述目标平面的第二运动状态分量信息;Determine the second motion state component information of the target object on the target plane according to the event data;
    根据所述第二运动状态分量信息与所述第一运动状态分量信息,生成所述目标物体的多维运动状态信息。Multi-dimensional motion state information of the target object is generated according to the second motion state component information and the first motion state component information.
  3. 根据权利要求2所述的检测方法,其中,所述事件数据包括所述目标平面中光强发生变化的像素点的坐标和光强变化信息;所述第二运动状态分量信息包括所述目标物体在所述目标平面中的位置坐标;The detection method according to claim 2, wherein the event data includes the coordinates of the pixel points in the target plane where the light intensity changes and light intensity change information; the second motion state component information includes the target object position coordinates in the target plane;
    所述根据所述事件数据确定所述目标物体在所述目标平面的第二运动状态分量信息的步骤,包括:The step of determining the second motion state component information of the target object on the target plane according to the event data includes:
    根据预设采样周期内的所述事件数据中各所述像素点的坐标和光强变化信息,生成事件帧;generating an event frame according to the coordinates and light intensity change information of each of the pixel points in the event data within the preset sampling period;
    根据所述事件帧确定所述目标物体在所述目标平面中的位置坐标。The position coordinates of the target object in the target plane are determined according to the event frame.
  4. 根据权利要求3所述的检测方法,其中,所述第二运动状态分量信息还包括所述目标物体在第二方向的第二速度分量和在第三方向的第三速度分量,所述第二方向、所述第三方向与所述目标平面平行且所述第二方向与所述第三方向垂直;The detection method according to claim 3, wherein the second motion state component information further includes a second velocity component of the target object in a second direction and a third velocity component in a third direction, the second direction, the third direction is parallel to the target plane, and the second direction is perpendicular to the third direction;
    所述根据所述事件数据确定所述目标物体在所述目标平面的第二运动状态分量信息的步骤,还包括:The step of determining the second motion state component information of the target object on the target plane according to the event data further includes:
    根据所述目标物体在所述目标平面中的位置坐标,确定所述目标物体在所述第二方向的第二偏移量和在所述第三方向的第三偏移量;determining a second offset of the target object in the second direction and a third offset of the target object in the third direction according to the position coordinates of the target object in the target plane;
    根据所述第二偏移量确定所述第二速度分量;determining the second velocity component based on the second offset;
    根据所述第三偏移量确定所述第三速度分量。The third velocity component is determined based on the third offset.
  5. 根据权利要求4所述的检测方法,其中,第一运动状态分量信息包括所述目标物体在所述第一方向的第一速度分量和距离参数;所述多维运动状态信息包括所述目标物体的三维速度和三维坐标;The detection method according to claim 4, wherein the first motion state component information includes a first velocity component and a distance parameter of the target object in the first direction; the multi-dimensional motion state information includes a 3D velocity and 3D coordinates;
    所述根据所述第二运动状态分量信息与所述第一运动状态分量信息,生成所述目标物体的多维运动状态信息的步骤,包括:The step of generating multi-dimensional motion state information of the target object according to the second motion state component information and the first motion state component information includes:
    确定对应于同一时间点的所述目标物体在所述目标平面中的位置坐标、所述第一速度分量、所述第二速度分量、所述第三速度分量和所述目标物体在所述第一方向的距离参数;Determine the position coordinates of the target object in the target plane corresponding to the same time point, the first velocity component, the second velocity component, the third velocity component and the target object in the first velocity component. distance parameter in one direction;
    根据对应于同一时间点的所述目标物体在所述目标平面中的位置坐标、所述第一速度分量、所述第二速度分量、所述第三速度分量和所述目标物体在所述第一方向的距离参数,确定所述三维速度和所述三维坐标。According to the position coordinates of the target object in the target plane corresponding to the same time point, the first velocity component, the second velocity component, the third velocity component and the target object in the first velocity A distance parameter in one direction determines the three-dimensional velocity and the three-dimensional coordinates.
  6. 根据权利要求2至5中任意一项所述的检测方法,其中,所述获取针对所述目标物体的雷达探测数据的步骤,包括:The detection method according to any one of claims 2 to 5, wherein the step of acquiring radar detection data for the target object comprises:
    获取所述目标物体在所述第一方向的第一速度分量和距离参数,以作为所述第一运动状态分量信息。A first velocity component and a distance parameter of the target object in the first direction are acquired as the first motion state component information.
  7. 根据权利要求1所述的检测方法,其中,所述雷达探测数据为激光点云数据;The detection method according to claim 1, wherein the radar detection data is laser point cloud data;
    所述将所述事件数据与所述雷达探测数据进行融合处理,生成所述目标物体的多维运动状态信息的步骤,包括:The step of performing fusion processing on the event data and the radar detection data to generate the multi-dimensional motion state information of the target object includes:
    根据所述激光点云数据生成三维图像;generating a three-dimensional image according to the laser point cloud data;
    根据预设采样周期内的所述事件数据,生成事件帧;generating an event frame according to the event data in the preset sampling period;
    将所述三维图像和所述事件帧输入神经网络进行处理,生成所述目标物体的多维运动状态信息。The three-dimensional image and the event frame are input into a neural network for processing to generate multi-dimensional motion state information of the target object.
  8. 根据权利要求1所述的检测方法,其中,所述雷达探测数据为激光点云数据;The detection method according to claim 1, wherein the radar detection data is laser point cloud data;
    所述将所述事件数据与所述雷达探测数据进行融合处理,生成所述目标物体的多维 运动状态信息的步骤,包括:The described event data is fused with the radar detection data, and the step of generating the multi-dimensional motion state information of the target object includes:
    对所述激光点云数据进行处理,生成所述激光点云数据的前视图和俯视图;processing the laser point cloud data to generate a front view and a top view of the laser point cloud data;
    根据预设采样周期内的所述事件数据,生成事件帧;generating an event frame according to the event data in the preset sampling period;
    将所述前视图、所述俯视图、所述事件帧输入所述神经网络进行处理,生成所述目标物体的多维运动状态信息。The front view, the top view, and the event frame are input into the neural network for processing to generate multi-dimensional motion state information of the target object.
  9. 根据权利要求1所述的检测方法,其中,所述雷达探测数据为激光点云数据;The detection method according to claim 1, wherein the radar detection data is laser point cloud data;
    所述将所述事件数据与所述雷达探测数据进行融合处理,生成所述目标物体的多维运动状态信息的步骤,包括:The step of performing fusion processing on the event data and the radar detection data to generate the multi-dimensional motion state information of the target object includes:
    根据所述事件数据确定至少一个目标区域,得到所述至少一个目标区域的第一坐标信息,每一个所述目标区域对应一个所述目标物体;Determine at least one target area according to the event data, obtain first coordinate information of the at least one target area, and each of the target areas corresponds to one of the target objects;
    根据所述第一坐标信息确定所述目标区域在所述激光点云数据中的第二坐标信息,生成所述目标物体的多维运动状态信息。The second coordinate information of the target area in the laser point cloud data is determined according to the first coordinate information, and multi-dimensional motion state information of the target object is generated.
  10. 根据权利要求1所述的检测方法,其中,所述雷达探测数据为激光点云数据;The detection method according to claim 1, wherein the radar detection data is laser point cloud data;
    所述检测方法还包括:获取至少一路RGB图像信号;The detection method further includes: acquiring at least one RGB image signal;
    所述将所述事件数据与所述雷达探测数据进行融合处理,生成所述目标物体的多维运动状态信息的步骤,包括:The step of performing fusion processing on the event data and the radar detection data to generate the multi-dimensional motion state information of the target object includes:
    将至少一路所述RGB图像信号和所述激光点云数据、所述事件数据进行融合处理,生成所述目标物体的多维运动状态信息。At least one channel of the RGB image signal, the laser point cloud data, and the event data are fused to generate multi-dimensional motion state information of the target object.
  11. 根据权利要求1至10中任意一项所述的检测方法,其中,所述获取事件数据的步骤,包括:The detection method according to any one of claims 1 to 10, wherein the step of acquiring event data comprises:
    响应于所述目标平面中像素点的光强度变化获取所述事件数据,所述事件数据包括所述目标平面中光强发生变化的像素点的坐标、光强变化信息以及时间信息。The event data is acquired in response to changes in the light intensity of the pixel points in the target plane, and the event data includes coordinates of the pixel points where the light intensity changes in the target plane, light intensity change information, and time information.
  12. 根据权利要求1至10中任意一项所述的检测方法,其中,所述检测方法还包括:The detection method according to any one of claims 1 to 10, wherein the detection method further comprises:
    输出所述多维运动状态信息。The multi-dimensional motion state information is output.
  13. 根据权利要求1至10中任意一项所述检测方法,其中,所述获取事件数据的步骤,包括:The detection method according to any one of claims 1 to 10, wherein the step of acquiring event data comprises:
    通过动态视觉传感器获取所述事件数据。The event data is acquired by a dynamic vision sensor.
  14. 一种目标物体的检测装置,包括:A device for detecting a target object, comprising:
    第一传感器,用于检测目标平面中的光强变化信息,以生成事件数据,所述光强变化信息用于确定所述目标平面中的至少一个目标物体;a first sensor for detecting light intensity change information in the target plane to generate event data, the light intensity change information being used to determine at least one target object in the target plane;
    雷达,用于获取针对所述目标物体的雷达探测数据,所述雷达探测数据为描述所述目标物体的运动状态的信息;a radar, for acquiring radar detection data for the target object, where the radar detection data is information describing the motion state of the target object;
    融合处理单元,用于将所述事件数据与所述雷达探测数据进行融合处理,生成所述目标物体的多维运动状态信息。The fusion processing unit is configured to perform fusion processing on the event data and the radar detection data to generate multi-dimensional motion state information of the target object.
  15. 根据权利要求14所述的检测装置,其中,所述第一传感器为动态视觉传感器;The detection device according to claim 14, wherein the first sensor is a dynamic vision sensor;
    所述动态视觉传感器用于检测所述目标平面中各像素点的光强的变化,生成所述事件数据;所述事件数据包括所述目标平面中光强发生变化的像素点的坐标和光强变化信息。The dynamic vision sensor is used to detect the change of the light intensity of each pixel in the target plane, and generate the event data; the event data includes the coordinates and light intensity of the pixel where the light intensity changes in the target plane change information.
  16. 根据权利要求14所述的检测装置,其中,所述雷达为脉冲多普勒雷达,所述雷达探测数据包括所述目标物体在第一方向的第一运动状态分量信息,所述第一方向垂直于所述目标平面;The detection device according to claim 14, wherein the radar is a pulse Doppler radar, and the radar detection data includes first motion state component information of the target object in a first direction, and the first direction is vertical at the target plane;
    所述脉冲多普勒雷达用于发送并接收脉冲信号,以确定所述目标物体在所述第一方向的所述第一运动状态分量信息。The pulse Doppler radar is used for sending and receiving pulse signals to determine the first motion state component information of the target object in the first direction.
  17. 根据权利要求16所述的检测装置,其中,所述融合处理单元用于:根据所述事件数据确定所述目标物体在所述目标平面的第二运动状态分量信息;根据所述第二运动状态分量信息与所述第一运动状态分量信息,生成所述目标物体的多维运动状态信息。The detection device according to claim 16, wherein the fusion processing unit is configured to: determine the second motion state component information of the target object on the target plane according to the event data; The component information and the first motion state component information generate multi-dimensional motion state information of the target object.
  18. 根据权利要求14所述的检测装置,其中,所述雷达为激光雷达,所述雷达探测数据为激光点云数据;The detection device according to claim 14, wherein the radar is a laser radar, and the radar detection data is laser point cloud data;
    所述激光雷达,用于发射激光束对至少一个目标物体进行探测,生成所述激光点云数据;The lidar is used for emitting a laser beam to detect at least one target object, and generating the laser point cloud data;
    所述融合处理单元包括第一图像信号处理器和第一神经网络;The fusion processing unit includes a first image signal processor and a first neural network;
    所述第一图像信号处理器用于:根据所述激光点云数据生成三维图像;根据预设采样周期内的所述事件数据,生成事件帧;The first image signal processor is used for: generating a three-dimensional image according to the laser point cloud data; generating an event frame according to the event data in a preset sampling period;
    所述第一神经网络用于对所述三维图像和所述事件帧进行处理,生成所述目标物体 的多维运动状态信息。The first neural network is used to process the three-dimensional image and the event frame to generate multi-dimensional motion state information of the target object.
  19. 根据权利要求14所述的检测装置,其中,所述雷达为激光雷达,所述雷达探测数据为激光点云数据;The detection device according to claim 14, wherein the radar is a laser radar, and the radar detection data is laser point cloud data;
    所述激光雷达,用于发射激光束对至少一个目标物体进行探测,生成所述激光点云数据;The lidar is used for emitting a laser beam to detect at least one target object, and generating the laser point cloud data;
    所述融合处理单元包括第二图像信号处理器和第二神经网络;The fusion processing unit includes a second image signal processor and a second neural network;
    所述第二图像信号处理器用于对所述激光点云数据进行处理,生成所述激光点云数据的前视图和俯视图;根据预设采样周期内的所述事件数据,生成事件帧;The second image signal processor is used for processing the laser point cloud data to generate a front view and a top view of the laser point cloud data; and generating an event frame according to the event data in a preset sampling period;
    所述第二神经网络用于对所述前视图、所述俯视图、所述事件帧进行处理,生成所述目标物体的多维运动状态信息。The second neural network is used for processing the front view, the top view and the event frame to generate multi-dimensional motion state information of the target object.
  20. 根据权利要求14所述的目标检测方法,其中,所述雷达为激光雷达,所述雷达探测数据为激光点云数据;The target detection method according to claim 14, wherein the radar is a laser radar, and the radar detection data is laser point cloud data;
    所述激光雷达,用于发射激光束对至少一个目标物体进行探测,生成所述激光点云数据;The lidar is used for emitting a laser beam to detect at least one target object, and generating the laser point cloud data;
    所述融合处理单元用于:根据所述事件数据确定至少一个目标区域,得到所述至少一个目标区域的第一坐标信息,每一个所述目标区域对应一个所述目标物体;根据所述第一坐标信息确定所述目标区域在所述激光点云数据中的第二坐标信息,生成所述目标物体的多维运动状态信息。The fusion processing unit is configured to: determine at least one target area according to the event data, obtain first coordinate information of the at least one target area, and each target area corresponds to one of the target objects; The coordinate information determines the second coordinate information of the target area in the laser point cloud data, and generates multi-dimensional motion state information of the target object.
  21. 根据权利要求14所述的检测方法,其中,所述雷达为激光雷达,所述雷达探测数据为激光点云数据;所述激光雷达,用于发射激光束对至少一个目标物体进行探测,生成所述激光点云数据;The detection method according to claim 14, wherein the radar is a lidar, and the radar detection data is laser point cloud data; the lidar is used to emit a laser beam to detect at least one target object, and generate all the laser point cloud data;
    所述检测装置还包括至少一个第二传感器;The detection device further includes at least one second sensor;
    所述第二传感器用于获取RGB图像,并生成RGB图像信号;The second sensor is used to acquire RGB images and generate RGB image signals;
    所述融合处理单元用于将至少一路所述RGB图像信号和所述激光点云数据、所述事件数据进行融合处理,生成所述目标物体的多维运动状态信息。The fusion processing unit is configured to perform fusion processing on at least one channel of the RGB image signal, the laser point cloud data, and the event data to generate multi-dimensional motion state information of the target object.
  22. 一种融合处理单元,应用于目标物体的检测装置,所述融合处理单元包括:A fusion processing unit, applied to a detection device of a target object, the fusion processing unit comprising:
    一个或多个处理器;one or more processors;
    存储装置,其上存储有一个或多个程序,当所述一个或多个程序被所述一个或多个 处理器执行,使得所述一个或多个处理器实现根据权利要求1至13中任意一项所述的目标物体的检测方法。A storage device having one or more programs stored thereon which, when executed by the one or more processors, cause the one or more processors to implement any one of claims 1 to 13 A method for detecting a target object.
  23. 一种计算机可读介质,其上存储有计算机程序,所述程序被处理器执行时实现根据权利要求1至13中任意一项所述的目标物体的检测方法。A computer-readable medium on which a computer program is stored, and when the program is executed by a processor, implements the method for detecting a target object according to any one of claims 1 to 13.
PCT/CN2021/141370 2020-12-25 2021-12-24 Method and apparatus for detecting target object, fusion processing unit, and medium WO2022135594A1 (en)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
CN202011560817.6 2020-12-25
CN202011562118.5 2020-12-25
CN202011562118.5A CN112816995B (en) 2020-12-25 2020-12-25 Target detection method and device, fusion processing unit and computer readable medium
CN202011560817.6A CN112666550B (en) 2020-12-25 2020-12-25 Moving object detection method and device, fusion processing unit and medium

Publications (1)

Publication Number Publication Date
WO2022135594A1 true WO2022135594A1 (en) 2022-06-30

Family

ID=82158850

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/141370 WO2022135594A1 (en) 2020-12-25 2021-12-24 Method and apparatus for detecting target object, fusion processing unit, and medium

Country Status (1)

Country Link
WO (1) WO2022135594A1 (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115422387A (en) * 2022-11-04 2022-12-02 山东矩阵软件工程股份有限公司 Point cloud data processing method and system based on multi-dimensional point cloud fusion data
CN116184427A (en) * 2022-12-21 2023-05-30 湖南迈克森伟电子科技有限公司 Distance detection system based on laser distance sensing technology
CN116912798A (en) * 2023-09-14 2023-10-20 南京航空航天大学 Cross-modal noise perception-based automatic driving event camera target detection method
CN117115926A (en) * 2023-10-25 2023-11-24 天津大树智能科技有限公司 Human body action standard judging method and device based on real-time image processing

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5604580A (en) * 1994-02-10 1997-02-18 Mitsubishi Denki Kabushiki Kaisha Optical radar apparatus for vehicle
CN106338616A (en) * 2016-08-02 2017-01-18 南京理工大学 Optical fiber array spatial filtering velocity measurement system
CN106908783A (en) * 2017-02-23 2017-06-30 苏州大学 Obstacle detection method based on multi-sensor information fusion
CN107036534A (en) * 2016-02-03 2017-08-11 北京振兴计量测试研究所 Method and system based on laser speckle measurement Vibration Targets displacement
CN111931752A (en) * 2020-10-13 2020-11-13 中航金城无人系统有限公司 Dynamic target detection method based on event camera
CN112666550A (en) * 2020-12-25 2021-04-16 北京灵汐科技有限公司 Moving object detection method and apparatus, fusion processing unit, and medium
CN112816995A (en) * 2020-12-25 2021-05-18 北京灵汐科技有限公司 Target detection method and device, fusion processing unit and computer readable medium

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5604580A (en) * 1994-02-10 1997-02-18 Mitsubishi Denki Kabushiki Kaisha Optical radar apparatus for vehicle
CN107036534A (en) * 2016-02-03 2017-08-11 北京振兴计量测试研究所 Method and system based on laser speckle measurement Vibration Targets displacement
CN106338616A (en) * 2016-08-02 2017-01-18 南京理工大学 Optical fiber array spatial filtering velocity measurement system
CN106908783A (en) * 2017-02-23 2017-06-30 苏州大学 Obstacle detection method based on multi-sensor information fusion
CN111931752A (en) * 2020-10-13 2020-11-13 中航金城无人系统有限公司 Dynamic target detection method based on event camera
CN112666550A (en) * 2020-12-25 2021-04-16 北京灵汐科技有限公司 Moving object detection method and apparatus, fusion processing unit, and medium
CN112816995A (en) * 2020-12-25 2021-05-18 北京灵汐科技有限公司 Target detection method and device, fusion processing unit and computer readable medium

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115422387A (en) * 2022-11-04 2022-12-02 山东矩阵软件工程股份有限公司 Point cloud data processing method and system based on multi-dimensional point cloud fusion data
CN116184427A (en) * 2022-12-21 2023-05-30 湖南迈克森伟电子科技有限公司 Distance detection system based on laser distance sensing technology
CN116184427B (en) * 2022-12-21 2023-12-29 湖南迈克森伟电子科技有限公司 Distance detection system based on laser distance sensing technology
CN116912798A (en) * 2023-09-14 2023-10-20 南京航空航天大学 Cross-modal noise perception-based automatic driving event camera target detection method
CN116912798B (en) * 2023-09-14 2023-12-19 南京航空航天大学 Cross-modal noise perception-based automatic driving event camera target detection method
CN117115926A (en) * 2023-10-25 2023-11-24 天津大树智能科技有限公司 Human body action standard judging method and device based on real-time image processing
CN117115926B (en) * 2023-10-25 2024-02-06 天津大树智能科技有限公司 Human body action standard judging method and device based on real-time image processing

Similar Documents

Publication Publication Date Title
WO2022135594A1 (en) Method and apparatus for detecting target object, fusion processing unit, and medium
US20210190497A1 (en) Simultaneous location and mapping (slam) using dual event cameras
Zaarane et al. Distance measurement system for autonomous vehicles using stereo camera
US20200334843A1 (en) Information processing apparatus, control method for same, non-transitory computer-readable storage medium, and vehicle driving support system
US9989630B2 (en) Structured-light based multipath cancellation in ToF imaging
US8180107B2 (en) Active coordinated tracking for multi-camera systems
WO2021023106A1 (en) Target recognition method and apparatus, and camera
US9501839B1 (en) Methods and systems for detecting moving objects in a sequence of image frames produced by sensors with inconsistent gain, offset, and dead pixels
KR102151815B1 (en) Method and Apparatus for Vehicle Detection Using Lidar Sensor and Camera Convergence
JP2020042819A (en) Method and apparatus for determining motion vector field, device, computer-readable storage medium, and vehicle
WO2021195886A1 (en) Distance determination method, mobile platform, and computer-readable storage medium
WO2022217988A1 (en) Sensor configuration scheme determination method and apparatus, computer device, storage medium, and program
Kang et al. Accurate fruit localisation using high resolution LiDAR-camera fusion and instance segmentation
JP2016085602A (en) Sensor information integrating method, and apparatus for implementing the same
CN113474819A (en) Information processing apparatus, information processing method, and program
CN112666550B (en) Moving object detection method and device, fusion processing unit and medium
CN112816995B (en) Target detection method and device, fusion processing unit and computer readable medium
JP2007200364A (en) Stereo calibration apparatus and stereo image monitoring apparatus using the same
Megalingam et al. Adding intelligence to the robotic coconut tree climber
CN113792645A (en) AI eyeball fusing image and laser radar
CN116136408A (en) Indoor navigation method, server, device and terminal
KR20180097004A (en) Method of position calculation between radar target lists and vision image ROI
WO2021097807A1 (en) Method and device for calibrating external parameters of detection device, and mobile platform
JP2021051524A (en) Work machine peripheral detection object position detection system, and work machine peripheral detection object position detection program
JP2021012043A (en) Information processing device for machine learning, information processing method for machine learning, and information processing program for machine learning

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21909589

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 02/11/2023)