WO2022033076A1 - 目标检测方法、装置、设备、存储介质及程序产品 - Google Patents
目标检测方法、装置、设备、存储介质及程序产品 Download PDFInfo
- Publication number
- WO2022033076A1 WO2022033076A1 PCT/CN2021/090359 CN2021090359W WO2022033076A1 WO 2022033076 A1 WO2022033076 A1 WO 2022033076A1 CN 2021090359 W CN2021090359 W CN 2021090359W WO 2022033076 A1 WO2022033076 A1 WO 2022033076A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- information
- dimensional
- pixel
- target
- Prior art date
Links
- 238000001514 detection method Methods 0.000 title claims abstract description 276
- 238000011176 pooling Methods 0.000 claims description 35
- 238000013528 artificial neural network Methods 0.000 claims description 33
- 238000012549 training Methods 0.000 claims description 20
- 238000004590 computer program Methods 0.000 claims description 15
- 238000012545 processing Methods 0.000 claims description 15
- 238000000605 extraction Methods 0.000 claims description 13
- 238000007689 inspection Methods 0.000 claims description 5
- 238000000034 method Methods 0.000 description 64
- 230000008569 process Effects 0.000 description 29
- 238000010586 diagram Methods 0.000 description 14
- 238000005516 engineering process Methods 0.000 description 10
- 230000003287 optical effect Effects 0.000 description 7
- 238000013135 deep learning Methods 0.000 description 6
- 238000006243 chemical reaction Methods 0.000 description 5
- 230000018109 developmental process Effects 0.000 description 5
- 230000006870 function Effects 0.000 description 4
- 230000009466 transformation Effects 0.000 description 4
- 238000004891 communication Methods 0.000 description 3
- 230000008878 coupling Effects 0.000 description 3
- 238000010168 coupling process Methods 0.000 description 3
- 238000005859 coupling reaction Methods 0.000 description 3
- 238000011161 development Methods 0.000 description 3
- 230000003190 augmentative effect Effects 0.000 description 2
- 230000008859 change Effects 0.000 description 2
- 239000013589 supplement Substances 0.000 description 2
- 238000012360 testing method Methods 0.000 description 2
- 230000006399 behavior Effects 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 230000015556 catabolic process Effects 0.000 description 1
- 238000006731 degradation reaction Methods 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 238000002372 labelling Methods 0.000 description 1
- 238000013507 mapping Methods 0.000 description 1
- 239000011159 matrix material Substances 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 239000000047 product Substances 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 230000000717 retained effect Effects 0.000 description 1
- 230000011218 segmentation Effects 0.000 description 1
- 238000009966 trimming Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/56—Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
- G06T7/73—Determining position or orientation of objects or cameras using feature-based methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/764—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10016—Video; Image sequence
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30248—Vehicle exterior or interior
- G06T2207/30252—Vehicle exterior; Vicinity of vehicle
Definitions
- the present disclosure is based on the Chinese patent application with the application number of 202010792241.X, the application date of August 8, 2020, and the application name of "a target detection method, device, equipment and storage medium", and requires the Chinese patent application Priority, the entire content of this Chinese patent application is hereby incorporated by reference into the present disclosure in its entirety.
- the present disclosure relates to the technical field of computer vision, and in particular, to a target detection method, apparatus, device, storage medium and program product.
- Target detection refers to the use of computer technology to detect and identify targets of interest in images or videos, such as common pedestrian detection, obstacle detection, etc.
- Target detection technology has been widely used in various fields, such as robotics, autonomous driving, and behavior recognition.
- the embodiments of the present disclosure provide at least one target detection solution.
- an embodiment of the present disclosure provides a target detection method, including:
- an embodiment of the present disclosure provides a target detection device, including:
- an acquisition module configured to acquire an image acquired by the image acquisition component and internal parameters of the image acquisition component; a determination module configured to determine each pixel in the acquired image based on the acquired image and the internal parameters
- the three-dimensional coordinate information of the point in the world coordinate system; the generating module is configured to generate and the collected image according to the three-dimensional coordinate information of each pixel point in the collected image in the world coordinate system.
- the three-dimensional information image corresponding to the acquired image; the sorting of the pixels in the three-dimensional information image is the same as the sorting of the pixels in the acquired image; the detection module is configured to determine the three-dimensional information image based on the three-dimensional information image.
- embodiments of the present disclosure provide an electronic device, including: a processor, a memory, and a bus, where the memory stores machine-readable instructions executable by the processor, and when the electronic device runs, the processing The processor and the memory communicate through a bus, and when the machine-readable instructions are executed by the processor, the steps of the target detection method according to the first aspect are performed.
- an embodiment of the present disclosure provides a computer-readable storage medium, where a computer program is stored on the computer-readable storage medium, and the computer program is executed by a processor to execute the target detection method described in the first aspect. step.
- an embodiment of the present disclosure provides a computer program product, including computer-readable code, and when the computer-readable code is executed in an electronic device, a processor in the electronic device executes the first The steps of the object detection method described in the aspect.
- a three-dimensional information image with the same image structure and the three-dimensional coordinate information of each pixel in the world coordinate system can be obtained based on the acquired image, based on the three-dimensional information image.
- the information image can complete the three-dimensional target detection for the target object.
- the image acquisition component has the advantages of high portability and low cost.
- the complete target object in the field of view can be obtained, including the target object with a small volume, so the three-dimensional target detection for the target object in the short-range area can be accurately completed.
- FIG. 1A shows a schematic diagram of the detection result of a target object in three-dimensional space
- FIG. 1B shows a schematic diagram of the detection result of the target object on the two-dimensional image
- FIG. 1C shows a flowchart of a target detection method provided by an embodiment of the present disclosure
- FIG. 2 shows a flowchart of a method for determining three-dimensional coordinate information of a pixel in a world coordinate system provided by an embodiment of the present disclosure
- FIG. 3 shows a schematic diagram of a scene for determining three-dimensional coordinate information of a pixel in a world coordinate system provided by an embodiment of the present disclosure
- FIG. 4 shows a flowchart of a first method for generating a three-dimensional information image provided by an embodiment of the present disclosure
- FIG. 5 shows a flowchart of a second method for generating a three-dimensional information image provided by an embodiment of the present disclosure
- FIG. 6 shows a flowchart of a method for determining three-dimensional detection information of a target object provided by an embodiment of the present disclosure
- FIG. 7 shows a flowchart of a method for determining three-dimensional detection information of a target object provided by an embodiment of the present disclosure
- FIG. 8 shows a schematic diagram of a neural network for determining three-dimensional detection information of a target object provided by an embodiment of the present disclosure
- FIG. 9A shows a schematic diagram of a training method of a neural network provided by an embodiment of the present disclosure
- FIG. 9B shows a schematic diagram of a training method of a neural network provided by an embodiment of the present disclosure.
- FIG. 10 shows a flowchart of a control method for a target vehicle provided by an embodiment of the present disclosure
- FIG. 11A shows a logic flow diagram of a target detection method provided by an embodiment of the present disclosure
- FIG. 11B shows a schematic diagram of an image to be detected provided by an embodiment of the present disclosure
- FIG. 11C shows a schematic diagram of a depth image provided by an embodiment of the present disclosure.
- 11D shows a schematic diagram of a three-dimensional information image provided by an embodiment of the present disclosure
- FIG. 12 shows a schematic structural diagram of a target detection apparatus provided by an embodiment of the present disclosure
- FIG. 13 shows a schematic diagram of an electronic device provided by an embodiment of the present disclosure.
- Object detection refers to the use of computer technology to detect and identify objects of interest in images or videos, such as common pedestrian detection and obstacle detection.
- two-dimensional target detection and three-dimensional target detection are included: the two-dimensional target detection result can mark the two-dimensional detection frame of the target object contained in the image, and the three-dimensional target detection result can mark the three-dimensional detection frame of the target object contained in the image.
- three-dimensional target detection is more complex and more significant.
- 3D object detection is an important task. This task needs to detect the coordinates, shape and orientation of the target in three-dimensional space. Due to the lack of depth information in image data, image-based 3D detection systems generally need to perform depth estimation on the target image to obtain the depth information of each pixel in the image, and then use the RGB image and the estimated depth map as the input of the system to calculate 3D information of the object in the image. As shown in FIGS. 1A and 1B , the detection results of the target object (car) in the three-dimensional space and the detection results on the two-dimensional image are respectively shown. The rectangular frame 11 is the detection result, and the rectangular frame 12 is the manual labeling result.
- the image-based three-dimensional detection methods in the related art mainly have the following shortcomings: on the one hand, the image data lacks corresponding depth information, and the three-dimensional information (position, shape, orientation) of the target cannot be effectively estimated; They belong to different coordinate systems, and directly using image data to calculate the results in three-dimensional space will produce large errors, resulting in serious performance degradation; on the other hand, using camera parameters can map depth data to three-dimensional space, but this method will Pixmaps are 3D point clouds. It will lead to additional problems: for example, the whole system will contain different forms of data (image data and point cloud data), so that the system must contain different modules to process these two kinds of data separately, which cannot be processed uniformly.
- a radar device When 3D target detection is performed on the target object based on the method of collecting point cloud images by the radar device, it is necessary to install a radar device for the object to be detected.
- a radar device is installed for a robot that performs 3D target detection.
- the method is more expensive and less portable.
- the radar device collects point cloud images for 3D target detection, because the radar device has a radar blind spot and the problem of low resolution, such a close-range radar blind spot, or a small target object, may Unable to generate valid point cloud data containing target correspondence. Therefore, when the radar device collects point cloud images for target detection, there are problems such as high cost, poor portability, and low accuracy when detecting objects in close-range areas or small volumes.
- an embodiment of the present disclosure provides a target detection method. After the image collected by the image collection component is obtained, each pixel in the collected image can be determined by using the collected image and the internal parameters of the image collection component. The three-dimensional coordinate information of the point in the world coordinate system, and then according to the collected image and the three-dimensional coordinate information of each pixel in the collected image under the world coordinate system, the pixel point sorting and the pixel points in the collected image are obtained. Sort consistent 3D informative images. Because the order of pixel points remains unchanged, the 3D information image can still retain the same image structure as the captured image. Based on this, the 3D detection information of the target object contained in the captured image in the world coordinate system can be effectively determined.
- the embodiment of the present disclosure performs target detection, after the image collected by the image collection component, the same image structure can be obtained based on the collected image, and the 3D coordinate information of each pixel in the world coordinate system is increased. 3D infographics. Based on the three-dimensional information image, three-dimensional target detection for the target object can be completed.
- the image acquisition component has the advantages of high portability and low cost; and compared with the point cloud data collected by the radar device, the image acquisition component can also obtain the complete target object within the field of view in the short-range area, including The target object is small in volume, so it can accurately complete the three-dimensional target detection for the target object in the close range.
- the execution body of the target detection method provided by the embodiments of the present disclosure is generally a computer device with a certain computing capability, and the computer device includes, for example, a terminal device or a server or other processing device.
- the object detection method may be implemented by the processor calling computer-readable instructions stored in the memory.
- the target detection method includes the following steps S101 to S104, wherein:
- Step S101 acquiring an image captured by an image capturing component and internal parameters of the image capturing component.
- the image acquisition component may include a visible light (red: Red; green: Green; blue: Blue, RGB) camera or a camera component such as an RGB camera that can acquire RGB images, and the corresponding acquired images may be RGB images.
- a visible light red: Red; green: Green; blue: Blue, RGB
- RGB visible light
- a camera component such as an RGB camera that can acquire RGB images
- the corresponding acquired images may be RGB images.
- the internal parameters of the image acquisition component may include some or all of the parameters in the camera internal parameter matrix for converting the image coordinate system to the camera coordinate system, which is not limited in this embodiment of the present disclosure.
- Step S102 based on the collected image and internal parameters, determine the three-dimensional coordinate information of each pixel in the collected image in the world coordinate system.
- an image coordinate system can be established based on the collected image, and the pixel coordinate value of each pixel in the image coordinate system can be determined based on the constructed image coordinate system, based on the conversion relationship between the image coordinate system and the camera coordinate system.
- the coordinate values of each pixel included in the acquired image along the X-axis and the Y-axis in the camera coordinate system can be determined.
- the coordinates of each pixel included in the acquired image along the X axis and the Y axis under the world coordinate system can be determined. value.
- the coordinate value of each pixel in the camera coordinate system can be directly used as the coordinate value of the pixel in the world coordinate system.
- the coordinate value of each pixel point along the Z-axis direction in the world coordinate system can be determined according to the depth information of the pixel point in the camera coordinate system.
- the depth image corresponding to the collected image can be determined according to the collected image and the pre-trained neural network for determining the depth image, so that each pixel in the collected image is at Depth information in the camera coordinate system. In this way, combining the pixel coordinate value of each pixel in the image coordinate system and the depth information of the pixel in the camera coordinate system, the three-dimensional coordinate information of the pixel in the world coordinate system can be determined, and the implementation process will be carried out later. elaborate.
- Step S103 according to the collected image and the three-dimensional coordinate information of each pixel in the collected image in the world coordinate system, generate a three-dimensional information image corresponding to the collected image;
- the order of the pixels in the three-dimensional information image is the same as the order of the pixels in the collected image.
- a plurality of pixels included in the collected image may form an image structure according to set information such as texture, tone, and order.
- the image structure can reflect the structure information corresponding to the target object to be detected contained in the collected image.
- the image structure of the collected image will not change. , that is, the shape of the target object contained in the image does not change. Therefore, when the order of the pixels in the 3D information image is the same as the order of the pixels in the collected image, the 3D information image can still retain the same image structure as the collected image. Based on this, the collected image can be effectively determined.
- the collected image when generating a three-dimensional information image corresponding to the collected image according to the collected image and the three-dimensional coordinate information of each pixel in the collected image in the world coordinate system, it may include:
- the three-dimensional information image is generated according to the three-dimensional coordinate information corresponding to each pixel point and the index information of the pixel point in the collected image; wherein, the channel information of each pixel point in the three-dimensional information image at least includes the pixel point in 3D coordinate information in the world coordinate system.
- the index information of each pixel in the collected image may represent the position of the pixel in the collected image, for example, the collected image contains m*n pixels, which can be obtained through the index information (i , j) to represent the index information of the pixel in the collected image, and (i, j) can represent that the pixel is located at row i and column j in the collected image.
- the 3D coordinate information corresponding to each pixel can be combined with the location of the pixel in The index information in the collected images is used to reconstruct a three-dimensional information image in the form of an image.
- the constructed three-dimensional information image and the collected image have the same image structure, that is, The shape of the contained target object remains unchanged.
- three-dimensional object detection can be performed on the target object included in the three-dimensional information image.
- the three-dimensional information image when generating the three-dimensional information image corresponding to the collected image, it is generated according to the index information of each pixel in the collected image, so the three-dimensional information image can still be retained and the collected image. same image structure.
- the three-dimensional information image also adds the three-dimensional coordinate information of the pixel in the world coordinate system for each pixel point, so it is possible to detect the target object in the world coordinate system based on the three-dimensional information image. 3D inspection information.
- Step S104 based on the three-dimensional information image, determine three-dimensional detection information of the target object contained in the collected image in the world coordinate system.
- the target objects include different shapes in different application scenarios.
- the target objects may include vehicles, pedestrians, and railings waiting for three-dimensional target detection.
- three-dimensional object detection can be performed on the target object based on the three-dimensional information image. Because the three-dimensional information image contains the same image structure as the acquired image, the three-dimensional detection information of the target object contained in the acquired image in the world coordinate system can be detected through the three-dimensional information image.
- the three-dimensional detection information of each target object in the world coordinate system may include the position coordinates of the center point of the target object in the world coordinate system, and the length, width and height of the target object in the world coordinate system, And the orientation angle of the target object in the world coordinate system.
- the orientation angle can be represented by the angle between the preset positive direction of the target object and the preset direction.
- the angle between the front of the vehicle and the preset direction can be used to represent the orientation of the vehicle. angle.
- the three-dimensional detection information of the target object may be represented by the position information of the three-dimensional (three-dimensional, 3D) detection frame corresponding to the target object.
- the length, width and height of the target object in the world coordinate system can be respectively represented by the length, width and height of the 3D detection frame corresponding to the target object
- the center point of the target object can be represented by the length, width and height of the 3D detection frame corresponding to the target object.
- the center point of the 3D detection frame is represented, and the orientation angle of the target object can be represented by the orientation angle of the 3D detection frame corresponding to the target object.
- the 3D detection frame corresponding to the target object can be represented by the circumscribed cuboid of the target object.
- a three-dimensional information image with the same image structure and the three-dimensional coordinate information of each pixel in the world coordinate system can be obtained based on the acquired image, based on the three-dimensional information image.
- the information image can complete the three-dimensional object detection for the target object.
- the image acquisition component has the advantages of high portability and low cost, and compared with the point cloud data collected by the radar device, the image acquisition component can also obtain the complete target object within the field of view in the short-range area. Including small-volume target objects, it can accurately complete the three-dimensional target detection for target objects in the close range.
- step S102 when determining the three-dimensional coordinate information of each pixel in the collected image in the world coordinate system based on the collected image and internal parameters, as shown in FIG. 2, the following steps S1021 to S1022 may be included:
- Step S1021 based on the collected image, generate a depth image corresponding to the collected image, where the depth image includes depth information of each pixel in the collected image;
- Step S1022 Determine the three-dimensional coordinate information of the pixel in the world coordinate system based on the two-dimensional coordinate information of each pixel in the collected image in the image coordinate system, the depth information and internal parameters of the pixel.
- the depth image corresponding to the collected image can be determined according to a pre-trained neural network for determining the depth image, so as to obtain each depth image in the collected image.
- the depth information of each pixel can be, for example, the depth information in the camera coordinate system.
- the neural network used to determine the depth image corresponding to the collected image can be obtained by training a large number of pre-collected sample images and the depth information of the set pixels marked for the sample image in the camera coordinate system.
- the embodiment does not limit the training process of the neural network for determining the depth image.
- the three-dimensional coordinate information of each pixel in the camera coordinate system can be determined first, and then the three-dimensional coordinate information of the pixel in the world coordinate system can be determined.
- the three-dimensional coordinate information of each pixel in the world coordinate system may include the coordinate value along the X-axis direction, the coordinate value along the Y-axis direction, and the coordinate value along the Z-axis direction under the world coordinate system.
- the embodiment of the present disclosure can make the camera coordinate system coincide with the world coordinate system, that is, make the coordinate origin of the camera coordinate system and the coordinate origin of the world coordinate system overlap, so that the X-axis of the camera coordinate system and the X-axis of the world coordinate system Coincidence makes the Y axis of the camera coordinate system coincide with the Y axis of the world coordinate system, and makes the Z axis of the camera coordinate system coincide with the Z axis of the world coordinate system.
- the pixel point P is the pixel point of the i-th row and the j-th column in the collected image.
- the three-dimensional coordinate information of the pixel point P in the world coordinate system can be determined according to the following formula (1):
- Z (i, j) represents the coordinate value of the pixel point P of the collected image along the Z-axis direction in the world coordinate system
- X (i, j) represents the pixel point P of the collected image in the world coordinate system.
- Y (i, j) represents the coordinate value of the pixel point P of the acquired image along the Y-axis direction in the world coordinate system
- u (i, j) represents the pixel point of the acquired image
- v (i, j) represents the coordinate value of the pixel point P of the collected image along the v-axis direction in the pixel coordinate system
- d (i, j) represents the acquisition The depth value of the pixel point P of the obtained image
- (Cx, Cy) represents the coordinate value of the light point C of the image acquisition component in the world coordinate system, where Cx represents the intersection of the optical axis of the image
- the camera parameter information used includes the coordinate value of the intersection of the optical axis of the image acquisition component and the acquired image along the X axis in the world coordinate system, and the image acquisition component.
- the optical center of the image acquisition component set on the target vehicle can be directly used as the origin, so that the world coordinate system and the camera coordinate system corresponding to the image acquisition component are coincident, so that the above formula can be directly used to determine each The three-dimensional coordinate information of a pixel in the world coordinate system.
- the depth information corresponding to each pixel of the collected image can be quickly predicted based on the collected image, and further based on the two-dimensional coordinate information of each pixel in the image coordinate system, the corresponding depth information, combined with the internal parameters of the image acquisition component, to quickly obtain the three-dimensional coordinate information of each pixel in the acquired image in the world coordinate system.
- a three-dimensional information image corresponding to the collected image can be generated based on the three-dimensional coordinate information of each pixel in the world coordinate system.
- step S103 when generating a three-dimensional information image according to the three-dimensional coordinate information corresponding to each pixel point and the index information of the pixel point in the collected image, as shown in FIG. 4 , the following steps S1031 to S1032 may be included:
- Step S1031 taking the three-dimensional coordinate information corresponding to each pixel as the multi-channel information corresponding to the pixel in the three-dimensional information image;
- Step S1032 Generate a three-dimensional information image based on the multi-channel information corresponding to the pixel in the three-dimensional information image and the index information of the pixel in the collected image.
- each pixel in the RGB image contains three-channel information in the RGB image, that is, the channel value on the R channel, the channel value on the G channel, and the channel on the B channel. value.
- the channel value of each pixel on the R channel, the channel value on the G channel, and the channel value on the B channel can represent the color information of the pixel in the RGB image.
- a three-dimensional information image is also composed of multiple pixels.
- the corresponding pixel point can be sequentially corresponding to the pixel point according to the index information of the pixel point in the collected image.
- the multi-channel information of each pixel in the three-dimensional information image in the three-dimensional information image includes the coordinate value of the pixel along the X-axis channel under the world coordinate system, and the coordinate value along the Y-axis channel under the world coordinate system.
- the three-dimensional information image contains the same number of pixels, and the sorting method of the pixels is unchanged. Therefore, the three-dimensional information image has the same image structure compared with the corresponding collected image. Therefore, the structure information of the target object contained in the collected image can be identified, so that it is convenient to perform three-dimensional object detection on the target object contained in the collected image based on the three-dimensional information image.
- step S103 when generating a three-dimensional information image according to the three-dimensional coordinate information corresponding to each pixel point and the index information of the pixel point in the collected image, as shown in FIG. 5 , may include the following steps S1033 to S1034:
- Step S1033 taking the three-dimensional coordinate information corresponding to each pixel point and the information of the pixel point in the collected image as the multi-channel information corresponding to the pixel point in the three-dimensional information image;
- Step S1034 Generate a three-dimensional information image based on the multi-channel information corresponding to the pixel in the three-dimensional information image and the index information of the pixel in the collected image.
- the pixel point can be defined as the pixel point according to the index information of the pixel point in the collected image.
- Add three-channel information composed of three-dimensional coordinate information and generate a three-dimensional information image corresponding to the collected image.
- each pixel of the three-dimensional information image obtained in this way can contain six channels.
- the information includes the channel value on the R channel, the channel value on the G channel, the channel value on the B channel, the coordinate value on the X-axis channel under the world coordinate system, and the coordinate value along the Y-axis channel under the world coordinate system. and the coordinate values along the Z-axis channel in the world coordinate system.
- the 3D information image generated in this way contains the same number of pixels and the same way of sorting the pixels. Therefore, compared with the corresponding collected image, the 3D information image has the same number of pixels as the collected image. image consistent with the image structure.
- the three-dimensional information image also retains the information of the collected image, such as the color information of the collected image, so as to facilitate accurate identification of the target object contained in the collected image based on the three-dimensional information image. 3D object detection.
- step S104 when determining the three-dimensional detection information of the target object contained in the collected image in the world coordinate system based on the three-dimensional information image, as shown in FIG. 6, the following steps S1041 to S1044 may be included:
- Step S1041 crop the three-dimensional information image based on the two-dimensional detection information of the target object contained in the collected image to obtain at least one three-dimensional information image block, and each three-dimensional information image block contains at least one target object.
- a pre-trained neural network for 2D target detection can be used to perform target detection on the collected image, so as to obtain the 2D detection information of the target object included in the collected image.
- the two-dimensional detection information of the target object may be the location area of the two-dimensional detection frame of the target object in the collected image.
- the three-dimensional information sample image block of the same size as the two-dimensional detection frame can be obtained by trimming the three-dimensional information image, so that the area that does not contain the target object can be filtered out. Therefore, the target detection can be directly performed on the three-dimensional information image block in the later stage, which can narrow the detection range and improve the detection efficiency.
- Step S1042 Perform feature extraction on each three-dimensional information image block to obtain multiple feature images corresponding to the three-dimensional information image block, and the multiple feature images include depth feature images representing depth information of the target object.
- multiple feature images corresponding to each three-dimensional information image block can be extracted based on the feature extraction network in the pre-trained neural network.
- the size of the three-dimensional information image blocks with different sizes can be adjusted so that the sizes of the three-dimensional information image blocks input to the feature extraction network are consistent.
- the feature extraction network can contain multiple convolution kernels, and each convolution kernel is used to extract a feature image corresponding to the three-dimensional information image block.
- the plurality of feature images may include a depth feature image used to characterize the depth information of the target object, a feature image used to characterize the length information of the target object, a feature image used to characterize the width information of the target object, and a feature image used to characterize the center point position of the target object. Feature image of information.
- Step S1043 Classify at least one 3D information image block based on the depth feature image corresponding to each 3D information image block, and determine a 3D object detection network corresponding to each type of 3D information image block.
- the depth information of the target object contained in each 3D information image block in the world coordinate system may be different, and multiple 3D information image blocks can be classified based on the depth feature image corresponding to each 3D information image block in advance. According to the depth information corresponding to the target object, a plurality of three-dimensional information image blocks are classified, and a three-dimensional target detection network corresponding to each type of three-dimensional information image is determined.
- the pre-trained neural network may include multiple 3D object detection networks, and each 3D object detection network can predict the 3D detection information of the target object contained in a class of 3D information image blocks, such as in the pre-trained neural network.
- Contains three target detection networks the first target detection network is used to detect 3D information image blocks with depth information greater than 0 and less than or equal to L1, and the second target detection network is used to detect depth information greater than L1 and less than or equal to L2.
- 3D information image patches the third object detection network is used to detect 3D information image patches with depth information greater than L3.
- each 3D object detection network can be enabled to detect 3D information image blocks with the same depth range.
- the differences in the 3D detection information corresponding to the target objects in the 3D information image blocks with the same depth range are small, so that the 3D object detection network can improve the detection accuracy during 3D object detection; on the other hand, when the 3D information image
- the 3D target detection can be performed simultaneously through a plurality of 3D target detection networks, so that the detection speed can be improved.
- the 3D object detection network corresponding to the 3D information image block can be determined.
- Step S1044 for each three-dimensional information image block, according to the three-dimensional object detection network corresponding to the three-dimensional information image block and a plurality of characteristic images corresponding to the three-dimensional information image block, determine that the target object in the three-dimensional information image block is in the world coordinate system. 3D inspection information under.
- the 3D target detection network When performing 3D target detection on the corresponding 3D information image block based on the 3D target detection network, it is necessary to consider multiple feature images corresponding to the 3D information image block, such as the above-mentioned depth feature image used to represent the depth information of the target object, A feature image used to characterize the length information of the target object, a feature image used to characterize the width information of the target object, and a feature image used to characterize the center point position information of the target object, etc. Each 3D target detection network can be based on these feature images.
- the three-dimensional detection information of the target object contained in the corresponding three-dimensional information image block is predicted.
- the three-dimensional information image can be cropped based on the two-dimensional detection information corresponding to the target object contained in the collected image to obtain a plurality of three-dimensional information image blocks.
- This method can filter out detections that do not contain the target object.
- 3D target detection is performed for 3D information image blocks
- multiple 3D target detection networks can be pre-built for simultaneous detection, which can improve detection accuracy and speed.
- step S1044 for each three-dimensional information image block, according to the three-dimensional object detection network corresponding to the three-dimensional information image block and a plurality of characteristic images corresponding to the three-dimensional information image block, it is determined that the target object in the three-dimensional information image block is in the
- the three-dimensional detection information in the world coordinate system may include the following steps S10441 to S10443:
- Step S10441 for each three-dimensional information image block, according to the set pooling size and pooling step size, perform maximum pooling processing on each feature image corresponding to the three-dimensional information image block, and obtain the feature image after pooling processing.
- the corresponding pooling value for each three-dimensional information image block, according to the set pooling size and pooling step size, perform maximum pooling processing on each feature image corresponding to the three-dimensional information image block, and obtain the feature image after pooling processing. The corresponding pooling value.
- each feature image contains an attribute feature of the target object contained in the 3D information image block, for example, it may contain the texture attribute feature, color attribute feature, depth attribute feature, and length attribute feature of the target object contained in the 3D information image block. , width attribute feature, center point position attribute feature, etc.
- a maximum pooling process may be performed to obtain a pooling value corresponding to the feature image after the pooling process. For example, take one of the feature images as an example.
- the feature image contains 4*4 feature values. According to the pooling size of 2*2 and the step size of 2, the maximum pooling process is performed, and 2*2 pooling values can be obtained. If Maximum pooling is performed according to the pooling size of the same size as the feature image, and 1*1 pooling values can be obtained.
- a binary mask image corresponding to the 3D information image block may be determined, and the binary mask image is in the region representing the target object The value of is 1, and the value of the area representing the non-target object is 0.
- each feature corresponding to the three-dimensional information image block can be first based on the binary mask image. The image is filtered, and the feature values representing the target object in each feature image are filtered out, and the feature values of non-target objects are changed to 0.
- the speed of the pooling process can be improved;
- the interference feature value of the region can be obtained, so that a more accurate pooling value can be obtained, so as to improve the accuracy of 3D target detection in the later stage.
- Step S10442 the pooled values corresponding to each feature image of the three-dimensional information image block form a target detection feature vector corresponding to the three-dimensional information image block.
- a target detection feature vector corresponding to the 3D information image block can be formed based on the pooled values of multiple feature images corresponding to the 3D information image block, and the 3D information image is represented by the target detection feature vector
- the comprehensive feature information of the target object contained in the block, the comprehensive feature information may include the above-mentioned texture attribute features, color attribute features, depth attribute features, length attribute features, width attribute features, and center point position attribute features, etc. .
- each three-dimensional information image block contains 10 feature images, and each feature image corresponds to 1*1 pooling value, then the target detection feature vector corresponding to the three-dimensional information image block contains 10 feature values; if Each feature image corresponds to 2*2 pooling values, and the target detection feature vector corresponding to the three-dimensional information image block contains 10*4 feature values.
- Step S10443 Based on the target detection feature vector corresponding to the 3D information image block and the 3D target detection network corresponding to the 3D information image block, determine the 3D detection information of the target object in the 3D information image block in the world coordinate system.
- the target detection feature vector corresponding to the three-dimensional information image block is input into the three-dimensional target detection network corresponding to the three-dimensional information image block, and the three-dimensional detection information of the target object contained in the three-dimensional information image block in the world coordinate system can be determined.
- a plurality of feature images 83 corresponding to the three-dimensional information image blocks 81 can be obtained.
- the pooling value 85 based on which the target detection feature vector corresponding to the three-dimensional information image block 81 is generated.
- the type prediction process can be performed on the pooling value 85, and the 3D object detection network 87 corresponding to each 3D information image block can be determined based on the pooling value representing the depth information of the target object, and the corresponding 3D information image block can be further
- the target detection feature vector is input to the corresponding 3D target detection network to complete the 3D target detection.
- the 3D detection information mentioned many times above is detected by a pre-trained neural network, and the neural network is obtained by training the sample images containing the labeled 3D detection information of the target sample object.
- a large number of sample images can be collected in advance, and each sample image can be labeled with the target sample object, and the labeled 3D detection information corresponding to the target sample object contained in each sample image can be determined.
- the labeled 3D detection information can be based on a preset.
- the three-dimensional coordinate information of the good target sample object in the world coordinate system is determined.
- the neural network is obtained by training the following steps, including steps S901 to S905:
- Step S901 acquiring the sample image collected by the image collecting component and the internal parameters of the image collecting component.
- This process is similar to the above-mentioned process of acquiring the acquired image and the internal parameters of the image acquisition component.
- the process description of the above-mentioned internal parameters of the image acquisition component please refer to the process description of the above-mentioned internal parameters of the image acquisition component.
- Step S902 based on the collected sample image and internal parameters, determine the three-dimensional coordinate information of each sample pixel in the collected sample image in the world coordinate system.
- This process is similar to the above-mentioned way of determining the three-dimensional coordinate information of each pixel in each acquired image in the world coordinate system.
- the technical details not disclosed in this process please refer to the above-mentioned determination of each acquired image.
- the process description of the three-dimensional coordinate information of each pixel in the world coordinate system is understood.
- Step S903 according to the collected sample image and the three-dimensional coordinate information of each sample pixel in the collected sample image in the world coordinate system, generate a three-dimensional information sample image corresponding to the collected sample image;
- the ordering of the sample pixels of is the same as the ordering of the sample pixels in the collected sample image.
- This process is similar to the above-mentioned method of generating a three-dimensional information image.
- technical details not disclosed in this process please refer to the above-mentioned description of the process of generating a three-dimensional information image.
- Step S904 based on the three-dimensional information sample image and the neural network to be trained, predict the three-dimensional detection information of the target sample object contained in the sample image in the world coordinate system.
- the neural network to be trained includes a variety of three-dimensional object detection networks. For the above step S904, based on the three-dimensional information sample image and the neural network to be trained, it is predicted that the target sample object contained in the sample image is in the world
- the three-dimensional detection information in the coordinate system may include the following steps S9041 to S9043:
- Step S9041 based on the two-dimensional detection information of the target sample object included in the sample image, the three-dimensional information sample image is trimmed to obtain at least one three-dimensional information sample image block, and each three-dimensional information image block includes at least one target object;
- Step S9042 performing feature extraction on at least one three-dimensional information sample image block to obtain multiple feature sample images corresponding to each three-dimensional information sample image block, and the multiple feature sample images include depth feature sample images representing depth information of the target sample object;
- Step S9043 classifying at least one three-dimensional information sample image block based on the depth feature sample images corresponding to the at least one three-dimensional information sample image respectively, and determining a three-dimensional target detection network corresponding to each three-dimensional information sample image block;
- Step S9044 for each three-dimensional information sample image block, predict the three-dimensional information sample image according to the three-dimensional target detection network corresponding to the three-dimensional information sample image block in the neural network and a plurality of characteristic sample images corresponding to the three-dimensional information sample image block.
- This process is similar to the above-mentioned way of predicting the 3D detection information of the target object in each 3D information image block in the world coordinate system.
- various 3D target detection networks can be obtained, and 3D target detection can be performed on 3D information image blocks with different depth information, thereby improving the detection accuracy and speed in the application process.
- Step S905 based on the predicted 3D detection information and the labeled 3D detection information, adjust the network parameter values in the neural network to be trained to obtain a neural network for determining the 3D detection information.
- the 3D detection information of the target sample object contained in each sample image can be predicted and obtained, and the loss value corresponding to the loss function of the neural network to be trained can be obtained based on the predicted 3D detection information and the actual labeled 3D detection information. , and then adjust the network parameter value based on the loss value to obtain a neural network for determining the three-dimensional detection information.
- the loss values corresponding to the predicted 3D detection information and the actual labeled 3D detection information may include a loss value for the size of the target sample object, a loss value for the center point of the target sample object, and a loss value for the target sample object
- the loss value of the orientation angle, etc. through multiple trainings to make the loss value less than the set loss threshold, or after the number of training times reaches the set number of training times, the adjustment of the network parameter values can be completed, and the trained neural network can be obtained.
- the target detection method provided by the embodiment of the present disclosure may be applied to the field of automatic driving, wherein the image acquisition component may be located on the target vehicle.
- the target detection method provided by the embodiment of the present disclosure further includes the following steps S1001 to S1002:
- Step S1001 based on the three-dimensional detection information of each target object, determine the distance information between the target object and the target vehicle;
- Step S1002 control the target vehicle to travel based on the three-dimensional detection information, distance information of each target object, and current pose data of the target vehicle.
- each target object Based on the three-dimensional detection information corresponding to each target object, it can include the size, orientation angle, and center point position coordinates of the target object in the world coordinate system, based on which the pose data of the target object in the world coordinate system can be represented. In addition, the distance information between the target object and the target vehicle can be obtained based on the position coordinates of the center point of each target object.
- the target vehicle Based on the three-dimensional detection information of each target object, the distance information to the target vehicle, and the current pose data of the target vehicle, the target vehicle can be controlled to avoid the target object as an obstacle.
- a world coordinate system can be established with the optical center of the image acquisition component as the origin, so that the distance between the center point of the target object and the origin in the world coordinate system can represent the distance between the target object and the origin in the world coordinate system. Distance information between target vehicles.
- the distance between the target object and the target vehicle may first be used to determine whether the target vehicle has entered the dangerous area corresponding to the target object, for example, when the distance is less than the preset safety distance, It can be determined that the target vehicle has entered the dangerous area corresponding to the target object, and further based on the three-dimensional pose data corresponding to the target object and the current pose data of the target vehicle, it can be determined whether a collision will occur when driving according to the current driving route. When it is determined that a collision will not occur, the vehicle can continue driving according to the original route, and when it is determined that a collision will occur, the driving route can be adjusted, or the vehicle can slow down to avoid obstacles.
- the distance information between each target object and the target vehicle can be obtained based on this, taking into account the three-dimensional detection of each target object
- the information can represent the pose data of the target object in the world coordinate system. Therefore, the driving of the target vehicle is controlled based on the three-dimensional detection information of the target object, the distance information from the target vehicle, and the current pose data of the target vehicle, which can improve the driving safety of the target vehicle.
- the embodiment of the present disclosure provides an image data coordinate system conversion method of an image-based three-dimensional detection system, which can maintain the image structure while converting the coordinate system, and further improve the accuracy of the detection system.
- the depth image of the image to be detected is calculated first, and then the internal parameters of the camera that captures the image are obtained; then the three-dimensional spatial position of each pixel is calculated by using the depth image and the internal parameters of the camera, and is organized into image data Finally, the three-dimensional information of the target is obtained by using image-oriented deep learning technology.
- FIG. 11A is a logical flowchart of a target detection method provided by an embodiment of the present disclosure. As shown in FIG. 11A , taking the image acquisition component as a camera as an example, the method includes at least the following steps:
- Step S1101 acquiring an image to be detected captured by a camera
- the image to be detected is a two-dimensional image of the target object, lacks corresponding depth information, and cannot effectively estimate the three-dimensional information (position, shape, orientation) of the target object.
- Step S1102 acquiring the depth image of the image to be detected
- the depth image of the image to be detected is shown in FIG. 11C , and the depth value of the part of the target object (car) is different from the depth value of the other parts.
- the missing depth information in image data can be compensated by image depth estimation methods. Using depth estimation to obtain the depth image of the image to be detected can effectively supplement the missing depth information in the two-dimensional image.
- depth estimation algorithms in the related art can generally meet this requirement, that is, to obtain camera parameters when the images to be detected are captured, and the embodiments of the present disclosure do not limit which depth estimation algorithms are sampled.
- Step S1103 acquiring camera parameters when the to-be-detected image is captured
- the camera parameters are internal parameters of the camera, which may include focal length and principal point.
- Step S1104 determining the three-dimensional coordinate information of each pixel in the image to be detected
- the index value (i, j) can indicate that the pixel is located in row i and column j in the image to be detected;
- the depth value d of the index value; and the camera internal parameters obtained in the previous step use formula (1) to calculate the coordinates of the pixel in the three-dimensional space, so as to obtain the three-dimensional coordinate information of all pixels in the image to be detected.
- Step S1105 generating a three-dimensional information image based on the three-dimensional coordinate information of each pixel;
- the three-dimensional coordinate information of each pixel in the image to be detected is organized into an image form through the three-dimensional information image, as shown in FIG. 11D .
- the calculated three-dimensional coordinates can be regarded as different channels and put back into the image, for example, to replace the original RGB channel.
- the pixel information after coordinate transformation is organized in the form of an image, thereby avoiding the introduction of point cloud data, so that there is only one data representation form of the image in the entire system, and the system is kept simple and efficient.
- step S1106 the neural network is used to detect the three-dimensional information image to obtain the detection result of the target object.
- 3D object detection is performed using deep learning techniques oriented to image data, such as estimating the pose of 3D objects. It is only necessary to use the image-oriented deep learning technology to estimate the three-dimensional information of the target, and the example of the present disclosure does not limit which kind of neural network is used.
- the embodiments of the present disclosure utilize the depth estimation method to obtain the depth image of the image to be detected, which can effectively supplement the missing depth information in the two-dimensional image.
- the embodiment of the present disclosure introduces coordinate system conversion, and establishes a one-to-one mapping from the image coordinate system to the three-dimensional world coordinate system through the internal parameters of the camera and the estimated depth image, eliminating the need for the image coordinate system and the three-dimensional world coordinate system. The ambiguity between them can greatly improve the detection performance of the system.
- the generated three-dimensional coordinate points are organized into an image representation form according to the coordinate index of the original image, and the image structure is maintained.
- the pixel information after coordinate transformation is organized in the form of images, thus avoiding the introduction of point cloud data, so that there is only one form of data representation in the entire system, keeping the system simple and efficient.
- the embodiments of the present disclosure only have the following beneficial effects: first, high precision: a method that does not use coordinate system transformation (or uses coordinate system transformation, but does not organize the transformed data into an image representation form) Compared with this system, the detection performance that can be obtained by this system is higher; secondly, the model training/testing process is simple: after other existing methods convert the image coordinate system to the 3D coordinate system, the pixels are regarded as point cloud data, which requires Neural networks with different structures are used to train the subsequent steps separately.
- the system uses data in the form of images from beginning to end, thus avoiding the conversion of data forms and making the overall training/testing process of the system easier; thirdly, it supports end-to-end training : Previous methods require training the model in stages.
- the neural network is trained using 2D image-oriented, and in the second stage, the neural network is trained using surface 3D point cloud.
- the two stages cannot interact, so the optimal solution cannot be obtained.
- the system can integrate two parts and use the neural network training for 2D images uniformly, thus supporting end-to-end training.
- the target detection method provided by the embodiments of the present disclosure can be applied to an automatic/assisted driving system based on image data.
- the target detection method provided by the embodiments of the present disclosure may be applied to an AR (Augmented Reality, augmented reality) system and/or a VR (Virtual Reality, virtual reality) system of a mobile terminal (such as a mobile phone), To achieve 3D object detection in AR systems and/or VR systems.
- AR Augmented Reality, augmented reality
- VR Virtual Reality, virtual reality
- the writing order of each step does not mean a strict execution order but constitutes any limitation on the implementation process, and the actual execution order of each step should be based on its functions and possible Internal logic is determined.
- the embodiment of the present disclosure also provides a target detection device corresponding to the target detection method. See implementation of the method.
- the target detection apparatus 1200 includes:
- an acquisition module 1201, configured to acquire an image acquired by an image acquisition component, and internal parameters of the image acquisition component;
- the determining module 1202 is configured to determine, based on the collected image and internal parameters, the three-dimensional coordinate information of each pixel in the collected image in the world coordinate system;
- the generating module 1203 is configured to generate a three-dimensional information image corresponding to the collected image according to the collected image and the three-dimensional coordinate information of each pixel in the collected image in the world coordinate system; the pixel point in the three-dimensional information image The ordering is the same as the ordering of the pixels in the collected image;
- the detection module 1204 is configured to determine, based on the three-dimensional information image, the three-dimensional detection information of the target object contained in the collected image in the world coordinate system.
- the target detection device 1200 further includes a control module 1205, and the image acquisition component is located on the target vehicle. After determining the three-dimensional detection information of the target object contained in the acquired image, the control module 1205 is configured to:
- the target vehicle is controlled to travel.
- the determining module 1202 is configured to:
- a depth image corresponding to the collected image is generated, and the depth image includes depth information corresponding to each pixel in the collected image;
- the depth information and internal parameters of each pixel determine the three-dimensional coordinate information of each pixel in the world coordinate system .
- the generating module 1203 is configured to:
- a three-dimensional information image is generated; the channel information of each pixel point in the three-dimensional information image contains at least each The three-dimensional coordinate information of the pixel point in the world coordinate system.
- the generating module 1203 is configured to:
- the three-dimensional information image is generated based on the multi-channel information corresponding to each pixel point in the three-dimensional information image and the index information of each pixel point in the collected image.
- the generating module 1203 is configured to:
- the three-dimensional information image is generated based on the multi-channel information corresponding to each pixel point in the three-dimensional information image and the index information of each pixel point in the collected image.
- the detection module 1204 is configured to:
- the three-dimensional information image is cropped to obtain at least one three-dimensional information image block, and each three-dimensional information image block contains at least one target object;
- each three-dimensional information image block Perform feature extraction on each three-dimensional information image block to obtain multiple feature images corresponding to each of the three-dimensional information image blocks, and the multiple feature images include depth feature images representing depth information of the target object;
- each 3D information image block For each 3D information image block, according to the 3D object detection network corresponding to each 3D information image block and a plurality of feature images corresponding to each 3D information image block, determine the 3D information image block in each 3D information image block.
- the 3D detection information of the target object in the world coordinate system.
- the detection module 1204 is configured to:
- each 3D information image block For each 3D information image block, according to the set pooling size and pooling step size, perform maximum pooling processing on each feature image corresponding to each 3D information image block to obtain each feature image pool The corresponding pooling value after processing;
- the pooled value corresponding to each feature image of each of the three-dimensional information image blocks is formed into a target detection feature vector corresponding to each of the three-dimensional information image blocks;
- the target detection feature vector corresponding to each of the three-dimensional information image blocks and the three-dimensional target detection network corresponding to each of the three-dimensional information image blocks it is determined that the target object in each of the three-dimensional information image blocks is in the world coordinate system 3D inspection information.
- the target detection apparatus 1200 further includes a training module 1206, and the training module 1206 is configured to:
- a neural network configured to detect 3D detection information is trained, and the neural network is trained using sample images containing labeled 3D detection information of the target sample object.
- an embodiment of the present disclosure further provides an electronic device 1300 .
- a schematic diagram of the electronic device provided by the embodiment of the present disclosure includes:
- the processor 131 and the memory 132 communicate through the bus 133, so that the processor 131 executes the following instructions : Acquire the image collected by the image acquisition component and the internal parameters of the image acquisition component; based on the acquired image and internal parameters, determine the three-dimensional coordinate information of each pixel in the acquired image in the world coordinate system; image and the three-dimensional coordinate information of each pixel in the collected image in the world coordinate system, to generate a three-dimensional information image corresponding to the collected image; the sorting of the pixels in the three-dimensional information image is the same as that in the collected image.
- the order of the pixel points is the same; based on the three-dimensional information image, the three-dimensional detection information of the target object contained in the collected image in the world coordinate system is determined.
- Embodiments of the present disclosure further provide a computer-readable storage medium, where a computer program is stored on the computer-readable storage medium, and when the computer program is run by a processor, the steps of the target detection method described in the above method embodiments are executed.
- the storage medium may be a volatile or non-volatile computer-readable storage medium.
- the computer program product of the target detection method provided by the embodiments of the present disclosure includes a computer-readable storage medium storing program codes, and the instructions included in the program codes can be used to execute the steps of the target detection method described in the above method embodiments. , see the above method examples.
- Embodiments of the present disclosure also provide a computer program, which implements any one of the methods in the foregoing embodiments when the computer program is executed by a processor.
- the computer program product can be implemented in hardware, software or a combination thereof.
- the computer program product is embodied as a computer storage medium, and in another optional embodiment, the computer program product is embodied as a software product, such as a software development kit (Software Development Kit, SDK) and the like.
- the units described as separate components may or may not be physically separated, and components displayed as units may or may not be physical units, that is, may be located in one place, or may be distributed to multiple network units. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution in this embodiment.
- each functional unit in each embodiment of the present disclosure may be integrated into one processing unit, or each unit may exist physically alone, or two or more units may be integrated into one unit.
- the functions, if implemented in the form of software functional units and sold or used as stand-alone products, may be stored in a processor-executable non-volatile computer-readable storage medium.
- the computer software products are stored in a storage medium, including Several instructions are used to cause a computer device (which may be a personal computer, a server, or a network device, etc.) to execute all or part of the steps of the methods described in various embodiments of the present disclosure.
- the aforementioned storage medium includes: U disk, mobile hard disk, read-only memory (Read-Only Memory, ROM), random access memory (Random Access Memory, RAM), magnetic disk or optical disk and other media that can store program codes .
- a three-dimensional information image with the same image structure and the three-dimensional coordinate information of each pixel in the world coordinate system can be obtained based on the acquired image, based on the three-dimensional information image.
- the information image can complete the three-dimensional target detection for the target object.
- the image acquisition component has the advantages of high portability and low cost.
- the complete target object in the field of view can be obtained, including the target object with a small volume, so the three-dimensional target detection for the target object in the short-range area can be accurately completed.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Evolutionary Computation (AREA)
- Data Mining & Analysis (AREA)
- Artificial Intelligence (AREA)
- Computing Systems (AREA)
- General Health & Medical Sciences (AREA)
- General Engineering & Computer Science (AREA)
- Life Sciences & Earth Sciences (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Multimedia (AREA)
- Computational Linguistics (AREA)
- Biophysics (AREA)
- Biomedical Technology (AREA)
- Molecular Biology (AREA)
- Mathematical Physics (AREA)
- Evolutionary Biology (AREA)
- Bioinformatics & Computational Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Databases & Information Systems (AREA)
- Medical Informatics (AREA)
- Image Analysis (AREA)
Abstract
Description
Claims (20)
- 一种目标检测方法,包括:获取图像采集部件采集的图像,以及所述图像采集部件的内部参数;基于采集到的图像和所述内部参数,确定所述采集到的图像中每个像素点在世界坐标系下的三维坐标信息;根据所述采集到的图像和所述采集到的图像中每个像素点在世界坐标系下的三维坐标信息,生成与所述采集到的图像对应的三维信息图像;所述三维信息图像中的像素点的排序与所述采集到的图像中的像素点的排序相同;基于所述三维信息图像,确定所述采集到的图像中包含的目标对象在所述世界坐标系下的三维检测信息。
- 根据权利要求1所述的目标检测方法,其中,所述图像采集部件位于目标车辆上,在确定所述采集到的图像中包含的目标对象的三维检测信息后,所述目标检测方法还包括:基于每个目标对象的三维检测信息,确定每个所述目标对象与所述目标车辆之间的距离信息;基于每个所述目标对象的所述三维检测信息、所述距离信息、以及所述目标车辆的当前位姿数据,控制所述目标车辆行驶。
- 根据权利要求1或2所述的目标检测方法,其中,所述基于采集到的图像和所述内部参数,确定所述采集到的图像中每个像素点在世界坐标系下的三维坐标信息,包括:基于所述采集到的图像,生成所述采集到的图像对应的深度图像,所述深度图像中包含所述采集到的图像中的每个像素点的深度信息;基于所述采集到的图像中每个像素点在图像坐标系下的二维坐标信息、每个所述像素点的深度信息以及所述内部参数,确定每个所述像素点在所述世界坐标系下的三维坐标信息。
- 根据权利要求1至3任一所述的目标检测方法,其中,所述根据所述采集到的图像和所述采集到的图像中每个像素点在世界坐标系下的三维坐标信息,生成与所述采集到的图像对应的三维信息图像,包括:按照所述采集到的图像中每个像素点对应的三维坐标信息,以及每个所述像素点在所述采集到的图像中的索引信息,生成所述三维信息图像;所述三维信息图像中的每个像素点的通道信息至少包含每个所述像素点在世界坐标系下的三维坐标信息。
- 根据权利要求4所述的目标检测方法,其中,所述按照所述采集到的图像中每个像素点对应的三维坐标信息,以及每个所述像素点在所述采集到的图像中的索引信息,生成所述三维信息图像,包括:将所述采集到的图像中每个像素点对应的三维坐标信息,作为每个所述像素点在所述三维信息图像中对应的多通道信息;基于每个所述像素点在所述三维信息图像中对应的多通道信息,以及每个所述像素点在所述采集到的图像中的索引信息,生成所述三维信息图像。
- 根据权利要求4所述的目标检测方法,其中,所述按照所述采集到的图像中每个像素点对应的三维坐标信息,以及每个所述像素点在所述采集到的图像中的索引信息,生成所述三维信息图像,包括:将所述采集到的图像中每个像素点对应的三维坐标信息以及每个所述像素点在所述采集到的图像中的信息,作为每个所述像素点在所述三维信息图像中对应的多通道信息;基于每个所述像素点在所述三维信息图像中对应的多通道信息,以及每个所述像素点在所述采集到的图像中的索引信息,生成所述三维信息图像。
- 根据权利要求1至6任一所述的目标检测方法,其中,所述基于所述三维信息图像,确定所述采集到的图像中包含的目标对象在所述世界坐标系下的三维检测信息,包括:基于所述采集到的图像中包含的目标对象的二维检测信息,对所述三维信息图像进行裁剪,得到至少一个三维信息图像块;其中,每个所述三维信息图像块中包含至少一个目标对象;针对每个所述三维信息图像块进行特征提取,得到每个所述三维信息图像块对应的多个特征图像,所述多个特征图像中包含表征每个所述目标对象深度信息的深度特征图像;基于每个所述三维信息图像块对应的深度特征图像,对所述至少一个三维信息图像块进行分类, 确定每种类别的三维信息图像块对应的三维目标检测网络;针对每个所述三维信息图像块,按照每个所述三维信息图像块对应的三维目标检测网络以及每个所述三维信息图像块对应的所述多个特征图像,确定每个所述三维信息图像块中的目标对象在所述世界坐标系下的三维检测信息。
- 根据权利要求7所述的目标检测方法,其中,所述针对每个三维信息图像块,按照每个所述三维信息图像块对应的三维目标检测网络以及每个所述三维信息图像块对应的所述多个特征图像,确定每个所述三维信息图像块中的目标对象在所述世界坐标系下的三维检测信息,包括:针对每个三维信息图像块,按照设定的池化尺寸和池化步长,对每个所述三维信息图像块对应的每个特征图像进行最大池化处理,得到每个所述特征图像池化处理后对应的池化值;将每个所述三维信息图像块的每个特征图像对应的池化值,组成每个所述三维信息图像块对应的目标检测特征向量;基于每个所述三维信息图像块对应的目标检测特征向量,以及每个所述三维信息图像块对应的三维目标检测网络,确定每个所述三维信息图像块中的目标对象在所述世界坐标系下的三维检测信息。
- 根据权利要求1至8任一所述的目标检测方法,其中,所述三维检测信息由神经网络检测得到,所述神经网络利用了包含目标样本对象的标注三维检测信息的样本图像训练得到。
- 一种目标检测装置,包括:获取模块,配置为获取图像采集部件采集的图像,以及所述图像采集部件的内部参数;确定模块,配置为基于采集到的图像和所述内部参数,确定所述采集到的图像中每个像素点在世界坐标系下的三维坐标信息;生成模块,配置为根据所述采集到的图像和所述采集到的图像中每个像素点在世界坐标系下的三维坐标信息,生成与所述采集到的图像对应的三维信息图像;所述三维信息图像中的像素点的排序与所述采集到的图像中的像素点的排序相同;检测模块,配置为基于所述三维信息图像,确定所述采集到的图像中包含的目标对象在所述世界坐标系下的三维检测信息。
- 根据权利要求10所述的目标检测装置,其中,所述目标检测装置还包括控制模块,所述图像采集部件位于目标车辆上,在确定所述采集到的图像中包含的目标对象的三维检测信息后,所述控制模块配置为:基于每个目标对象的三维检测信息,确定每个所述目标对象与所述目标车辆之间的距离信息;基于每个所述目标对象的所述三维检测信息、所述距离信息、以及所述目标车辆的当前位姿数据,控制所述目标车辆行驶。
- 根据权利要求10或11所述的目标检测装置,其中,所述确定模块配置为:基于所述采集到的图像,生成所述采集到的图像对应的深度图像,所述深度图像中包含所述采集到的图像中的每个像素点的深度信息;基于所述采集到的图像中每个像素点在图像坐标系下的二维坐标信息、每个所述像素点的深度信息以及所述内部参数,确定每个所述像素点在所述世界坐标系下的三维坐标信息。
- 根据权利要求10至12任一所述的目标检测装置,其中,所述生成模块配置为:按照所述采集到的图像中每个像素点对应的三维坐标信息,以及每个所述像素点在所述采集到的图像中的索引信息,生成所述三维信息图像;所述三维信息图像中的每个像素点的通道信息至少包含每个所述像素点在世界坐标系下的三维坐标信息。
- 根据权利要求13所述的目标检测装置,其中,所述生成模块配置为:将所述采集到的图像中每个像素点对应的三维坐标信息,作为每个所述像素点在所述三维信息图像中对应的多通道信息;基于每个所述像素点在所述三维信息图像中对应的多通道信息,以及每个所述像素点在所述采集到的图像中的索引信息,生成所述三维信息图像。
- 根据权利要求13所述的目标检测装置,其中,所述生成模块配置为:将所述采集到的图像中每个像素点对应的三维坐标信息以及每个所述像素点在所述采集到的图像中的信息,作为每个所述像素点在所述三维信息图像中对应的多通道信息;基于每个所述像素点在所述三维信息图像中对应的多通道信息,以及每个所述像素点在所述采集到的图像中的索引信息,生成所述三维信息图像。
- 根据权利要求10至15任一所述的目标检测装置,其中,所述检测模块配置为:基于所述采集到的图像中包含的目标对象的二维检测信息,对所述三维信息图像进行裁剪,得到至少一个三维信息图像块;其中,每个所述三维信息图像块中包含至少一个目标对象;针对每个所述三维信息图像块进行特征提取,得到每个所述三维信息图像块对应的多个特征图像,所述多个特征图像中包含表征每个所述目标对象深度信息的深度特征图像;基于每个所述三维信息图像块对应的深度特征图像,对所述至少一个三维信息图像块进行分类,确定每种类别的三维信息图像块对应的三维目标检测网络;针对每个所述三维信息图像块,按照每个所述三维信息图像块对应的三维目标检测网络以及每个所述三维信息图像块对应的所述多个特征图像,确定每个所述三维信息图像块中的目标对象在所述世界坐标系下的三维检测信息。
- 根据权利要求16所述的目标检测装置,其中,所述检测模块配置为:针对每个三维信息图像块,按照设定的池化尺寸和池化步长,对每个所述三维信息图像块对应的每个特征图像进行最大池化处理,得到每个所述特征图像池化处理后对应的池化值;将每个所述三维信息图像块的每个特征图像对应的池化值,组成每个所述三维信息图像块对应的目标检测特征向量;基于每个所述三维信息图像块对应的目标检测特征向量,以及每个所述三维信息图像块对应的三维目标检测网络,确定每个所述三维信息图像块中的目标对象在所述世界坐标系下的三维检测信息。
- 一种电子设备,包括:处理器、存储器和总线,所述存储器存储有所述处理器可执行的机器可读指令,当电子设备运行时,所述处理器与所述存储器之间通过总线通信,所述机器可读指令被所述处理器执行时执行如权利要求1至9任一所述的目标检测方法的步骤。
- 一种计算机可读存储介质,所述计算机可读存储介质上存储有计算机程序,所述计算机程序被处理器运行时执行如权利要求1至9任一所述的目标检测方法的步骤。
- 一种计算机程序产品,包括计算机可读代码,在所述计算机可读代码在电子设备中运行的情况下,所述电子设备中的处理器执行如权利要求1至9任一项所述的目标检测方法的步骤。
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020217042833A KR20220024193A (ko) | 2020-08-08 | 2021-04-27 | 타깃 검출 방법, 장치, 기기, 저장 매체 및 프로그램 제품 |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010792241.X | 2020-08-08 | ||
CN202010792241.XA CN111931643A (zh) | 2020-08-08 | 2020-08-08 | 一种目标检测方法、装置、电子设备及存储介质 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2022033076A1 true WO2022033076A1 (zh) | 2022-02-17 |
Family
ID=73308121
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2021/090359 WO2022033076A1 (zh) | 2020-08-08 | 2021-04-27 | 目标检测方法、装置、设备、存储介质及程序产品 |
Country Status (3)
Country | Link |
---|---|
KR (1) | KR20220024193A (zh) |
CN (1) | CN111931643A (zh) |
WO (1) | WO2022033076A1 (zh) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114655207A (zh) * | 2022-05-13 | 2022-06-24 | 中汽创智科技有限公司 | 一种数据处理方法、装置、设备及存储介质 |
CN115100423A (zh) * | 2022-06-17 | 2022-09-23 | 四川省寰宇众恒科技有限公司 | 一种基于视图采集数据实现实时定位系统及方法 |
CN115115687A (zh) * | 2022-06-24 | 2022-09-27 | 合众新能源汽车有限公司 | 车道线测量方法及装置 |
CN117308967A (zh) * | 2023-11-30 | 2023-12-29 | 中船(北京)智能装备科技有限公司 | 一种目标对象位置信息的确定方法、装置及设备 |
Families Citing this family (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111931643A (zh) * | 2020-08-08 | 2020-11-13 | 商汤集团有限公司 | 一种目标检测方法、装置、电子设备及存储介质 |
CN112926395A (zh) * | 2021-01-27 | 2021-06-08 | 上海商汤临港智能科技有限公司 | 目标检测方法、装置、计算机设备及存储介质 |
CN112907757A (zh) * | 2021-04-08 | 2021-06-04 | 深圳市慧鲤科技有限公司 | 一种导航提示方法、装置、电子设备及存储介质 |
KR102591835B1 (ko) * | 2021-08-13 | 2023-10-24 | 한국전자통신연구원 | 딥러닝 기반 의상 속성 분류 장치 및 방법 |
CN115035492B (zh) * | 2022-06-21 | 2024-01-23 | 苏州浪潮智能科技有限公司 | 车辆识别方法、装置、设备和存储介质 |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20120219183A1 (en) * | 2011-02-24 | 2012-08-30 | Daishi Mori | 3D Object Detecting Apparatus and 3D Object Detecting Method |
CN106875444A (zh) * | 2017-01-19 | 2017-06-20 | 浙江大华技术股份有限公司 | 一种目标物定位方法及装置 |
CN111274943A (zh) * | 2020-01-19 | 2020-06-12 | 深圳市商汤科技有限公司 | 一种检测方法、装置、电子设备及存储介质 |
CN111382613A (zh) * | 2018-12-28 | 2020-07-07 | 中国移动通信集团辽宁有限公司 | 图像处理方法、装置、设备和介质 |
CN111931643A (zh) * | 2020-08-08 | 2020-11-13 | 商汤集团有限公司 | 一种目标检测方法、装置、电子设备及存储介质 |
Family Cites Families (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10304191B1 (en) * | 2016-10-11 | 2019-05-28 | Zoox, Inc. | Three dimensional bounding box estimation from two dimensional images |
CN110826357B (zh) * | 2018-08-07 | 2022-07-26 | 北京市商汤科技开发有限公司 | 对象三维检测及智能驾驶控制的方法、装置、介质及设备 |
CN109671102B (zh) * | 2018-12-03 | 2021-02-05 | 华中科技大学 | 一种基于深度特征融合卷积神经网络的综合式目标跟踪方法 |
CN109784194B (zh) * | 2018-12-20 | 2021-11-23 | 北京图森智途科技有限公司 | 目标检测网络构建方法和训练方法、目标检测方法 |
CN109961522B (zh) * | 2019-04-02 | 2023-05-05 | 阿波罗智联(北京)科技有限公司 | 图像投射方法、装置、设备和存储介质 |
CN110427797B (zh) * | 2019-05-28 | 2023-09-15 | 东南大学 | 一种基于几何条件限制的三维车辆检测方法 |
CN110689008A (zh) * | 2019-09-17 | 2020-01-14 | 大连理工大学 | 一种面向单目图像的基于三维重建的三维物体检测方法 |
-
2020
- 2020-08-08 CN CN202010792241.XA patent/CN111931643A/zh active Pending
-
2021
- 2021-04-27 KR KR1020217042833A patent/KR20220024193A/ko unknown
- 2021-04-27 WO PCT/CN2021/090359 patent/WO2022033076A1/zh active Application Filing
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20120219183A1 (en) * | 2011-02-24 | 2012-08-30 | Daishi Mori | 3D Object Detecting Apparatus and 3D Object Detecting Method |
CN106875444A (zh) * | 2017-01-19 | 2017-06-20 | 浙江大华技术股份有限公司 | 一种目标物定位方法及装置 |
CN111382613A (zh) * | 2018-12-28 | 2020-07-07 | 中国移动通信集团辽宁有限公司 | 图像处理方法、装置、设备和介质 |
CN111274943A (zh) * | 2020-01-19 | 2020-06-12 | 深圳市商汤科技有限公司 | 一种检测方法、装置、电子设备及存储介质 |
CN111931643A (zh) * | 2020-08-08 | 2020-11-13 | 商汤集团有限公司 | 一种目标检测方法、装置、电子设备及存储介质 |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114655207A (zh) * | 2022-05-13 | 2022-06-24 | 中汽创智科技有限公司 | 一种数据处理方法、装置、设备及存储介质 |
CN115100423A (zh) * | 2022-06-17 | 2022-09-23 | 四川省寰宇众恒科技有限公司 | 一种基于视图采集数据实现实时定位系统及方法 |
CN115100423B (zh) * | 2022-06-17 | 2023-10-10 | 四川省寰宇众恒科技有限公司 | 一种基于视图采集数据实现实时定位系统及方法 |
CN115115687A (zh) * | 2022-06-24 | 2022-09-27 | 合众新能源汽车有限公司 | 车道线测量方法及装置 |
CN117308967A (zh) * | 2023-11-30 | 2023-12-29 | 中船(北京)智能装备科技有限公司 | 一种目标对象位置信息的确定方法、装置及设备 |
CN117308967B (zh) * | 2023-11-30 | 2024-02-02 | 中船(北京)智能装备科技有限公司 | 一种目标对象位置信息的确定方法、装置及设备 |
Also Published As
Publication number | Publication date |
---|---|
KR20220024193A (ko) | 2022-03-03 |
CN111931643A (zh) | 2020-11-13 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2022033076A1 (zh) | 目标检测方法、装置、设备、存储介质及程序产品 | |
CN107742311B (zh) | 一种视觉定位的方法及装置 | |
CN111328396B (zh) | 用于图像中的对象的姿态估计和模型检索 | |
CN112287860B (zh) | 物体识别模型的训练方法及装置、物体识别方法及系统 | |
CN109903331B (zh) | 一种基于rgb-d相机的卷积神经网络目标检测方法 | |
CN109145928B (zh) | 一种基于图像的车头朝向识别方法及装置 | |
CN111080693A (zh) | 一种基于YOLOv3的机器人自主分类抓取方法 | |
CN111340797A (zh) | 一种激光雷达与双目相机数据融合检测方法及系统 | |
CN112528878A (zh) | 检测车道线的方法、装置、终端设备及可读存储介质 | |
CN111553949B (zh) | 基于单帧rgb-d图像深度学习对不规则工件的定位抓取方法 | |
CN106971185B (zh) | 一种基于全卷积网络的车牌定位方法及装置 | |
CN110751097B (zh) | 一种半监督的三维点云手势关键点检测方法 | |
CN112287859A (zh) | 物体识别方法、装置和系统,计算机可读存储介质 | |
CN111144349A (zh) | 一种室内视觉重定位方法及系统 | |
CN111382658B (zh) | 一种基于图像灰度梯度一致性的自然环境下道路交通标志检测方法 | |
CN112395962A (zh) | 数据增广方法及装置、物体识别方法及系统 | |
CN114219855A (zh) | 点云法向量的估计方法、装置、计算机设备和存储介质 | |
CN115272691A (zh) | 一种钢筋绑扎状态检测模型的训练方法、识别方法及设备 | |
JP7336653B2 (ja) | ディープラーニングを利用した屋内位置測位方法 | |
CN113658274B (zh) | 用于灵长类动物种群行为分析的个体间距自动计算方法 | |
CN116051736A (zh) | 一种三维重建方法、装置、边缘设备和存储介质 | |
CN116052120A (zh) | 基于图像增强和多传感器融合的挖掘机夜间物体检测方法 | |
CN115240150A (zh) | 基于单目相机的车道偏离预警方法、系统、设备及介质 | |
CN117576494A (zh) | 特征地图生成方法、装置、存储介质和计算机设备 | |
CN113052118A (zh) | 基于高速快球摄像机实现场景变换视频分析检测的方法、系统、装置、处理器及存储介质 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
ENP | Entry into the national phase |
Ref document number: 2021565778 Country of ref document: JP Kind code of ref document: A |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 21855132 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: JP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 11.07.2023) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 21855132 Country of ref document: EP Kind code of ref document: A1 |