WO2022166309A1 - Method and apparatus for processing image data of image sensor - Google Patents

Method and apparatus for processing image data of image sensor Download PDF

Info

Publication number
WO2022166309A1
WO2022166309A1 PCT/CN2021/131698 CN2021131698W WO2022166309A1 WO 2022166309 A1 WO2022166309 A1 WO 2022166309A1 CN 2021131698 W CN2021131698 W CN 2021131698W WO 2022166309 A1 WO2022166309 A1 WO 2022166309A1
Authority
WO
WIPO (PCT)
Prior art keywords
image data
image
processing
sub
image sensor
Prior art date
Application number
PCT/CN2021/131698
Other languages
French (fr)
Chinese (zh)
Inventor
李文斌
段小祥
Original Assignee
华为技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 华为技术有限公司 filed Critical 华为技术有限公司
Publication of WO2022166309A1 publication Critical patent/WO2022166309A1/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/70Circuitry for compensating brightness variation in the scene
    • H04N23/73Circuitry for compensating brightness variation in the scene by influencing the exposure time
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/70Circuitry for compensating brightness variation in the scene

Definitions

  • the present application relates to the field of intelligent networked vehicles, and in particular, to a method and device for processing image sensor image data.
  • Perceptual recognition plays an important role in the advanced driving assistance system (ADAS) and autonomous driving system (Autonomous Driving) of intelligent networked vehicles.
  • ADAS advanced driving assistance system
  • Autonomous Driving autonomous driving system
  • the intelligent networked vehicle is equipped with a variety of image sensors.
  • image sensors Common examples of in-vehicle image sensors for intelligent perception recognition are cameras (camera), LiDAR (Light Detection and Ranging, LiDAR), millimeter wave radar (millimeter wave), etc.
  • the "image data" obtained by them is rich in information, so Intelligent connected vehicles can realize the recognition function through these image data.
  • the image data reaches the advanced driving assistance system or the autonomous driving system from the image sensor, and after the steps of image processing, algorithm recognition, and driving decision-making, it is finally transformed into the operation control instructions for the various actuators of the vehicle.
  • Driving controls In this process, from the perspective of safety, reducing the end-to-end delay from the generation of image data by image sensors to the execution of operations by vehicle actuators is the goal of advanced driver assistance systems and autonomous driving systems.
  • Image sensors include many image sensors that sense "scanning", such as lidars and cameras. Such a scanning image sensor does not generate all the image data of a frame of images simultaneously in the acquisition area, but sequentially generates all the image data of a frame of images over a period of time, for example by scanning in rows or columns.
  • the vehicle-mounted Image Signal Processor (ISP) and subsequent algorithm platforms process the frame of image after waiting for all the image data of the frame to be received. This processing method has a great impact on the end-to-end delay.
  • ISP Image Signal Processor
  • the present application provides a method and apparatus for processing image sensor image data, which can reduce the end-to-end delay from the image sensor generating the image data to the vehicle actuator performing the operation.
  • a method for processing image data of an image sensor comprising the steps of: receiving first image data from an image sensor, the first image data being that the image sensor scans an acquisition area within one scan period One image data among a plurality of image data that can be generated by the corresponding physical area, the collection area represents the collection range of the image sensor; image processing is performed on the first image data to obtain second image data; and the first image data is output. 2. Image data.
  • the first image data generated by the image sensor can be processed in advance.
  • An image processing flow of image data thereby reducing the end-to-end delay from the generation of image data by an image sensor to the execution of operations by vehicle actuators.
  • image processing can be performed on the part of the first image data without waiting for all the first image data of the entire frame of image to be received.
  • the image processing flow of the image data is started only when an image data is generated.
  • the image sensor is a camera, and the image is a two-dimensional plane image.
  • the acquisition area of the camera is often referred to as the target surface.
  • the frame rate of the camera is generally 30Hz (Hertz, Hertz).
  • the exposure time of the camera from the first line of the target surface to the last line of the target surface to form one frame of image is 33ms (milisecond, millisecond).
  • the processing of the first image data can be performed immediately, without waiting for all the images of the whole frame to be received.
  • first image data This can advance the image processing process of the first image data of the plane image, save the processing time of the first image data of each frame of plane image, thereby reducing the end-to-end time from the generation of the first image data by the camera to the operation of the vehicle actuator. extension.
  • the image sensor is a lidar
  • the image is a three-dimensional point cloud.
  • the acquisition area of the lidar is often referred to as the scan area.
  • the frame rate of lidar is generally 10Hz or 20Hz. That is to say, the lidar scans from the first column of the scanning area to the last column of the scanning area, so that the duration of forming the first image data of a frame of point cloud is usually 100ms or 50ms. Similar to processing the image data of a plane image, it is usually necessary to wait for the first image data of the whole frame of point cloud to be generated and transmitted before processing the received first image data of the whole frame of point cloud. However, with the method for processing image data of the present application, for example, after receiving the first first image data of a frame of point cloud, the first image data can be processed immediately without waiting for the whole frame of points to be received. All first image data of the cloud.
  • the collection area includes a plurality of sub-collection areas; the performing image processing on the first image data includes: when the first data set A included in the first data set A is received After one image data is obtained, image processing is performed in units of all the first image data included in the first image data group A, where the first image data group A is the physical image corresponding to the image sensor scanning one of the sub-collection areas. The set of the first image data generated by the region.
  • each first image data group By defining each first image data group in each sub-collection area of the collection area, the plurality of first image data of one frame of image are sorted out in advance. After receiving the first image data included in the first data group A, image processing is performed using all the first image data included in the first image data group A as a unit, that is, by using the first image data group as a unit to perform image processing processing, the image processing flow of the first image data can be advanced, and at the same time, the complexity of the subsequent processing and the increase of the subsequent processing time caused thereby can be avoided.
  • each first image data set is defined by the sub-acquisition regions, which can be applied to image sensors of different resolutions.
  • the size and quantity of the plurality of sub-collection regions are preset.
  • the number of sub-collection regions is selected from 2 to 4.
  • each sub-collection area is equal.
  • the method further includes the following steps: receiving a division strategy, and presetting the size and quantity of the plurality of sub-collection areas according to the division strategy.
  • the process of image processing of multiple first image data can be flexibly adjusted according to the actual needs of the application scenario, so that the method for processing image sensor image data of the present application Adapt to various application scenarios.
  • the sub-collection area is a rectangle, and the size of the sub-collection area is defined by coordinates of four corners of the rectangle.
  • each sub-collection area can be flexibly divided in a particularly simple and intuitive manner.
  • a method for processing image data of an image sensor comprising the steps of: sequentially receiving second image data, the second image data is obtained by performing image processing on the first image data, the first The image data is one image data among a plurality of image data that can be generated by the image sensor scanning the physical area corresponding to the acquisition area in one scan period, and the acquisition area represents the acquisition range of the image sensor; Extract feature data from image data; perform fusion recognition processing on each feature data; and output image recognition results.
  • the feature data of the second image data originating from different parts of the acquisition area are respectively extracted, and the fusion recognition is performed in the fusion recognition process.
  • Each feature data extracted from all the second image data can ensure the stability and accuracy of the image recognition result of the fusion recognition.
  • the image sensor is a camera, and the image is a two-dimensional plane image.
  • the image sensor is a lidar, and the image is a three-dimensional point cloud.
  • the second image data is grouped; the respective feature data is extracted from each group of the second image data.
  • the second image data By grouping the second image data, a plurality of second image data of one frame of image are sorted out in advance.
  • the image recognition process of the image data can be advanced and the subsequent processing can be avoided excessively. complexity and consequent increase in subsequent processing time.
  • the quantity of the second image data of each group of the second image data is preset.
  • the collection area includes a plurality of sub-collection areas
  • the second image data is grouped according to the plurality of sub-collection areas
  • a group of second image data is image-processed
  • a set of first image data is a set of the first image data generated by the image sensor scanning a physical area corresponding to one of the sub-acquisition areas.
  • the method further includes the following steps: receiving a division strategy, and presetting the size and number of the plurality of sub-collection areas according to the received division strategy.
  • the size and number of multiple sub-collection areas are pre-set by receiving the division strategy, so as to adjust the grouping of image data, and the process of extracting feature data from multiple image data can be flexibly adjusted according to the actual needs of the application scenario, so that the present application
  • the method of processing image sensor image data is suitable for various application scenarios.
  • the sub-collection area is a rectangle, and the size of the sub-collection area is defined by coordinates of four corners of the rectangle.
  • each sub-collection area can be flexibly divided in a particularly simple and intuitive manner.
  • the extracting feature data includes convolution processing and pooling (pooling, also known as pooling) processing.
  • the convolution processing and the pooling processing are alternately performed more than once.
  • the convolution process includes one or more sub-convolution processes.
  • the number of sub-convolution processes in the convolution process is a natural number selected from 1 to 3.
  • the fusion identification includes feature fusion processing.
  • each feature data extracted from the second image data originating from each acquisition area can be effectively fused into the form of feature data of the entire frame image.
  • the fusion identification further includes full connection processing.
  • the feature fusion process includes a series of feature fusion (concat) processes.
  • the full connection processing includes one or more sub-full connection processing.
  • an image data processing device comprising: a receiving module configured to receive first image data from an image sensor, where the first image data is corresponding to a scanning and acquisition area of the image sensor in one scanning period One image data among a plurality of image data that can be generated by the physical area, the collection area represents the collection range of the image sensor; and an image processing module, configured to perform image processing on the first image data to obtain second image data, and for outputting the second image data.
  • the collection area includes a plurality of sub-collection areas; the performing image processing on the first image data includes: the image processing module is further configured to, when the receiving After the module receives the first image data contained in the first data set A, it performs image processing in units of all the first image data contained in the first image data set A, where the first image data set A is the image The set of the first image data generated by the sensor scanning a physical area corresponding to the sub-collection area.
  • the size and quantity of the plurality of sub-collection regions are preset.
  • the sub-collection area is a rectangle, and the size of the sub-collection area is defined by coordinates of four corners of the rectangle.
  • the receiving module is further configured to receive a division strategy
  • the image data processing module is further configured to preset the number of sub-collection regions according to the division strategy. size and quantity.
  • the device of the third aspect of the present application can execute the method of the first aspect, the advantages and benefits of the device of the third aspect are similar to those of the first aspect, and the relevant descriptions of the first aspect are referred to and will not be repeated here.
  • an image recognition device comprising: a receiving module configured to sequentially receive second image data, wherein the second image data is obtained by performing image processing on the first image data, and the first image data is The image sensor scans one image data among multiple image data that can be generated by scanning the physical area corresponding to the acquisition area in one scan period, and the acquisition area represents the acquisition range of the image sensor; the feature extraction module is used to sequentially select from all the image data. feature data is extracted from the second image data; and a fusion recognition module is used to perform fusion recognition processing on each feature data and output image recognition results.
  • the quantity of the second image data of each group of the second image data is preset.
  • the collection area includes a plurality of sub-collection areas
  • the second image data is grouped according to the plurality of sub-collection areas
  • a group of second image data is image-processed
  • a set of first image data is a set of the first image data generated by the image sensor scanning a physical area corresponding to one of the sub-acquisition areas.
  • the size and quantity of the plurality of sub-collection regions are preset.
  • the receiving module is further configured to receive a division strategy
  • the feature extraction module is further configured to preset the size and quantity of the plurality of sub-collection regions according to the division strategy.
  • the sub-collection area is a rectangle
  • the size of the second sub-collection area is defined by coordinates of four corners of the rectangle.
  • the feature extraction module includes a convolution layer and a pooling layer.
  • the convolutional layer includes one or more sub-convolutional layers.
  • the number of sub-convolutional layers in the convolutional layer is a natural number selected from 1 to 3.
  • the fusion identification module includes a feature fusion layer.
  • the fusion identification module further includes a fully connected layer.
  • the feature fusion processing module includes a series of feature fusion (concat) layers.
  • the fully-connected layer includes one or more sub-fully-connected layers.
  • the receiving module is further configured to receive a division strategy
  • the feature extraction module is further configured to preset the plurality of second sub-collections according to the division strategy The size and number of regions.
  • the apparatus of the fourth aspect of the present application can perform the method of the second aspect, the advantages and benefits of the apparatus of the fourth aspect are similar to those of the second aspect, and the relevant descriptions of the second aspect are referred to, and are not repeated here.
  • an image sensor image data processing system including any image data processing apparatus as in the third aspect and any image recognition apparatus as in the fourth aspect.
  • the image sensor image data processing system further includes a division management module, configured to provide a division strategy to the image data processing apparatus and the image recognition apparatus, the division The strategy is used to preset the size and quantity of the plurality of sub-collection areas of the collection area.
  • system of the fifth aspect includes the apparatus of the third aspect and the fourth aspect
  • advantages and benefits of the system of the fifth aspect will include the advantages and benefits of the third aspect and the fourth aspect, with reference to the third aspect and the fourth aspect. The description will not be repeated here.
  • a driving system which includes any image sensor image data processing system and a driving decision-making unit in the fifth aspect; wherein the driving decision-making unit is connected to the image sensor image data processing system for The image recognition result output by the image sensor image data processing system executes behavioral decision-making and motion planning and outputs operation instructions.
  • the driving system of the present application can process the process of image data in advance, save processing time, thereby reducing the end-to-end time from the image sensor generating the image data to the vehicle actuator performing the operation end delay.
  • the driving system is an advanced driving assistance system. In another possible implementation, the driving system is an autonomous driving system.
  • a vehicle which includes an image sensor connected in sequence, any one of the driving systems of the sixth aspect, an electronic control unit, and an actuator; wherein the image sensor is used to perceive the vehicle environment in a scanning manner and outputting first image data; the electronic control unit is configured to control the actuator to perform operations according to the operating instructions of the driving system.
  • the vehicle of the present application can process the process in advance, save processing time, and thus reduce the end-to-end delay from the image sensor generating the image data to the vehicle actuator performing the operation.
  • a computing device comprising: at least one processor; and at least one memory connected in connection with the processing and storing program instructions, the program instructions when executed by the at least one processor
  • the at least one processor is caused to perform the method of any one of the first and third aspects.
  • the processor in the computing device of the present application can execute any one of the above-mentioned methods for processing image sensor image data in the first and second aspects, the advantages and benefits of the computing device are also similar to the first and second aspects For the advantages and benefits, refer to the relevant descriptions of the first aspect and the second aspect, which will not be repeated here.
  • a computer-readable storage medium having program instructions stored thereon, the program instructions, when executed by a computer, cause the computer to perform any one of the first and second aspects above to process image sensor images method of data.
  • the computer-readable storage medium of the present application can enable a computer to perform any one of the above-mentioned methods for processing image sensor image data in the first aspect and the second aspect
  • the advantages and benefits of the computer-readable storage medium are also similar to those of the first aspect
  • Fig. 1 is the schematic diagram of the image data processing scheme in the prior art
  • FIG. 2 is a schematic diagram of an image data processing solution of the prior art and an embodiment of the present application, wherein the solution of the prior art is located at the upper part, and the solution of the embodiment of the present application is located at the lower part;
  • FIG. 3 is a schematic structural diagram of an image sensor image data processing system according to an embodiment of the present application.
  • FIG. 4 is a schematic structural diagram of an image data processing apparatus according to an embodiment of the present application.
  • FIG. 5 is a schematic structural diagram of an image recognition apparatus according to an embodiment of the present application.
  • FIG. 6 is a schematic flowchart of a method for processing image data according to an embodiment of the present application.
  • FIG. 7 is a schematic diagram of dividing a collection area by a dividing strategy in an embodiment of the present application.
  • FIG. 8 is a schematic diagram of the division of each sub-collection area of the image sensor collection area in FIG. 7;
  • FIG. 9 is a schematic flowchart of a feature extraction and fusion identification process in a method for processing image sensor image data according to an embodiment of the present application.
  • FIG. 10 is a schematic diagram of an image data processing solution of the prior art and another embodiment of the present application, wherein the solution of the prior art is located at the upper part, and the solution of the embodiment of the present application is located at the lower part;
  • FIG. 11 is a schematic diagram of an image data processing solution of the prior art and another embodiment of the present application, wherein the solution of the prior art is located at the upper part, and the solution of the embodiment of the present application is located at the lower part;
  • FIG. 12 is a schematic flowchart of a method for processing image sensor image data according to another embodiment of the present application.
  • FIG. 13 is a schematic structural diagram of a driving system according to an embodiment of the present application.
  • FIG. 14 is a schematic structural diagram of a vehicle according to an embodiment of the present application.
  • FIG. 15 is a schematic structural diagram of a computing device according to an embodiment of the present application.
  • first, second, third or “region A, region B, region C” and other similar terms in the description and claims are only used to distinguish similar objects, and do not indicate a specific ordering of objects, and may It is understood that the specific order or sequence may be interchanged, where permitted, to enable the embodiments of the application described herein to be practiced in sequences other than those illustrated or described herein.
  • first image data and second image data used in this application are both image data, but the “second image data” is obtained by performing image processing on the "first image data”, so the term “first image data” is used.
  • One" and “Second” distinguish the two.
  • imaging sensor as used in the specification and claims includes cameras and lidars.
  • the image sensor is used to scan a plurality of image data that can be generated from the physical area corresponding to the acquisition area in one scan cycle.
  • the acquisition area represents the acquisition range of the image sensor, and the acquisition area of the camera is also called the target surface.
  • the collection area of the lidar is also called the scanning area.
  • image processing refers to techniques of analyzing images with computing devices to achieve desired results.
  • the image processing applicable to the image data of different types of images is different.
  • image processing of image data for two-dimensional planar images may include, but is not limited to, black level compensation, lens shading correction, bad pixel correction, demosaic ), Bayesian domain denoising, white balance correction (automatic white balance), color correction (color correction), Gamma correction and color space conversion (RGB to YUV).
  • Image processing for the image data of the three-dimensional point cloud may include, but is not limited to, filtering, down sampling, and outlier removal. Therefore, the "image processing" involved in the method embodiments of the present application may include one or more sub-image processing, and the "image processing module" involved in the apparatus embodiments of the present application may include one or more sub-image processing modules.
  • feature extraction or “extraction of feature data” used in the description and claims refers to constructing a variety of informative but not redundant techniques for the remaining feature data.
  • the feature extraction applicable to the image data of different types of images is different.
  • feature extraction for image data of two-dimensional planar images may include, but is not limited to, convolution and pooling.
  • Feature extraction for image data of three-dimensional point clouds may include, but is not limited to, feature vector extraction.
  • fusion and recognition used in the description and claims refers to a technology that, after all feature data extracted from image data is fused into a whole, performs recognition analysis based on the whole and outputs an image recognition result.
  • the fusion recognition applicable to the image data of different types of images is different.
  • fusion recognition of image data for two-dimensional planar images may include, but is not limited to, feature fusion, fully connected, and output.
  • Fusion identification of image data for 3D point clouds may include, but is not limited to, feature point matching.
  • feature fusion used in the specification and claims is a process of fusing all feature data extracted from image data of the entire acquisition area into one.
  • the feature fusion involved in this application is an early fusion (early fusion) performed before the image recognition result is obtained, which may include but not limited to serial feature fusion (concat) and parallel fusion (add).
  • the recognition algorithm processing will be performed on the second image data (that is, the first image data after image processing) until time t3. Complete the recognition algorithm processing of the whole frame image.
  • the inventors of the present application found that such a method of processing image sensor image data has the following drawbacks.
  • the image processing process must wait until the image sensor scans the entire acquisition area to generate all the first image data and transfer them, and the image recognition process must wait until all the first image data has been image-processed and transformed into the second image data. start.
  • Each of these waiting times delays the processing flow of the image sensor image data, resulting in a large end-to-end delay from the generation of the first image data by the image sensor to the execution of the operation by the vehicle actuator.
  • the image processing and the recognition algorithm processing are performed in units of the first or second image data of the whole frame of image, it is impossible to process the received first or second image data while receiving the first or second image data of the frame.
  • the first or second image data is processed, which reduces the speed of processing the image sensor image data and increases the end-to-end latency.
  • the inventors of the present application found that such a method of processing image sensor image data has the following drawbacks.
  • the image processing process must wait until the image sensor scans the entire acquisition area to generate all the first image data and transfer them, and the image recognition process must wait until all the first image data has been image-processed and transformed into the second image data. start.
  • Each of these waiting times delays the processing flow of the image sensor image data, resulting in a large end-to-end delay from the generation of the first image data by the image sensor to the execution of the operation by the vehicle actuator.
  • the image processing and the recognition algorithm processing are performed in units of the first or second image data of the whole frame of image, it is impossible to process the received first or second image data while receiving the first or second image data of the frame.
  • the first or second image data is processed, which reduces the speed of processing the image sensor image data and increases the end-to-end latency.
  • an embodiment of the present application provides a method for processing image data of an image sensor, including: receiving first image data from an image sensor, where the first image data is scanned by the image sensor within one scan period One image data among a plurality of image data that can be generated by the physical area corresponding to the acquisition area, the acquisition area represents the acquisition range of the image sensor; image processing is performed on the first image data to obtain second image data; the second image data.
  • FIG. 2 shows the existing solution (located in the upper part of FIG. 2 ) and the solution of the present application (located in the lower part of FIG. 2 ) in the same time axis.
  • the image sensor scans its acquisition area and sequentially generates a plurality of image data.
  • the scanning area of the image sensor is pre-divided into three sub-acquisition areas A, B and C, when all the image data generated by the scanning sub-acquisition area A is received (ie at time t11 ), that is, when receiving
  • the processing method of the present application starts to perform image processing before all image data generated by scanning the entire acquisition area (ie, at time t1).
  • the method for processing image sensor image data of the present application can be applied to processing image data generated by various "scanning" image sensors, and can be applied to vehicles, robots and other devices having such image sensors.
  • the robot uses the camera to perceive the environment and plan and execute the corresponding movement according to the image recognition results of the camera's image data.
  • an image recognition result can be obtained more quickly, so that a corresponding movement can be made in response to the result more quickly. Therefore, using the method for processing image sensor image data of the present application can make the motion of the motion robot more agile.
  • the vehicle uses lidar to perceive road conditions and plan and execute corresponding autonomous driving operations based on the image recognition results of lidar's point cloud image data.
  • the image recognition result can be obtained more quickly, so that the corresponding driving operation can be performed in response to the result more quickly. Therefore, using the method for processing image sensor image data of the present application can make the automatic driving of the intelligent networked vehicle safer.
  • the method for processing image sensor image data of the present application is to improve the solution for processing one frame of image, the time for processing one frame of image is shortened, thereby shortening the end-to-end delay. It should be understood that the method of processing image sensor image data of the present application is also applicable to processing video including multiple frames of images. By using the method for processing image sensor image data of the present application for each frame of images in the video, the processing time of the video can be shortened.
  • FIG. 3 optional modules of the device are represented by dashed boxes, that is, the optional modules may be omitted in other embodiments.
  • FIG. 3 exemplarily shows an image sensor image data processing system 1001 according to an embodiment of the present application, including an image data processing apparatus 1100 and an image recognition apparatus 1200 that are connected to each other.
  • Figure 3 shows an optional partition management module 1300 in a dashed box.
  • the division management module 1300 is respectively connected with the image data processing apparatus 1100 and the image recognition apparatus 1200, and is used for providing division policies to both.
  • the division strategy may be used to preset the size and quantity of the multiple sub-capturing regions of the image sensor's capturing region.
  • the division strategy provided by the division management module 1300 enables the collection area to be divided into three sub-collection areas A, B, and C equally.
  • FIG. 3 additionally shows image sensor 2000 .
  • the image sensor 2000 scans the acquisition area to sequentially generate multiple first image data constituting one frame of image, and sequentially sends the multiple first image data to the image data processing device 1100 in the image sensor image data transmission processing system 1001 of the present application.
  • the image data processing apparatus 1100 includes a receiving module 1110 and an image processing module 1120 which are connected to each other.
  • the receiving module 1110 of the image data processing apparatus 1100 is configured to receive a plurality of first image data sent by the image sensor 2000 , and may also be configured to receive a division strategy provided by the division management module 1300 .
  • the image processing module 1120 pre-divides the acquisition area into three sub-acquisition areas A, B and C of equal size according to the division strategy received by the receiving module 1110, the first image received by the receiving module 1110
  • the data can be divided into 3 groups of first image data, so that when the image sensor scans the 3 different sub-acquisition regions, a group of first image data is respectively generated.
  • the image processing module 1120 can thus be further configured to perform image processing on a set of first image data when the set of first image data has been received.
  • the image processing module 1120 is configured to start performing image processing on the image data when the receiving module 1110 receives all the first image data generated from the sub-collection area A at time t11. Then, at time t12, when the receiving module 1110 receives all the first image data generated from the sub-collection area B, the image processing module 1120 starts to perform image processing on these image data. Finally, at time t13, when the receiving module 1110 receives all the first image data generated from the sub-collection area C, the image processing module 1120 starts to perform image processing on these image data.
  • the image processing module 1120 may include a plurality of image processing sub-modules (not shown). Similarly, image processing may include multiple image sub-processing.
  • the image data processing apparatus 1100 may be, for example, an image signal processor.
  • the image recognition apparatus 1200 includes a receiving module 1210 , a feature extraction module 1220 and a fusion recognition module 1230 which are connected in sequence.
  • the receiving module 1210 of the image recognition apparatus 1200 is configured to receive a plurality of second image data.
  • the plurality of second image data are obtained by performing image processing on the plurality of first image data by the image data processing apparatus 1100, and the plurality of first image data are sequentially generated by the image sensor 3000 scanning the acquisition area.
  • the receiving module 1210 is further configured to receive the partition policy provided by the partition management module 1300 .
  • the feature extraction module 1220 is configured to sequentially extract feature data from the received second image data during the period when the receiving module receives the plurality of second image data. Since the feature extraction module 1220 groups the second image data according to the division strategy received by the receiving module 1210, that is, the acquisition area is pre-divided into three sub-acquisition areas A, B, and C of equal size, so the second image data received by the receiving module 1210 The image data can also be similarly divided into 3 different sets of second image data, so that each second image data originating from the sub-acquisition areas A, B, C belong to different sets of second image data, respectively.
  • the feature extraction module 1220 is configured to extract feature data from a set of second image data when the set of second image data has been received.
  • the feature extraction module 1220 starts to extract feature data from the set of second image data together.
  • the feature extraction module 1220 starts to extract the second image data from these second image data together. Extract feature data.
  • the feature extraction module 1220 performs feature extraction on the second image data originating from the sub-acquisition area B and the sub-acquisition area C, respectively, in a similar manner.
  • the fusion identification module 1230 is configured to first fuse all the characteristic data into a whole after the characteristic extraction module 1220 extracts characteristic data from all the second image data originating from the sub-collection areas A, B, and C, respectively, Then the whole is recognized, and finally the image recognition result is output at time t4.
  • both the feature extraction module 1220 and the fusion recognition module 1230 are involved in image recognition processing, in practice, they are usually set in the same entity, for example, in the algorithm platform of the vehicle.
  • the image processing module 1120 and the feature extraction module 1220 respectively process the first image data generated from a sub-acquisition area or the second image data from a sub-acquisition area, that is, their processing objects are from First or second image data of a sub-acquisition area.
  • the fusion recognition module 1230 of this embodiment is configured to perform fusion recognition from all the feature data of the second image data originating from the entire collection area as a whole, that is, the processing object is all the second image data originating from the whole collection area.
  • Feature data for image data is configured to perform fusion recognition from all the feature data of the second image data originating from the entire collection area as a whole, that is, the processing object is all the second image data originating from the whole collection area.
  • the feature extraction module 1220 may include one or more convolutional layers (not shown) and one or more pooling layers (not shown).
  • the fusion recognition module 1230 may include one or more feature fusion layers (not shown), one or more fully connected layers (not shown), and an output layer (not shown). These layers may also each include one or more sublayers (not shown).
  • FIG. 6 exemplarily shows a schematic flowchart of a method for processing image data according to an embodiment of the present application, which includes the following steps S101 to S107:
  • step S101 the division management module 1300 provides a division policy. Specifically, the division strategy is used to preset the number of sub-collection areas of the collection area to 3 and to preset the size of each sub-collection area to be equal to each other.
  • the image data processing apparatus 1100 and the image recognition apparatus 1200 respectively receive the division strategy, and pre-set each sub-collection area of the collection area according to the division policy, thereby grouping the first image data and the second image data respectively.
  • the image data processing device 1100 and the receiving module 1110 of the image recognition device 1200 and the receiving module 1210 of the image recognition device 1200 receive the division strategy respectively, and the corresponding image processing module 1120 and the feature extraction module 1220 divide the collection area according to the division strategy.
  • a first sub-collection area and a second sub-collection area are formed, and the first and second sub-collection areas are preset as sub-collection areas A, B, and C of equal size to each other.
  • the second sub-acquisition areas of the image data processing apparatus 1100 and the recognition apparatus 1200 are all sub-acquisition areas A, B and C.
  • step S103 the image data processing apparatus 1100 receives a plurality of first image data sequentially generated by the image sensor scanning the acquisition area. Specifically, the receiving module 1110 of the image data processing apparatus 1100 receives a plurality of first image data sequentially generated and sent by the image sensor 3000 by scanning the acquisition area.
  • step S104 when the image data processing apparatus 1100 has received a set of first image data generated from a first sub-collection area, it performs image processing on the set of first image data, and outputs corresponding second image data .
  • the image processing module 1120 when the receiving module 1110 receives a set of first image data generated from a first sub-collection area, the image processing module 1120 performs image processing on the set of first image data, and sequentially outputs the set of first images The corresponding second image data obtained by performing the image processing.
  • step S105 the image recognition apparatus 1200 receives a plurality of second image data sequentially output by the image data processing apparatus 1100.
  • the receiving module 1210 of the image recognition apparatus 1200 receives a plurality of second image data sequentially output by the image data processing apparatus 1100 .
  • step S106 the image recognition apparatus 1200 extracts feature data from a set of second image data originating from a second sub-capture area when the image recognition apparatus 1200 has received the set of second image data. Specifically, when the receiving module 1210 receives all the second image data originating from a second sub-collection area, the feature extraction module 1220 of the image recognition apparatus 1200 extracts feature data from the image data together.
  • step S107 the image recognition device 1200 performs fusion recognition processing on the feature data extracted from the plurality of second image data, and outputs an image recognition result. Specifically, after the feature extraction module 1220 of the image recognition device 1200 completes the extraction of feature data from the second image data originating from the sub-collection areas A, B, and C, respectively, the fusion recognition module 1230 of the image recognition device 1200 fuses all the feature data As a whole, then identify the whole, and finally output the image recognition result.
  • steps S101 to S107 are not arranged in chronological order.
  • the image processing module 1120 in step S104 may perform image processing on the first image data generated by the sub-collection area A deal with. Therefore, step S103 and step S104 may overlap in time. Similarly, step S105 and step S105 may coincide in time.
  • steps S104 and S106 are correspondingly repeated 3 times for processing respectively First image data from each sub-acquisition area and second image data from each sub-acquisition area are generated, as shown in FIG. 2 .
  • the execution object of the image processing step S104 and the feature extraction step S106 in this embodiment is the first or second image data generated from or originating from a sub-acquisition area of the acquisition area.
  • the execution object of the fusion identification step S107 is all feature data extracted from the second image data originating from the entire collection area.
  • equation (1) can be converted into:
  • t1 is the start time of performing image processing on the first image data generated from the entire collection area in the existing solution, and t1 is also the image processing time on the first image data generated from the sub-collection area C in this embodiment of the present application
  • t2 is the moment when the image processing ends and the algorithm processing starts in the existing scheme
  • t3 is the moment when the algorithm processing ends in the existing scheme
  • t4 is the moment when the fusion recognition ends in the present embodiment of the present application
  • T0 is the current
  • the duration of feature extraction that belongs to a part of algorithm processing in the scheme is also the duration of feature processing in this embodiment of the present application
  • T1 is the duration of fusion recognition that belongs to a part of algorithm processing in the existing scheme, which is also the duration of this embodiment of the present application The duration of fusion recognition in .
  • the delay benefit in this embodiment of the present application benefits from the advance of the image processing flow, that is, in this embodiment, the processing of one frame has already started before time t1. Image processing is performed on 2/3 of the first image data of the image, thereby reducing the end-to-end delay.
  • the delay benefit in this embodiment of the present application benefits from the advance of the image recognition process, that is, in this embodiment, 2/3 of the second image data of the frame image has been processed before time t1. feature extraction, thereby further reducing the end-to-end delay.
  • FIG. 7 exemplarily shows a schematic diagram of dividing the acquisition area of the image sensor 3000 using the division strategy provided by the division management module 1300 .
  • a frame of image is generated by the camera in its target surface (also referred to as an acquisition area in the present invention) by line exposure from top to bottom.
  • the collection area is divided into three rectangular sub-collection areas A, B, and C in the direction from top to bottom, and the sizes of each sub-collection area are set to be equal.
  • the image sensor scans the sub-acquisition area A first and thus generates the first image data of the sub-acquisition area A first, and then transmits the image data generated from the sub-acquisition area A to the image sensor image data processing system 1001 first.
  • first image data The image sensor image data processing system 1001 therefore firstly performs image processing and feature extraction on the image data generated from the sub-acquisition area A, then processes the first image data generated from the sub-acquisition area B, and finally processes the image data generated from the sub-acquisition area B.
  • the first image data of the sub-acquisition area C is processed.
  • the acquisition area of the camera can also be divided by the same division strategy for processing the second frame of images generated by subsequent scans, up to the Z-th frame image, where Z is any integer greater than 2. That is to say, the present application can process images of multiple frames, and thus can process objects composed of images of multiple frames, such as videos.
  • FIG. 8 is a schematic diagram of division of each sub-collection area of the image sensor collection area in FIG. 7 . It is assumed that the resolution of the acquisition area shown by FIG. 7 is 1920*1080.
  • the upper left corner of the collection area is set as the imaginary coordinate origin, and each sub-collection area A, B and C can be easily defined by the coordinates of the upper left corner, lower left corner, upper right corner and lower right corner of the rectangle. Please refer to Table 1 for the specific coordinates. and Figure 8.
  • each sub-collection area of the collection area can be limited in a simple and flexible manner, and it is convenient to adjust the number of the sub-collection areas and the size of each sub-collection area according to the actual situation.
  • FIG. 9 exemplarily shows a schematic diagram of a flow chart of feature extraction and fusion recognition in a method for processing image data according to an embodiment of the present application.
  • the parts involved in the feature extraction step S106 and the part involved in the fusion identification step S107 are respectively marked with different dashed boxes.
  • conv1 indicates the first convolution process that can be performed by the first convolution layer of the feature extraction module 1220
  • conv2 indicates the second convolution process that can be performed by the second convolution layer of the feature extraction module
  • conv3 indicates that the feature The third convolution process performed by the third convolution layer of the extraction module
  • pool1 to pool3 each indicate the first to third pooling processes that can be performed by the first to third pooling layers of the feature extraction module 1220, respectively
  • fc1 and fc2 each indicate the first and second fully connected processes that may be performed by the first and second fully connected layers of the fusion recognition module 1230, respectively
  • concat indicates a series of feature fusion processes that may be performed by the series feature fusion layer of the fusion recognition module 1230.
  • conv2_1 indicates the first sub-convolution process of the first convolution process that can be performed by the first sub-convolution layer of the first convolution layer
  • conv2_2 indicates the second sub-convolution process that can be performed by the first convolution layer.
  • the second subconvolution process of the first convolution process performed by the subconvolution layer.
  • conv3_1 to conv3_3 each indicate the first to third sub-convolution processes of the third convolution process that can be performed by the first to third sub-convolution layers of the third convolution layer, respectively.
  • step S106 For convenience of display, the processes involved in step S106 are not arranged in time sequence in FIG. 10 .
  • each convolution and pooling of the image data of the sub-collection area A located in the left column is first performed from top to bottom processing, and then perform each convolution and pooling processing of the image data of the sub-collection area B in the middle column from top to bottom, and finally perform each of the image data of the sub-collection area C in the right column from top to bottom.
  • Convolution and pooling In other words, the convolution and pooling of the second image data of each sub-acquisition area A, B and C are not performed simultaneously, but are sequentially performed from left to right in units of the second image data of a sub-collection area .
  • the fusion recognition module 1230 when the fusion recognition module 1230 performs the fusion recognition step S107, it sequentially performs a series of feature fusion processing, first full connection processing, second full connection processing and image recognition in order from top to bottom result output.
  • the image data processing apparatus 1100 takes a set of first image data generated from a sub-collection area as the processing object, and the image recognition apparatus 1200 uses all the second image data from the entire collection area as a processing object.
  • the method for image data processing according to this embodiment of the present application includes: during receiving a plurality of first image data generated from scanning the entire acquisition area, sequentially processing a group of images generated from a sub-acquisition area Image processing is performed on the first image data to output the corresponding second image data; when the image processing is completed on all the first image data generated from the three sub-collection areas and all the second image data are output, all the second image data are processed.
  • Feature extraction and fusion recognition are performed as a whole. Therefore, the delay benefit of this solution is only 2/3 of the image processing time, that is, this solution only benefits from the advance of image processing of the first image data generated from the sub-collection areas A and B.
  • the sizes of the various sub-collection regions may be defined to be unequal to each other.
  • the size of the sub-collection area D is set to be 1/2 of the size of the sub-collection area E.
  • the image data processing apparatus or the image recognition apparatus may perform image processing or feature extraction on the received first or second image data at a start time that is not the time at which the image data is received.
  • the image data processing device performs image processing on the first image data generated from the sub-collection area D at time t16, that is, after receiving the first image data generated from the sub-collection area D. Between time t14 when all the first image data of the sub-capture area D is acquired and time t15 when all the first image data of the sub-acquisition area E are received.
  • the start time when the image data processing apparatus performs image processing on all the first image data generated from the sub-capturing area E is the completion time when the image processing is performed on all the first image data generated from the sub-capturing area D. Therefore, the delay benefit of this solution is only 1/3 of the image processing time, that is, this solution only benefits from the advance of image processing of all the first image data generated from the sub-collection area D.
  • FIG. 2 , FIG. 10 and FIG. 11 it can be understood that the setting of the start time for image processing performed by the image data processing apparatus and feature extraction performed by the image recognition apparatus both have an impact on the delay benefit of the present application.
  • the setting of the number and size of the sub-collection areas of the collection area will also have an impact on the delay benefit of the present application.
  • the size and number of sub-collection areas of the collection area may be directly pre-configured according to preset values. Therefore, the division management module 1300 may be omitted, or steps S101 and S102 may be omitted. In this case, in combination with the solution shown in the lower part of FIG. 10 , the image data processing apparatus 1100 and the image recognition apparatus 1200 execute the following steps S201 - S205 correspondingly according to preset values, as shown in FIG. 12 .
  • step S201 the image data processing apparatus 1101 receives a plurality of first image data sequentially generated by the image sensor scanning the acquisition area. Specifically, the receiving module 1111 of the image data processing apparatus 1101 receives a plurality of first image data sequentially generated by the image sensor scanning the acquisition area.
  • step S202 when the image data processing device 1101 has received a set of first image data, image processing is performed on the set of first image data, and a set of first image data is generated from a preset first image data in the acquisition area.
  • Sub-collection area Specifically, according to a predetermined value, the image processing module 1121 of the image data processing apparatus 1100 presets each of the first sub-collection areas A, B and C of the collection area.
  • the image processing module 1121 of the image data processing device 1100 performs image processing on the first image data, and sequentially outputs the set of first image data.
  • the corresponding second image data is obtained by performing the image processing on one image data.
  • step S203 the image recognition device 1201 receives a plurality of second image data sequentially output by the image data processing device 1101. Specifically, these second image data are received by the receiving module 1211 of the image recognition apparatus 1201 .
  • step S204 when receiving the plurality of second image data, the image recognition device 1201 extracts feature data from the plurality of second image data. Specifically, the feature extraction module 1221 of the image recognition apparatus 1201 extracts feature data from the plurality of second image data together when the receiving module 1211 receives the plurality of second image data originating from the entire acquisition area.
  • step S205 the image recognition device 1201 performs fusion recognition processing on the feature data extracted from the plurality of second image data, and outputs an image recognition result.
  • the feature extraction module 1221 of the image recognition device 1201 extracts feature data from a plurality of second image data
  • the fusion recognition module 1231 of the image recognition device 1201 fuses the plurality of feature data obtained thereby, performs recognition, and finally outputs Image recognition results.
  • the first image data and the second image data of different frames can be grouped in different ways, for example, different division strategies or preset values are used to pre-set the acquisition area of the camera when scanning different frames of images.
  • the number and size of sub-collection regions in one such embodiment, in processing the image data of the first frame of image, the division method shown in FIG. 7 can be used, so that the acquisition area is divided into three sub-acquisition areas of equal size.
  • the division strategy adopted in Figure 11 can be used, that is, the acquisition area is divided into two sub-acquisition areas with unequal sizes, and the size of one sub-acquisition area is twice that of the other. .
  • the first image data and the second image data of the same frame can also be grouped in different ways.
  • each sub-collection area of the collection area is set in another way.
  • the acquisition area in the image data processing flow, the acquisition area is divided into 4 equal sub-acquisition areas F, G, H, and I, while in the image recognition flow, the acquisition area is divided into 2 equal sub-acquisition areas
  • the sub-collection areas J and K of wherein the sub-collection area J coincides with the intersection of the sub-collection areas F and G, and the sub-collection area K coincides with the intersection of the sub-collection areas H and I.
  • each sub-acquisition area of the image may not be defined by the coordinates of the four corners, but may be defined by a rotation angle, for example, a rotation of 0° to 90° may be used
  • the angular extent defines a sub-collection area of the point cloud.
  • the number of layers and sub-layers of the feature extraction module and the fusion recognition module is adjustable.
  • the number of times that each process and each sub-process of feature extraction and fusion identification is performed can also be adjusted.
  • the image data processing apparatus receives first image data from an image sensor, where the first image data is among a plurality of image data that can be generated by the image sensor scanning a physical area corresponding to the acquisition area in one scan period An image data, the acquisition area represents the acquisition range of the image sensor.
  • the image data processing device performs image processing on the first image data to obtain second image data.
  • the image data processing device outputs the second image data. That is, in this embodiment, in the image processing flow, the first image data is not processed in a grouped manner.
  • the image recognition device sequentially receives second image data, where the second image data is obtained by performing image processing on the first image data, and the first image data corresponds to the area scanned by the image sensor in one scan period.
  • One of the plurality of image data that can be generated by the physical area of the acquisition area represents the acquisition range of the image sensor.
  • the image recognition device sequentially extracts feature data from the second image data.
  • the image recognition device performs fusion recognition processing on each feature data.
  • the image recognition device outputs an image recognition result. That is, in this embodiment, in the image recognition flow, the second image data is not processed in groups.
  • FIG. 13 exemplarily shows a schematic structural diagram of a driving system 3001 according to an embodiment of the present application.
  • the driving system 3001 is an advanced driving assistance system (ADAS), which includes an image sensor image data processing system 1001 and a driving decision unit 3100 .
  • ADAS advanced driving assistance system
  • the image sensor image data processing system 1001 can be connected in communication with the camera 2001 outside the driving system 3001, process and recognize a plurality of first image data sequentially generated by the camera 2001 scanning the acquisition area, and output the image recognition result.
  • the driving decision unit 3100 is connected in communication with the image sensor image data processing system 1001, and is used for executing behavior decision and motion planning and outputting operation instructions according to the image recognition result output by the image sensor image data processing system 1001.
  • Fig. 14 exemplarily shows a schematic structural diagram of an intelligent networked vehicle V according to an embodiment of the present application.
  • the intelligent networked vehicle V includes a camera 2001 usually set in the front of the car, a driving system 3001 set in the car, an electronic control unit 4001 and an actuator 5001 such as a braking mechanism.
  • the camera 2001 perceives the vehicle environment in a manner of scanning its acquisition area in rows, and sequentially outputs a plurality of first image data.
  • the driving system 3001 is connected in communication with the camera 2001 for outputting operation instructions according to a plurality of first image data from the camera 2001 .
  • the Electronic Control Unit (ECU) 4001 is connected in communication with the driving system 3001, and is used to control the actuator 5001 to perform operations according to the operation commands from the driving system 3001, for example, control the braking mechanism to perform braking according to the braking command of the driving system operate.
  • FIG. 15 is an exemplary structural diagram of a computing device 1500 provided by an embodiment of the present application.
  • the computing device 1500 includes: a processor 1510 , a memory 1520 , a communication interface 1530 , and a bus 1540 .
  • the communication interface 1530 in the computing device 1500 shown in FIG. 15 may be used to perform communication with other devices.
  • the processor 1510 can be connected with the memory 1520 .
  • the memory 1520 may be used to store the program codes and data. Therefore, the memory 1520 may be a storage unit inside the processor 1510 , or an external storage unit independent from the processor 1510 , or may include a storage unit inside the processor 1510 and an external storage unit independent from the processor 1510 . part.
  • computing device 1500 may also include bus 1540 .
  • the memory 1520 and the communication interface 1530 may be connected to the processor 1510 through the bus 1540 .
  • the bus 1540 may be a peripheral component interconnect standard (Peripheral Component Interconnect, PCI) bus or an Extended Industry Standard Architecture (Extended Industry Standard Architecture, EISA) bus or the like.
  • PCI peripheral component interconnect standard
  • EISA Extended Industry Standard Architecture
  • the bus 1540 can be divided into an address bus, a data bus, a control bus, and the like. For ease of representation, only one line is shown in FIG. 15, but it does not mean that there is only one bus or one type of bus.
  • the processor 1510 may adopt a central processing unit (central processing unit, CPU).
  • the processor may also be other general-purpose processors, digital signal processors (DSPs), application specific integrated circuits (ASICs), off-the-shelf programmable gate arrays (FPGAs) or other Programmable logic devices, discrete gate or transistor logic devices, discrete hardware components, etc.
  • DSPs digital signal processors
  • ASICs application specific integrated circuits
  • FPGAs off-the-shelf programmable gate arrays
  • a general purpose processor may be a microprocessor or the processor may be any conventional processor or the like.
  • the processor 1510 uses one or more integrated circuits to execute related programs to implement the technical solutions provided by the embodiments of the present application.
  • the memory 1520 may include read only memory and random access memory and provides instructions and data to the processor 1510 .
  • a portion of the processor 1510 may also include non-volatile random access memory.
  • the processor 1510 may also store device type information.
  • the processor 1510 executes the computer-implemented instructions in the memory 1520 to perform the operation steps of any of the above methods for processing image sensor image data.
  • the communication interface 1530 and the bus 1540 are omitted.
  • the computing device 1500 may correspond to the corresponding subjects in executing the methods according to the various embodiments of the present application, and the above-mentioned and other operations and/or functions of the various units in the computing device 1500 are respectively for realizing the present invention.
  • the corresponding procedures of each method in the implementation manner are not repeated here.
  • the disclosed systems, devices and methods may be implemented in other manners.
  • the apparatus implementations described above are only exemplary.
  • the division of the units is only a logical function division.
  • there may be other division strategies for example, multiple units or components may be combined or Can be integrated into another system, or some features can be ignored, or not implemented.
  • the shown or discussed mutual coupling or direct coupling or communication connection may be through some interfaces, indirect coupling or communication connection of devices or units, and may be in electrical, mechanical or other forms.
  • the units described as separate components may or may not be physically separated, and components displayed as units may or may not be physical units, that is, may be located in one place, or may be distributed to multiple network units. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution in this implementation manner.
  • each functional unit in each embodiment of the present application may be integrated into one processing unit, or each unit may exist physically alone, or two or more units may be integrated into one unit.
  • the functions, if implemented in the form of software functional units and sold or used as independent products, may be stored in a computer-readable storage medium.
  • the technical solution of the present application can be embodied in the form of a software product in essence, or the part that contributes to the prior art or the part of the technical solution.
  • the computer software product is stored in a storage medium, including Several instructions are used to cause a computer device (which may be a personal computer, a server, or a network device, etc.) to execute all or part of the steps of the methods described in the various embodiments of the present application.
  • the aforementioned storage medium includes: U disk, mobile hard disk, read-only memory (Read-Only Memory, ROM), random access memory (Random Access Memory, RAM), magnetic disk or optical disk and other media that can store program codes .
  • Embodiments of the present application further provide a computer-readable storage medium on which a computer program is stored, and when the program is executed by a processor, is used to execute a method for processing image data of an image sensor, and the method includes the methods described in the foregoing embodiments. at least one of the methods described.
  • the computer storage medium of the embodiments of the present application may adopt any combination of one or more computer-readable mediums.
  • the computer-readable medium may be a computer-readable signal medium or a computer-readable storage medium.
  • the computer readable storage medium can be, for example, but not limited to, an electrical, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus or device, or any combination of the above. More specific examples (a non-exhaustive list) of computer readable storage media include: electrical connections having one or more wires, portable computer disks, hard disks, random access memory (RAM), read only memory (ROM), Erasable Programmable Read Only Memory (EPROM or Flash), fiber optics, portable compact disk read only memory (CD-ROM), optical storage, magnetic storage, cloud, or any suitable combination of the above.
  • a computer-readable storage medium can be any tangible medium that contains or stores a program that can be used by or in conjunction with an instruction execution system, apparatus, or device.
  • a computer-readable signal medium may include a propagated data signal in baseband or as part of a carrier wave with computer-readable program code embodied thereon. Such propagated data signals may take a variety of forms, including but not limited to electromagnetic signals, optical signals, or any suitable combination of the foregoing.
  • a computer-readable signal medium can also be any computer-readable medium other than a computer-readable storage medium that can transmit, propagate, or transport the program for use by or in connection with the instruction execution system, apparatus, or device .
  • Program code embodied on a computer readable medium may be transmitted using any suitable medium including, but not limited to, wireless, wireline, optical fiber cable, RF, etc., or any suitable combination of the foregoing.
  • Computer program code for performing the operations of the present application may be written in one or more programming languages, including object-oriented programming languages—such as Java, Smalltalk, C++, but also conventional Procedural programming language - such as the "C" language or similar programming language.
  • the program code may execute on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer, or on the remote computer or server.
  • the remote computer may be connected to the user's computer through any kind of network, including a local area network (LAN) or wide area network (WAN), or may be connected to an external computer (eg, through the Internet using an Internet service provider) connect).
  • LAN local area network
  • WAN wide area network
  • Internet service provider an external computer

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Image Processing (AREA)

Abstract

The present application relates to a method and apparatus for processing image data of an image sensor. The method comprises: receiving first image data from an image sensor, the first image data being one of a plurality of pieces of image data generated by the image sensor scanning a physical region corresponding to an acquisition region in one scanning period, and the acquisition region representing an acquisition range of the image sensor; performing image processing on the first image data to obtain second image data; and outputting the second image data. The method and apparatus for processing image data of an image sensor of the present application can advance the processing flow for image data generated by the image sensor, thereby reducing the end-to-end delay from the image sensor generating image data to a vehicle actuator executing an operation. The method and apparatus for processing image data of an image sensor of the present application are particularly suitable for advanced driving assistance systems (ADASs), robot systems, and the like.

Description

处理图像传感器图像数据的方法和装置Method and apparatus for processing image sensor image data 技术领域technical field
本申请涉及智能网联车领域,尤其涉及一种处理图像传感器图像数据的方法和装置。The present application relates to the field of intelligent networked vehicles, and in particular, to a method and device for processing image sensor image data.
背景技术Background technique
感知识别在智能网联车的高级驾驶辅助系统(Advanced Driving Assistance System,ADAS)和自主驾驶系统(Autonomous Driving)中起重要作用。为了实现感知识别功能,智能网联车搭载多种图像传感器。用于智能感知识别的车载图像传感器的常见例子为摄像机(camera)、激光雷达(Light Detection and Ranging,LiDAR)、毫米波雷达(millimeter wave)等,它们获得的“图像数据”信息量丰富,因此智能网联车可以通过这些图像数据实现识别功能。Perceptual recognition plays an important role in the advanced driving assistance system (ADAS) and autonomous driving system (Autonomous Driving) of intelligent networked vehicles. In order to realize the perception and recognition function, the intelligent networked vehicle is equipped with a variety of image sensors. Common examples of in-vehicle image sensors for intelligent perception recognition are cameras (camera), LiDAR (Light Detection and Ranging, LiDAR), millimeter wave radar (millimeter wave), etc. The "image data" obtained by them is rich in information, so Intelligent connected vehicles can realize the recognition function through these image data.
在智能网联车中,图像数据从图像传感器到达高级驾驶辅助系统或自主驾驶系统,经过图像处理、算法识别、驾驶决策等步骤后,最终转变为对车辆各个执行器的操作控制指令,实现对驾驶的控制。在这个过程中,出于安全角度考虑,降低从图像传感器生成图像数据到车辆执行器执行操作的端到端时延,是高级驾驶辅助系统和自主驾驶系统不断追求的目标。In the ICV, the image data reaches the advanced driving assistance system or the autonomous driving system from the image sensor, and after the steps of image processing, algorithm recognition, and driving decision-making, it is finally transformed into the operation control instructions for the various actuators of the vehicle. Driving controls. In this process, from the perspective of safety, reducing the end-to-end delay from the generation of image data by image sensors to the execution of operations by vehicle actuators is the goal of advanced driver assistance systems and autonomous driving systems.
车载传感器图像传感器中具有许多感知方式为“扫描式”的图像传感器,例如激光雷达和摄像机。这种扫描式图像传感器不是在采集区域中同时生成一帧图像的所有图像数据,而是例如以按行或按列扫描的方式,在一段时间内依次生成一帧图像的所有图像数据。目前,车载的图像信号处理器(Image Signal Processor,ISP)以及后续的算法平台在等待接收完一帧图像的所有图像数据后,对该帧图像进行处理。这种处理方式对端到端时延有很大影响。In-vehicle sensors Image sensors include many image sensors that sense "scanning", such as lidars and cameras. Such a scanning image sensor does not generate all the image data of a frame of images simultaneously in the acquisition area, but sequentially generates all the image data of a frame of images over a period of time, for example by scanning in rows or columns. At present, the vehicle-mounted Image Signal Processor (ISP) and subsequent algorithm platforms process the frame of image after waiting for all the image data of the frame to be received. This processing method has a great impact on the end-to-end delay.
发明内容SUMMARY OF THE INVENTION
有鉴于此,本申请提供处理图像传感器图像数据的方法和装置,其能够降低从图像传感器生成图像数据到车辆执行器执行操作的端到端时延。In view of this, the present application provides a method and apparatus for processing image sensor image data, which can reduce the end-to-end delay from the image sensor generating the image data to the vehicle actuator performing the operation.
第一方面,提供一种处理图像传感器图像数据的方法,该方法包括以下步骤:接收来自图像传感器的第一图像数据,所述第一图像数据为所述图像传感器在一个扫描周期内扫描采集区域对应的物理区域可生成的多个图像数据中一个图像数据,所述采集区域表示所述图像传感器的采集范围;对所述第一图像数据进行图像处理得到第二图像数据;以及输出所述第二图像数据。In a first aspect, a method for processing image data of an image sensor is provided, the method comprising the steps of: receiving first image data from an image sensor, the first image data being that the image sensor scans an acquisition area within one scan period One image data among a plurality of image data that can be generated by the corresponding physical area, the collection area represents the collection range of the image sensor; image processing is performed on the first image data to obtain second image data; and the first image data is output. 2. Image data.
通过一边接收图像传感器在一个扫描周期内扫描采集区域对应的物理区域可生成的多个图像数据中一个图像数据,一边对已接收的第一图像数据进行图像处理,能够提前对图像传感器生成的第一图像数据的图像处理流程,从而降低从图像传感器生成图像数据到车辆执行器执行操作的端到端时延。换句话说,当接收到扫描采集区域所能够生成的一帧图像的一部分第一图像数据时,就可以对该部分第一图像数 据执行图像处理,而不必等待接收到整帧的图像的所有第一图像数据时才开始图像数据的图像处理流程。By performing image processing on the received first image data while receiving one image data among a plurality of image data that can be generated by the image sensor scanning the physical area corresponding to the acquisition area in one scan period, the first image data generated by the image sensor can be processed in advance. An image processing flow of image data, thereby reducing the end-to-end delay from the generation of image data by an image sensor to the execution of operations by vehicle actuators. In other words, when a part of the first image data of a frame of images that can be generated by scanning the acquisition area is received, image processing can be performed on the part of the first image data without waiting for all the first image data of the entire frame of image to be received. The image processing flow of the image data is started only when an image data is generated.
在一种可能的实现方式中,所述图像传感器是摄像机,所述图像是二维平面图像。In one possible implementation, the image sensor is a camera, and the image is a two-dimensional plane image.
摄像机的采集区域通常被称为靶面。摄像机的帧率一般为30Hz(Hertz,赫兹)。换句话说,摄像机从靶面的第一行曝光到靶面的最后一行以形成一帧图像的时长为33ms(milisecond,毫秒)。通常需要等待整帧图像的所有第一图像数据生成并传输完成后,才对接收到的整帧图像的图像数据执行处理。即从摄像机开始扫描靶面后,至少需要等待33ms才能处理由此生成的第一图像数据。然而,通过本申请的处理图像数据的方法,可以例如在接收了一帧图像的第一个第一图像数据后,就立即对该第一图像数据执行处理,而不必等到接收整帧图像的所有第一图像数据。这能够将平面图像的第一图像数据的图像处理流程提前,节省每帧平面图像的第一图像数据的处理时长,从而降低从摄像机生成第一图像数据到车辆执行器执行操作的端到端时延。The acquisition area of the camera is often referred to as the target surface. The frame rate of the camera is generally 30Hz (Hertz, Hertz). In other words, the exposure time of the camera from the first line of the target surface to the last line of the target surface to form one frame of image is 33ms (milisecond, millisecond). Generally, it is necessary to wait for the completion of the generation and transmission of all the first image data of the whole frame of image before performing processing on the received image data of the whole frame of image. That is, after the camera starts to scan the target surface, it takes at least 33ms to process the first image data thus generated. However, with the method for processing image data of the present application, for example, after receiving the first first image data of a frame of image, the processing of the first image data can be performed immediately, without waiting for all the images of the whole frame to be received. first image data. This can advance the image processing process of the first image data of the plane image, save the processing time of the first image data of each frame of plane image, thereby reducing the end-to-end time from the generation of the first image data by the camera to the operation of the vehicle actuator. extension.
在一种可能的实现方式中,所述图像传感器是激光雷达,所述图像是三维点云。激光雷达的采集区域通常被称为扫描区域。In a possible implementation, the image sensor is a lidar, and the image is a three-dimensional point cloud. The acquisition area of the lidar is often referred to as the scan area.
激光雷达的帧率一般为10Hz或20Hz。也就是说,激光雷达从扫描区域的第一列扫描到扫描区域的最后一列,从而形成一帧点云的第一图像数据的时长通常为100ms或50ms。与处理平面图像的图像数据类似地,通常需要等待整帧点云的第一图像数据生成并传输完成后,才对接收到的整帧点云的第一图像数据执行处理。然而,通过本申请的处理图像数据的方法,可以例如在接收了一帧点云的第一个第一图像数据后,就可以立即对该第一图像数据执行处理,而不必等到接收整帧点云的所有第一图像数据。这能够将点云的第一图像数据的图像处理流程提前,节省每帧点云的第一图像数据的处理时长,从而降低从激光雷达生成第一图像数据到车辆执行器执行操作的端到端时延。The frame rate of lidar is generally 10Hz or 20Hz. That is to say, the lidar scans from the first column of the scanning area to the last column of the scanning area, so that the duration of forming the first image data of a frame of point cloud is usually 100ms or 50ms. Similar to processing the image data of a plane image, it is usually necessary to wait for the first image data of the whole frame of point cloud to be generated and transmitted before processing the received first image data of the whole frame of point cloud. However, with the method for processing image data of the present application, for example, after receiving the first first image data of a frame of point cloud, the first image data can be processed immediately without waiting for the whole frame of points to be received. All first image data of the cloud. This can advance the image processing flow of the first image data of the point cloud, save the processing time of the first image data of each frame of the point cloud, thereby reducing the end-to-end process from the generation of the first image data by the lidar to the operation performed by the vehicle actuator. time delay.
结合第一方面,在一种可能的实现方式中,所述采集区域包括多个子采集区域;所述对所述第一图像数据进行图像处理包括:当接收到第一数据组A所包含的第一图像数据后,以第一图像数据组A所包含的全部第一图像数据为单位进行图像处理,所述第一图像数据组A是所述图像传感器扫描一个所述子采集区域所对应的物理区域而生成的所述第一图像数据的集合。With reference to the first aspect, in a possible implementation manner, the collection area includes a plurality of sub-collection areas; the performing image processing on the first image data includes: when the first data set A included in the first data set A is received After one image data is obtained, image processing is performed in units of all the first image data included in the first image data group A, where the first image data group A is the physical image corresponding to the image sensor scanning one of the sub-collection areas. The set of the first image data generated by the region.
通过以采集区域的各子采集区域限定各第一图像数据组的方式,提前对一帧图像的多个第一图像数据进行梳理。当接收到第一数据组A所包含的第一图像数据后,以第一图像数据组A所包含的全部第一图像数据为单位进行图像处理,即通过以第一图像数据组作为单位进行图像处理,可以在提前第一图像数据的图像处理流程的同时,避免过分增加后续处理的复杂性和由此导致的后续处理时间的增加。此外,通过子采集区域限定各第一图像数据组,能够适用于不同分辨率的图像传感器。By defining each first image data group in each sub-collection area of the collection area, the plurality of first image data of one frame of image are sorted out in advance. After receiving the first image data included in the first data group A, image processing is performed using all the first image data included in the first image data group A as a unit, that is, by using the first image data group as a unit to perform image processing processing, the image processing flow of the first image data can be advanced, and at the same time, the complexity of the subsequent processing and the increase of the subsequent processing time caused thereby can be avoided. In addition, each first image data set is defined by the sub-acquisition regions, which can be applied to image sensors of different resolutions.
结合第一方面,在一种可能的实现方式中,所述多个子采集区域的大小和数量为预先设定。With reference to the first aspect, in a possible implementation manner, the size and quantity of the plurality of sub-collection regions are preset.
在一种可能的实现方式中,子采集区域的数量选自2至4。In a possible implementation, the number of sub-collection regions is selected from 2 to 4.
在一种可能的实现方式中,各个子采集区域的大小相等。In a possible implementation manner, the sizes of each sub-collection area are equal.
结合第一方面,在一种可能的实现方式中,所述方法还包括以下步骤:接收划分策略,以及根据所述划分策略预先设定所述多个子采集区域的大小和数量。With reference to the first aspect, in a possible implementation manner, the method further includes the following steps: receiving a division strategy, and presetting the size and quantity of the plurality of sub-collection areas according to the division strategy.
通过接收划分策略来预先设定多个子采集区域的大小和数量,可以根据应用场景的实际需求灵活调整多个第一图像数据的图像处理的过程,从而使得本申请的处理图像传感器图像数据的方法适配各种应用场景。By receiving the division strategy to pre-set the size and number of multiple sub-collection areas, the process of image processing of multiple first image data can be flexibly adjusted according to the actual needs of the application scenario, so that the method for processing image sensor image data of the present application Adapt to various application scenarios.
结合第一方面,在一种可能的实现方式中,所述子采集区域为矩形,所述子采集区域的大小由矩形的四个角的坐标限定。With reference to the first aspect, in a possible implementation manner, the sub-collection area is a rectangle, and the size of the sub-collection area is defined by coordinates of four corners of the rectangle.
通过所述子采集区域的四个角的坐标限定所述子采集区域的大小,能够以尤其简单和直观的方式灵活划分各个子采集区域。By defining the size of the sub-collection area by the coordinates of the four corners of the sub-collection area, each sub-collection area can be flexibly divided in a particularly simple and intuitive manner.
第二方面,提供一种处理图像传感器图像数据的方法,该方法包括以下步骤:依次接收第二图像数据,所述第二图像数据是对第一图像数据进行图像处理得到的,所述第一图像数据为所述图像传感器在一个扫描周期内扫描采集区域对应的物理区域可生成的多个图像数据中一个图像数据,所述采集区域表示所述图像传感器的采集范围;依次从所述第二图像数据提取特征数据;对各个特征数据进行融合识别处理;以及输出图像识别结果。In a second aspect, there is provided a method for processing image data of an image sensor, the method comprising the steps of: sequentially receiving second image data, the second image data is obtained by performing image processing on the first image data, the first The image data is one image data among a plurality of image data that can be generated by the image sensor scanning the physical area corresponding to the acquisition area in one scan period, and the acquisition area represents the acquisition range of the image sensor; Extract feature data from image data; perform fusion recognition processing on each feature data; and output image recognition results.
通过一边接收第二图像数据期间,一边依次从已接收的第二图像数据提取特征数据,能够提前第二图像数据的图像识别流程,从而降低从图像传感器生成第一图像数据到车辆执行器执行操作的端到端时延。换句话说,当接收到一帧图像的一部分第二图像数据时,就可以对该帧图像的该部分第二图像数据执行作为识别算法处理过程一部分的提取特征数据,而不必等到接收到整帧图像的所有第二图像数据时才开始进行特征提取。By sequentially extracting feature data from the received second image data while receiving the second image data, it is possible to advance the image recognition process of the second image data, thereby reducing the time from the generation of the first image data by the image sensor to the time when the vehicle actuator performs operations. end-to-end delay. In other words, when a part of the second image data of a frame of image is received, the extraction of feature data as part of the processing process of the recognition algorithm can be performed on the part of the second image data of the frame image, without waiting for the whole frame to be received Feature extraction starts only when all of the second image data of the image is performed.
此外,通过将传统的识别算法处理过程区分为特征提取过程和融合识别过程,例如在特征提取过程中分别提取源自采集区域不同部分的第二图像数据的特征数据,而在融合识别中融合识别从所有第二图像数据提取的各特征数据,能够保证融合识别的图像识别结果的稳定性和准确性。In addition, by dividing the traditional recognition algorithm processing process into a feature extraction process and a fusion recognition process, for example, in the feature extraction process, the feature data of the second image data originating from different parts of the acquisition area are respectively extracted, and the fusion recognition is performed in the fusion recognition process. Each feature data extracted from all the second image data can ensure the stability and accuracy of the image recognition result of the fusion recognition.
结合第二方面,在一种可能的实现方式中,所述图像传感器是摄像机,所述图像是二维平面图像。结合第二方面,在一种可能的实现方式中,所述图像传感器是激光雷达,所述图像是三维点云。With reference to the second aspect, in a possible implementation manner, the image sensor is a camera, and the image is a two-dimensional plane image. With reference to the second aspect, in a possible implementation manner, the image sensor is a lidar, and the image is a three-dimensional point cloud.
结合第二方面,在一种可能的实现方式中,对所述第二图像数据分组;所述各个特征数据是从各组所述第二图像数据中提取。With reference to the second aspect, in a possible implementation manner, the second image data is grouped; the respective feature data is extracted from each group of the second image data.
通过对所述第二图像数据分组,提前对一帧图像的多个第二图像数据进行梳理。通过从各组所述第二图像数据中提取所述各个特征数据,即以一组第二图像数据作为单位进行提取特征数据,可以在提前图像数据的图像识别流程的同时,避免过分增加后续处理的复杂性和由此导致的后续处理时间的增加。By grouping the second image data, a plurality of second image data of one frame of image are sorted out in advance. By extracting the respective feature data from each set of the second image data, that is, extracting the feature data with a set of second image data as a unit, the image recognition process of the image data can be advanced and the subsequent processing can be avoided excessively. complexity and consequent increase in subsequent processing time.
结合第二方面,在一种可能的实现方式中,所述各组所述第二图像数据的第二图像数据的数量为预先设定。With reference to the second aspect, in a possible implementation manner, the quantity of the second image data of each group of the second image data is preset.
结合第二方面,在一种可能的实现方式中,所述采集区域包括多个子采集区域,依据所述多个子采集区对所述第二图像数据分组,一组第二图像数据是经过图像处理的一组第一图像数据,该组第一图像是所述图像传感器扫描一个所述子采集区 域所对应的物理区域而生成的所述第一图像数据的集合。With reference to the second aspect, in a possible implementation manner, the collection area includes a plurality of sub-collection areas, the second image data is grouped according to the plurality of sub-collection areas, and a group of second image data is image-processed A set of first image data is a set of the first image data generated by the image sensor scanning a physical area corresponding to one of the sub-acquisition areas.
结合第二方面,在一种可能的实现方式中,所述方法还包括以下步骤:接收划分策略,以及根据所述接收划分策略预先设定所述多个子采集区域的大小和数量。With reference to the second aspect, in a possible implementation manner, the method further includes the following steps: receiving a division strategy, and presetting the size and number of the plurality of sub-collection areas according to the received division strategy.
通过接收划分策略来预先设定多个子采集区域的大小和数量,从而调整对跌图像数据的分组,可以根据应用场景的实际需求灵活调整从多个图像数据提取特征数据的过程,从而使得本申请的处理图像传感器图像数据的方法适配各种应用场景。The size and number of multiple sub-collection areas are pre-set by receiving the division strategy, so as to adjust the grouping of image data, and the process of extracting feature data from multiple image data can be flexibly adjusted according to the actual needs of the application scenario, so that the present application The method of processing image sensor image data is suitable for various application scenarios.
结合第二方面,在一种可能的实现方式中,所述子采集区域为矩形,所述子采集区域的大小由矩形的四个角的坐标限定。With reference to the second aspect, in a possible implementation manner, the sub-collection area is a rectangle, and the size of the sub-collection area is defined by coordinates of four corners of the rectangle.
通过所述子采集区域的四个角的坐标限定所述区域的大小,能够以尤其简单和直观的方式灵活划分各个子采集区域。By defining the size of the sub-collection area by the coordinates of the four corners of the sub-collection area, each sub-collection area can be flexibly divided in a particularly simple and intuitive manner.
结合第二方面,在一种可能的实现方式中,所述提取特征数据包括卷积处理和池化(pooling,又称汇聚)处理。With reference to the second aspect, in a possible implementation manner, the extracting feature data includes convolution processing and pooling (pooling, also known as pooling) processing.
在一种可能的实现方式中,所述卷积处理和所述池化处理交替执行一次以上。在一种可能的实现方式中,卷积处理包括一或多个子卷积处理。在一种可能的实现方式中,卷积处理中子卷积处理的个数为选自1至3的自然数。In a possible implementation manner, the convolution processing and the pooling processing are alternately performed more than once. In one possible implementation, the convolution process includes one or more sub-convolution processes. In a possible implementation manner, the number of sub-convolution processes in the convolution process is a natural number selected from 1 to 3.
结合第二方面,在一种可能的实现方式中,所述融合识别包括特征融合处理。With reference to the second aspect, in a possible implementation manner, the fusion identification includes feature fusion processing.
通过特征融合处理,能够有效地将从源自各个采集区域的第二图像数据中提取的各特征数据融合成整帧图像的特征数据的形式。Through the feature fusion process, each feature data extracted from the second image data originating from each acquisition area can be effectively fused into the form of feature data of the entire frame image.
结合第二方面,在一种可能的实现方式中,所述融合识别还包括全连接处理。With reference to the second aspect, in a possible implementation manner, the fusion identification further includes full connection processing.
通过全连接处理,能够对整帧图像的图像数据的特征执行全局分析识别。Through the full connection processing, it is possible to perform global analysis and recognition on the characteristics of the image data of the entire frame of image.
在一种可能的实现方式中,特征融合处理包括系列特征融合(concat)处理。在一种可能的实现方式中,全连接处理包括一次或多次子全连接处理。In one possible implementation, the feature fusion process includes a series of feature fusion (concat) processes. In a possible implementation manner, the full connection processing includes one or more sub-full connection processing.
第三方面,提供一种图像数据处理装置,包括:接收模块,用于接收来自图像传感器的第一图像数据,所述第一图像数据为所述图像传感器在一个扫描周期内扫描采集区域对应的物理区域可生成的多个图像数据中一个图像数据,所述采集区域表示所述图像传感器的采集范围;以及图像处理模块,用于对所述第一图像数据进行图像处理得到第二图像数据,以及用于输出所述第二图像数据。In a third aspect, an image data processing device is provided, comprising: a receiving module configured to receive first image data from an image sensor, where the first image data is corresponding to a scanning and acquisition area of the image sensor in one scanning period One image data among a plurality of image data that can be generated by the physical area, the collection area represents the collection range of the image sensor; and an image processing module, configured to perform image processing on the first image data to obtain second image data, and for outputting the second image data.
结合第三方面,在一种可能的实现方式中,所述采集区域包括多个子采集区域;所述对所述第一图像数据进行图像处理包括:所述图像处理模块还用于当所述接收模块接收到第一数据组A所包含的第一图像数据后,以第一图像数据组A所包含的全部第一图像数据为单位进行图像处理,所述第一图像数据组A是所述图像传感器扫描一个所述子采集区域所对应的物理区域而生成的所述第一图像数据的集合。With reference to the third aspect, in a possible implementation manner, the collection area includes a plurality of sub-collection areas; the performing image processing on the first image data includes: the image processing module is further configured to, when the receiving After the module receives the first image data contained in the first data set A, it performs image processing in units of all the first image data contained in the first image data set A, where the first image data set A is the image The set of the first image data generated by the sensor scanning a physical area corresponding to the sub-collection area.
结合第三方面,在一种可能的实现方式中,所述多个子采集区域的大小和数量为预先设定。With reference to the third aspect, in a possible implementation manner, the size and quantity of the plurality of sub-collection regions are preset.
结合第三方面,在一种可能的实现方式中,所述子采集区域为矩形,所述子采集区域的大小由矩形的四个角的坐标限定。With reference to the third aspect, in a possible implementation manner, the sub-collection area is a rectangle, and the size of the sub-collection area is defined by coordinates of four corners of the rectangle.
结合第三方面,在一种可能的实现方式中,所述接收模块还用于接收划分 策略,以及所述图像数据处理模块还用于根据所述划分策略预先设定所述多个子采集区域的大小和数量。With reference to the third aspect, in a possible implementation manner, the receiving module is further configured to receive a division strategy, and the image data processing module is further configured to preset the number of sub-collection regions according to the division strategy. size and quantity.
由于本申请第三方面的装置能够执行第一方面的方法,因此第三方面的装置的优点及益处类似于第一方面的优点及益处,参照第一方面的有关描述,在此不再赘述。Since the device of the third aspect of the present application can execute the method of the first aspect, the advantages and benefits of the device of the third aspect are similar to those of the first aspect, and the relevant descriptions of the first aspect are referred to and will not be repeated here.
第四方面,提供一种图像识别装置,包括:接收模块,用于依次接收第二图像数据,所述第二图像数据是对第一图像数据进行图像处理得到的,所述第一图像数据为所述图像传感器在一个扫描周期内扫描采集区域对应的物理区域可生成的多个图像数据中一个图像数据,所述采集区域表示所述图像传感器的采集范围;特征提取模块,用于依次从所述第二图像数据提取特征数据;以及融合识别模块,用于对各个特征数据进行融合识别处理,以及用于输出图像识别结果。In a fourth aspect, an image recognition device is provided, comprising: a receiving module configured to sequentially receive second image data, wherein the second image data is obtained by performing image processing on the first image data, and the first image data is The image sensor scans one image data among multiple image data that can be generated by scanning the physical area corresponding to the acquisition area in one scan period, and the acquisition area represents the acquisition range of the image sensor; the feature extraction module is used to sequentially select from all the image data. feature data is extracted from the second image data; and a fusion recognition module is used to perform fusion recognition processing on each feature data and output image recognition results.
结合第四方面,在一种可能的实现方式中,所述各组所述第二图像数据的第二图像数据的数量为预先设定。With reference to the fourth aspect, in a possible implementation manner, the quantity of the second image data of each group of the second image data is preset.
结合第四方面,在一种可能的实现方式中,所述采集区域包括多个子采集区域,依据所述多个子采集区对所述第二图像数据分组,一组第二图像数据是经过图像处理的一组第一图像数据,该组第一图像是所述图像传感器扫描一个所述子采集区域所对应的物理区域而生成的所述第一图像数据的集合。With reference to the fourth aspect, in a possible implementation manner, the collection area includes a plurality of sub-collection areas, the second image data is grouped according to the plurality of sub-collection areas, and a group of second image data is image-processed A set of first image data is a set of the first image data generated by the image sensor scanning a physical area corresponding to one of the sub-acquisition areas.
结合第四方面,在一种可能的实现方式中,所述多个子采集区域的大小和数量为预先设定。With reference to the fourth aspect, in a possible implementation manner, the size and quantity of the plurality of sub-collection regions are preset.
结合第四方面,在一种可能的实现方式中,所述接收模块进一步用于接收划分策略,并且特征提取模块进一步用于根据划分策略预先设定所述多个子采集区域的大小和数量。With reference to the fourth aspect, in a possible implementation manner, the receiving module is further configured to receive a division strategy, and the feature extraction module is further configured to preset the size and quantity of the plurality of sub-collection regions according to the division strategy.
结合第四方面,在一种可能的实现方式中,所述子采集区域为矩形,所述第二子采集区域的大小由矩形的四个角的坐标限定。With reference to the fourth aspect, in a possible implementation manner, the sub-collection area is a rectangle, and the size of the second sub-collection area is defined by coordinates of four corners of the rectangle.
结合第四方面,在一种可能的实现方式中,所述特征提取模块包括卷积层和池化层。With reference to the fourth aspect, in a possible implementation manner, the feature extraction module includes a convolution layer and a pooling layer.
在一种可能的实现方式中,所述卷积层和池化层交替配置一个以上。在一种可能的实现方式中,卷积层包括一个或多个子卷积层。在一种可能的实现方式中,卷积层中子卷积层的个数为选自1至3的自然数。In a possible implementation manner, more than one convolution layer and pooling layer are alternately configured. In one possible implementation, the convolutional layer includes one or more sub-convolutional layers. In a possible implementation manner, the number of sub-convolutional layers in the convolutional layer is a natural number selected from 1 to 3.
结合第四方面,在一种可能的实现方式中,所述融合识别模块包括特征融合层。With reference to the fourth aspect, in a possible implementation manner, the fusion identification module includes a feature fusion layer.
结合第四方面,在一种可能的实现方式中,所述融合识别模块还包括全连接层。With reference to the fourth aspect, in a possible implementation manner, the fusion identification module further includes a fully connected layer.
在一种可能的实现方式中,特征融合处理模块包括系列特征融合(concat)层。在一种可能的实现方式中,全连接层包括一个或多个子全连接层。In a possible implementation, the feature fusion processing module includes a series of feature fusion (concat) layers. In a possible implementation manner, the fully-connected layer includes one or more sub-fully-connected layers.
结合第四方面,在一种可能的实现方式中,所述接收模块还用于接收划分策略,以及所述特征提取模块还用于根据所述划分策略预先设定所述多个第二子采集区域的大小和数量。With reference to the fourth aspect, in a possible implementation manner, the receiving module is further configured to receive a division strategy, and the feature extraction module is further configured to preset the plurality of second sub-collections according to the division strategy The size and number of regions.
由于本申请第四方面的装置能够执行第二方面的方法,因此第四方面的装 置的优点及益处类似于第二方面的优点及益处,参照第二方面的有关描述,在此不再赘述。Since the apparatus of the fourth aspect of the present application can perform the method of the second aspect, the advantages and benefits of the apparatus of the fourth aspect are similar to those of the second aspect, and the relevant descriptions of the second aspect are referred to, and are not repeated here.
第五方面,提供一种图像传感器图像数据处理系统,包括如第三方面中的任一种图像数据处理装置和如第四方面中的任一种图像识别装置。In a fifth aspect, an image sensor image data processing system is provided, including any image data processing apparatus as in the third aspect and any image recognition apparatus as in the fourth aspect.
结合第五方面,在一种可能的实现方式中,所述图像传感器图像数据处理系统还包括划分管理模块,用于向所述图像数据处理装置和所述图像识别装置提供划分策略,所述划分策略用于预先设定所述采集区域的多个子采集区域的大小和数量。With reference to the fifth aspect, in a possible implementation manner, the image sensor image data processing system further includes a division management module, configured to provide a division strategy to the image data processing apparatus and the image recognition apparatus, the division The strategy is used to preset the size and quantity of the plurality of sub-collection areas of the collection area.
由于第五方面的系统包括第三方面和第四方面的装置,因此第五方面的系统的优点及益处将包括第三方面和第四方面的优点及益处,参照第三方面和第四方面的有关描述,在此不再赘述。Since the system of the fifth aspect includes the apparatus of the third aspect and the fourth aspect, the advantages and benefits of the system of the fifth aspect will include the advantages and benefits of the third aspect and the fourth aspect, with reference to the third aspect and the fourth aspect. The description will not be repeated here.
第六方面,提供一种驾驶系统,其包括上述第五方面任一种图像传感器图像数据处理系统和驾驶决策单元;其中所述驾驶决策单元与所述图像传感器图像数据处理系统连接,用于依据所述图像传感器图像数据处理系统输出的图像识别结果执行行为决策和运动规划并输出操作指令。In a sixth aspect, a driving system is provided, which includes any image sensor image data processing system and a driving decision-making unit in the fifth aspect; wherein the driving decision-making unit is connected to the image sensor image data processing system for The image recognition result output by the image sensor image data processing system executes behavioral decision-making and motion planning and outputs operation instructions.
通过采用第五方面的任一种图像传感器图像数据处理系统,本申请的驾驶系统能够提前处理图像数据的流程,节省处理时长,从而降低从图像传感器生成图像数据到车辆执行器执行操作的端到端时延。By adopting any of the image sensor image data processing systems of the fifth aspect, the driving system of the present application can process the process of image data in advance, save processing time, thereby reducing the end-to-end time from the image sensor generating the image data to the vehicle actuator performing the operation end delay.
在一种可能的实现方式中,所述驾驶系统是高级驾驶辅助系统。在另一种可能的实现方式中,所述驾驶系统是自主驾驶系统。In one possible implementation, the driving system is an advanced driving assistance system. In another possible implementation, the driving system is an autonomous driving system.
第七方面,提供一种车辆,其包括依次连接的图像传感器、上述第六方面的任一种驾驶系统、电子控制单元和执行器;其中所述图像传感器用于以扫描的方式感知车辆环境并输出第一图像数据;所述电子控制单元用于依据所述驾驶系统的操作指令控制执行器执行操作。In a seventh aspect, a vehicle is provided, which includes an image sensor connected in sequence, any one of the driving systems of the sixth aspect, an electronic control unit, and an actuator; wherein the image sensor is used to perceive the vehicle environment in a scanning manner and outputting first image data; the electronic control unit is configured to control the actuator to perform operations according to the operating instructions of the driving system.
通过采用上述第六方面的任一种驾驶系统,本申请的车辆能够提前处理流程,节省处理时长,从而降低从图像传感器生成图像数据到车辆执行器执行操作的端到端时延。By adopting any of the driving systems in the sixth aspect, the vehicle of the present application can process the process in advance, save processing time, and thus reduce the end-to-end delay from the image sensor generating the image data to the vehicle actuator performing the operation.
第八方面,提供一种计算设备,其包括:至少一个处理器;以及至少一个存储器,其与所述处理连接连接并存储有程序指令,所述程序指令当被所述至少一个处理器执行时使得所述至少一个处理器执行第一方面和第三方面中任一项所述的方法。In an eighth aspect, there is provided a computing device comprising: at least one processor; and at least one memory connected in connection with the processing and storing program instructions, the program instructions when executed by the at least one processor The at least one processor is caused to perform the method of any one of the first and third aspects.
由于本申请的计算设备中的处理器可以执行上述第一方面和第二方面中任一种处理图像传感器图像数据的方法,因此该计算设备的优点及益处也类似于第一方面和第二方面的优点及益处,参照第一方面和第二方面的有关描述,在此不再赘述。Since the processor in the computing device of the present application can execute any one of the above-mentioned methods for processing image sensor image data in the first and second aspects, the advantages and benefits of the computing device are also similar to the first and second aspects For the advantages and benefits, refer to the relevant descriptions of the first aspect and the second aspect, which will not be repeated here.
第九方面,提供一种计算机可读存储介质,其上存储有程序指令,所述程序指令当被计算机执行时使得所述计算机执行上述第一方面和第二方面中任一种处理图像传感器图像数据的方法。In a ninth aspect, there is provided a computer-readable storage medium having program instructions stored thereon, the program instructions, when executed by a computer, cause the computer to perform any one of the first and second aspects above to process image sensor images method of data.
由于本申请的计算机可读存储介质可以使得计算机执行上述第一方面和第二方面中任一种处理图像传感器图像数据的方法,因此该计算机可读存储介质的优点及益处也类似于第一方面和第二方面优点及益处,参照第一方面和第二方面的有关 描述,在此不再赘述。Since the computer-readable storage medium of the present application can enable a computer to perform any one of the above-mentioned methods for processing image sensor image data in the first aspect and the second aspect, the advantages and benefits of the computer-readable storage medium are also similar to those of the first aspect For the advantages and benefits of the second aspect, reference is made to the relevant descriptions of the first aspect and the second aspect, which will not be repeated here.
附图说明Description of drawings
以下参照附图来进一步说明本申请的各个特征和各个特征之间的联系。附图均为示例性的,一些特征并不以实际比例示出,并且一些附图中可能省略了本申请所涉及领域的惯常的且对于本申请非必要的特征,或是额外示出了对于本申请非必要的特征,附图所示的各个特征的组合并不用以限制本申请。另外,在本说明书全文中,相同的附图标记所指代的内容也是相同的。具体的附图说明如下:The various features of the present application and the connections between the various features are further explained below with reference to the accompanying drawings. The drawings are exemplary, some features are not shown to scale, and some of the drawings may omit features that are customary in the field to which the application relates and not essential to the application, or additionally show The non-essential features of the present application, and the combination of individual features shown in the drawings are not intended to limit the present application. In addition, the same reference numerals refer to the same contents throughout the present specification. The specific drawings are described as follows:
图1是现有技术中的图像数据处理方案的示意图;Fig. 1 is the schematic diagram of the image data processing scheme in the prior art;
图2是现有技术和本申请一个实施方式的图像数据处理方案的示意图,其中现有技术的方案位于上部,本申请的实施方式的方案位于下部;2 is a schematic diagram of an image data processing solution of the prior art and an embodiment of the present application, wherein the solution of the prior art is located at the upper part, and the solution of the embodiment of the present application is located at the lower part;
图3是根据本申请一个实施方式的图像传感器图像数据处理系统的结构示意图;3 is a schematic structural diagram of an image sensor image data processing system according to an embodiment of the present application;
图4是根据本申请一个实施方式的图像数据处理装置的结构示意图;4 is a schematic structural diagram of an image data processing apparatus according to an embodiment of the present application;
图5是根据本申请一个实施方式的图像识别装置的结构示意图;5 is a schematic structural diagram of an image recognition apparatus according to an embodiment of the present application;
图6是根据本申请一个实施方式的处理图像数据的方法的流程示意图;6 is a schematic flowchart of a method for processing image data according to an embodiment of the present application;
图7是用本申请实施方式中的划分策略划分采集区域的示意图;7 is a schematic diagram of dividing a collection area by a dividing strategy in an embodiment of the present application;
图8是图7中图像传感器采集区域的各个子采集区域的划分示意图;FIG. 8 is a schematic diagram of the division of each sub-collection area of the image sensor collection area in FIG. 7;
图9是根据本申请一个实施方式的处理图像传感器图像数据的方法中的特征提取和融合识别过程的流程示意图;以及9 is a schematic flowchart of a feature extraction and fusion identification process in a method for processing image sensor image data according to an embodiment of the present application; and
图10是现有技术和本申请另一个实施方式的图像数据处理方案的示意图,其中现有技术的方案位于上部,本申请的实施方式的方案位于下部;10 is a schematic diagram of an image data processing solution of the prior art and another embodiment of the present application, wherein the solution of the prior art is located at the upper part, and the solution of the embodiment of the present application is located at the lower part;
图11是现有技术和本申请另一个实施方式的图像数据处理方案的示意图,其中现有技术的方案位于上部,本申请的实施方式的方案位于下部;11 is a schematic diagram of an image data processing solution of the prior art and another embodiment of the present application, wherein the solution of the prior art is located at the upper part, and the solution of the embodiment of the present application is located at the lower part;
图12是根据本申请另一个实施方式的处理图像传感器图像数据的方法的流程示意图;12 is a schematic flowchart of a method for processing image sensor image data according to another embodiment of the present application;
图13是根据本申请一个实施方式的驾驶系统的结构示意图;13 is a schematic structural diagram of a driving system according to an embodiment of the present application;
图14是根据本申请一个实施方式的车辆的结构示意图;以及FIG. 14 is a schematic structural diagram of a vehicle according to an embodiment of the present application; and
图15是本申请实施方式的一种计算设备的结构示意图。FIG. 15 is a schematic structural diagram of a computing device according to an embodiment of the present application.
具体实施方式Detailed ways
<定义><definition>
在以下的描述中,所涉及的表示步骤的标号,如S101、S102……等,并不表示一定会按此步骤执行,在允许的情况下可以互换前后步骤的顺序,或同时执行。In the following description, the reference numerals representing steps, such as S101, S102, etc., do not necessarily mean that this step will be performed, and the order of the preceding and following steps may be interchanged or performed simultaneously if permitted.
说明书和权利要求书中的词语“第一、第二、第三”或“区域A、区域B、区域C”等类似用语,仅用于区别类似的对象,不指示针对对象的特定排序,可以理解地,在允许的情况下可以互换特定的顺序或先后次序,以使这里描述的本申请实方式能够以除了在这里图示或描述的以外的顺序实施。例如,本申请中使用术语“第一图像数据”和“第二图像数据”均为图像数据,但“第二图像数据”是对“第一图像 数据”进行图像处理得到的,因此使用“第一”和“第二”将这两者进行区分。The words "first, second, third" or "region A, region B, region C" and other similar terms in the description and claims are only used to distinguish similar objects, and do not indicate a specific ordering of objects, and may It is understood that the specific order or sequence may be interchanged, where permitted, to enable the embodiments of the application described herein to be practiced in sequences other than those illustrated or described herein. For example, the terms "first image data" and "second image data" used in this application are both image data, but the "second image data" is obtained by performing image processing on the "first image data", so the term "first image data" is used. One" and "Second" distinguish the two.
说明书和权利要求书中使用的术语“包括”不应解释为限制于其后列出的内容;它不排除其它的元件或步骤。因此,其应当诠释为指定所提到的所述特征、整体、步骤或部件的存在,但并不排除存在或添加一个或更多其它特征、整体、步骤或部件及其组群。因此,表述“包括装置A和B的设备”不应局限为仅由部件A和B组成的设备。The term "comprising" used in the description and claims should not be interpreted as being limited to what is listed thereafter; it does not exclude other elements or steps. Accordingly, it should be interpreted as specifying the presence of said features, integers, steps or components mentioned, but not excluding the presence or addition of one or more other features, integers, steps or components and groups thereof. Therefore, the expression "apparatus comprising means A and B" should not be limited to apparatuses consisting of parts A and B only.
说明书中提到的“一个实施方式”或“实施方式”意味着与该实施方式结合描述的特定特征、结构或特性包括在本申请的至少一个实施方式中。因此,在本说明书各处出现的用语“在一个实施方式中”或“在实施方式中”并不一定都指同一实施方式,但可以指同一实施方式。此外,在一个或多个实施方式中,能够以任何适当的方式组合各特定特征、结构或特性,如从本公开对本领域的普通技术人员显而易见的那样。Reference in the specification to "one embodiment" or "an embodiment" means that a particular feature, structure, or characteristic described in connection with the embodiment is included in at least one embodiment of the present application. Thus, the appearances of the terms "in one embodiment" or "in an embodiment" in various places in this specification are not necessarily all referring to the same embodiment, but may refer to the same embodiment. Furthermore, the particular features, structures or characteristics can be combined in any suitable manner, as would be apparent to one of ordinary skill in the art from this disclosure, in one or more embodiments.
说明书中使用的词语“可选的”、“可选地”、“可选”,指的是其所修饰的特征是在一些实施方式中是可以省略的,但存在于一些替代实施方式中。The words "optional", "optionally", and "optional" are used in the specification to mean that the feature they modify may be omitted in some embodiments, but in some alternative embodiments.
说明书和权利要求书中使用的术语“图像传感器(imaging sensor)”包括摄像机和激光雷达。图像传感器用于在一个扫描周期内扫描采集区域对应的物理区域可生成的多个图像数据。采集区域表示图像传感器的采集范围,摄像头的采集区域又称为靶面。激光雷达的采集区域又称为扫描区域。The term "imaging sensor" as used in the specification and claims includes cameras and lidars. The image sensor is used to scan a plurality of image data that can be generated from the physical area corresponding to the acquisition area in one scan cycle. The acquisition area represents the acquisition range of the image sensor, and the acquisition area of the camera is also called the target surface. The collection area of the lidar is also called the scanning area.
说明书和权利要求书中使用的术语“图像处理(image processing)”是指用计算设备对图像进行分析处理,以达到所需结果的技术。不同种类型的图像的图像数据适用的图像处理是不同的。例如,用于二维平面图像的图像数据的图像处理可以包括但不限于黑电平补偿(black level compensation)、镜头矫正(lens shading correction)、坏像素矫正(bad pixel correction)、颜色插值(demosaic)、贝叶斯域去噪、白平衡矫正(automatic white balance)、色彩矫正(color correction)、Gamma校正和色彩空间转换(RGB转换为YUV)。用于三维点云的图像数据的图像处理可以包括但不限于滤波(filter)、下采样(down sample)和去离群噪声(outlier removal)。因此,本申请的方法实施方式中涉及的“图像处理”可以包括一个或多个子图像处理,本申请的装置实施方式中涉及的“图像处理模块”可以包括一个或多个子图像处理模块。The term "image processing" as used in the description and claims refers to techniques of analyzing images with computing devices to achieve desired results. The image processing applicable to the image data of different types of images is different. For example, image processing of image data for two-dimensional planar images may include, but is not limited to, black level compensation, lens shading correction, bad pixel correction, demosaic ), Bayesian domain denoising, white balance correction (automatic white balance), color correction (color correction), Gamma correction and color space conversion (RGB to YUV). Image processing for the image data of the three-dimensional point cloud may include, but is not limited to, filtering, down sampling, and outlier removal. Therefore, the "image processing" involved in the method embodiments of the present application may include one or more sub-image processing, and the "image processing module" involved in the apparatus embodiments of the present application may include one or more sub-image processing modules.
说明书和权利要求书中使用的术语“特征提取(feature extraction)”或“提取特征数据(extraction of feature data)”是指从经过图像处理后的图像数据中构建出多种含有资讯性而不冗余的特征数据的技术。不同类型的图像的图像数据适用的特征提取是不同的。例如,用于二维平面图像的图像数据的特征提取可以包括但不限于卷积(convolution)和池化(pooling)。用于三维点云的图像数据的特征提取可以包括但不限于特征向量提取。The term "feature extraction" or "extraction of feature data" used in the description and claims refers to constructing a variety of informative but not redundant techniques for the remaining feature data. The feature extraction applicable to the image data of different types of images is different. For example, feature extraction for image data of two-dimensional planar images may include, but is not limited to, convolution and pooling. Feature extraction for image data of three-dimensional point clouds may include, but is not limited to, feature vector extraction.
说明书和权利要求书中使用的术语“融合识别(fusion and recognition)”是指将从图像数据中提取的所有特征数据融合为一整体后,基于该整体执行识别分析并输出图像识别结果的技术。不同种类型的图像的图像数据适用的融合识别是不同的。例如,用于二维平面图像的图像数据的融合识别可以包括但不限于特征融合(feature fusion)、全连接(fully connected)、输出(output)。用于三维点云的图像数据的融合识别 可以包括但不限于特征点匹配(feature point matching)。The term "fusion and recognition" used in the description and claims refers to a technology that, after all feature data extracted from image data is fused into a whole, performs recognition analysis based on the whole and outputs an image recognition result. The fusion recognition applicable to the image data of different types of images is different. For example, fusion recognition of image data for two-dimensional planar images may include, but is not limited to, feature fusion, fully connected, and output. Fusion identification of image data for 3D point clouds may include, but is not limited to, feature point matching.
说明书和权利要求书中使用的术语“特征融合(feature fusion)”是将从整个采集区域的图像数据的中提取的所有特征数据融合成一体的过程。本申请涉及的特征融合是得到图像识别结果前执行的早融合(early fusion),可以包括但不限于系列特征融合(concat)和并行融合(add)。The term "feature fusion" used in the specification and claims is a process of fusing all feature data extracted from image data of the entire acquisition area into one. The feature fusion involved in this application is an early fusion (early fusion) performed before the image recognition result is obtained, which may include but not limited to serial feature fusion (concat) and parallel fusion (add).
<技术背景><Technical Background>
目前,车载图像传感器中具有许多感知方式为“扫描式”的图像图像传感器。这种扫描式图像图像传感器不是在采集区域中同时生成一帧图像的所有第一图像数据,而是例如以按行或按列扫描采集区域的方式,在一段时间内依次生成一帧图像的所有第一图像数据。如图1所示,在现有的处理图像传感器图像数据的方法中,从t0时刻处开始,由图像传感器生成并传输一帧图像第一个第一图像数据,直到t1时刻处,该帧图像的最后一个第一图像数据被传输完成时,才会对整帧图像的所有第一图像数据执行图像处理。然后在t2时刻处,在对整帧图像的第一图像数据的图像处理完成时,才会开始对第二图像数据(即经过图像处理的第一图像数据)执行识别算法处理,直至t3时刻处完成整帧图像的识别算法处理。At present, there are many image sensors with a "scanning" sensing method in the vehicle-mounted image sensor. Instead of simultaneously generating all the first image data of a frame of image in the acquisition area, such a scanning image image sensor generates all the first image data of a frame of image sequentially over a period of time, for example by scanning the acquisition area in rows or columns. first image data. As shown in Figure 1, in the existing method for processing image sensor image data, starting from time t0, the image sensor generates and transmits the first first image data of a frame of image, until time t1, the frame of image The image processing is performed on all the first image data of the whole frame of image only when the last first image data of the frame is transmitted. Then at time t2, when the image processing of the first image data of the whole frame of image is completed, the recognition algorithm processing will be performed on the second image data (that is, the first image data after image processing) until time t3. Complete the recognition algorithm processing of the whole frame image.
本申请的发明人发现,这样的处理图像传感器图像数据的方法具有如下缺陷。图像处理流程必须等到图像传感器扫描整个采集区域生成所有第一图像数据并将它们传输完成时才能开始,图像识别流程也相应地必须等到所有第一图像数据被图像处理而转变为第二图像数据才能开始。这些等待时间各自拖延了图像传感器图像数据的处理流程,导致从图像传感器生成第一图像数据到车辆执行器执行操作的端到端时延较大。换句话说,由于图像处理和识别算法处理是以整帧图像的第一或第二图像数据为单位执行的,无法在一边接收该帧的第一或第二图像数据时一边对已被接收的第一或第二图像数据进行处理,这降低了处理图像传感器图像数据的速度,增加了端到端时延。The inventors of the present application found that such a method of processing image sensor image data has the following drawbacks. The image processing process must wait until the image sensor scans the entire acquisition area to generate all the first image data and transfer them, and the image recognition process must wait until all the first image data has been image-processed and transformed into the second image data. start. Each of these waiting times delays the processing flow of the image sensor image data, resulting in a large end-to-end delay from the generation of the first image data by the image sensor to the execution of the operation by the vehicle actuator. In other words, since the image processing and the recognition algorithm processing are performed in units of the first or second image data of the whole frame of image, it is impossible to process the received first or second image data while receiving the first or second image data of the frame. The first or second image data is processed, which reduces the speed of processing the image sensor image data and increases the end-to-end latency.
本申请的发明人发现,这样的处理图像传感器图像数据的方法具有如下缺陷。图像处理流程必须等到图像传感器扫描整个采集区域生成所有第一图像数据并将它们传输完成时才能开始,图像识别流程也相应地必须等到所有第一图像数据被图像处理而转变为第二图像数据才能开始。这些等待时间各自拖延了图像传感器图像数据的处理流程,导致从图像传感器生成第一图像数据到车辆执行器执行操作的端到端时延较大。换句话说,由于图像处理和识别算法处理是以整帧图像的第一或第二图像数据为单位执行的,无法在一边接收该帧的第一或第二图像数据时一边对已被接收的第一或第二图像数据进行处理,这降低了处理图像传感器图像数据的速度,增加了端到端时延。The inventors of the present application found that such a method of processing image sensor image data has the following drawbacks. The image processing process must wait until the image sensor scans the entire acquisition area to generate all the first image data and transfer them, and the image recognition process must wait until all the first image data has been image-processed and transformed into the second image data. start. Each of these waiting times delays the processing flow of the image sensor image data, resulting in a large end-to-end delay from the generation of the first image data by the image sensor to the execution of the operation by the vehicle actuator. In other words, since the image processing and the recognition algorithm processing are performed in units of the first or second image data of the whole frame of image, it is impossible to process the received first or second image data while receiving the first or second image data of the frame. The first or second image data is processed, which reduces the speed of processing the image sensor image data and increases the end-to-end latency.
<技术构思><Technical idea>
有鉴于此,本申请的一个实施方式提供一种处理图像传感器图像数据的方法,包括:接收来自图像传感器的第一图像数据,所述第一图像数据为所述图像传感器在一个扫描周期内扫描采集区域对应的物理区域可生成的多个图像数据中一个图像数据,所述采集区域表示所述图像传感器的采集范围;对所述第一图像数据进行图像处理得到第二图像数据;以及输出所述第二图像数据。In view of this, an embodiment of the present application provides a method for processing image data of an image sensor, including: receiving first image data from an image sensor, where the first image data is scanned by the image sensor within one scan period One image data among a plurality of image data that can be generated by the physical area corresponding to the acquisition area, the acquisition area represents the acquisition range of the image sensor; image processing is performed on the first image data to obtain second image data; the second image data.
以下结合图2进一步说明本申请的处理图像传感器图像数据的方法的构思。为了直观起见,图2在相同的时间轴中示出了现有方案(位于图2上部)和本申请方案(位于图2下部)。The concept of the method for processing image sensor image data of the present application is further described below with reference to FIG. 2 . For the sake of intuition, FIG. 2 shows the existing solution (located in the upper part of FIG. 2 ) and the solution of the present application (located in the lower part of FIG. 2 ) in the same time axis.
在图2下部示出的本申请的处理图像传感器图像数据的方法的实施方式中,图像传感器扫描其采集区域并依次生成多个图像数据。在图像传感器的扫描区域被预先划分为三个子采集区域A、B和C的情况下,将当接收到由扫描子采集区域A生成的所有图像数据时(即t11时刻处),即在接收到由扫描整个采集区域生成的所有图像数据(即t1时刻处)之前,本申请的处理方法就开始执行图像处理。这将处理流程从现有技术中的t1时刻提前至本申请的t11时刻,由此使得本申请中的图像处理结束时刻相比于现有方案中的图像处理结束时刻提前了,因而最后算法处理的结束时刻也必然会提前,从而能够更快地得到图像识别结果,最终缩短端到端时延。In the embodiment of the method for processing image sensor image data of the present application shown in the lower part of FIG. 2 , the image sensor scans its acquisition area and sequentially generates a plurality of image data. In the case where the scanning area of the image sensor is pre-divided into three sub-acquisition areas A, B and C, when all the image data generated by the scanning sub-acquisition area A is received (ie at time t11 ), that is, when receiving The processing method of the present application starts to perform image processing before all image data generated by scanning the entire acquisition area (ie, at time t1). This advances the processing flow from time t1 in the prior art to time t11 in the present application, so that the end time of image processing in the present application is earlier than the end time of image processing in the existing solution, so the final algorithm processing The end time of the image recognition will inevitably be advanced, so that the image recognition results can be obtained faster, and the end-to-end delay can be shortened.
<实施场景><Implementation Scenario>
本申请的处理图像传感器图像数据的方法可以适用于处理各种“扫描式”图像传感器生成的图像数据,并且能够应用于具有这种图像传感器的车辆、机器人等设备。The method for processing image sensor image data of the present application can be applied to processing image data generated by various "scanning" image sensors, and can be applied to vehicles, robots and other devices having such image sensors.
例如,在具有摄像机的运动机器人场景中,机器人利用摄像机来感知环境并根据摄像机的图像数据的图像识别结果来规划并执行相应运动。当这样的机器人应用本申请的处理图像传感器图像数据的方法,能够更加迅速地得到图像识别结果,从而更快地应对该结果做出相应运动。因此,采用本申请的处理图像传感器图像数据的方法能够使得运动机器人的动作更加敏捷。For example, in a motion robot scene with a camera, the robot uses the camera to perceive the environment and plan and execute the corresponding movement according to the image recognition results of the camera's image data. When such a robot applies the method for processing image sensor image data of the present application, an image recognition result can be obtained more quickly, so that a corresponding movement can be made in response to the result more quickly. Therefore, using the method for processing image sensor image data of the present application can make the motion of the motion robot more agile.
例如,在具有激光雷达的智能网联车场景中,车辆利用激光雷达来感知路况并根据激光雷达的点云图像数据的图像识别结果来规划并执行相应自动驾驶操作。当这样的机器人应用本申请的处理图像传感器图像数据的方法,能够更加迅速地得到图像识别结果,从而更快地应对该结果做出相应驾驶操作。因此,采用本申请的处理图像传感器图像数据的方法能够使得智能网联车的自动驾驶更加安全。For example, in an ICV scenario with lidar, the vehicle uses lidar to perceive road conditions and plan and execute corresponding autonomous driving operations based on the image recognition results of lidar's point cloud image data. When such a robot applies the method for processing image sensor image data of the present application, the image recognition result can be obtained more quickly, so that the corresponding driving operation can be performed in response to the result more quickly. Therefore, using the method for processing image sensor image data of the present application can make the automatic driving of the intelligent networked vehicle safer.
此外,虽然本申请的处理图像传感器图像数据的方法是针对处理一帧图像的方案进行改进,缩短处理一帧图像的时间,从而缩短端到端时延。应当理解,本申请的处理图像传感器图像数据的方法还适用于处理包括多帧图像的视频。通过分别对视频中的各帧图像采用本申请的处理图像传感器图像数据的方法,可以缩短视频的处理时长。In addition, although the method for processing image sensor image data of the present application is to improve the solution for processing one frame of image, the time for processing one frame of image is shortened, thereby shortening the end-to-end delay. It should be understood that the method of processing image sensor image data of the present application is also applicable to processing video including multiple frames of images. By using the method for processing image sensor image data of the present application for each frame of images in the video, the processing time of the video can be shortened.
<实施方式><Embodiment>
以下将参照图2至图9对本申请的图像传感器图像数据处理系统的一个实施方式和处理图像数据的方法的一个实施方式执行详细的说明。在图3中,用虚线框表示装置的可选模块,即该可选模块在其他一些实施方式中是可以省略的。One embodiment of the image sensor image data processing system and one embodiment of the method of processing image data of the present application will be described in detail below with reference to FIGS. 2 to 9 . In FIG. 3 , optional modules of the device are represented by dashed boxes, that is, the optional modules may be omitted in other embodiments.
图3示例性地示出了根据本申请一个实施方式的图像传感器图像数据处理系统1001,包括相互连接的图像数据处理装置1100和图像识别装置1200。FIG. 3 exemplarily shows an image sensor image data processing system 1001 according to an embodiment of the present application, including an image data processing apparatus 1100 and an image recognition apparatus 1200 that are connected to each other.
图3以虚线框示出了可选的划分管理模块1300。划分管理模块1300分别与图像数据处理装置1100和图像识别装置1200连接,用于向这两者提供划分策略。划分策略可以用于预先设定图像传感器的采集区域的多个子采集区域的大小和数量。 在本实施方式中,划分管理模块1300提供的划分策略能够使得采集区域被等分为3个子采集区域A、B和C。Figure 3 shows an optional partition management module 1300 in a dashed box. The division management module 1300 is respectively connected with the image data processing apparatus 1100 and the image recognition apparatus 1200, and is used for providing division policies to both. The division strategy may be used to preset the size and quantity of the multiple sub-capturing regions of the image sensor's capturing region. In this embodiment, the division strategy provided by the division management module 1300 enables the collection area to be divided into three sub-collection areas A, B, and C equally.
图3还额外示出了图像传感器2000。图像传感器2000扫描采集区域依次生成构成一帧图像的多个第一图像数据,并向本申请的图像传感器图像数据传输处理系统1001中的图像数据处理装置1100依次发送多个第一图像数据。FIG. 3 additionally shows image sensor 2000 . The image sensor 2000 scans the acquisition area to sequentially generate multiple first image data constituting one frame of image, and sequentially sends the multiple first image data to the image data processing device 1100 in the image sensor image data transmission processing system 1001 of the present application.
如图4所示,图像数据处理装置1100包括相互连接的接收模块1110和图像处理模块1120。As shown in FIG. 4 , the image data processing apparatus 1100 includes a receiving module 1110 and an image processing module 1120 which are connected to each other.
在本实施方式中,图像数据处理装置1100的接收模块1110用于接收图像传感器2000发送的多个第一图像数据,并且还可以用于接收划分管理模块1300提供的划分策略。In this embodiment, the receiving module 1110 of the image data processing apparatus 1100 is configured to receive a plurality of first image data sent by the image sensor 2000 , and may also be configured to receive a division strategy provided by the division management module 1300 .
在本实施方式中,由于图像处理模块1120依据接收模块1110接收到的划分策略将采集区域预先划分为3个大小相等的子采集区域A、B和C,因此接收模块1110接收到的第一图像数据可以被分为3组的第一图像数据,使得图像传感器扫描3个不同子采集区域时分别生成一组第一图像数据。图像处理模块1120由此可以进一步用于当已接收到一组第一图像数据时,对该组第一图像数据进行图像处理。In this embodiment, since the image processing module 1120 pre-divides the acquisition area into three sub-acquisition areas A, B and C of equal size according to the division strategy received by the receiving module 1110, the first image received by the receiving module 1110 The data can be divided into 3 groups of first image data, so that when the image sensor scans the 3 different sub-acquisition regions, a group of first image data is respectively generated. The image processing module 1120 can thus be further configured to perform image processing on a set of first image data when the set of first image data has been received.
具体地,参照图2,图像处理模块1120用于在t11时刻处,当接收模块1110接收到生成自子采集区域A的所有第一图像数据时,就开始对这些图像数据执行图像处理。然后,在t12时刻处,当接收模块1110接收到生成自子采集区域B的所有第一图像数据时,图像处理模块1120就开始对这些图像数据执行图像处理。最后在t13时刻处,当接收模块1110接收到生成自子采集区域C的所有第一图像数据时,图像处理模块1120就开始对这些图像数据执行图像处理。Specifically, referring to FIG. 2 , the image processing module 1120 is configured to start performing image processing on the image data when the receiving module 1110 receives all the first image data generated from the sub-collection area A at time t11. Then, at time t12, when the receiving module 1110 receives all the first image data generated from the sub-collection area B, the image processing module 1120 starts to perform image processing on these image data. Finally, at time t13, when the receiving module 1110 receives all the first image data generated from the sub-collection area C, the image processing module 1120 starts to perform image processing on these image data.
在本申请中,图像处理模块1120可以包括多个图像处理子模块(未示出)。类似地,图像处理可以包括多个图像子处理。在实际中,图像数据处理装置1100例如可以是图像信号处理器。In this application, the image processing module 1120 may include a plurality of image processing sub-modules (not shown). Similarly, image processing may include multiple image sub-processing. In practice, the image data processing apparatus 1100 may be, for example, an image signal processor.
如图5所示,图像识别装置1200包括依次连接的接收模块1210、特征提取模块1220和融合识别模块1230。As shown in FIG. 5 , the image recognition apparatus 1200 includes a receiving module 1210 , a feature extraction module 1220 and a fusion recognition module 1230 which are connected in sequence.
在本实施方式中,图像识别装置1200的接收模块1210用于接收多个第二图像数据。多个第二图像数据是图像数据处理装置1100对多个第一图像数据进行图像处理得到的,而多个第一图像数据是图像传感器3000扫描采集区域所依次生成的。此外,接收模块1210还用于接收划分管理模块1300提供的划分策略。In this embodiment, the receiving module 1210 of the image recognition apparatus 1200 is configured to receive a plurality of second image data. The plurality of second image data are obtained by performing image processing on the plurality of first image data by the image data processing apparatus 1100, and the plurality of first image data are sequentially generated by the image sensor 3000 scanning the acquisition area. In addition, the receiving module 1210 is further configured to receive the partition policy provided by the partition management module 1300 .
在本实施方式中,特征提取模块1220用于在所述接收模块接收所述多个第二图像数据期间,依次从已接收的第二图像数据提取特征数据。由于特征提取模块1220依据接收模块1210接收到的划分策略对第二图像数据分组,即将采集区域预先划分为3个大小相等的子采集区域A、B、C,因此接收模块1210接收到的第二图像数据也可以类似地被分为3组不同的第二图像数据,使得源自于子采集区域A、B、C的各第二图像数据分别属于不同组的第二图像数据。特征提取模块1220用于当已接收到一组第二图像数据时,从该组第二图像数据组提取特征数据。In this embodiment, the feature extraction module 1220 is configured to sequentially extract feature data from the received second image data during the period when the receiving module receives the plurality of second image data. Since the feature extraction module 1220 groups the second image data according to the division strategy received by the receiving module 1210, that is, the acquisition area is pre-divided into three sub-acquisition areas A, B, and C of equal size, so the second image data received by the receiving module 1210 The image data can also be similarly divided into 3 different sets of second image data, so that each second image data originating from the sub-acquisition areas A, B, C belong to different sets of second image data, respectively. The feature extraction module 1220 is configured to extract feature data from a set of second image data when the set of second image data has been received.
具体地,参照图2,特征提取模块1220用于在接收模块1210接收到源自于子采集区域A的一组第二图像数据时,就开始一起从该组第二图像数据中提取特征 数据。换句话说,当生成自子采集区域A的所有第一图像数据经过图像处理转变为相应的各第二图像数据并被接收模块1210接收时,特征提取模块1220就开始一起从这些第二图像数据中提取特征数据。随后,特征提取模块1220以类似的方式分别对源自于子采集区域B和子采集区域C的第二图像数据执行特征提取。Specifically, referring to FIG. 2 , when the receiving module 1210 receives a set of second image data from the sub-collection area A, the feature extraction module 1220 starts to extract feature data from the set of second image data together. In other words, when all the first image data generated from the sub-collection area A are transformed into corresponding second image data through image processing and received by the receiving module 1210, the feature extraction module 1220 starts to extract the second image data from these second image data together. Extract feature data. Subsequently, the feature extraction module 1220 performs feature extraction on the second image data originating from the sub-acquisition area B and the sub-acquisition area C, respectively, in a similar manner.
在本实施方式中,融合识别模块1230用于在特征提取模块1220分别从源自与子采集区域A、B和C的所有第二图像数据提取特征数据之后,将所有特征数据先融合为整体,然后在对该整体进行识别,最后在t4时刻处输出图像识别结果。In this embodiment, the fusion identification module 1230 is configured to first fuse all the characteristic data into a whole after the characteristic extraction module 1220 extracts characteristic data from all the second image data originating from the sub-collection areas A, B, and C, respectively, Then the whole is recognized, and finally the image recognition result is output at time t4.
由于特征提取模块1220和融合识别模块1230均涉及对图像的识别处理,因此在实际中,通常被设定在同一实体中,例如设定在车辆的算法平台中。Since both the feature extraction module 1220 and the fusion recognition module 1230 are involved in image recognition processing, in practice, they are usually set in the same entity, for example, in the algorithm platform of the vehicle.
图像处理模块1120和特征提取模块1220在本实施方式中分别对生成自一子采集区域的第一图像数据或源自一子采集区域的第二图像数据进行处理,即它们的处理对象是来自于一个子采集区域的第一或第二图像数据。本实施方式的融合识别模块1230则用于将从源自于整个采集区域的第二图像数据的所有特征数据作为整体来执行融合识别,即其处理对象是源自于整个采集区域的所有第二图像数据的特征数据。In this embodiment, the image processing module 1120 and the feature extraction module 1220 respectively process the first image data generated from a sub-acquisition area or the second image data from a sub-acquisition area, that is, their processing objects are from First or second image data of a sub-acquisition area. The fusion recognition module 1230 of this embodiment is configured to perform fusion recognition from all the feature data of the second image data originating from the entire collection area as a whole, that is, the processing object is all the second image data originating from the whole collection area. Feature data for image data.
此外,特征提取模块1220可以包括一个或多个卷积层(未示出)和一个或多个池化层(未示出)。融合识别模块1230可以包括一个或多个特征融合层(未示出)、一个或多个全连接层(未示出)和输出层(未示出)。这些层也可以各自包括一个或多个子层(未示出)。将在以下对于特征提取模块1220和融合识别模块1230中各种层执行的各个步骤执行更进一步的描述,因此这些模块内的各种层的结构将在以下的相关描述中变得更加清楚。Additionally, the feature extraction module 1220 may include one or more convolutional layers (not shown) and one or more pooling layers (not shown). The fusion recognition module 1230 may include one or more feature fusion layers (not shown), one or more fully connected layers (not shown), and an output layer (not shown). These layers may also each include one or more sublayers (not shown). The various steps performed by the various layers in the feature extraction module 1220 and the fusion identification module 1230 will be further described below, so the structure of the various layers within these modules will become clearer in the related description below.
以下结合上述参考图3至图5的描述,参考图6对本申请的处理图像传感器图像数据的方法的一个实施方式进行详细说明。An embodiment of the method for processing image sensor image data of the present application will be described in detail below with reference to FIG. 6 in conjunction with the above descriptions with reference to FIGS. 3 to 5 .
图6示例性地示出了根据本申请一个实施方式的处理图像数据的方法的流程示意图,其包括以下步骤S101至S107:FIG. 6 exemplarily shows a schematic flowchart of a method for processing image data according to an embodiment of the present application, which includes the following steps S101 to S107:
在步骤S101中,划分管理模块1300提供划分策略。具体地,该划分策略用于预先设定采集区域的子采集区域的数量为3并用于预先设定各个子采集区域的大小为彼此相等。In step S101, the division management module 1300 provides a division policy. Specifically, the division strategy is used to preset the number of sub-collection areas of the collection area to 3 and to preset the size of each sub-collection area to be equal to each other.
在步骤S102中,图像数据处理装置1100和图像识别装置1200分别接收划分策略,并分别依据划分策略预先设定采集区域的各子采集区域,从而分别对第一图像数据和第二图像数据分组。具体地,图像数据处理装置1100和图像识别装置1200的接收模块1110和图像识别装置1200的接收模块1210分别接收划分策略,相应的图像处理模块1120和特征提取模块1220分别依据划分策略将采集区域划分成第一子采集区域和第二子采集区域,并且将第一和第二子采集区域预先设定为大小彼此相等子采集区域A、B和C。换句话说,由于所采用的划分策略相同,图像数据处理装置1100的和识别装置1200的第二子采集区域均为子采集区域A、B和C。In step S102, the image data processing apparatus 1100 and the image recognition apparatus 1200 respectively receive the division strategy, and pre-set each sub-collection area of the collection area according to the division policy, thereby grouping the first image data and the second image data respectively. Specifically, the image data processing device 1100 and the receiving module 1110 of the image recognition device 1200 and the receiving module 1210 of the image recognition device 1200 receive the division strategy respectively, and the corresponding image processing module 1120 and the feature extraction module 1220 divide the collection area according to the division strategy. A first sub-collection area and a second sub-collection area are formed, and the first and second sub-collection areas are preset as sub-collection areas A, B, and C of equal size to each other. In other words, since the adopted division strategy is the same, the second sub-acquisition areas of the image data processing apparatus 1100 and the recognition apparatus 1200 are all sub-acquisition areas A, B and C.
在步骤S103中,图像数据处理装置1100接收图像传感器扫描采集区域所依次生成的多个第一图像数据。具体地,由图像数据处理装置1100的接收模块1110接收图像传感器3000扫描采集区域依次生成并依次发送的多个第一图像数据。In step S103, the image data processing apparatus 1100 receives a plurality of first image data sequentially generated by the image sensor scanning the acquisition area. Specifically, the receiving module 1110 of the image data processing apparatus 1100 receives a plurality of first image data sequentially generated and sent by the image sensor 3000 by scanning the acquisition area.
在步骤S104中,图像数据处理装置1100在已接收到生成自一第一子采集 区域的一组第一图像数据时,对该组第一图像数据执行图像处理,输出相应的各第二图像数据。具体地,图像处理模块1120在接收模块1110接收到生成自一第一子采集区域的一组第一图像数据时,对该组第一图像数据执行图像处理,并依次输出对该组第一图像数据进行所述图像处理得到的相应的各第二图像数据。In step S104, when the image data processing apparatus 1100 has received a set of first image data generated from a first sub-collection area, it performs image processing on the set of first image data, and outputs corresponding second image data . Specifically, when the receiving module 1110 receives a set of first image data generated from a first sub-collection area, the image processing module 1120 performs image processing on the set of first image data, and sequentially outputs the set of first images The corresponding second image data obtained by performing the image processing.
在步骤S105中,图像识别装置1200接收图像数据处理装置1100依次输出的多个第二图像数据。具体地,由图像识别装置1200的接收模块1210接收图像数据处理装置1100依次输出的多个第二图像数据。In step S105, the image recognition apparatus 1200 receives a plurality of second image data sequentially output by the image data processing apparatus 1100. Specifically, the receiving module 1210 of the image recognition apparatus 1200 receives a plurality of second image data sequentially output by the image data processing apparatus 1100 .
在步骤S106中,图像识别装置1200在已接收到源自一第二子采集区域的一组第二图像数据时,从该组第二图像数据提取特征数据。具体地,图像识别装置1200的特征提取模块1220在接收模块1210接收到源自一第二子采集区域的所有第二图像数据时,一起从这些图像数据中提取特征数据。In step S106, the image recognition apparatus 1200 extracts feature data from a set of second image data originating from a second sub-capture area when the image recognition apparatus 1200 has received the set of second image data. Specifically, when the receiving module 1210 receives all the second image data originating from a second sub-collection area, the feature extraction module 1220 of the image recognition apparatus 1200 extracts feature data from the image data together.
在步骤S107中,图像识别装置1200对从多个第二图像数据提取的特征数据进行融合识别处理,输出图像识别结果。具体地,在图像识别装置1200的特征提取模块1220分别完成从源自于子采集区域A、B和C第二图像数据提取特征数据后,图像识别装置1200的融合识别模块1230将所有特征数据融合为一体,然后对该整体进行识别,最后输出图像识别结果。In step S107, the image recognition device 1200 performs fusion recognition processing on the feature data extracted from the plurality of second image data, and outputs an image recognition result. Specifically, after the feature extraction module 1220 of the image recognition device 1200 completes the extraction of feature data from the second image data originating from the sub-collection areas A, B, and C, respectively, the fusion recognition module 1230 of the image recognition device 1200 fuses all the feature data As a whole, then identify the whole, and finally output the image recognition result.
应当理解,上述步骤S101至S107并非按时间顺序排列。例如,在本实施方式中,当步骤S103中接收模块1110正在接收由子采集区域B生成的第一图像数据时,步骤S104中图像处理模块1120可以对由子采集区域A生成的第一图像数据进行图像处理。因此,步骤S103和步骤S104可以在时间上重合。类似地,步骤S105和步骤S105可以在时间上重合。此外,在本实施方式中,由于采集区域被划分为3个子采集区域,随着图像传感器按顺序扫描子采集区域A、B和C,步骤S104和步骤S106相应地重复了3次,以分别处理生成自各子采集区域的第一图像数据和源自各子采集区域的第二图像数据,如图2所示。It should be understood that the above steps S101 to S107 are not arranged in chronological order. For example, in this embodiment, when the receiving module 1110 is receiving the first image data generated by the sub-collection area B in step S103, the image processing module 1120 in step S104 may perform image processing on the first image data generated by the sub-collection area A deal with. Therefore, step S103 and step S104 may overlap in time. Similarly, step S105 and step S105 may coincide in time. In addition, in this embodiment, since the acquisition area is divided into 3 sub-acquisition areas, as the image sensor scans the sub-acquisition areas A, B, and C in sequence, steps S104 and S106 are correspondingly repeated 3 times for processing respectively First image data from each sub-acquisition area and second image data from each sub-acquisition area are generated, as shown in FIG. 2 .
此外,应当理解,图像处理步骤S104和特征提取步骤S106在本实施方式中的执行对象为生成自或源自采集区域的一个子采集区的第一或第二图像数据。而融合识别步骤S107在本实施方式中的执行对象为从源自整个采集区域的第二图像数据提取的所有特征数据。In addition, it should be understood that the execution object of the image processing step S104 and the feature extraction step S106 in this embodiment is the first or second image data generated from or originating from a sub-acquisition area of the acquisition area. In this embodiment, the execution object of the fusion identification step S107 is all feature data extracted from the second image data originating from the entire collection area.
再次参考图2,可以知道,上述系统和方法实施方式相比于现有方案所实现的时延收益ΔT可以如此计算:Referring to FIG. 2 again, it can be known that the delay benefit ΔT achieved by the above system and method implementations compared to the existing solution can be calculated as follows:
ΔT=t3-t4              式(1);ΔT=t3-t4 Equation (1);
以t1时刻作为基准,式(1)可以转换为:Taking time t1 as the benchmark, equation (1) can be converted into:
ΔT=(t3-t1)-(t2-t1)/3-T0/3-T1        式(2);ΔT=(t3-t1)-(t2-t1)/3-T0/3-T1 Formula (2);
t3=t2+T0+T1,由此转换式(2),可得t3=t2+T0+T1, thus converting formula (2), we can get
ΔT=2/3*(t2-t1+T0)         式(3);ΔT=2/3*(t2-t1+T0) Equation (3);
其中,t1是现有方案中对生成自整个采集区域的第一图像数据进行图像处理的开始时刻,t1也是本申请的本实施方式中对生成自子采集区域C的第一图像数据进行图像处理的开始时刻;t2是现有方案中图像处理结束、算法处理开始的时刻;t3是现有方案中算法处理结束的时刻;t4是本申请的本实施方式中融合识别结束的时刻; T0是现有方案中属于算法处理一部分的特征提取的时长,也是本申请的本实施方式中特征处理的时长;以及T1是现有方案中属于算法处理一部分的融合识别的时长,也是本申请的本实施方式中融合识别的时长。Wherein, t1 is the start time of performing image processing on the first image data generated from the entire collection area in the existing solution, and t1 is also the image processing time on the first image data generated from the sub-collection area C in this embodiment of the present application t2 is the moment when the image processing ends and the algorithm processing starts in the existing scheme; t3 is the moment when the algorithm processing ends in the existing scheme; t4 is the moment when the fusion recognition ends in the present embodiment of the present application; T0 is the current The duration of feature extraction that belongs to a part of algorithm processing in the scheme is also the duration of feature processing in this embodiment of the present application; and T1 is the duration of fusion recognition that belongs to a part of algorithm processing in the existing scheme, which is also the duration of this embodiment of the present application The duration of fusion recognition in .
通过图2和式(3)可以清楚知道,本申请的本实施方式中的时延收益一方面得益于图像处理流程的提前,即在本实施方式中,在t1时刻之前已经开始对一帧图像的2/3的第一图像数据进行图像处理,从而降低端到端时延。本申请的本实施方式中的时延收益另一方面得益于图像识别流程的提前,即在本实施方式中,在t1时刻之前已经开始对该帧图像的2/3的第二图像数据进行特征提取,从而进一步降低了端到端时延。It can be clearly seen from FIG. 2 and Equation (3) that, on the one hand, the delay benefit in this embodiment of the present application benefits from the advance of the image processing flow, that is, in this embodiment, the processing of one frame has already started before time t1. Image processing is performed on 2/3 of the first image data of the image, thereby reducing the end-to-end delay. On the other hand, the delay benefit in this embodiment of the present application benefits from the advance of the image recognition process, that is, in this embodiment, 2/3 of the second image data of the frame image has been processed before time t1. feature extraction, thereby further reducing the end-to-end delay.
以下参照图7-9对本申请的系统和方法实施方式中的一些模块和步骤进行更详细的说明。Some modules and steps in the system and method embodiments of the present application will be described in more detail below with reference to FIGS. 7-9 .
图7示例性地示出了用划分管理模块1300提供的划分策略的划分图像传感器3000的采集区域的示意图。以摄像机为例,如图7所示,一帧图像是由摄像机在其靶面(在本发明中也称为采集区域)内从上到下按行曝光生成到的。根据划分管理模块1300提供的划分策略,该采集区域在从上到下的方向上被划分为3个矩形的子采集区域A、B和C,并且各个子采集区域的尺寸被设定为相等。FIG. 7 exemplarily shows a schematic diagram of dividing the acquisition area of the image sensor 3000 using the division strategy provided by the division management module 1300 . Taking a camera as an example, as shown in FIG. 7 , a frame of image is generated by the camera in its target surface (also referred to as an acquisition area in the present invention) by line exposure from top to bottom. According to the division strategy provided by the division management module 1300, the collection area is divided into three rectangular sub-collection areas A, B, and C in the direction from top to bottom, and the sizes of each sub-collection area are set to be equal.
在图7中,图像传感器最先扫描子采集区域A并因此最先生成该子采集区域A的第一图像数据,然后最先向图像传感器图像数据处理系统1001传输生成自该子采集区域A的第一图像数据。图像传感器图像数据处理系统1001因此最先对生成自该子采集区域A的图像数据进行图像处理和特征提取,然后是对生成自子采集区域B的第一图像数据进行处理,最后是对生成自子采集区域C的第一图像数据进行处理。在此之后,还可以以相同的划分策略划分摄像机的采集区域,以用于处理后续扫描生成的第二帧图像,直至第Z帧图像,Z为大于2的任意整数。也就是说,本申请可以处理多帧图像,因此可以处理由多帧图像构成的对象,例如视频。In FIG. 7 , the image sensor scans the sub-acquisition area A first and thus generates the first image data of the sub-acquisition area A first, and then transmits the image data generated from the sub-acquisition area A to the image sensor image data processing system 1001 first. first image data. The image sensor image data processing system 1001 therefore firstly performs image processing and feature extraction on the image data generated from the sub-acquisition area A, then processes the first image data generated from the sub-acquisition area B, and finally processes the image data generated from the sub-acquisition area B. The first image data of the sub-acquisition area C is processed. After that, the acquisition area of the camera can also be divided by the same division strategy for processing the second frame of images generated by subsequent scans, up to the Z-th frame image, where Z is any integer greater than 2. That is to say, the present application can process images of multiple frames, and thus can process objects composed of images of multiple frames, such as videos.
图8是图7中图像传感器采集区域的各个子采集区域的划分示意图。假设由图7示出的采集区域的分辨率为1920*1080。将采集区域的左上角设定为假想坐标原点,各个子采集区域A、B和C就可以容易地被矩形的左上角、左下角、右上角和右下角的各坐标限定,具体坐标参见表1和图8。FIG. 8 is a schematic diagram of division of each sub-collection area of the image sensor collection area in FIG. 7 . It is assumed that the resolution of the acquisition area shown by FIG. 7 is 1920*1080. The upper left corner of the collection area is set as the imaginary coordinate origin, and each sub-collection area A, B and C can be easily defined by the coordinates of the upper left corner, lower left corner, upper right corner and lower right corner of the rectangle. Please refer to Table 1 for the specific coordinates. and Figure 8.
表1Table 1
左上角的坐标 右上角的坐标 左下角的坐标 右下角的坐标The coordinates of the upper left corner The coordinates of the upper right corner The coordinates of the lower left corner The coordinates of the lower right corner
子采集区域A(0,0) (1919,0) (0,359) (1919,359)Sub-collection area A(0,0) (1919,0) (0,359) (1919,359)
子采集区域B(0,360) (1919,360) (0,719) (1919,719)Sub-collection area B(0,360) (1919,360) (0,719) (1919,719)
子采集区域C(0,720) (1919,720) (0,1079) (1919,1079)Sub-collection area C(0,720) (1919,720) (0,1079) (1919,1079)
这样就能够以简单灵活的方式限定采集区域的各个子采集区域,方便根据实际情况对子采集区域的数量和各个子采集区域的大小进行调整。In this way, each sub-collection area of the collection area can be limited in a simple and flexible manner, and it is convenient to adjust the number of the sub-collection areas and the size of each sub-collection area according to the actual situation.
下面参考图9进一步描述特征提取步骤S106和融合识别步骤S107以及相关步骤执行体特征提取模块1220和融合识别模块1230的结构。The structure of the feature extraction step S106 and the fusion identification step S107 and the related step executor feature extraction module 1220 and the fusion identification module 1230 are further described below with reference to FIG. 9 .
图9示例性地示出了根据本申请一个实施方式的处理图像数据的方法中的特征提取和融合识别流程示意图。涉及特征提取步骤S106和涉及融合识别步骤S107 的部分分别用不同的虚线框标识。其中,conv1指示可以由特征提取模块1220的第一卷积层执行的第一卷积处理,conv2指示可以由特征提取模块的第二卷积层执行的第二卷积处理,conv3指示可以由特征提取模块的第三卷积层执行的第三卷积处理;类似地,pool1至pool3各自指示可以分别由特征提取模块1220的第一至第三池化层执行的第一至第三池化处理,fc1和fc2各自指示可以分别由融合识别模块1230的第一和第二全连接层执行的第一和第二全连接处理。concat指示可以由融合识别模块1230的系列特征融合层执行的系列特征融合处理。output指示可以由融合识别模块1230的输出层执行图像识别结果的输出。此外,在图9中,conv2_1指示可以由第一卷积层的第一子卷积层执行的第一卷积处理的第一子卷积处理,conv2_2指示可以由第一卷积层的第二子卷积层执行的第一卷积处理的第二子卷积处理。类似地,conv3_1至conv3_3各自指示可以分别由第三卷积层的第一至第三子卷积层执行的第三卷积处理的第一至第三子卷积处理。FIG. 9 exemplarily shows a schematic diagram of a flow chart of feature extraction and fusion recognition in a method for processing image data according to an embodiment of the present application. The parts involved in the feature extraction step S106 and the part involved in the fusion identification step S107 are respectively marked with different dashed boxes. Wherein, conv1 indicates the first convolution process that can be performed by the first convolution layer of the feature extraction module 1220, conv2 indicates the second convolution process that can be performed by the second convolution layer of the feature extraction module, and conv3 indicates that the feature The third convolution process performed by the third convolution layer of the extraction module; similarly, pool1 to pool3 each indicate the first to third pooling processes that can be performed by the first to third pooling layers of the feature extraction module 1220, respectively , fc1 and fc2 each indicate the first and second fully connected processes that may be performed by the first and second fully connected layers of the fusion recognition module 1230, respectively. concat indicates a series of feature fusion processes that may be performed by the series feature fusion layer of the fusion recognition module 1230. output indicates that the output of the image recognition result may be performed by the output layer of the fusion recognition module 1230 . Furthermore, in FIG. 9 , conv2_1 indicates the first sub-convolution process of the first convolution process that can be performed by the first sub-convolution layer of the first convolution layer, and conv2_2 indicates the second sub-convolution process that can be performed by the first convolution layer. The second subconvolution process of the first convolution process performed by the subconvolution layer. Similarly, conv3_1 to conv3_3 each indicate the first to third sub-convolution processes of the third convolution process that can be performed by the first to third sub-convolution layers of the third convolution layer, respectively.
为了便于显示,在图10中没有按时序排列步骤S106涉及的流程。具体地,在特征提取模块1220执行特征提取步骤S106时,如S106的虚线框中所示,首先从上到下执行位于左一列的子采集区域A的所述图像数据的各个卷积和池化处理,然后再从上到下地执行中间列的子采集区域B的所述图像数据的各个卷积和池化处理,最后从上到下地执行右一列的子采集区域C的所述图像数据的各个卷积和池化处理。换句话说,各子采集区域A、B和C的第二图像数据的卷积和池化不是同时进行的,而是以一子采集区域的第二图像数据为单位依次从左到右进行的。For convenience of display, the processes involved in step S106 are not arranged in time sequence in FIG. 10 . Specifically, when the feature extraction module 1220 performs the feature extraction step S106, as shown in the dashed box in S106, each convolution and pooling of the image data of the sub-collection area A located in the left column is first performed from top to bottom processing, and then perform each convolution and pooling processing of the image data of the sub-collection area B in the middle column from top to bottom, and finally perform each of the image data of the sub-collection area C in the right column from top to bottom. Convolution and pooling. In other words, the convolution and pooling of the second image data of each sub-acquisition area A, B and C are not performed simultaneously, but are sequentially performed from left to right in units of the second image data of a sub-collection area .
如图10的S107虚线框所示,在融合识别模块1230执行融合识别步骤S107时,按照从上到下的顺序依次执行系列特征融合处理、第一全连接处理、第二全连接处理和图像识别结果的输出。As shown in the dotted box in S107 of FIG. 10 , when the fusion recognition module 1230 performs the fusion recognition step S107, it sequentially performs a series of feature fusion processing, first full connection processing, second full connection processing and image recognition in order from top to bottom result output.
为了便于理解,以上详细地描述了本申请的图像传感器图像数据处理系统的一个实施方式和用于图像数据处理的方法的一个实施方式。然而,不应理解为本申请的系统和方法仅限于上述实施方式的特征组合。For ease of understanding, an embodiment of an image sensor image data processing system and an embodiment of a method for image data processing of the present application are described above in detail. However, it should not be understood that the systems and methods of the present application are limited to the feature combinations of the above-described embodiments.
在本申请其他一些实施方式中,图像数据处理装置1100以生成自一子采集区域的一组第一图像数据作为处理对象,而图像识别装置1200以源自于整个采集区域的所有第二图像数据作为处理对象。如图10下部所示,本申请该实施方式的用于图像数据处理的方法包括:在接收生成自扫描整个采集区域的多个第一图像数据期间,依次对生成自一子采集区域的一组第一图像数据执行图像处理,以输出相应的第二图像数据;在完成对生成自3个子采集区域的所有第一图像数据进行图像处理并输出所有第二图像数据时,将所有第二图像数据作为整体执行特征提取和融合识别。由此,本方案的时延收益仅为2/3的图像处理时长,即本方案仅得益于生成自子采集区域A和B的第一图像数据的图像处理的提前。In some other embodiments of the present application, the image data processing apparatus 1100 takes a set of first image data generated from a sub-collection area as the processing object, and the image recognition apparatus 1200 uses all the second image data from the entire collection area as a processing object. As shown in the lower part of FIG. 10 , the method for image data processing according to this embodiment of the present application includes: during receiving a plurality of first image data generated from scanning the entire acquisition area, sequentially processing a group of images generated from a sub-acquisition area Image processing is performed on the first image data to output the corresponding second image data; when the image processing is completed on all the first image data generated from the three sub-collection areas and all the second image data are output, all the second image data are processed. Feature extraction and fusion recognition are performed as a whole. Therefore, the delay benefit of this solution is only 2/3 of the image processing time, that is, this solution only benefits from the advance of image processing of the first image data generated from the sub-collection areas A and B.
在其他一些实施方式中,各个子采集区域的大小可以被限定为彼此不相等。例如,在如图11所示的本申请另一实施方式中,子采集区域D的大小被设定为是子采集区域E的大小的1/2。In some other embodiments, the sizes of the various sub-collection regions may be defined to be unequal to each other. For example, in another embodiment of the present application as shown in FIG. 11 , the size of the sub-collection area D is set to be 1/2 of the size of the sub-collection area E.
在其他一些实施方式中,图像数据处理装置或图像识别装置对接收到的第一或第二图像数据进行图像处理或特征提取的开始时刻可以不是接收到该图像数据 的时刻。例如,在如图11下部所示的本申请另一实施方式中,图像数据处理装置对生成自子采集区域D的第一图像数据进行图像处理是在t16时刻,即是在接收到生成自子采集区域D的所有第一图像数据的时刻t14和在接收到子采集区域E的所有第一图像数据的时刻t15之间。如此,使得图像数据处理装置对生成自子采集区域E的所有第一图像数据进行图像处理的开始时刻是对生成自子采集区域D的所有第一图像数据进行图像处理的完成时刻。由此,本方案的时延收益仅为1/3的图像处理时长,即本方案仅得益于生成自子采集区域D的所有第一图像数据的图像处理的提前。In some other implementations, the image data processing apparatus or the image recognition apparatus may perform image processing or feature extraction on the received first or second image data at a start time that is not the time at which the image data is received. For example, in another embodiment of the present application shown in the lower part of FIG. 11 , the image data processing device performs image processing on the first image data generated from the sub-collection area D at time t16, that is, after receiving the first image data generated from the sub-collection area D. Between time t14 when all the first image data of the sub-capture area D is acquired and time t15 when all the first image data of the sub-acquisition area E are received. In this way, the start time when the image data processing apparatus performs image processing on all the first image data generated from the sub-capturing area E is the completion time when the image processing is performed on all the first image data generated from the sub-capturing area D. Therefore, the delay benefit of this solution is only 1/3 of the image processing time, that is, this solution only benefits from the advance of image processing of all the first image data generated from the sub-collection area D.
因此,参照图2、图10和图11,可以理解,图像数据处理装置执行图像处理和图像识别装置执行特征提取的开始时刻的设定均对本申请的时延收益产生影响。此外,对采集区域的子采集区域的数量和大小的设定也会对本申请的时延收益产生影响。Therefore, referring to FIG. 2 , FIG. 10 and FIG. 11 , it can be understood that the setting of the start time for image processing performed by the image data processing apparatus and feature extraction performed by the image recognition apparatus both have an impact on the delay benefit of the present application. In addition, the setting of the number and size of the sub-collection areas of the collection area will also have an impact on the delay benefit of the present application.
在其他一些实施方式中,采集区域的子采集区域的大小和数量可以根据预设值直接预先配置,因此,可以省略划分管理模块1300,或省略步骤S101和S102。在这种情况下,结合图10下部所示的方案,图像数据处理装置1100和图像识别装置1200根据预设值相应执行以下步骤S201-S205,如图12所示。In some other embodiments, the size and number of sub-collection areas of the collection area may be directly pre-configured according to preset values. Therefore, the division management module 1300 may be omitted, or steps S101 and S102 may be omitted. In this case, in combination with the solution shown in the lower part of FIG. 10 , the image data processing apparatus 1100 and the image recognition apparatus 1200 execute the following steps S201 - S205 correspondingly according to preset values, as shown in FIG. 12 .
在步骤S201中,图像数据处理装置1101接收图像传感器扫描采集区域所依次生成的多个第一图像数据。具体地,图像数据处理装置1101的接收模块1111接收图像传感器扫描采集区域所依次生成的多个第一图像数据。In step S201, the image data processing apparatus 1101 receives a plurality of first image data sequentially generated by the image sensor scanning the acquisition area. Specifically, the receiving module 1111 of the image data processing apparatus 1101 receives a plurality of first image data sequentially generated by the image sensor scanning the acquisition area.
在步骤S202中,图像数据处理装置1101在已接收到一组第一图像数据时,对该组第一图像数据进行图像处理,一组第一图像数据生成自采集区域的一预设的第一子采集区域。具体地,根据预定值,图像数据处理装置1100的图像处理模块1121预先设定采集区域的各第一子采集区域A、B和C。图像数据处理装置1100的图像处理模块1121在接收模块1111接收到生成自一第一子采集区域的一组第一图像数据时,对该第一图像数据执行图像处理,并依次输出对该组第一图像数据进行所述图像处理得到的相应的各第二图像数据。In step S202, when the image data processing device 1101 has received a set of first image data, image processing is performed on the set of first image data, and a set of first image data is generated from a preset first image data in the acquisition area. Sub-collection area. Specifically, according to a predetermined value, the image processing module 1121 of the image data processing apparatus 1100 presets each of the first sub-collection areas A, B and C of the collection area. When the receiving module 1111 receives a set of first image data generated from a first sub-collection area, the image processing module 1121 of the image data processing device 1100 performs image processing on the first image data, and sequentially outputs the set of first image data. The corresponding second image data is obtained by performing the image processing on one image data.
在步骤S203中,图像识别装置1201接收图像数据处理装置1101依次输出的多个第二图像数据。具体地,这些第二图像数据由图像识别装置1201的接收模块1211接收。In step S203, the image recognition device 1201 receives a plurality of second image data sequentially output by the image data processing device 1101. Specifically, these second image data are received by the receiving module 1211 of the image recognition apparatus 1201 .
在步骤S204中,图像识别装置1201在接收到多个第二图像数据时,从多个第二图像数据中提取特征数据。具体地,图像识别装置1201的特征提取模块1221在接收模块1211接收到源自于整个采集区域的多个第二图像数据时,一起从多个第二图像数据中提取特征数据。In step S204, when receiving the plurality of second image data, the image recognition device 1201 extracts feature data from the plurality of second image data. Specifically, the feature extraction module 1221 of the image recognition apparatus 1201 extracts feature data from the plurality of second image data together when the receiving module 1211 receives the plurality of second image data originating from the entire acquisition area.
在步骤S205中,图像识别装置1201对从多个第二图像数据提取的特征数据进行融合识别处理,输出图像识别结果。具体地,在图像识别装置1201的特征提取模块1221多个第二图像数据提取特征数据,图像识别装置1201的融合识别模块1231将由此得到的多个特征数据融合为一体,然后进行识别,最后输出图像识别结果。In step S205, the image recognition device 1201 performs fusion recognition processing on the feature data extracted from the plurality of second image data, and outputs an image recognition result. Specifically, the feature extraction module 1221 of the image recognition device 1201 extracts feature data from a plurality of second image data, and the fusion recognition module 1231 of the image recognition device 1201 fuses the plurality of feature data obtained thereby, performs recognition, and finally outputs Image recognition results.
在其他一些实施方式中,可以以不同的方式对不同帧的第一图像数据和第二图像数据分组,例如以不同的划分策略或预设值预先设定摄像头扫描不同帧图像时的采集区域的子采集区域的数量和大小。在一个这样的实施方式中,在处理第一帧图 像的图像数据中,可以使用图7示出的的划分方式,使得采集区域被划分为大小相等的3个子采集区域。而在处理第二帧图像的图像数据中,可以使用图11采用的划分策略,即使得采集区域被划分为大小不相等的两个子采集区域,其中一个子采集区域的大小是另一个的2倍。In some other embodiments, the first image data and the second image data of different frames can be grouped in different ways, for example, different division strategies or preset values are used to pre-set the acquisition area of the camera when scanning different frames of images. The number and size of sub-collection regions. In one such embodiment, in processing the image data of the first frame of image, the division method shown in FIG. 7 can be used, so that the acquisition area is divided into three sub-acquisition areas of equal size. In processing the image data of the second frame of images, the division strategy adopted in Figure 11 can be used, that is, the acquisition area is divided into two sub-acquisition areas with unequal sizes, and the size of one sub-acquisition area is twice that of the other. .
在其他一些实施方式中,还可以以不同的方式对同一帧的第一图像数据和第二图像数据分组,例如,在图像数据处理流程中以一种方式设定采集区域的各子采集区域,而在图像识别流程中以另一种方式设定采集区域的各子采集区域。在一个这样的实施方式中,在图像数据处理流程中,采集区域被划分为4个相等的子采集区域F、G、H和I,而在图像识别流程中,采集区域被划分为2个相等的子采集区域J和K,其中子采集区域J与子采集区域F和G的交集重合,子采集区域K与子采集区域H和I的交集重合。这样既可以使得图像数据处理流程得以提前,又能够保证图像识别流程不过分复杂,避免增加图像识别流程的时间。In some other implementation manners, the first image data and the second image data of the same frame can also be grouped in different ways. In the image recognition process, each sub-collection area of the collection area is set in another way. In one such embodiment, in the image data processing flow, the acquisition area is divided into 4 equal sub-acquisition areas F, G, H, and I, while in the image recognition flow, the acquisition area is divided into 2 equal sub-acquisition areas The sub-collection areas J and K of , wherein the sub-collection area J coincides with the intersection of the sub-collection areas F and G, and the sub-collection area K coincides with the intersection of the sub-collection areas H and I. In this way, the image data processing process can be advanced, and the image recognition process can be ensured not to be overly complicated, so as to avoid increasing the time of the image recognition process.
在其他一些实施方式中,尤其是图像传感器例如为激光雷达的实施方式中,可以不用四角的坐标限定图像的各子采集区域,而是用旋转角度限定,例如可以用0°至90°的旋转角度范围限定点云的一子采集区域。In some other implementations, especially in implementations where the image sensor is, for example, a lidar, each sub-acquisition area of the image may not be defined by the coordinates of the four corners, but may be defined by a rotation angle, for example, a rotation of 0° to 90° may be used The angular extent defines a sub-collection area of the point cloud.
在其他一些实施方式中,特征提取模块和融合识别模块的各层和各子层的数量是可调整的。相应地,在其他一些实施方式中,特征提取和融合识别的各处理和各子处理被执行的次数也是可调整的。In some other embodiments, the number of layers and sub-layers of the feature extraction module and the fusion recognition module is adjustable. Correspondingly, in some other embodiments, the number of times that each process and each sub-process of feature extraction and fusion identification is performed can also be adjusted.
在一些实施方式中,图像数据处理装置接收来自图像传感器的第一图像数据,该第一图像数据为所述图像传感器在一个扫描周期内扫描采集区域对应的物理区域可生成的多个图像数据中一个图像数据,该采集区域表示所述图像传感器的采集范围。图像数据处理装置对所述第一图像数据进行图像处理得到第二图像数据。图像数据处理装置输出所述第二图像数据。也就是说,在本实施方式中,在图像处理流程中,没有以分组的方式处理第一图像数据。In some embodiments, the image data processing apparatus receives first image data from an image sensor, where the first image data is among a plurality of image data that can be generated by the image sensor scanning a physical area corresponding to the acquisition area in one scan period An image data, the acquisition area represents the acquisition range of the image sensor. The image data processing device performs image processing on the first image data to obtain second image data. The image data processing device outputs the second image data. That is, in this embodiment, in the image processing flow, the first image data is not processed in a grouped manner.
在一些实施方式中,图像识别装置依次接收第二图像数据,该第二图像数据是对第一图像数据进行图像处理得到的,该第一图像数据为图像传感器在一个扫描周期内扫描采集区域对应的物理区域可生成的多个图像数据中一个图像数据,该采集区域表示所述图像传感器的采集范围。图像识别装置依次从所述第二图像数据提取特征数据。图像识别装置对各个特征数据进行融合识别处理。图像识别装置输出图像识别结果。也就是说,在本实施方式中,在图像识别流程中,没有以分组的方式处理第二图像数据。In some embodiments, the image recognition device sequentially receives second image data, where the second image data is obtained by performing image processing on the first image data, and the first image data corresponds to the area scanned by the image sensor in one scan period. One of the plurality of image data that can be generated by the physical area of the acquisition area represents the acquisition range of the image sensor. The image recognition device sequentially extracts feature data from the second image data. The image recognition device performs fusion recognition processing on each feature data. The image recognition device outputs an image recognition result. That is, in this embodiment, in the image recognition flow, the second image data is not processed in groups.
图13示例性地示出了根据本申请一个实施方式的驾驶系统3001的结构示意图。驾驶系统3001是高级驾驶辅助系统(ADAS),其包括图像传感器图像数据处理系统1001和驾驶决策单元3100。图像传感器图像数据处理系统1001可以与驾驶系统3001外的摄像机2001通信连接,处理和识别由摄像机2001扫描采集区域所依次生成的多个第一图像数据并且输出图像识别结果。驾驶决策单元3100与图像传感器图像数据处理系统1001通信连接,用于依据图像传感器图像数据处理系统1001输出的图像识别结果执行行为决策和运动规划并输出操作指令。FIG. 13 exemplarily shows a schematic structural diagram of a driving system 3001 according to an embodiment of the present application. The driving system 3001 is an advanced driving assistance system (ADAS), which includes an image sensor image data processing system 1001 and a driving decision unit 3100 . The image sensor image data processing system 1001 can be connected in communication with the camera 2001 outside the driving system 3001, process and recognize a plurality of first image data sequentially generated by the camera 2001 scanning the acquisition area, and output the image recognition result. The driving decision unit 3100 is connected in communication with the image sensor image data processing system 1001, and is used for executing behavior decision and motion planning and outputting operation instructions according to the image recognition result output by the image sensor image data processing system 1001.
图14示例性地示出了根据本申请一个实施方式的智能网联车V的结构示 意图。智能网联车V包括通常设定在车前的摄像机2001、设定在车内的驾驶系统3001、电子控制单元4001和例如为制动机构的执行器5001。摄像机2001以按行扫描其采集区域的方式感知车辆环境并依次输出多个第一图像数据。驾驶系统3001与摄像机2001通信连接,用于根据来自摄像机2001的多个第一图像数据输出操作指令。电子控制单元(Electronic Control Unit,ECU)4001与驾驶系统3001通信连接,用于依据来自驾驶系统3001的操作指令控执行器5001执行操作,例如依据驾驶系统的制动指令控制制动机构执行制动操作。Fig. 14 exemplarily shows a schematic structural diagram of an intelligent networked vehicle V according to an embodiment of the present application. The intelligent networked vehicle V includes a camera 2001 usually set in the front of the car, a driving system 3001 set in the car, an electronic control unit 4001 and an actuator 5001 such as a braking mechanism. The camera 2001 perceives the vehicle environment in a manner of scanning its acquisition area in rows, and sequentially outputs a plurality of first image data. The driving system 3001 is connected in communication with the camera 2001 for outputting operation instructions according to a plurality of first image data from the camera 2001 . The Electronic Control Unit (ECU) 4001 is connected in communication with the driving system 3001, and is used to control the actuator 5001 to perform operations according to the operation commands from the driving system 3001, for example, control the braking mechanism to perform braking according to the braking command of the driving system operate.
图15是本申请实施方式提供的一种计算设备1500的结构性示例性图。该计算设备1500包括:处理器1510、存储器1520、通信接口1530、总线1540。FIG. 15 is an exemplary structural diagram of a computing device 1500 provided by an embodiment of the present application. The computing device 1500 includes: a processor 1510 , a memory 1520 , a communication interface 1530 , and a bus 1540 .
应理解,图15所示的计算设备1500中的通信接口1530可以用于与其他设备之间执行通信。It should be understood that the communication interface 1530 in the computing device 1500 shown in FIG. 15 may be used to perform communication with other devices.
其中,该处理器1510可以与存储器1520连接。该存储器1520可以用于存储该程序代码和数据。因此,该存储器1520可以是处理器1510内部的存储单元,也可以是与处理器1510独立的外部存储单元,还可以是包括处理器1510内部的存储单元和与处理器1510独立的外部存储单元的部件。Wherein, the processor 1510 can be connected with the memory 1520 . The memory 1520 may be used to store the program codes and data. Therefore, the memory 1520 may be a storage unit inside the processor 1510 , or an external storage unit independent from the processor 1510 , or may include a storage unit inside the processor 1510 and an external storage unit independent from the processor 1510 . part.
可选的,计算设备1500还可以包括总线1540。其中,存储器1520、通信接口1530可以通过总线1540与处理器1510连接。总线1540可以是外设部件互连标准(Peripheral Component Interconnect,PCI)总线或扩展工业标准结构(Extended Industry Standard Architecture,EISA)总线等。所述总线1540可以分为地址总线、数据总线、控制总线等。为便于表示,图15中仅用一条线表示,但并不表示仅有一根总线或一种类型的总线。Optionally, computing device 1500 may also include bus 1540 . The memory 1520 and the communication interface 1530 may be connected to the processor 1510 through the bus 1540 . The bus 1540 may be a peripheral component interconnect standard (Peripheral Component Interconnect, PCI) bus or an Extended Industry Standard Architecture (Extended Industry Standard Architecture, EISA) bus or the like. The bus 1540 can be divided into an address bus, a data bus, a control bus, and the like. For ease of representation, only one line is shown in FIG. 15, but it does not mean that there is only one bus or one type of bus.
应理解,在本申请实施方式中,该处理器1510可以采用中央处理单元(central processing unit,CPU)。该处理器还可以是其它通用处理器、数字信号处理器(digital signal processor,DSP)、专用集成电路(application specific integrated circuit,ASIC)、现成可编程门阵列(field programmable gate Array,FPGA)或者其它可编程逻辑器件、分立门或者晶体管逻辑器件、分立硬件组件等。通用处理器可以是微处理器或者该处理器也可以是任何常规的处理器等。或者该处理器1510采用一个或多个集成电路,用于执行相关程序,以实现本申请实施方式所提供的技术方案。It should be understood that, in the embodiments of the present application, the processor 1510 may adopt a central processing unit (central processing unit, CPU). The processor may also be other general-purpose processors, digital signal processors (DSPs), application specific integrated circuits (ASICs), off-the-shelf programmable gate arrays (FPGAs) or other Programmable logic devices, discrete gate or transistor logic devices, discrete hardware components, etc. A general purpose processor may be a microprocessor or the processor may be any conventional processor or the like. Alternatively, the processor 1510 uses one or more integrated circuits to execute related programs to implement the technical solutions provided by the embodiments of the present application.
该存储器1520可以包括只读存储器和随机存取存储器,并向处理器1510提供指令和数据。处理器1510的一部分还可以包括非易失性随机存取存储器。例如,处理器1510还可以存储设备类型的信息。The memory 1520 may include read only memory and random access memory and provides instructions and data to the processor 1510 . A portion of the processor 1510 may also include non-volatile random access memory. For example, the processor 1510 may also store device type information.
在计算设备1500运行时,所述处理器1510执行所述存储器1520中的计算机执行指令执行上述任一种处理图像传感器图像数据的方法的操作步骤。When the computing device 1500 is running, the processor 1510 executes the computer-implemented instructions in the memory 1520 to perform the operation steps of any of the above methods for processing image sensor image data.
在其他一些实施方式中,省略了通信接口1530和总线1540。In other embodiments, the communication interface 1530 and the bus 1540 are omitted.
应理解,根据本申请实施方式的计算设备1500可以对应于执行根据本申请各实施方式的方法中的相应主体,并且计算设备1500中的各个单元的上述和其它操作和/或功能分别为了实现本实施方式各方法的相应流程,为了简洁,在此不再赘述。It should be understood that the computing device 1500 according to the embodiments of the present application may correspond to the corresponding subjects in executing the methods according to the various embodiments of the present application, and the above-mentioned and other operations and/or functions of the various units in the computing device 1500 are respectively for realizing the present invention. For the sake of brevity, the corresponding procedures of each method in the implementation manner are not repeated here.
本领域普通技术人员可以意识到,结合本文中所公开的实施方式描述的各示例的单元及算法步骤,能够以电子硬件、或者计算机软件和电子硬件的结合来实现。 这些功能究竟以硬件还是软件方式来执行,取决于技术方案的特定应用和设计约束条件。专业技术人员可以对每个特定的应用来使用不同方法来实现所描述的功能,但是这种实现不应认为超出本申请的范围。Those of ordinary skill in the art can realize that the units and algorithm steps of each example described in conjunction with the embodiments disclosed herein can be implemented in electronic hardware, or a combination of computer software and electronic hardware. Whether these functions are performed in hardware or software depends on the specific application and design constraints of the technical solution. Skilled artisans may implement the described functionality using different methods for each particular application, but such implementations should not be considered beyond the scope of this application.
所属领域的技术人员可以清楚地了解到,为描述的方便和简洁,上述描述的系统、装置和单元的具体工作过程,可以参考前述方法实施方式中的对应过程,在此不再赘述。Those skilled in the art can clearly understand that, for the convenience and brevity of description, the specific working process of the above-described systems, devices and units may refer to the corresponding processes in the foregoing method implementations, which will not be repeated here.
在本申请所提供的几个实施方式中,应该理解到,所揭露的系统、装置和方法,可以通过其它的方式实现。例如,以上所描述的装置实施方式仅仅是示例性的,例如,所述单元的划分,仅仅为一种逻辑功能划分,实际实现时可以有另外的划分策略,例如多个单元或组件可以结合或者可以集成到另一个系统,或一些特征可以忽略,或不执行。另一点,所显示或讨论的相互之间的耦合或直接耦合或通信连接可以是通过一些接口,装置或单元的间接耦合或通信连接,可以是电性,机械或其它的形式。In the several embodiments provided in this application, it should be understood that the disclosed systems, devices and methods may be implemented in other manners. For example, the apparatus implementations described above are only exemplary. For example, the division of the units is only a logical function division. In actual implementation, there may be other division strategies, for example, multiple units or components may be combined or Can be integrated into another system, or some features can be ignored, or not implemented. On the other hand, the shown or discussed mutual coupling or direct coupling or communication connection may be through some interfaces, indirect coupling or communication connection of devices or units, and may be in electrical, mechanical or other forms.
所述作为分离部件说明的单元可以是或者也可以不是物理上分开的,作为单元显示的部件可以是或者也可以不是物理单元,即可以位于一个地方,或者也可以分布到多个网络单元上。可以根据实际的需要选择其中的部分或者全部单元来实现本实施方式方案的目的。The units described as separate components may or may not be physically separated, and components displayed as units may or may not be physical units, that is, may be located in one place, or may be distributed to multiple network units. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution in this implementation manner.
另外,在本申请各个实施方式中的各功能单元可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中。In addition, each functional unit in each embodiment of the present application may be integrated into one processing unit, or each unit may exist physically alone, or two or more units may be integrated into one unit.
所述功能如果以软件功能单元的形式实现并作为独立的产品销售或使用时,可以存储在一个计算机可读取存储介质中。基于这样的理解,本申请的技术方案本质上或者说对现有技术做出贡献的部分或者该技术方案的部分可以以软件产品的形式体现出来,该计算机软件产品存储在一个存储介质中,包括若干指令用以使得一台计算机设备(可以是个人计算机,服务器,或者网络设备等)执行本申请各个实施方式所述方法的全部或部分步骤。而前述的存储介质包括:U盘、移动硬盘、只读存储器(Read-Only Memory,ROM)、随机存取存储器(Random Access Memory,RAM)、磁碟或者光盘等各种可以存储程序代码的介质。The functions, if implemented in the form of software functional units and sold or used as independent products, may be stored in a computer-readable storage medium. Based on this understanding, the technical solution of the present application can be embodied in the form of a software product in essence, or the part that contributes to the prior art or the part of the technical solution. The computer software product is stored in a storage medium, including Several instructions are used to cause a computer device (which may be a personal computer, a server, or a network device, etc.) to execute all or part of the steps of the methods described in the various embodiments of the present application. The aforementioned storage medium includes: U disk, mobile hard disk, read-only memory (Read-Only Memory, ROM), random access memory (Random Access Memory, RAM), magnetic disk or optical disk and other media that can store program codes .
本申请实施方式还提供了一种计算机可读存储介质,其上存储有计算机程序,该程序被处理器执行时用于执行一种处理图像传感器图像数据的方法,该方法包括上述各个实施方式所描述的方法中的至少之一。Embodiments of the present application further provide a computer-readable storage medium on which a computer program is stored, and when the program is executed by a processor, is used to execute a method for processing image data of an image sensor, and the method includes the methods described in the foregoing embodiments. at least one of the methods described.
本申请实施方式的计算机存储介质,可以采用一个或多个计算机可读的介质的任意组合。计算机可读介质可以是计算机可读信号介质或者计算机可读存储介质。计算机可读存储介质例如可以是,但不限于,电、磁、光、电磁、红外线、或半导体的系统、装置或器件,或者任意以上的组合。计算机可读存储介质的更具体的例子(非穷举的列表)包括:具有一个或多个导线的电连接、便携式计算机磁盘、硬盘、随机存取存储器(RAM)、只读存储器(ROM)、可擦式可编程只读存储器(EPROM或闪存)、光纤、便携式紧凑磁盘只读存储器(CD-ROM)、光存储器件、磁存储器件、云(cloud)或者上述的任意合适的组合。在本文件中,计算机可读存储介质可以是任何包含或存储程序的有形介质,该程序可以被指令执行系统、装置或者器件使用或者与其结合使用。The computer storage medium of the embodiments of the present application may adopt any combination of one or more computer-readable mediums. The computer-readable medium may be a computer-readable signal medium or a computer-readable storage medium. The computer readable storage medium can be, for example, but not limited to, an electrical, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus or device, or any combination of the above. More specific examples (a non-exhaustive list) of computer readable storage media include: electrical connections having one or more wires, portable computer disks, hard disks, random access memory (RAM), read only memory (ROM), Erasable Programmable Read Only Memory (EPROM or Flash), fiber optics, portable compact disk read only memory (CD-ROM), optical storage, magnetic storage, cloud, or any suitable combination of the above. In this document, a computer-readable storage medium can be any tangible medium that contains or stores a program that can be used by or in conjunction with an instruction execution system, apparatus, or device.
计算机可读的信号介质可以包括在基带中或者作为载波一部分传播的数 据信号,其中承载了计算机可读的程序代码。这种传播的数据信号可以采用多种形式,包括但不限于电磁信号、光信号或上述的任意合适的组合。计算机可读的信号介质还可以是计算机可读存储介质以外的任何计算机可读介质,该计算机可读介质可以发送、传播或者传输用于由指令执行系统、装置或者器件使用或者与其结合使用的程序。A computer-readable signal medium may include a propagated data signal in baseband or as part of a carrier wave with computer-readable program code embodied thereon. Such propagated data signals may take a variety of forms, including but not limited to electromagnetic signals, optical signals, or any suitable combination of the foregoing. A computer-readable signal medium can also be any computer-readable medium other than a computer-readable storage medium that can transmit, propagate, or transport the program for use by or in connection with the instruction execution system, apparatus, or device .
计算机可读介质上包含的程序代码可以用任何适当的介质传输,包括、但不限于无线、电线、光缆、RF等等,或者上述的任意合适的组合。Program code embodied on a computer readable medium may be transmitted using any suitable medium including, but not limited to, wireless, wireline, optical fiber cable, RF, etc., or any suitable combination of the foregoing.
可以以一种或多种程序设计语言或其组合来编写用于执行本申请操作的计算机程序代码,所述程序设计语言包括面向对象的程序设计语言—诸如Java、Smalltalk、C++,还包括常规的过程式程序设计语言—诸如“C”语言或类似的程序设计语言。程序代码可以地在用户计算机上执行、部分地在用户计算机上执行、作为一个独立的软件包执行、部分在用户计算机上部分在远程计算机上执行、或者在远程计算机或服务器上执行。在涉及远程计算机的情形中,远程计算机可以通过任意种类的网络,包括局域网(LAN)或广域网(WAN),连接到用户计算机,或者,可以连接到外部计算机(例如利用因特网服务提供商来通过因特网连接)。Computer program code for performing the operations of the present application may be written in one or more programming languages, including object-oriented programming languages—such as Java, Smalltalk, C++, but also conventional Procedural programming language - such as the "C" language or similar programming language. The program code may execute on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer, or on the remote computer or server. In the case of a remote computer, the remote computer may be connected to the user's computer through any kind of network, including a local area network (LAN) or wide area network (WAN), or may be connected to an external computer (eg, through the Internet using an Internet service provider) connect).
注意,上述仅为本申请的较佳实施方式及所运用的技术原理。本领域技术人员会理解,本申请不限于这里所述的特定实施方式,对本领域技术人员来说能够进行各种明显的变化、重新调整和替代而不会脱离本申请的保护范围。因此,虽然通过以上实施方式对本申请进行了较为详细的说明,但是本申请不仅仅限于以上实施方式,在不脱离本申请的构思的情况下,还可以包括更多其他等效实施方式,均属于本申请的保护范畴。Note that the above are only the preferred embodiments of the present application and the applied technical principles. Those skilled in the art will understand that the present application is not limited to the specific embodiments described herein, and various obvious changes, readjustments and substitutions can be made by those skilled in the art without departing from the protection scope of the present application. Therefore, although the present application has been described in detail through the above embodiments, the present application is not limited to the above embodiments, and can also include other equivalent embodiments without departing from the concept of the present application, all of which belong to The scope of protection of this application.

Claims (17)

  1. 一种处理图像传感器图像数据的方法,其特征在于,包括:A method for processing image sensor image data, comprising:
    接收来自图像传感器的第一图像数据,所述第一图像数据为所述图像传感器在一个扫描周期内扫描采集区域对应的物理区域可生成的多个图像数据中一个图像数据,所述采集区域表示所述图像传感器的采集范围;Receive first image data from an image sensor, where the first image data is one image data among multiple image data that can be generated by the image sensor scanning a physical area corresponding to the acquisition area in one scan period, and the acquisition area represents the acquisition range of the image sensor;
    对所述第一图像数据进行图像处理得到第二图像数据;以及performing image processing on the first image data to obtain second image data; and
    输出所述第二图像数据。The second image data is output.
  2. 根据权利要求1所述的方法,其特征在于,所述采集区域包括多个子采集区域;The method according to claim 1, wherein the collection area includes a plurality of sub-collection areas;
    所述对所述第一图像数据进行图像处理包括:The performing image processing on the first image data includes:
    当接收到第一数据组A所包含的第一图像数据后,以第一图像数据组A所包含的全部第一图像数据为单位进行图像处理,所述第一图像数据组A是所述图像传感器扫描一个所述子采集区域所对应的物理区域而生成的所述第一图像数据的集合。After receiving the first image data included in the first data group A, image processing is performed in units of all the first image data included in the first image data group A, where the first image data group A is the image The set of the first image data generated by the sensor scanning a physical area corresponding to the sub-collection area.
  3. 根据权利要求2所述的方法,其特征在于,所述多个子采集区域的大小和数量为预先设定。The method according to claim 2, wherein the size and quantity of the plurality of sub-collection regions are preset.
  4. 一种处理图像传感器图像数据的方法,其特征在于,包括:A method for processing image sensor image data, comprising:
    依次接收第二图像数据,所述第二图像数据是对第一图像数据进行图像处理得到的,所述第一图像数据为所述图像传感器在一个扫描周期内扫描采集区域对应的物理区域可生成的多个图像数据中一个图像数据,所述采集区域表示所述图像传感器的采集范围;Receive second image data in sequence, where the second image data is obtained by performing image processing on the first image data, and the first image data can be generated by the image sensor scanning the physical area corresponding to the acquisition area within one scan cycle One image data among the plurality of image data, the collection area represents the collection range of the image sensor;
    依次从所述第二图像数据提取特征数据;extracting feature data sequentially from the second image data;
    对各个特征数据进行融合识别处理;以及Perform fusion identification processing on each feature data; and
    输出图像识别结果。Output image recognition results.
  5. 根据权利要求4所述的方法,其特征在于,对所述第二图像数据分组;The method of claim 4, wherein the second image data is grouped;
    所述各个特征数据是从各组所述第二图像数据中提取。The respective feature data are extracted from the respective sets of the second image data.
  6. 根据权利要求5所述的方法,其特征在于,所述各组所述第二图像数据的第二图像数据的数量为预先设定。The method according to claim 5, wherein the quantity of the second image data of each group of the second image data is preset.
  7. 一种图像数据处理装置,其特征在于,包括:An image data processing device, comprising:
    接收模块,用于接收来自图像传感器的第一图像数据,所述第一图像数据为所述图像传感器在一个扫描周期内扫描采集区域对应的物理区域可生成的多个图像数据中一个图像数据,所述采集区域表示所述图像传感器的采集范围;以及a receiving module, configured to receive first image data from an image sensor, where the first image data is one image data among a plurality of image data that can be generated by the image sensor scanning a physical area corresponding to the acquisition area in one scan period, the acquisition area represents the acquisition range of the image sensor; and
    图像处理模块,用于对所述第一图像数据进行图像处理得到第二图像数据,以及用于输出所述第二图像数据。An image processing module, configured to perform image processing on the first image data to obtain second image data, and to output the second image data.
  8. 根据权利要求7所述的装置,其特征在于,所述采集区域包括多个子采集区域;The device according to claim 7, wherein the collection area includes a plurality of sub-collection areas;
    所述对所述第一图像数据进行图像处理包括:The performing image processing on the first image data includes:
    所述图像处理模块还用于当所述接收模块接收到第一数据组A所包含的第一图像数据后,以第一图像数据组A所包含的全部第一图像数据为单位进行图像处理,所述第一图像数据组A是所述图像传感器扫描一个所述子采集区域所对应的物理区域而生成的所述第一图像数据的集合。The image processing module is further configured to perform image processing in units of all the first image data included in the first image data group A after the receiving module receives the first image data included in the first data group A, The first image data group A is a collection of the first image data generated by the image sensor scanning a physical area corresponding to one of the sub-acquisition areas.
  9. 根据权利要求8所述的装置,其特征在于,所述多个子采集区域的大小和数量为预先设定。The device according to claim 8, wherein the size and quantity of the plurality of sub-collection regions are preset.
  10. 一种图像识别装置,其特征在于,包括:An image recognition device, comprising:
    接收模块,用于依次接收第二图像数据,所述第二图像数据是对第一图像数据进行图像处理得到的,所述第一图像数据为所述图像传感器在一个扫描周期内扫描采集区域对应的物理区域可生成的多个图像数据中一个图像数据,所述采集区域表示所述图像传感器的采集范围;A receiving module, configured to sequentially receive second image data, the second image data is obtained by performing image processing on the first image data, and the first image data corresponds to the scanning and collecting area of the image sensor in one scanning period One image data among a plurality of image data that can be generated by the physical area of the image sensor, and the acquisition area represents the acquisition range of the image sensor;
    特征提取模块,用于依次从所述第二图像数据提取特征数据;以及a feature extraction module for sequentially extracting feature data from the second image data; and
    融合识别模块,用于对各个特征数据进行融合识别处理,以及用于输出图像识别结果。The fusion recognition module is used to perform fusion recognition processing on each feature data, and to output image recognition results.
  11. 根据权利要求10所述的装置,其特征在于,所述特征提取模块还用于对所述第二图像数据分组;The apparatus according to claim 10, wherein the feature extraction module is further configured to group the second image data;
    所述各个特征数据是从各组所述第二图像数据中提取。The respective feature data are extracted from the respective sets of the second image data.
  12. 根据权利要求11所述的装置,其特征在于,所述各组所述第二图像数据的第二图像数据的数量为预先设定。The apparatus according to claim 11, wherein the quantity of the second image data of each group of the second image data is preset.
  13. 一种图像传感器图像数据处理系统,其特征在于,包括:An image sensor image data processing system, comprising:
    如权利要求7至9中任一项所述的图像数据处理装置;以及The image data processing apparatus according to any one of claims 7 to 9; and
    如权利要求10至12中任一项所述的图像识别装置。The image recognition device according to any one of claims 10 to 12.
  14. 一种驾驶系统,其特征在于,包括如权利要求13所述的图像传感器图像数据处理系统和驾驶决策单元;A driving system, comprising the image sensor image data processing system and the driving decision unit as claimed in claim 13;
    其中所述驾驶决策单元与所述图像传感器图像数据处理系统连接,用于依据所述图像传感器图像数据处理系统输出的图像识别结果执行行为决策和运动规划,并输出操作指令。The driving decision-making unit is connected to the image sensor image data processing system, and is configured to execute behavior decision-making and motion planning according to the image recognition result output by the image sensor image data processing system, and output operation instructions.
  15. 一种车辆,其特征在于,包括依次连接的图像传感器、如权利要求14所述的驾驶系统、电子控制单元和执行器;其中A vehicle, characterized by comprising sequentially connected image sensors, a driving system as claimed in claim 14, an electronic control unit and an actuator; wherein
    所述图像传感器用于以扫描的方式感知车辆环境并输出第一图像数据;The image sensor is used to perceive the vehicle environment in a scanning manner and output first image data;
    所述电子控制单元用于依据所述驾驶系统的操作指令控制所述执行器执行操作。The electronic control unit is used for controlling the actuator to perform an operation according to an operation instruction of the driving system.
  16. 一种计算设备,其特征在于,包括:A computing device, comprising:
    至少一个处理器;以及at least one processor; and
    至少一个存储器,其与所述处理连接并存储有程序指令,所述程序指令当被所述至少一个处理器执行时使得所述至少一个处理器执行权利要求1至3和权利要求4至6中任一项所述的方法。at least one memory connected to the processing and storing program instructions which, when executed by the at least one processor, cause the at least one processor to perform the functions of claims 1 to 3 and 4 to 6 The method of any one.
  17. 一种计算机可读存储介质,其上存储有程序指令,其特征在于,所述程序指令当被计算机执行时使得所述计算机执行权利要求1至3和权利要求4至6中任一项所述的方法。A computer-readable storage medium having program instructions stored thereon, wherein the program instructions, when executed by a computer, cause the computer to perform the execution of any one of claims 1 to 3 and claims 4 to 6 Methods.
PCT/CN2021/131698 2021-02-07 2021-11-19 Method and apparatus for processing image data of image sensor WO2022166309A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202110169585.XA CN114915731A (en) 2021-02-07 2021-02-07 Method and apparatus for processing image data of image sensor
CN202110169585.X 2021-02-07

Publications (1)

Publication Number Publication Date
WO2022166309A1 true WO2022166309A1 (en) 2022-08-11

Family

ID=82741838

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/131698 WO2022166309A1 (en) 2021-02-07 2021-11-19 Method and apparatus for processing image data of image sensor

Country Status (2)

Country Link
CN (1) CN114915731A (en)
WO (1) WO2022166309A1 (en)

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2005092007A (en) * 2003-09-19 2005-04-07 Ricoh Co Ltd Image processing system, image processing method, program and information recording medium
CN101719985A (en) * 2009-11-17 2010-06-02 北京中星微电子有限公司 Method and device for collecting and processing images
CN102956027A (en) * 2011-08-30 2013-03-06 安凯(广州)微电子技术有限公司 Image enhancement processing system and method based on camera image processing chip
CN104902193A (en) * 2015-05-19 2015-09-09 上海集成电路研发中心有限公司 Method for performing segmentation processing and display for image data based on FPGA

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101819677B (en) * 2010-04-12 2012-06-13 中国科学院长春光学精密机械与物理研究所 Fusion processing system of multi-sensor images
CN103021069B (en) * 2012-11-21 2015-02-18 深圳市兆图电子有限公司 High-speed note image acquisition processing system and acquisition processing method thereof
DE102016202948A1 (en) * 2016-02-25 2017-08-31 Robert Bosch Gmbh Method and device for determining an image of an environment of a vehicle
CN108470166A (en) * 2018-04-04 2018-08-31 北京天目智联科技有限公司 A kind of biological characteristic 3D 4 D datas recognition methods and system based on laser scanning
CN112753212A (en) * 2018-09-26 2021-05-04 祖克斯有限公司 Image scan line time stamping
CN110833429B (en) * 2019-12-03 2024-03-26 上海联影医疗科技股份有限公司 Computed tomography imaging method, device, equipment and storage medium

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2005092007A (en) * 2003-09-19 2005-04-07 Ricoh Co Ltd Image processing system, image processing method, program and information recording medium
CN101719985A (en) * 2009-11-17 2010-06-02 北京中星微电子有限公司 Method and device for collecting and processing images
CN102956027A (en) * 2011-08-30 2013-03-06 安凯(广州)微电子技术有限公司 Image enhancement processing system and method based on camera image processing chip
CN104902193A (en) * 2015-05-19 2015-09-09 上海集成电路研发中心有限公司 Method for performing segmentation processing and display for image data based on FPGA

Also Published As

Publication number Publication date
CN114915731A (en) 2022-08-16

Similar Documents

Publication Publication Date Title
JP7380180B2 (en) Solid-state imaging device, imaging device, imaging method, and imaging program
Heimberger et al. Computer vision in automated parking systems: Design, implementation and challenges
CN107945112B (en) Panoramic image splicing method and device
US8964025B2 (en) Visual obstruction removal with image capture
US10003740B2 (en) Increasing spatial resolution of panoramic video captured by a camera array
US11457143B2 (en) Sensor device, electronic device, sensor system and control method
JP4258539B2 (en) Multiple angle of view camera
US10445402B1 (en) Fast and energy-efficient region of interest pooling for object detection with convolutional neural network
JP2022510329A (en) Parking space detection method, equipment and electronic devices
JP6316976B2 (en) In-vehicle image recognition device
TW201535318A (en) Image transformation and multi-view output systems and methods
WO2023024697A1 (en) Image stitching method and electronic device
CN113724303B (en) Point cloud and image matching method and device, electronic equipment and storage medium
US11403742B2 (en) Image processing device, image processing method, and recording medium for generating bird&#39;s eye synthetic image
CN113329219B (en) Multi-output parameter dynamically configurable depth camera
US11443151B2 (en) Driving assistant system, electronic device, and operation method thereof
US20240029303A1 (en) Three-dimensional target detection method and apparatus
CN108364320B (en) Camera calibration method, terminal device and computer readable storage medium
WO2021101732A1 (en) Joint rolling shutter correction and image deblurring
CN114821506A (en) Multi-view semantic segmentation method and device, electronic equipment and storage medium
US11689820B2 (en) Combining grayscale scanned images with color image to create high-resolution overlay images in vehicles
CN115049820A (en) Determination method and device of occlusion region and training method of segmentation model
JP2019091169A (en) Image processor, method of controlling image processor, and program
EP4050553A1 (en) Method and device for restoring image obtained from array camera
WO2022166309A1 (en) Method and apparatus for processing image data of image sensor

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21924305

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 21924305

Country of ref document: EP

Kind code of ref document: A1