WO2021008500A1 - Image processing method and apparatus - Google Patents

Image processing method and apparatus Download PDF

Info

Publication number
WO2021008500A1
WO2021008500A1 PCT/CN2020/101717 CN2020101717W WO2021008500A1 WO 2021008500 A1 WO2021008500 A1 WO 2021008500A1 CN 2020101717 W CN2020101717 W CN 2020101717W WO 2021008500 A1 WO2021008500 A1 WO 2021008500A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
vehicle
area
frame
image area
Prior art date
Application number
PCT/CN2020/101717
Other languages
French (fr)
Chinese (zh)
Inventor
刘兴业
谢伟伦
Original Assignee
华为技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 华为技术有限公司 filed Critical 华为技术有限公司
Publication of WO2021008500A1 publication Critical patent/WO2021008500A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • G06T3/4053Scaling of whole images or parts thereof, e.g. expanding or contracting based on super-resolution, i.e. the output image resolution being higher than the sensor resolution
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting

Definitions

  • This application relates to the field of automatic driving or assisted driving technology, and in particular to image processing methods and devices.
  • Image super resolution technology has important application value in surveillance equipment, satellite imagery and autonomous driving technology.
  • Image super-resolution technology includes single-frame super-division technology and multi-frame super-division technology.
  • Single-frame super-division refers to recovering a high-resolution image from a low-resolution image.
  • Multi-frame super-division refers to recovering a high-resolution image from multiple low-resolution images.
  • the embodiments of the present application provide an image processing method and device, which help improve the effect of image compensation.
  • an image processing method is provided, which is applied to an in-vehicle device.
  • the method includes: firstly, acquiring a multi-frame image, the multi-frame image including image information of the surrounding road of the vehicle where the in-vehicle device is located; then, acquiring the multi-frame The first image area in each frame of the image; wherein the multiple first image areas (where each frame of the image includes a first image area) of the multi-frame image correspond to the first scene; then, the multiple first image areas A super-division operation is performed on an image area.
  • the super-division operation is specifically a multi-frame super-division.
  • the feature information of the image information in the multi-frame image can be combined for image compensation, which is helpful compared with the technical solution of using a single-frame super-division for image compensation in the prior art.
  • To improve the image compensation effect since in the present technical solution, multiple image regions corresponding to the same scene in the multi-frame image are subjected to multi-frame super-division, instead of performing super-division operations on the multi-frame image itself, it helps to reduce the super-division The complexity of the operation, thereby speeding up the processing rate of super-division.
  • the processing result of the image processing method provided by the present technical solution is applied to an assisted automatic driving path planning scene, it is helpful to improve the accuracy of the automatic driving path planning.
  • the first scene may be understood as the road conditions around the vehicle or the spatial area where one or more objects in the driving field of view are located.
  • the first image area may be part or all of the area in one frame of image.
  • the first image area may or may not contain the image information of the target object.
  • the target object may be predefined, of course, the embodiment of the present application is not limited to this.
  • the method further includes: determining that the image information of the target object exists in the image area obtained by the hyperdivision operation.
  • the method further includes: detecting the relative position between the vehicle and the target object.
  • the relative position includes a relative distance and a relative angle, and the relative angle includes an azimuth angle and/or a pitch angle.
  • the relative position may be used to assist automatic driving path planning.
  • the first image area is an area with a confidence level lower than or equal to the first threshold; or, the first image area corresponds to the driveable area of the vehicle A space area where the distance from the vehicle is greater than or equal to the second threshold; or, the first image area is an area at a preset position.
  • This possible design provides several features of the first image area. In specific implementation, the first image area can be determined based on one of these features.
  • the multi-frame image includes a first image and a second image; acquiring the first image area in each frame of the multi-frame image includes: acquiring the first image area in the second image.
  • Acquiring the first image area in the second image includes: according to the first image area in the first image (which may specifically include: the position and size of the first image area in the first image in the first image) and the vehicle Obtaining the first image area in the second image (which may specifically include: obtaining the position and size of the first image area in the second image in the second image). That is to say, the embodiments of the present application support a technical solution to infer the first image area in the first image from the first image area in the first image based on the vehicle body information, which helps to improve the accuracy of the superdivision calculation .
  • the vehicle body information (including the first vehicle body information and the second vehicle body information below) can be the vehicle information directly detected by sensors and other equipment installed in the vehicle, or it can be the detection of these sensors and other equipment. Information about the vehicle obtained by processing the received information.
  • the first vehicle body information may include at least one of the first relative distance, the second relative distance, and the first vehicle steering angle.
  • the first relative distance is the relative distance between the vehicle and the space area corresponding to the first image area when the first image is taken.
  • the second relative distance is the relative distance between the vehicle and the space area corresponding to the first image area when the second image is taken.
  • the first vehicle steering angle is the angle between the direction of the vehicle in the time interval of shooting the first image and the second image.
  • the first image is a reference image in the multi-frame super division
  • the second image is any non-reference image in the multi-frame super division
  • performing a super-division operation on the plurality of first image regions includes: performing a super-division operation on all the plurality of first image regions after scene alignment.
  • the multiple frames of images include a third image and a fourth image; before the superdivision operation is performed on the multiple first image regions after scene alignment, the method further includes: according to the second image of the vehicle Car body information, performing scene alignment of the plurality of first image areas. That is to say, the embodiments of the present application support a technical solution for realizing scene alignment based on vehicle body information, which helps to improve the accuracy of super-division calculation.
  • the second vehicle body information includes at least one of the first relative angle, the second relative angle, and the second vehicle steering angle.
  • the first relative angle is the relative angle between the vehicle and the space area corresponding to the first image area when the third image is taken.
  • the second relative angle is the relative angle between the vehicle and the space area corresponding to the first image area when the fourth image is taken.
  • the second steering angle of the vehicle is the angle between the direction of the vehicle in the time interval of shooting the third image and the fourth image.
  • the third image and the first image or the second image may be the same or different, the fourth image and the first image or the second image may be the same or different, and the third image and the fourth image are different.
  • the multi-frame images are consecutive multi-frame images in time series. In this way, it is convenient to handle.
  • the time interval between the shooting moment of the first frame image and the shooting moment of the last frame image in the multi-frame images is less than or equal to the third threshold. In this way, it helps to improve the accuracy of the super-division calculation.
  • the method further includes: acquiring a second image area in each frame of the multi-frame image; wherein the multiple second image areas of the multi-frame image correspond to the second scene; then, Perform a hyperdivision operation on the plurality of second image regions.
  • this application supports the solution of including multiple image regions to be detected in one frame of image. There can be overlap or no overlap between the first image area and the second image area. The first scene is different from the second scene.
  • an image processing device which can be used to execute any method provided in the first aspect or any possible design of the first aspect.
  • the device may be an in-vehicle device or a chip.
  • the device may be divided into functional modules according to the method provided in the first aspect or any of the possible designs of the first aspect.
  • each functional module may be divided corresponding to each function, or Integrate two or more functions into one processing module.
  • the device may include a memory and a processor.
  • the memory is used to store computer programs.
  • the processor is used to invoke the computer program to execute the first aspect or the method provided by any possible design of the first aspect.
  • a computer-readable storage medium such as a non-transitory computer-readable storage medium.
  • a computer program (or instruction) is stored thereon, and when the computer program (or instruction) runs on a computer, the computer executes any method provided by the first aspect or any possible design of the first aspect .
  • a computer program product which, when running on a computer, enables any method provided in the first aspect or any possible design of the first aspect to be executed.
  • any image processing device, computer storage medium, computer program product or system provided above can be applied to the corresponding method provided above. Therefore, the beneficial effects that can be achieved can refer to the corresponding The beneficial effects of the method are not repeated here.
  • FIG. 1 is a schematic structural diagram of a computer system applicable to an embodiment of the present application
  • FIG. 2 is a schematic diagram of a result of target detection applicable to an embodiment of the present application
  • FIG. 3 is a schematic diagram of an image segmentation applicable to an embodiment of the present application.
  • FIG. 5 is a schematic flowchart of another image processing method provided by an embodiment of the application.
  • FIG. 6 is a schematic diagram of a vehicle body information in a vehicle body coordinate system provided by an embodiment of the application.
  • FIG. 7 is a schematic diagram of obtaining ROI2 based on ROI1 according to an embodiment of the application.
  • FIG. 8 is a schematic diagram of a scene of ROI1 and ROI2 provided by an embodiment of the application.
  • FIG. 9 is a schematic diagram of a view angle change when a vehicle turns right according to an embodiment of the application.
  • FIG. 10 is a schematic diagram of obtaining an alignment angle of a scene according to an embodiment of the application.
  • FIG. 11 is a schematic diagram of mapping ROI2 to a plane where ROI1 is located according to an embodiment of the application;
  • FIG. 12 is a schematic flowchart of another image processing method provided by an embodiment of the application.
  • FIG. 13 is a schematic diagram of a drivable area provided by an embodiment of the application.
  • FIG. 14 is a schematic diagram of a first image area provided by an embodiment of this application.
  • 15 is a schematic flowchart of another image processing method provided by an embodiment of the application.
  • FIG. 16 is a schematic diagram of another first image area provided by an embodiment of this application.
  • FIG. 17 is a schematic structural diagram of a vehicle-mounted device provided by an embodiment of the application.
  • FIG. 1 it is a schematic structural diagram of a computer system applicable to the embodiments of the present application.
  • the computer system may be located on the vehicle, and the computer system may include the vehicle-mounted equipment 101, and the equipment/device/network connected directly or indirectly with the vehicle-mounted equipment.
  • the vehicle-mounted device 101 includes a processor 103, and the processor 103 is coupled to a system bus 105.
  • the processor 103 may be one or more processors, where each processor may include one or more processor cores.
  • a display adapter (video adapter) 107 can drive the display 109, and the display 109 is coupled to the system bus 105.
  • the system bus 105 is coupled to an input/output (I/O) bus 113 through a bus bridge 111.
  • the I/O interface 115 is coupled to the I/O bus.
  • the I/O interface 115 communicates with a variety of I/O devices, such as input devices 117 (such as keyboard, mouse, touch screen, etc.), media tray 121 (such as compact disc read-only memory, CD -ROM), multimedia interface, etc.).
  • the transceiver 123 can send and/or receive radio communication signals), the camera 155 (can capture scene and dynamic digital video images), and an external universal serial bus (USB) interface 125.
  • the interface connected to the I/O interface 115 may be a USB interface.
  • the processor 103 may be any traditional processor, including a reduced instruction set computer (RISC) processor, a complex instruction set computer (CISC) processor, or a combination of the foregoing.
  • the processor may be a dedicated device such as an application specific integrated circuit (ASIC).
  • the processor 103 may be a neural network processor or a combination of a neural network processor and the foregoing traditional processors.
  • the processor 103 may be a central processing unit (CPU).
  • the camera 155 may be any camera used to collect images, for example, it may be a monocular camera or a binocular camera.
  • the number of cameras can be one or more, and each camera can be located in the front, rear, or side of the vehicle. For the convenience of description, the following specific examples all take the camera located directly in front of the vehicle as an example.
  • the camera 155 may be used to collect information about the surrounding environment (including surrounding roads, etc.) of the vehicle.
  • the camera 155 may include a software module, and the software module may be used to record the shooting time of the image taken by the camera. Alternatively, the module for recording the shooting time may also be a piece of hardware connected to the camera 155.
  • the position of the camera 155 relative to the vehicle may be fixed. In another example, the position of the camera 155 relative to the vehicle may be changed, for example, the camera 155 may perform rotation shooting.
  • the in-vehicle device 101 may be located far away from the autonomous driving vehicle, and may wirelessly communicate with the autonomous driving vehicle.
  • some of the processes described herein are executed on a processor provided in an autonomous vehicle, and others are executed by a remote processor, including taking actions required to perform a single manipulation.
  • the in-vehicle device 101 may communicate with a software deployment server (deploying server) 149 through a network interface 129.
  • the network interface 129 is a hardware network interface, such as a network card.
  • the network 127 may be an external network, such as the Internet, or an internal network, such as an Ethernet or a virtual private network (virtual private network, VPN).
  • the network 127 may also be a wireless network, such as a WiFi network, a cellular network, and so on.
  • the hard disk drive interface is coupled to the system bus 105.
  • the hardware drive interface is connected with the hard drive.
  • the system memory 135 and the system bus 105 are coupled.
  • the data running in the system memory 135 may include the operating system 137 and application programs 143 of the in-vehicle device 101.
  • the operating system includes a shell 139 and a kernel (kernel) 141.
  • Shell 139 is an interface between the user and the kernel of the operating system.
  • the shell is the outermost layer of the operating system. The shell manages the interaction between the user and the operating system, waits for the user's input, interprets the user's input to the operating system, and processes the output of various operating systems.
  • the kernel 141 is composed of those parts of the operating system for managing memory, files, peripherals, and system resources. Directly interact with hardware, the operating system kernel usually runs processes and provides inter-process communication, providing CPU time slice management, interrupts, memory management, IO management, and so on.
  • the application program 143 includes programs related to controlling the automatic driving of the vehicle. For example, a program for processing an image containing image information on a vehicle road acquired by an on-vehicle device, such as a program for implementing the image processing method provided by the embodiment of the present application. For another example, the program that manages the interaction between autonomous vehicles and road obstacles, the program that controls the route or speed of autonomous vehicles, and the process of interaction between autonomous vehicles and other autonomous vehicles on the road.
  • the application program 143 also exists on the system of the software deployment server 149. In one embodiment, when the application program 143 needs to be executed, the in-vehicle device 101 may download the application program 143 from the software deployment server 149.
  • the sensor 153 is associated with the in-vehicle device 101.
  • the sensor 153 is used to detect the environment around the in-vehicle device 101.
  • the sensor 153 can detect animals, vehicles, obstacles, and crosswalks.
  • the sensor can also detect the environment around objects such as animals, vehicles, obstacles, and crosswalks, such as: the environment around the animals, for example, when the animals appear around them. Other animals, weather conditions, the brightness of the surrounding environment, etc.
  • the sensor may be a camera, an infrared sensor, a chemical detector, a microphone, etc.
  • the senor 153 may include a speed sensor, used to measure the speed information (such as speed, acceleration, etc.) of the own vehicle (that is, the vehicle in which the computer system shown in FIG. 1 is located); an angle sensor, used to measure the direction information of the vehicle , And the relative angle between the vehicle and the objects/objects around the vehicle.
  • a speed sensor used to measure the speed information (such as speed, acceleration, etc.) of the own vehicle (that is, the vehicle in which the computer system shown in FIG. 1 is located)
  • an angle sensor used to measure the direction information of the vehicle , And the relative angle between the vehicle and the objects/objects around the vehicle.
  • FIG. 1 is only an example, which does not constitute a limitation on the computer system applicable to the embodiments of the present application.
  • one or more devices connected to the vehicle-mounted equipment shown in FIG. 1 may be integrated with the vehicle-mounted equipment, for example, a camera is integrated with the vehicle-mounted equipment.
  • Multi-frame super-division uses the image information of the non-reference image in the multi-frame image to process the image information of the reference image in the multi-frame image, such as compensating for edge features, sharpening features, etc., to obtain a frame of image.
  • the resolution of this image is higher than the resolution of the reference image.
  • the reference image may be any one of the multiple frames of images.
  • Non-reference images are all images in the multi-frame images except for reference images.
  • the object may also be called a road object or obstacle or road obstacle.
  • the objects may be people, vehicles, traffic lights, traffic signs (such as speed limit signs, etc.), telephone poles, trash cans, foreign objects, etc. on the roads surrounding the vehicle.
  • foreign objects refer to objects that should not have appeared on the road, such as boxes and tires left on the road.
  • the target object is the object that the vehicle-mounted equipment needs to recognize.
  • the target object may be predefined or indicated by the user, which is not limited in the embodiment of the present application.
  • the target object may include: people, cars, traffic lights, and so on.
  • Both target detection and image segmentation are image processing techniques.
  • the task of target detection is to find out the area where the image information of all the targets of interest in the image are located, and to determine the size of the area and the location of the area in the image.
  • the target of interest can be pre-defined or user-defined.
  • the target of interest may refer to the target object.
  • Different regions obtained by target detection that is, regions where the image information of different targets are located
  • FIG. 2 it is a schematic diagram of a target detection result.
  • the target of interest is a vehicle as an example for illustration.
  • the area defined by each rectangular box in FIG. 2 is the area where the image information of an interested target is located.
  • Image segmentation is a computer vision task that marks designated areas in an image according to the content of the image. In short, it is to determine the image information of which objects in a frame of image, and the position of the image information of the object in the image. Specifically, the purpose of image segmentation is to determine which object pixel each pixel in the image represents. Image segmentation can include semantic segmentation, instance segmentation, and so on. There is usually no overlap between image regions obtained by image segmentation. As shown in Figure 3, it is a schematic diagram of image segmentation. Each connected region in Figure 3 represents a region obtained by image segmentation.
  • the area in the actual scene that is, the area that exists objectively
  • the image of the area in the actual scene that is, the area in the image or picture
  • the image area to be detected refers to an area in the image that contains the image information of the target object with a high probability (for example, the probability is greater than a preset threshold). This is the definition of the image area to be detected in the embodiments of this application.
  • Feature 1 The confidence level is lower than or equal to the first threshold.
  • Feature 2 Corresponding to a space area where the distance between the vehicle and the vehicle in the travelable area is greater than or equal to the second threshold.
  • Feature 3 The position in the image to which it belongs is a preset position.
  • the image area to be detected is an area with a confidence level lower than or equal to the first threshold, or it corresponds to a space area in the drivable area of the vehicle whose distance from the vehicle is greater than or equal to the second threshold.
  • the image area or, is the area at the preset position in the image.
  • the image area to be detected may contain image information of one or more objects.
  • the image area to be detected may contain the image information of one or more target objects, or may not contain the image information of the target object.
  • the image area to be detected can be part or all of the area in a frame of image.
  • a frame of image can contain one or more image regions to be detected.
  • the first image area and the second image area described in the embodiments of the present application are both image areas to be detected.
  • words such as “exemplary” or “for example” are used as examples, illustrations, or illustrations. Any embodiment or design solution described as “exemplary” or “for example” in the embodiments of the present application should not be construed as being more preferable or advantageous than other embodiments or design solutions. To be precise, words such as “exemplary” or “for example” are used to present related concepts in a specific manner.
  • At least one refers to one or more.
  • Multiple means two or more.
  • "and/or” is merely an association relationship describing associated objects, indicating that there can be three types of relationships, for example, A and/or B, which can indicate that A exists alone, and both A and A B, there are three cases of B alone.
  • the character "/" in this text generally indicates that the associated objects before and after are in an "or" relationship.
  • FIG. 4 it is a schematic flowchart of an image processing method provided by an embodiment of this application.
  • the method can include:
  • the vehicle-mounted device acquires N frames of images, where the N-frame images include image information of the surrounding roads of the vehicle where the vehicle-mounted device is located.
  • N is an integer greater than or equal to 2.
  • the N frames of images may be images taken by a camera (such as the aforementioned camera 155) installed in the vehicle.
  • the surrounding roads of the vehicle may include one or more of the front road, the rear road, and the side road of the vehicle.
  • Any one of the N frames of images may be an image taken when the vehicle is in a stationary state, or may be an image taken when the vehicle is in a moving state (such as going straight, changing lanes, or turning).
  • the N frames of images are consecutive N frames of images in time series.
  • the N frames of images are N frames of images continuously captured by the camera installed in the vehicle.
  • the on-vehicle device may determine the "N frame images" in S101 in the images captured by the camera based on the sliding window N.
  • the embodiment of the present application does not limit the value of N.
  • the value of N may be the same or different.
  • the N frames of images in S101 may or may not overlap. For example, suppose that the images taken by the camera are sorted according to the order of shooting time to obtain the following sequence: image 1, image 2, image 3...image n.
  • the N frames of images in S101 may be images 2 to 4
  • the N frames of images in S101 may be images 3 to 4 6.
  • the vehicle-mounted device acquires the first image area in each frame of the N1 frame of images in the N frame of images.
  • the N1 first image regions of the N1 frame image correspond to the first scene. Among them, 2 ⁇ N1 ⁇ N, N1 is an integer. Different first image areas are image areas in different images.
  • the first image area is an image area to be detected.
  • the image area to be detected please refer to the above.
  • the first scenario is part of the road conditions or part of the driving field of view around the vehicle.
  • the first scene can be understood as the road conditions around the vehicle or the spatial area where one or more objects in the driving field of view are located.
  • the spatial area where an object is located refers to the area containing the object.
  • the first scene may be the space area where the traffic light is located, and the first image area may be the traffic light in each frame of the N1 frame image.
  • the image area where the image is located.
  • the first scene may be the space area where the vehicle is located, and the first image area may be the image of the vehicle in each frame of the N1 frame image. Image area.
  • the vehicle-mounted device may independently determine the first image area in the image.
  • the first image area in the image is determined according to at least one of the aforementioned features 1 to 3.
  • the vehicle-mounted device may first determine the first image area in the image, for example, determine the part according to at least one of the above-mentioned features 1 to 3 The first image area in the image; based on this inference, the first image area in the other images in the N frames of images is obtained.
  • the first image area in the image for example, determine the part according to at least one of the above-mentioned features 1 to 3
  • the first image area in the image based on this inference, the first image area in the other images in the N frames of images is obtained.
  • the shooting time interval of two adjacent frames of images in the N frames of images is less than or equal to a third threshold.
  • the shooting time interval of two frames of images refers to the time period between the moments when the two frames are shot.
  • the embodiment of the present application does not limit the specific value and value method of the third threshold. In this way, when the vehicle speed is faster, it helps to improve the image compensation effect of multi-frame super-division.
  • the vehicle speed when the vehicle speed is fast, if the time interval between two adjacent frames of images is relatively large, there may not be an image area corresponding to the same scene in the two frames of images.
  • a frame of image that appears includes: image information of traffic light 1 and image information of vehicles 1 to 3, and does not include image information of other objects; and the next frame of the image includes: image information of traffic light 2 and vehicle
  • the image information of 4 to 5 does not include the image information of other objects; then, there is no image area corresponding to the same scene in the two frames.
  • This may lead to the inability to use the image information in other frame images to compensate the image information in the image area corresponding to the scene, resulting in poor image compensation effect for multi-frame superdivision. Therefore, this optional implementation manner helps to improve the image compensation effect of multi-frame superdivision.
  • the shooting time interval between the first frame image and the last frame image in the N frames of images is less than or equal to a threshold.
  • a threshold In this way, when the vehicle speed is faster, it helps to improve the image compensation effect of multi-frame super-division.
  • S103 The vehicle-mounted device executes scene alignment of the N1 first image regions.
  • the step of aligning the first image area in the N1 frame of image may be an optional step.
  • the background and foreground of different images taken by the camera installed on the vehicle may change. In this case, the scene alignment can be performed before the superdivision calculation in S104 is executed.
  • Scene alignment is to ensure that the same or similar foreground and background are included in the image area corresponding to the same scene in multiple frames of images.
  • the image area corresponding to the same scene in the multi-frame image can be scaled to a uniform size, and some or all of the scene in the multi-frame image can be rotated by parameters such as angle, so as to ensure that the multi-frame corresponds to the
  • the image area of the scene contains uniform or similar foreground and background.
  • the in-vehicle device performs a hyperdivision operation on the N1 first image regions after scene alignment to obtain the first target image region.
  • This super-division operation can also be called a multi-frame super-division operation.
  • the in-vehicle device processes the first image area in the reference image after the scene is aligned (such as compensating for edge features, sharpening features, etc.) according to the first image area in the non-reference image after the scene is aligned to obtain the first image area.
  • the target image area wherein, the resolution of the first target image area is higher than the resolution of the first image area in the reference image.
  • the specific process can refer to the prior art.
  • the embodiment of the present application does not limit the application scenario of the first target image area, for example, it may be applied to a target object detection scenario.
  • the above method may further include the following step S105:
  • the vehicle-mounted device determines whether there is a target object in the first target image area. If it exists, the relative position of the target object and the vehicle is determined. If it does not exist, it ends.
  • step S105 can refer to the prior art.
  • the in-vehicle device performs target detection or image segmentation on the first target image area to determine whether the target object is contained in the first target image area, and if it exists, determines the relative position of the target object and the vehicle.
  • the super-division operation is specifically multi-frame super-division, which can combine the characteristic information of the image information in the multi-frame image to perform image compensation, which is similar to the prior art using single-frame super-division to perform image compensation. Compared with the compensation technical solution, it helps to improve the image compensation effect.
  • multiple image areas corresponding to the same scene in the multi-frame image are subjected to multi-frame super-division, instead of performing super-division operations on the multi-frame image itself, it is helpful to reduce the super-division. The complexity of the operation, thereby speeding up the processing rate of super-division.
  • the processing result of the image processing method provided by the embodiment of the present application is applied to assist automatic driving path planning, it helps to improve the accuracy of automatic driving path planning.
  • the method may further include: the vehicle-mounted device acquires the second image area in each frame of the N2 frames of the N frame of images.
  • the N2 second image regions of the N2 frame image correspond to the second scene. 2 ⁇ N2 ⁇ N, N2 is an integer. Different second image areas are image areas in different images.
  • the method further includes: steps S102' to S105'. S102' ⁇ S105' replace "first image area” in S102 ⁇ S105 with "second image area", "N1" with "N2", and "first target image area” with "second target” Image area".
  • the second image area is an image area to be detected that is different from the first image area.
  • the first image area and the second image area may or may not overlap partially.
  • the embodiment of the present application does not limit the magnitude relationship between N1 and N2.
  • the N1 frame image and the N2 frame image may or may not include the same image.
  • both the N1 frame image and the N2 frame image include a reference image (that is, a reference image used in the super-division operation).
  • a reference image that is, a reference image used in the super-division operation.
  • the N1 frame image can be the first to fifth frame images
  • the first image area can be the area where the image information of the traffic light 1 is located
  • the N2 frame image can be the first to seventh frame images
  • the second image area may be the area where the image information of the vehicle 1 is located.
  • FIG. 5 it is a schematic flowchart of an image processing method provided by an embodiment of this application.
  • the method can include:
  • the vehicle-mounted device performs preprocessing (such as image segmentation or target detection, etc.) on the first image in the N frames of images according to the candidate type set to obtain at least one candidate image area included in the first image, and each candidate image area Recognition results and the confidence level of each recognition result.
  • the first image may be any one of the N frames of images.
  • the first image is a reference image, and the reference image refers to a reference image in a multi-frame super division.
  • the candidate type set is a set composed of at least one candidate type.
  • the candidate type is the type of target object that the vehicle-mounted device needs to recognize (that is, the type of object that the vehicle-mounted device is interested in).
  • the type of object can be understood as what the object is. For example, if an object is a person, the type of the object is a person; if an object is a traffic light, the type of the object is a traffic light.
  • the candidate type set may include: people, cars, traffic lights, and so on.
  • the candidate image area is an area that contains image information of objects of candidate types.
  • the candidate type set is a set composed of people, cars, and traffic lights
  • the candidate image area includes: the image area where the image information of the person is located, the image area where the image information of the car is located, and the image where the image information of the traffic light is located. area.
  • the candidate image area when "preprocessing" is target detection, taking the first image as the image shown in FIG. 2 as an example, the area defined by the rectangular frame in FIG. 2 can be used as a candidate image area.
  • the "preprocessing" is image segmentation, taking the first image as the image shown in FIG. 3 as an example, each connected region shown in FIG. 3 can be used as a candidate image region.
  • the area occupied by the at least one candidate image area described in S202 may be part or all of the area in the first image.
  • the recognition result of the candidate image area may include: which kind of target object the image information in the candidate image area is. Optionally, it may also include the relative position between the target object and the vehicle.
  • the confidence of the recognition result of a candidate image area can be referred to as the confidence of the candidate image area, which can be understood as the accuracy of the recognition result of the candidate image area.
  • the image area taken by the object in the image taken by the camera in the vehicle is larger and clearer; if the object is closer to the vehicle Farther, the image area occupied by the object in the image taken by the camera is smaller and blurry. Therefore, in the same image, the confidence of the recognition result of the first candidate image area (that is, the candidate image area containing the image information of the object that is close to the vehicle) is generally higher than that of the second candidate image area ( That is, the confidence level of the recognition result of the candidate image region containing the image information of the object that is far away from the vehicle.
  • the recognition result of a candidate image area is: when the image information of the object included in the candidate image area is image information of a person, if the object is closer to the vehicle, the probability that the object is actually a "person” will be higher , That is, the confidence of the recognition result is high; if the object is far from the vehicle, the probability that the object is actually a "person” will be low, that is, the confidence of the recognition result is low, for example, the object may actually be a telephone pole Wait.
  • the vehicle-mounted device acquires at least one area to be detected in the first image.
  • the area to be detected is a candidate image area whose confidence is less than or equal to the first threshold.
  • the at least one area to be detected includes a first image area.
  • acquiring the first image area in the first image may include: acquiring the position of the first image area in the first image and the size of the first image area.
  • the at least one area to be detected includes the first image area as an example for description.
  • the at least one area to be detected may further include a second image area and the like.
  • the recognition result of the candidate image area obtained by the vehicle-mounted device in S202 has a higher accuracy. Therefore, in the subsequent steps, the vehicle-mounted device can no longer repeatedly detect (or identify) the image information included in these candidate image regions, that is, in the subsequent steps, only the confidence level is lower than or equal to the first.
  • the detection (or recognition) of the image information included in the candidate image area with a threshold helps reduce the detection complexity, thereby improving the detection efficiency.
  • the vehicle-mounted device acquires the speed v of the vehicle, the shooting time interval T between the first image and the second image, the relative angle ⁇ between the vehicle and the space area corresponding to the first image area when the first image is taken, and the second image is taken.
  • the speed of the vehicle can be variable or constant.
  • v, T, ⁇ 1, ⁇ 2, and ⁇ can all be measured by corresponding sensors in the vehicle, or the original information used to obtain some or all of these parameters can be measured by corresponding sensors in the vehicle.
  • the original information may be the heading of the vehicle at time t1 and the heading of the vehicle at time t2.
  • the in-vehicle device determines, according to v, T, ⁇ 1, and ⁇ 2, the relative distance R1 between the vehicle and the space area corresponding to the first image area when the first image is taken, and the distance between the vehicle and the first image area when the second image is taken.
  • the vehicle body coordinate system is X1 axis and Y1 axis to form the X1-Y1 coordinate system
  • the position of the vehicle is the origin of the X1-Y1 coordinate system
  • the Y1 axis is the forward direction of the vehicle.
  • the X1 axis is the vehicle tangential to the right direction, and the image taken by the camera is the first image; at t2, the vehicle body coordinate system is the X2 axis and the Y2 axis to form the X2-Y2 coordinate system, and the position of the vehicle is the X2-Y2 coordinate system
  • the origin of, and the Y2 axis is the forward direction of the vehicle, the X2 axis is the tangential right direction of the vehicle, and the image taken by the camera is the second image.
  • ⁇ 1 is the angle between the first image area in the first image and the Y1 axis
  • ⁇ 2 is the angle between the second image area in the second image and the Y2 axis.
  • the positions of ⁇ , R1 and R2 in the corresponding vehicle body coordinate system can be as shown in FIG. 6.
  • the dashed-dotted area in Figure 6 is the assumed imaging area of the front-view camera of the vehicle, where imaging area 1 is the imaging area at time t1 (ie the first image), and imaging area 2 is the imaging area at time t2 (ie the second image) ).
  • imaging area 1 is the imaging area at time t1 (ie the first image)
  • imaging area 2 is the imaging area at time t2 (ie the second image)
  • the imaging area will not be horizontal and vertical as shown in Figure 6.
  • it is assumed that the imaging area of the camera is a rectangular area in front of the vehicle.
  • R1 and R2 are determined based on the vehicle body coordinate system in FIG. 6. In actual implementation, R1 and R2 can also be determined based on other coordinate systems (such as the world coordinate system, camera coordinate system, etc.). This application is implemented The example does not limit this.
  • the vehicle-mounted device determines the first image area in the second image according to the first image area, R1, R2, and ⁇ in the first image.
  • the first image and the second image may be any two frames of the N frames of images described in S201.
  • the first image is a reference image (such as the first image) in the N frames of images
  • the second image is any non-reference image of the N frames of images. That is to say, the embodiments of the present application support the technical solution of "determining the first image area in any other frame of non-reference image according to the first image area in the reference image".
  • the first image and the second image are images that are adjacent in time series. That is to say, the embodiment of the present application supports the technical solution of "determining the first image area in a frame image according to the first image area in the previous frame image of the frame image".
  • ROI1 is the first image area in the first image. In the current example, it is a rectangle whose center point has been roughed.
  • ROI2 is the first image region in the second image, and S206 is specifically: determining the position and size of ROI2 in the second image according to the position and size of ROI1 in the first image, R1, R2, and ⁇ . The following describes how the position and size of ROI2 in the second image are determined.
  • is the steering angle of the vehicle from time t1 to time t2, it can be determined that the rotation angle of ROI2 relative to ROI1 is also ⁇ . It is assumed that the center points of ROI1 and ROI2 do not change at t1 and t2, that is, in the world coordinate system, the coordinates of the center point of ROI1 and the center point of ROI2 are the same. Therefore, the coordinates of the center point of ROI1 can be converted from the X1-Y1 coordinate system to the world coordinate system, and then to the X2-Y2 coordinate system, so that the coordinates of the center point of ROI2 in the X2-Y2 coordinate system can be obtained. So far, the position of ROI2 in the second image can be obtained.
  • ROI1' is obtained, as shown in Figure 7. Since in this example, compared to t1, when the vehicle travels to t2, it is closer to the spatial area corresponding to the first image area. Therefore, according to the imaging principle, theoretically a larger frame is required at t2. Frame the area with the same amount of information as at t1. Based on this, on the basis of ROI1', the length of ROI1' can be scaled (enlarged in this example) according to the ratio of R1 and R2 (optionally, according to a certain weight) to obtain the length of ROI2; Similarly, the width of ROI2 can be obtained. So far, the size of ROI2 can be obtained.
  • the size of the first image area in the second image is It must be larger than the size of the first image area in the first image (such as the area shown by the rectangular box in figure (b)) in order to make the first image area in the second image contain the person riding a motorcycle All image information.
  • the above S204 to S206 are described by taking the first image area in the second image inferred from the first image area in the first image as an example. Accordingly, for different second images, by executing S204-S206 one or more times, the first image area included in part or all of the N frames of images can be obtained.
  • the above parameters v, T, ⁇ 1, ⁇ 2, ⁇ , R1, and R2 are collectively referred to as vehicle body information.
  • the vehicle body information can be the information of the vehicle (such as the above v, T, ⁇ 1, ⁇ 2, and ⁇ , etc.) detected directly by the sensors and other equipment installed in the vehicle, or it can be the information detected by these sensors and other equipment
  • the information of the vehicle (such as R1 and R2, etc.) obtained by processing.
  • the above S204 to S206 are only examples of "a first image area in the second image obtained by reasoning based on vehicle body information" provided by the embodiment of this application, which is not applicable to the "reduction based on vehicle body information in the embodiment of this application".
  • the specific implementation of "the first image area in the second image” constitutes a limitation. In specific implementation, it is possible to implement "the first image area in the second image is obtained by reasoning" based on more or less vehicle body information than the vehicle body information listed above.
  • S207 The vehicle-mounted device executes scene alignment of the N1 first image regions.
  • the alignment process involves more scenes, and different scenarios can use different solutions for scene alignment.
  • the camera is a front-view camera and is installed directly in front of the vehicle as an example to describe the alignment process of the first image area in the two frames of images.
  • Example 1 The scene of the vehicle moving straight ahead.
  • the front-view camera collects the first image and the second image at different moments.
  • the first image and the second image include the first image regions ROI1 and ROI2, respectively. If the center points of ROI1 and ROI2 are exactly in front of the vehicle, it has been ensured that ROI2 contains the same information of the scene (including foreground and background) as ROI1 when acquiring the first image area, so the scenes are aligned, which can be specifically Scale ROI1 (in this case, zoom in) to the size of ROI2.
  • Example 2 A scene where the vehicle turns to the front right.
  • the front-view camera collects the first image and the second image at different moments (ie t1 and t2).
  • the first image and the second image include the first image regions ROI1 and ROI2, respectively .
  • ROI1 and ROI2 are at the front right of the vehicle, when the vehicle turns to the right (that is, the origin of the X2-Y2 coordinate system is at the upper right of the X1-Y1 coordinate system), the angle of view will inevitably include more information on the left side of the ROI, as shown in Figure 9. Show.
  • the image information of the object included in the first image area is a rectangle as an example for description.
  • the vehicle body coordinate system is the X1-Y1 coordinate system.
  • the vehicle body coordinate system is the X2-Y2 coordinate system.
  • the position of the camera is the origin of the coordinate system.
  • the amount of information contained in the scene in the first image area in the first image and the second image will be different, so that in addition to the ROI1 and ROI2 sizes, the direction of the scene in the field of view must also be aligned once, that is to say , It is necessary to obtain the scene alignment angle between the first image area in the first image and the first image area in the second image.
  • the scene alignment angle may be ⁇ in FIG. 10.
  • the scene in ROI2 can be mapped according to the scene alignment angle ⁇ .
  • Fig. 11 shows ROI2 is mapped to the plane where ROI1 is located by taking ⁇ as the angle and pixels as the unit to obtain ROI2’.
  • the plane where ROI2’ is located is the plane where ROI1 is located.
  • compare ROI1 to ROI2' for zooming in this example, zooming in
  • the example in S207 takes the alignment of the first image area in the first image and the second image as an example for description.
  • the first image and the second image are only for distinguishing any two images in the N1 image.
  • the first image and the second image in the alignment process may correspond to the same or different from the first image and the second image in the process of determining the first image area.
  • example 2 is only an example of "a realization of scene alignment based on vehicle body information" provided by the embodiment of this application, which is not applicable to the realization of "realization of scene alignment based on vehicle body information” in the embodiment of this application.
  • the way constitutes limitation. In specific implementation, it is possible to achieve scene alignment based on more or less vehicle body information than the vehicle body information listed in Example 2.
  • S208 to S209 The above S104 to S105 can be referred to, of course, the embodiment of the present application is not limited thereto.
  • a candidate image area with a confidence level lower than the first threshold in one frame of image is selected as the first image area, and the first image area in the multi-frame image is super-divided. .
  • the first image area in the multi-frame image is super-divided.
  • the first image area in one frame of image and the vehicle body information is determined, so that when performing super-division operations, it is helpful for vehicle equipment Get more spatial information, thereby improving the accuracy of super-division operations.
  • FIG. 12 it is a schematic flowchart of an image processing method provided by an embodiment of this application.
  • the method can include:
  • the vehicle-mounted device performs preprocessing (such as image segmentation, etc.) on the first image in the N frames of images to obtain image information of the drivable area in the first image.
  • preprocessing such as image segmentation, etc.
  • the drivable area is the area between the first objects that are away from the vehicle in all directions in the field of view.
  • the area enclosed by the black line in FIG. 13 represents a schematic diagram of a drivable area.
  • the in-vehicle device uses the image information corresponding to the space area in the drivable area whose distance to the vehicle is greater than or equal to the second threshold as the image information in the second image The first image area.
  • the in-vehicle device may first determine the position of the vehicle in which it is located in the image captured by the camera, and then use the image area in the second image whose distance from the position is greater than or equal to a threshold as the second image The first image area.
  • the ratio between the second threshold and the threshold is equal to the ratio between the spatial distance and the image distance (that is, the image distance mapped from the spatial distance to the image).
  • the image captured by the camera does not contain the image information of the vehicle (that is, the vehicle in which the on-board equipment is located in this embodiment), it can be based on the location information of the camera in the vehicle, or optionally based on The motion state of the vehicle (such as turning or going straight) determines the position of the vehicle in the image.
  • the corresponding position of the vehicle in the image may be the middle of the lower boundary of the picture, as shown in FIG. 14.
  • Figure 14 also illustrates the first image area in this case.
  • the corresponding position of the vehicle in the image captured by the camera may be the lower right corner of the picture.
  • the method for determining the corresponding position of the vehicle in the image captured by the camera is not limited to this, for example, reference may be made to the prior art.
  • S304-S309 The foregoing S204-S209 can be referred to, of course, the embodiment of the present application is not limited thereto.
  • the vehicle-mounted device can obtain the first image area included in each frame of the N1 frames of the N frames of images.
  • an image area corresponding to a space area in the drivable area of the vehicle whose distance from the vehicle is greater than or equal to the second threshold is taken as the first image Region, and super-divide the first image region in the multi-frame image.
  • the drivable area is defined as the "area between the first object of the vehicle in various directions in the field of view", there is no guarantee that the determined drivable area does not contain objects (or targets).
  • this embodiment is proposed based on this. In this way, it is helpful to improve the accuracy of detecting the target object on the basis of the existing technology.
  • the first image area in one frame of image and the vehicle body information the first image area in another frame of image is determined, so that when performing super-division operations, it is helpful for vehicle equipment Get more spatial information, thereby improving the accuracy of super-division operations.
  • FIG. 15 it is a schematic flowchart of an image processing method provided by an embodiment of this application.
  • the method can include:
  • the in-vehicle device uses the area at the preset position in the first image in the N frames of images as the first image area.
  • the position of the first image area in the first image may be the same or different.
  • the target object is a traffic light
  • the image information of the traffic light is usually above in a frame of image
  • the area (such as the upper two-fifths of the area) at the upper preset position in one frame of image can be taken as The first image area.
  • a schematic diagram of the first image area is shown in FIG. 16.
  • the target object is a speed limit sign on an expressway
  • the image information of the speed limit sign on an expressway is usually on the right side of an image
  • the right side of an image can be pre-defined. Set the location area (such as the area on the right side) as the first image area.
  • S403 to S408 The foregoing S204 to S209 can be referred to, of course, the embodiment of the present application is not limited thereto.
  • the vehicle-mounted device can obtain the first image area included in each frame of the N1 frames of the N frames of images.
  • an area at a preset position in one frame of image is used as the first image area, and super-division processing is performed on the first image area in multiple frames of images.
  • This method is relatively simple and convenient to implement.
  • the first image area in another frame of image is determined, so that when performing super-division operations, it is helpful for vehicle equipment Get more spatial information, thereby improving the accuracy of super-division operations.
  • the embodiments of the present application may divide the in-vehicle equipment into functional modules according to the foregoing method examples.
  • each functional module may be divided corresponding to each function, or two or more functions may be integrated into one processing module.
  • the above-mentioned integrated modules can be implemented in the form of hardware or software functional modules. It should be noted that the division of modules in the embodiments of the present application is illustrative, and is only a logical function division, and there may be other division methods in actual implementation.
  • a device provided by an embodiment of this application may specifically be a schematic structural diagram of a vehicle-mounted device 170.
  • the vehicle-mounted device 170 may be used to execute the steps performed by the vehicle-mounted device in the method shown in FIG. 4, FIG. 5, FIG. 12, or FIG.
  • the in-vehicle device 170 may include: a first acquisition module 1701, a second acquisition module 1702, and a super-division module 1703.
  • the first acquisition module 1701 is configured to acquire a multi-frame image
  • the multi-frame image includes image information of the surrounding road of the vehicle where the on-board device is located.
  • the second acquisition module 1702 is configured to acquire the first image area in each frame of the multi-frame image; wherein, the multiple first image areas of the multi-frame image correspond to the first scene.
  • the super-division module 1703 is configured to perform super-division operations on the multiple first image regions.
  • the first acquisition module 1701 may be used to perform S101
  • the second acquisition module 1702 may be used to perform S102
  • the super-division module 1703 may be used to perform S104.
  • the in-vehicle device 170 further includes: a determining module 1704, configured to determine that the image information of the target object exists in the image area obtained by the super-division operation.
  • a determining module 1704 configured to determine that the image information of the target object exists in the image area obtained by the super-division operation.
  • the determining module 1704 may be used to perform S105.
  • the first image area is an area with a confidence level lower than or equal to a first threshold; or, the first image area corresponds to the difference between the drivable area of the vehicle and the vehicle.
  • the multi-frame image includes a first image and a second image.
  • the second acquiring module 1702 is specifically configured to acquire the first image area in the second image.
  • Obtaining the first image area in the second image includes: obtaining the first image area in the second image according to the first image area in the first image and the first body information of the vehicle.
  • the first vehicle body information may include at least one of a first relative distance, a second relative distance, and a first vehicle steering angle.
  • the first relative distance is the relative distance between the vehicle and the space area corresponding to the first image area when the first image is taken
  • the second relative distance is the distance between the vehicle and the first image area when the second image is taken.
  • the relative distance between the spatial regions, the first vehicle steering angle is the angle between the direction of the vehicle in the time interval of shooting the first image and the second image.
  • the second acquisition module 1702 may be used to perform S204 and S205.
  • the first vehicle body information may also include vehicle height parameters, such as the height of the camera from the ground and/or the vehicle body height.
  • the super-division module 1703 is specifically configured to perform super-division operations on multiple first image regions after scene alignment.
  • the super-division module 1703 may be used to execute S104 in FIG. 4, S208 in FIG. 5, S308 in FIG. 12, or S407 in FIG. 15.
  • the multi-frame image includes a third image and a fourth image.
  • the in-vehicle device 170 further includes an alignment module 1705, configured to perform scene alignment of the plurality of first image regions according to the second body information of the vehicle.
  • the second vehicle body information may include at least one of a first relative angle, a second relative angle, and a second vehicle steering angle.
  • the first relative angle is the relative angle between the vehicle and the space area corresponding to the first image area when the third image is taken
  • the second relative angle is the relative angle between the vehicle and the first image area when the fourth image is taken.
  • the second vehicle steering angle is the angle between the direction of the vehicle in the time interval of shooting the third image and the fourth image.
  • the second vehicle body information may also include vehicle height parameters, such as the height of the camera from the ground and/or the vehicle body height.
  • the multi-frame images are consecutive multi-frame images in time series.
  • the time interval between the shooting moment of the first frame image and the shooting moment of the last frame image in the multi-frame image is less than or equal to a third threshold.
  • the second obtaining module 1702 is further configured to: obtain a second image area in each frame of the multi-frame image; wherein, the multiple second image areas of the multi-frame image correspond to the second scene.
  • the super-division module 1703 is further configured to perform super-division operations on the multiple second image regions.
  • the above-mentioned first acquisition module 1701 may be implemented through the I/O interface 115 in FIG. 1.
  • At least one of the above-mentioned second acquisition module 1702, super-division module 1703, determination module 1704, and alignment module 1705 One can be implemented by the processor 103 in FIG. 1 calling the application program 143.
  • the program can be stored in a computer-readable storage medium.
  • the aforementioned storage medium may be a read-only memory, a random access memory, and the like.
  • the above-mentioned processing unit or processor may be a central processing unit, a general-purpose processor, an application specific integrated circuit (ASIC), a microprocessor (digital signal processor, DSP), a field programmable gate array (field programmable gate array, FPGA) or other programmable logic devices, transistor logic devices, hardware components, or any combination thereof.
  • the embodiments of the present application also provide a computer program product containing instructions, which when the instructions are run on a computer, cause the computer to execute any one of the methods in the foregoing embodiments.
  • the computer program product includes one or more computer instructions. When the computer program instructions are loaded and executed on the computer, the processes or functions described in the embodiments of the present application are generated in whole or in part.
  • the computer may be a general-purpose computer, a special-purpose computer, a computer network, or other programmable devices.
  • Computer instructions may be stored in a computer-readable storage medium, or transmitted from one computer-readable storage medium to another computer-readable storage medium.
  • computer instructions may be transmitted from a website, computer, server, or data center through a cable (such as Coaxial cable, optical fiber, digital subscriber line (digital subscriber line, DSL) or wireless (such as infrared, wireless, microwave, etc.) transmission to another website site, computer, server, or data center.
  • the computer-readable storage medium may be any available medium that can be accessed by a computer or may include one or more data storage devices such as a server or a data center that can be integrated with the medium.
  • the usable medium may be a magnetic medium (for example, a floppy disk, a hard disk, and a magnetic tape), an optical medium (for example, a DVD), or a semiconductor medium (for example, a solid state disk (SSD)).
  • the foregoing devices for storing computer instructions or computer programs provided in the embodiments of the present application are non-transitory. .

Landscapes

  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Traffic Control Systems (AREA)
  • Image Analysis (AREA)

Abstract

Disclosed are an image processing method and apparatus, relating to the technical field of piloted driving or assisted driving and contributing to enhancing an image compensation effect. The method is applied to a vehicle-mounted device. The method may comprise: acquiring a plurality of frames of images, wherein the plurality of frames of images comprise image information of roads surrounding a vehicle on which the vehicle-mounted device is mounted (S101); acquiring a first image region in each frame of image of the plurality of frames of images, wherein a plurality of first image regions of the plurality of frames of images correspond to a first scene (S102); and performing a super-resolution operation on the plurality of first image regions (S104). The method can be used for target detection and tracking in assisted driving and piloted driving.

Description

图像处理方法和装置Image processing method and device
本申请要求于2019年07月12日提交国家知识产权局、申请号为201910633070.3、申请名称为“图像处理方法和装置”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。This application claims the priority of a Chinese patent application filed with the State Intellectual Property Office with the application number 201910633070.3 and the application name "Image Processing Method and Apparatus" on July 12, 2019, the entire content of which is incorporated into this application by reference.
技术领域Technical field
本申请涉及自动驾驶或辅助驾驶技术领域,尤其涉及图像处理方法和装置。This application relates to the field of automatic driving or assisted driving technology, and in particular to image processing methods and devices.
背景技术Background technique
图像超分辨率(image super resolution)技术在监控设备、卫星图像和自动驾驶技术领域都有重要的应用价值。图像超分辨率技术包括单帧超分技术和多帧超分技术。单帧超分是指由一帧低分辨率图像恢复出一帧高分辨率图像。多帧超分是指由多帧低分辨率图像恢复出一帧高分辨率图像。Image super resolution technology has important application value in surveillance equipment, satellite imagery and autonomous driving technology. Image super-resolution technology includes single-frame super-division technology and multi-frame super-division technology. Single-frame super-division refers to recovering a high-resolution image from a low-resolution image. Multi-frame super-division refers to recovering a high-resolution image from multiple low-resolution images.
在自动驾驶技术领域中,通常需要对道路上的对象进行检测,例如检测一个对象是人或交通灯等,以及该对象与车辆之间的距离等,以辅助自动驾驶路径规划。目前,通常对所拍摄的车辆周边道路的图像进行单帧超分,得到高分辨率图像,并基于该高分辨率图像对道路上的对象进行检测。由于单帧超分仅对一帧图像中的对象的边缘特征、锐化特征等进行补偿,而一帧图像中的特征信息较少,因此图像补偿效果较差。In the field of automatic driving technology, it is usually necessary to detect objects on the road, such as detecting whether an object is a person or a traffic light, and the distance between the object and the vehicle, etc., to assist automatic driving path planning. Currently, a single-frame superdivision is usually performed on the captured image of the road surrounding the vehicle to obtain a high-resolution image, and objects on the road are detected based on the high-resolution image. Since the single-frame super-division only compensates for the edge features, sharpening features, etc. of the object in one frame of image, while the feature information in one frame of image is less, the image compensation effect is poor.
发明内容Summary of the invention
本申请实施例提供了图像处理方法和装置,有助于提高图像补偿效果。The embodiments of the present application provide an image processing method and device, which help improve the effect of image compensation.
第一方面,提供一种图像处理方法,应用于车载设备,该方法包括:首先,获取多帧图像,该多帧图像包括该车载设备所在车辆的周边道路的图像信息;然后,获取该多帧图像的每帧图像中的第一图像区域;其中,该多帧图像的多个第一图像区域(其中每帧图像包括一个第一图像区域)对应于第一场景;接着,对该多个第一图像区域进行超分运算。该超分运算具体为多帧超分,这样,可以结合多帧图像中的图像信息的特征信息进行图像补偿,与现有技术中使用单帧超分进行图像补偿的技术方案相比,有助于提高图像补偿效果。另外,由于本技术方案中,对该多帧图像中对应于同一场景的多个图像区域进行多帧超分,而非对该多帧图像本身进行超分运算,因此,有助于降低超分运算的复杂度,从而加快超分处理速率。进一步地,当本技术方案提供的图像处理方法的处理结果应用于辅助自动驾驶路径规划场景时,有助于提高自动驾驶路径规划的精确度。In a first aspect, an image processing method is provided, which is applied to an in-vehicle device. The method includes: firstly, acquiring a multi-frame image, the multi-frame image including image information of the surrounding road of the vehicle where the in-vehicle device is located; then, acquiring the multi-frame The first image area in each frame of the image; wherein the multiple first image areas (where each frame of the image includes a first image area) of the multi-frame image correspond to the first scene; then, the multiple first image areas A super-division operation is performed on an image area. The super-division operation is specifically a multi-frame super-division. In this way, the feature information of the image information in the multi-frame image can be combined for image compensation, which is helpful compared with the technical solution of using a single-frame super-division for image compensation in the prior art. To improve the image compensation effect. In addition, since in the present technical solution, multiple image regions corresponding to the same scene in the multi-frame image are subjected to multi-frame super-division, instead of performing super-division operations on the multi-frame image itself, it helps to reduce the super-division The complexity of the operation, thereby speeding up the processing rate of super-division. Further, when the processing result of the image processing method provided by the present technical solution is applied to an assisted automatic driving path planning scene, it is helpful to improve the accuracy of the automatic driving path planning.
作为示例,第一场景可以理解为该车辆周边的路况或者行驶视野中的一个或多个对象所在的空间区域。第一图像区域可以是一帧图像中的部分或全部区域。第一图像区域中可以包含或不包含目标对象的图像信息。目标对象可以是预定义的,当然本申请实施例不限于此。As an example, the first scene may be understood as the road conditions around the vehicle or the spatial area where one or more objects in the driving field of view are located. The first image area may be part or all of the area in one frame of image. The first image area may or may not contain the image information of the target object. The target object may be predefined, of course, the embodiment of the present application is not limited to this.
在一种可能的设计中,该方法还包括:确定超分运算得到的图像区域中存在目标对象的图像信息。In a possible design, the method further includes: determining that the image information of the target object exists in the image area obtained by the hyperdivision operation.
在一种可能的设计中,该方法还包括:检测该车辆与该目标对象之间的相对位置。相对位置包括相对距离和相对角度,相对角度包括方位角和/或俯仰角。在一个示例中, 该相对位置可以用于辅助自动驾驶路径规划。In a possible design, the method further includes: detecting the relative position between the vehicle and the target object. The relative position includes a relative distance and a relative angle, and the relative angle includes an azimuth angle and/or a pitch angle. In one example, the relative position may be used to assist automatic driving path planning.
在一种可能的设计中,对于该多帧图像的每帧图像:第一图像区域是置信度低于或等于第一阈值的区域;或者,第一图像区域对应于该车辆的可行驶区域中与该车辆之间的距离大于或等于第二阈值的空间区域;或者,第一图像区域是预设位置的区域。该可能的设计提供了几种第一图像区域所具有的特征,具体实现时,可以基于这些特征之一确定第一图像区域。In a possible design, for each image of the multi-frame image: the first image area is an area with a confidence level lower than or equal to the first threshold; or, the first image area corresponds to the driveable area of the vehicle A space area where the distance from the vehicle is greater than or equal to the second threshold; or, the first image area is an area at a preset position. This possible design provides several features of the first image area. In specific implementation, the first image area can be determined based on one of these features.
在一种可能的设计中,该多帧图像包括第一图像和第二图像;获取该多帧图像的每帧图像中的第一图像区域,包括:获取第二图像中的第一图像区域。获取第二图像中的第一图像区域,包括:根据第一图像中的第一图像区域(具体可以包括:第一图像中的第一图像区域在第一图像中的位置和大小)和该车辆的第一车体信息,获取第二图像中的第一图像区域(具体可以包括:获取第二图像中的第一图像区域在第二图像中的位置和大小)。也就是说,本申请实施例支持基于车体信息,由第一图像中的第一图像区域推理得到第一图像中的第一图像区域的技术方案,这有助于提高超分运算的精确度。In a possible design, the multi-frame image includes a first image and a second image; acquiring the first image area in each frame of the multi-frame image includes: acquiring the first image area in the second image. Acquiring the first image area in the second image includes: according to the first image area in the first image (which may specifically include: the position and size of the first image area in the first image in the first image) and the vehicle Obtaining the first image area in the second image (which may specifically include: obtaining the position and size of the first image area in the second image in the second image). That is to say, the embodiments of the present application support a technical solution to infer the first image area in the first image from the first image area in the first image based on the vehicle body information, which helps to improve the accuracy of the superdivision calculation .
其中,车体信息(包括第一车体信息和下文中的第二车体信息),可以是直接通过车辆中安装的传感器等设备检测得到的车辆的信息,也可以是对这些传感器等设备检测到的信息进行处理得到的该车辆的信息。Among them, the vehicle body information (including the first vehicle body information and the second vehicle body information below) can be the vehicle information directly detected by sensors and other equipment installed in the vehicle, or it can be the detection of these sensors and other equipment. Information about the vehicle obtained by processing the received information.
在一种可能的设计中,第一车体信息可以包括:第一相对距离、第二相对距离和第一车辆转向角度中的至少一种。第一相对距离是拍摄第一图像时该车辆与第一图像区域所对应的空间区域之间的相对距离。第二相对距离是拍摄第二图像时该车辆与第一图像区域所对应的空间区域之间的相对距离。第一车辆转向角度是在第一图像和第二图像的拍摄时间间隔内,该车辆的朝向之间的夹角。In a possible design, the first vehicle body information may include at least one of the first relative distance, the second relative distance, and the first vehicle steering angle. The first relative distance is the relative distance between the vehicle and the space area corresponding to the first image area when the first image is taken. The second relative distance is the relative distance between the vehicle and the space area corresponding to the first image area when the second image is taken. The first vehicle steering angle is the angle between the direction of the vehicle in the time interval of shooting the first image and the second image.
在一种可能的设计中,第一图像是多帧超分中的参考图像,第二图像是多帧超分中的任意一帧非参考图像。In a possible design, the first image is a reference image in the multi-frame super division, and the second image is any non-reference image in the multi-frame super division.
在一种可能的设计中,对该多个第一图像区域进行超分运算,包括:对景物对齐后的所该多个第一图像区域进行超分运算。这是在考虑到车辆在运行过程中,前景和背景均可能变化,而提出的技术方案。基于此,有助于提高超分运算的精确度。In a possible design, performing a super-division operation on the plurality of first image regions includes: performing a super-division operation on all the plurality of first image regions after scene alignment. This is a technical solution that takes into account that the foreground and background may change during the operation of the vehicle. Based on this, it helps to improve the accuracy of the super-division calculation.
在一种可能的设计中,多帧图像包括第三图像和第四图像;在对景物对齐后的该多个第一图像区域进行超分运算之前,该方法还包括:根据该车辆的第二车体信息,执行该多个第一图像区域的景物对齐。也就是说,本申请实施例支持基于车体信息实现景物对齐的技术方案,这有助于提高超分运算的精确度。In a possible design, the multiple frames of images include a third image and a fourth image; before the superdivision operation is performed on the multiple first image regions after scene alignment, the method further includes: according to the second image of the vehicle Car body information, performing scene alignment of the plurality of first image areas. That is to say, the embodiments of the present application support a technical solution for realizing scene alignment based on vehicle body information, which helps to improve the accuracy of super-division calculation.
在一种可能的设计中,第二车体信息包括第一相对角度、第二相对角度和第二车辆转向角度中的至少一种。其中,第一相对角度是拍摄第三图像时该车辆与第一图像区域所对应的空间区域之间的相对角度。第二相对角度是拍摄第四图像时该车辆与第一图像区域所对应的空间区域之间的相对角度。第二车辆转向角度是在第三图像和第四图像的拍摄时间间隔内,该车辆的朝向之间的夹角。In a possible design, the second vehicle body information includes at least one of the first relative angle, the second relative angle, and the second vehicle steering angle. The first relative angle is the relative angle between the vehicle and the space area corresponding to the first image area when the third image is taken. The second relative angle is the relative angle between the vehicle and the space area corresponding to the first image area when the fourth image is taken. The second steering angle of the vehicle is the angle between the direction of the vehicle in the time interval of shooting the third image and the fourth image.
其中,第三图像与第一图像或第二图像可以相同或不同,第四图像与第一图像或第二图像可以相同或不同,且第三图像和第四图像不同。The third image and the first image or the second image may be the same or different, the fourth image and the first image or the second image may be the same or different, and the third image and the fourth image are different.
在一种可能的设计中,该多帧图像是时序上连续的多帧图像。这样,方便处理。In a possible design, the multi-frame images are consecutive multi-frame images in time series. In this way, it is convenient to handle.
在一种可能的设计中,该多帧图像中的首帧图像的拍摄时刻与末帧图像的拍摄时刻之间的时间间隔小于或等于第三阈值。这样,有助于提高超分运算精确度。In a possible design, the time interval between the shooting moment of the first frame image and the shooting moment of the last frame image in the multi-frame images is less than or equal to the third threshold. In this way, it helps to improve the accuracy of the super-division calculation.
在一种可能的设计中,该方法还包括:获取该多帧图像的每帧图像中的第二图像区域;其中,该多帧图像的多个第二图像区域对应于第二场景;然后,对该多个第二图像区域进行超分运算。也就是说,本申请支持一帧图像中包括多个待检测图像区域的方案。第一图像区域与第二图像区域之间可以有交叠或无交叠。第一场景与第二场景不同。In a possible design, the method further includes: acquiring a second image area in each frame of the multi-frame image; wherein the multiple second image areas of the multi-frame image correspond to the second scene; then, Perform a hyperdivision operation on the plurality of second image regions. In other words, this application supports the solution of including multiple image regions to be detected in one frame of image. There can be overlap or no overlap between the first image area and the second image area. The first scene is different from the second scene.
第二方面,提供了一种图像处理装置,该装置可用于执行上述第一方面或第一方面的任一种可能的设计提供的任一种方法。示例的,该装置可以是车载设备或芯片等。In a second aspect, an image processing device is provided, which can be used to execute any method provided in the first aspect or any possible design of the first aspect. For example, the device may be an in-vehicle device or a chip.
在一种可能的设计中,可以根据上述第一方面或第一方面的任一种可能的设计提供的方法对该装置进行功能模块的划分,例如,可以对应各个功能划分各个功能模块,也可以将两个或两个以上的功能集成在一个处理模块中。In a possible design, the device may be divided into functional modules according to the method provided in the first aspect or any of the possible designs of the first aspect. For example, each functional module may be divided corresponding to each function, or Integrate two or more functions into one processing module.
在一种可能的设计中,该装置可以包括存储器和处理器。存储器用于存储计算机程序。处理器用于调用该计算机程序,以执行第一方面或第一方面的任一种可能的设计提供的方法。In one possible design, the device may include a memory and a processor. The memory is used to store computer programs. The processor is used to invoke the computer program to execute the first aspect or the method provided by any possible design of the first aspect.
第三方面,提供了一种计算机可读存储介质,如计算机非瞬态的可读存储介质。其上储存有计算机程序(或指令),当该计算机程序(或指令)在计算机上运行时,使得该计算机执行上述第一方面或第一方面的任一种可能的设计提供的任一种方法。In a third aspect, a computer-readable storage medium is provided, such as a non-transitory computer-readable storage medium. A computer program (or instruction) is stored thereon, and when the computer program (or instruction) runs on a computer, the computer executes any method provided by the first aspect or any possible design of the first aspect .
第四方面,提供了一种计算机程序产品,当其在计算机上运行时,使得第一方面或第一方面的任一种可能的设计提供的任一种方法被执行。In a fourth aspect, a computer program product is provided, which, when running on a computer, enables any method provided in the first aspect or any possible design of the first aspect to be executed.
可以理解的是,上述提供的任一种图像处理装置、计算机存储介质、计算机程序产品或系统等均可以应用于上文所提供的对应的方法,因此,其所能达到的有益效果可参考对应的方法中的有益效果,此处不再赘述。It is understandable that any image processing device, computer storage medium, computer program product or system provided above can be applied to the corresponding method provided above. Therefore, the beneficial effects that can be achieved can refer to the corresponding The beneficial effects of the method are not repeated here.
附图说明Description of the drawings
图1为可适用于本申请实施例的一种计算机系统的结构示意图;FIG. 1 is a schematic structural diagram of a computer system applicable to an embodiment of the present application;
图2为可适用于本申请实施例的一种目标检测的结果示意图;FIG. 2 is a schematic diagram of a result of target detection applicable to an embodiment of the present application;
图3为可适用于本申请实施例的一种图像分割的示意图;FIG. 3 is a schematic diagram of an image segmentation applicable to an embodiment of the present application;
图4为本申请实施例提供的一种图像处理方法的流程示意图;4 is a schematic flowchart of an image processing method provided by an embodiment of this application;
图5为本申请实施例提供的另一种图像处理方法的流程示意图;FIG. 5 is a schematic flowchart of another image processing method provided by an embodiment of the application;
图6为本申请实施例提供的一种车体信息在车体坐标系中的示意图;6 is a schematic diagram of a vehicle body information in a vehicle body coordinate system provided by an embodiment of the application;
图7为本申请实施例提供的一种基于ROI1获得ROI2的示意图;FIG. 7 is a schematic diagram of obtaining ROI2 based on ROI1 according to an embodiment of the application;
图8为本申请实施例提供的一种ROI1和ROI2的场景示意图;FIG. 8 is a schematic diagram of a scene of ROI1 and ROI2 provided by an embodiment of the application;
图9为本申请实施例提供的一种车辆右转弯时视角变换的示意图;FIG. 9 is a schematic diagram of a view angle change when a vehicle turns right according to an embodiment of the application;
图10为本申请实施例提供的一种获取景物对齐角度的示意图;FIG. 10 is a schematic diagram of obtaining an alignment angle of a scene according to an embodiment of the application;
图11为本申请实施例提供的一种将ROI2映射到ROI1所在的平面的示意图;FIG. 11 is a schematic diagram of mapping ROI2 to a plane where ROI1 is located according to an embodiment of the application;
图12为本申请实施例提供的另一种图像处理方法的流程示意图;FIG. 12 is a schematic flowchart of another image processing method provided by an embodiment of the application;
图13为本申请实施例提供的一种可行驶区域的示意图;FIG. 13 is a schematic diagram of a drivable area provided by an embodiment of the application;
图14为本申请实施例提供的一种第一图像区域的示意图;FIG. 14 is a schematic diagram of a first image area provided by an embodiment of this application;
图15为本申请实施例提供的另一种图像处理方法的流程示意图;15 is a schematic flowchart of another image processing method provided by an embodiment of the application;
图16为本申请实施例提供的另一种第一图像区域的示意图;FIG. 16 is a schematic diagram of another first image area provided by an embodiment of this application;
图17为本申请实施例提供的一种车载设备的结构示意图。FIG. 17 is a schematic structural diagram of a vehicle-mounted device provided by an embodiment of the application.
具体实施方式Detailed ways
如图1所示,为可适用于本申请实施例的一种计算机系统的结构示意图。其中,该计算机系统可以位于车辆上,该计算机系统可以包括车载设备101,以及与车载设备直接或间接连接的设备/器件/网络等。参见图1,车载设备101包括处理器103,处理器103和系统总线105耦合。处理器103可以是一个或者多个处理器,其中每个处理器都可以包括一个或多个处理器核。显示适配器(video adapter)107,显示适配器可以驱动显示器109,显示器109和系统总线105耦合。系统总线105通过总线桥111和输入输出(input/output,I/O)总线113耦合。I/O接口115和I/O总线耦合。I/O接口115和多种I/O设备进行通信,比如输入设备117(如键盘、鼠标、触摸屏等),多媒体盘(media tray)121(如只读光盘(compact disc read-only memory,CD-ROM)、多媒体接口等)。收发器123(可以发送和/或接受无线电通信信号),摄像头155(可以捕捉景田和动态数字视频图像)和外部通用串行总线(universal serial bus,USB)接口125。其中,可选地,和I/O接口115相连接的接口可以是USB接口。As shown in FIG. 1, it is a schematic structural diagram of a computer system applicable to the embodiments of the present application. Wherein, the computer system may be located on the vehicle, and the computer system may include the vehicle-mounted equipment 101, and the equipment/device/network connected directly or indirectly with the vehicle-mounted equipment. Referring to FIG. 1, the vehicle-mounted device 101 includes a processor 103, and the processor 103 is coupled to a system bus 105. The processor 103 may be one or more processors, where each processor may include one or more processor cores. A display adapter (video adapter) 107 can drive the display 109, and the display 109 is coupled to the system bus 105. The system bus 105 is coupled to an input/output (I/O) bus 113 through a bus bridge 111. The I/O interface 115 is coupled to the I/O bus. The I/O interface 115 communicates with a variety of I/O devices, such as input devices 117 (such as keyboard, mouse, touch screen, etc.), media tray 121 (such as compact disc read-only memory, CD -ROM), multimedia interface, etc.). The transceiver 123 (can send and/or receive radio communication signals), the camera 155 (can capture scene and dynamic digital video images), and an external universal serial bus (USB) interface 125. Wherein, optionally, the interface connected to the I/O interface 115 may be a USB interface.
其中,处理器103可以是任何传统处理器,包括精简指令集计算(reduced instruction set computer,RISC)处理器、复杂指令集计算(complex instruction set computer,CISC)处理器或上述的组合。可选地,处理器可以是诸如专用集成电路(application specific integrated circuit,ASIC)的专用装置。可选地,处理器103可以是神经网络处理器或者是神经网络处理器和上述传统处理器的组合。例如,处理器103可以是中央处理器(central processing unit,CPU)。The processor 103 may be any traditional processor, including a reduced instruction set computer (RISC) processor, a complex instruction set computer (CISC) processor, or a combination of the foregoing. Optionally, the processor may be a dedicated device such as an application specific integrated circuit (ASIC). Optionally, the processor 103 may be a neural network processor or a combination of a neural network processor and the foregoing traditional processors. For example, the processor 103 may be a central processing unit (CPU).
摄像头155可以是任何用于采集图像的摄像头,例如,可以是单目摄像头或双目摄像头等。摄像头的数量可以有一个或多个,每个摄像头可以位于车辆的前方、后方或侧方等。为了方便描述,下文中的具体示例中均是以摄像头位于车辆内的正前方为例进行说明的。在本申请实施例中,摄像头155可以用于采集该车辆的周边环境(包括周边道路等)的信息。在一个示例中,摄像头155中可以包含一个软件模块,该软件模块可以用于记录摄像头所拍摄的图像的拍摄时间。或者,用于记录拍摄时间的模块也可以是一个与摄像头155连接的硬件。在一个示例中,摄像头155相对于车辆的位置可以是固定不变的。在另一个示例中,摄像头155相对于车辆的位置可以是变化的,例如摄像头155可以进行旋转拍摄。The camera 155 may be any camera used to collect images, for example, it may be a monocular camera or a binocular camera. The number of cameras can be one or more, and each camera can be located in the front, rear, or side of the vehicle. For the convenience of description, the following specific examples all take the camera located directly in front of the vehicle as an example. In the embodiment of the present application, the camera 155 may be used to collect information about the surrounding environment (including surrounding roads, etc.) of the vehicle. In an example, the camera 155 may include a software module, and the software module may be used to record the shooting time of the image taken by the camera. Alternatively, the module for recording the shooting time may also be a piece of hardware connected to the camera 155. In one example, the position of the camera 155 relative to the vehicle may be fixed. In another example, the position of the camera 155 relative to the vehicle may be changed, for example, the camera 155 may perform rotation shooting.
可选地,在本文所述的各种实施例中,车载设备101可位于远离自动驾驶车辆的地方,并且可与自动驾驶车辆无线通信。在其它方面,本文所述的一些过程在设置在自动驾驶车辆内的处理器上执行,其它由远程处理器执行,包括采取执行单个操纵所需的动作。Optionally, in various embodiments described herein, the in-vehicle device 101 may be located far away from the autonomous driving vehicle, and may wirelessly communicate with the autonomous driving vehicle. In other respects, some of the processes described herein are executed on a processor provided in an autonomous vehicle, and others are executed by a remote processor, including taking actions required to perform a single manipulation.
车载设备101可以通过网络接口129和软件部署服务器(deploying server)149通信。网络接口129是硬件网络接口,比如,网卡。网络127可以是外部网络,比如因特网,也可以是内部网络,比如以太网或者虚拟私人网络(virtual private network,VPN)。可选地,网络127还可以是无线网络,比如WiFi网络,蜂窝网络等。The in-vehicle device 101 may communicate with a software deployment server (deploying server) 149 through a network interface 129. The network interface 129 is a hardware network interface, such as a network card. The network 127 may be an external network, such as the Internet, or an internal network, such as an Ethernet or a virtual private network (virtual private network, VPN). Optionally, the network 127 may also be a wireless network, such as a WiFi network, a cellular network, and so on.
硬盘驱动接口和系统总线105耦合。硬件驱动接口和硬盘驱动器相连接。系统内 存135和系统总线105耦合。运行在系统内存135的数据可以包括车载设备101的操作系统137和应用程序143。The hard disk drive interface is coupled to the system bus 105. The hardware drive interface is connected with the hard drive. The system memory 135 and the system bus 105 are coupled. The data running in the system memory 135 may include the operating system 137 and application programs 143 of the in-vehicle device 101.
操作系统包括shell 139和内核(kernel)141。shell 139是介于使用者和操作系统之内核间的一个接口。shell是操作系统最外面的一层。shell管理使用者与操作系统之间的交互,等待使用者的输入,向操作系统解释使用者的输入,并且处理各种各样的操作系统的输出结果。The operating system includes a shell 139 and a kernel (kernel) 141. Shell 139 is an interface between the user and the kernel of the operating system. The shell is the outermost layer of the operating system. The shell manages the interaction between the user and the operating system, waits for the user's input, interprets the user's input to the operating system, and processes the output of various operating systems.
内核141由操作系统中用于管理存储器、文件、外设和系统资源的那些部分组成。直接与硬件交互,操作系统内核通常运行进程,并提供进程间的通信,提供CPU时间片管理、中断、内存管理、IO管理等等。The kernel 141 is composed of those parts of the operating system for managing memory, files, peripherals, and system resources. Directly interact with hardware, the operating system kernel usually runs processes and provides inter-process communication, providing CPU time slice management, interrupts, memory management, IO management, and so on.
应用程序143包括控制车辆自动驾驶相关的程序。例如,对车载设备所获取到包含车辆道路上的图像信息的图像进行处理程序,如用于实现本申请实施例所提供的图像处理方法的程序。又如,管理自动驾驶车辆和路上障碍物交互的程序,控制自动驾驶车辆路线或者速度的程序,控制自动驾驶车辆和路上其他自动驾驶车辆交互的程等。应用程序143也存在于软件部署服务器149的系统上。在一个实施例中,在需要执行应用程序143时,车载设备101可以从软件部署服务器149下载应用程序143。The application program 143 includes programs related to controlling the automatic driving of the vehicle. For example, a program for processing an image containing image information on a vehicle road acquired by an on-vehicle device, such as a program for implementing the image processing method provided by the embodiment of the present application. For another example, the program that manages the interaction between autonomous vehicles and road obstacles, the program that controls the route or speed of autonomous vehicles, and the process of interaction between autonomous vehicles and other autonomous vehicles on the road. The application program 143 also exists on the system of the software deployment server 149. In one embodiment, when the application program 143 needs to be executed, the in-vehicle device 101 may download the application program 143 from the software deployment server 149.
传感器153和车载设备101关联。传感器153用于探测车载设备101周围的环境。举例来说,传感器153可以探测动物、车辆、障碍物和人行横道等,进一步传感器还可以探测上述动物、车辆、障碍物和人行横道等物体周围的环境,比如:动物周围的环境,例如,动物周围出现的其他动物,天气条件,周围环境的光亮度等。可选地,如果车载设备101位于自动驾驶的车辆上,传感器可以是摄像头,红外线感应器,化学检测器,麦克风等。可选地,传感器153可以包括速度传感器,用于测量本车辆(即图1所示的计算机系统所在的车辆)的速度信息(如速度、加速度等);角度传感器,用于测量车辆的方向信息,以及车辆与车辆周边的物体/对象之间的相对角度等。The sensor 153 is associated with the in-vehicle device 101. The sensor 153 is used to detect the environment around the in-vehicle device 101. For example, the sensor 153 can detect animals, vehicles, obstacles, and crosswalks. Further, the sensor can also detect the environment around objects such as animals, vehicles, obstacles, and crosswalks, such as: the environment around the animals, for example, when the animals appear around them. Other animals, weather conditions, the brightness of the surrounding environment, etc. Optionally, if the in-vehicle device 101 is located on an autonomous vehicle, the sensor may be a camera, an infrared sensor, a chemical detector, a microphone, etc. Optionally, the sensor 153 may include a speed sensor, used to measure the speed information (such as speed, acceleration, etc.) of the own vehicle (that is, the vehicle in which the computer system shown in FIG. 1 is located); an angle sensor, used to measure the direction information of the vehicle , And the relative angle between the vehicle and the objects/objects around the vehicle.
需要说明的是,图1所示的计算机系统仅为示例,其不对本申请实施例可适用的计算机系统构成限定。例如,图1中所示意的与车载设备所连接的一个或多个器件可以与车载设备集成在一起,例如摄像头与车载设备集成在一起等。It should be noted that the computer system shown in FIG. 1 is only an example, which does not constitute a limitation on the computer system applicable to the embodiments of the present application. For example, one or more devices connected to the vehicle-mounted equipment shown in FIG. 1 may be integrated with the vehicle-mounted equipment, for example, a camera is integrated with the vehicle-mounted equipment.
以下对本申请实施例中所涉及的部分术语或技术进行解释说明:The following explains some terms or technologies involved in the embodiments of this application:
1)、多帧超分、参考图像、非参考图像1), multi-frame super-division, reference image, non-reference image
多帧超分是利用多帧图像中的非参考图像的图像信息,对该多帧图像中的参考图像的图像信息进行处理,如对边缘特征、锐化特征等进行补偿,得到一帧图像,该图像的分辨率高于参考图像的分辨率。其中,参考图像可以是该多帧图像中的任意一帧图像。非参考图像是该多帧图像中的除参考图像之外的所有图像。Multi-frame super-division uses the image information of the non-reference image in the multi-frame image to process the image information of the reference image in the multi-frame image, such as compensating for edge features, sharpening features, etc., to obtain a frame of image. The resolution of this image is higher than the resolution of the reference image. Wherein, the reference image may be any one of the multiple frames of images. Non-reference images are all images in the multi-frame images except for reference images.
2)、对象、目标对象2), object, target object
对象,也可以称作道路对象或者障碍物或者道路障碍物等。在本申请实施例中,对象可以是车辆周边道路上的人、车辆、交通灯、交通指示牌(如限速指示牌等)、电线杆、垃圾桶、异物等。其中,异物是指本不应该出现在道路上的物体,如遗落在道路上的箱子、轮胎等。The object may also be called a road object or obstacle or road obstacle. In the embodiments of the present application, the objects may be people, vehicles, traffic lights, traffic signs (such as speed limit signs, etc.), telephone poles, trash cans, foreign objects, etc. on the roads surrounding the vehicle. Among them, foreign objects refer to objects that should not have appeared on the road, such as boxes and tires left on the road.
目标对象,是车载设备需要识别的对象。目标对象可以是预定义的,或者可以是用户指示的,本申请实施例对此不进行限定。示例的,在自动驾驶场景中,目标对象 可以包括:人、车、交通灯等。The target object is the object that the vehicle-mounted equipment needs to recognize. The target object may be predefined or indicated by the user, which is not limited in the embodiment of the present application. For example, in an autonomous driving scene, the target object may include: people, cars, traffic lights, and so on.
3)、目标检测、图像分割3), target detection, image segmentation
目标检测和图像分割均是图像处理技术。Both target detection and image segmentation are image processing techniques.
目标检测的任务是找出图像中所有感兴趣的目标的图像信息所在的区域,并确定该区域的大小和该区域在该图像中的位置。感兴趣的目标可以是预定义的,也可以是用户确定的。在本申请实施例中,在本申请实施例中,感兴趣的目标可以是指目标对象。目标检测所得到的不同区域(即不同目标的图像信息所在的区域)之间可能有交叠,也可能没有交叠。如图2所示,为一种目标检测的结果示意图。图2中是以感兴趣的目标是车辆为例进行说明的。图2中的每个矩形框所限定的区域为一个感兴趣的目标的图像信息所在的区域。The task of target detection is to find out the area where the image information of all the targets of interest in the image are located, and to determine the size of the area and the location of the area in the image. The target of interest can be pre-defined or user-defined. In the embodiments of the present application, in the embodiments of the present application, the target of interest may refer to the target object. Different regions obtained by target detection (that is, regions where the image information of different targets are located) may or may not overlap. As shown in Figure 2, it is a schematic diagram of a target detection result. In Fig. 2, the target of interest is a vehicle as an example for illustration. The area defined by each rectangular box in FIG. 2 is the area where the image information of an interested target is located.
图像分割,是根据图像内容对图像中的指定区域进行标记的计算机视觉任务。简言之,就是确定一帧图像中有哪些物体的图像信息,以及该物体的图像信息在该图像中的位置等。具体而言,图像分割的目的是确定图像中的每一个像素所表示的是哪个物体的像素。图像分割可以包括:语义分割、实例分割等。图像分割所得到的图像区域之间通常没有交叠。如图3所示,为一种图像分割的示意图。图3中的每个连通区域表示图像分割所得到的一个区域。Image segmentation is a computer vision task that marks designated areas in an image according to the content of the image. In short, it is to determine the image information of which objects in a frame of image, and the position of the image information of the object in the image. Specifically, the purpose of image segmentation is to determine which object pixel each pixel in the image represents. Image segmentation can include semantic segmentation, instance segmentation, and so on. There is usually no overlap between image regions obtained by image segmentation. As shown in Figure 3, it is a schematic diagram of image segmentation. Each connected region in Figure 3 represents a region obtained by image segmentation.
4)、图像区域、空间区域4), image area, space area
为了清楚地区分实际场景中的区域(即客观存在的区域)与实际场景中的区域的图像(即图像或图片中的区域),在本申请中,将实际场景中的区域称为“空间区域”,并将空间区域的图像称为“图像区域”。In order to clearly distinguish the area in the actual scene (that is, the area that exists objectively) and the image of the area in the actual scene (that is, the area in the image or picture), in this application, the area in the actual scene is called "spatial area" ", and call the image of the spatial area "image area".
5)、待检测图像区域、第一图像区域、第二图像区域5), the image area to be detected, the first image area, the second image area
待检测图像区域,是指图像中包含目标对象的图像信息的概率较高(如该概率大于预设阈值)的区域。此为本申请实施例给出的待检测图像区域的定义,但是,在实际实现时,确定图像中的待检测图像区域时,通常不需要直接确定一个图像区域中包含目标对象的图像信息的概率,以及不需要在车载设备中设置该预设阈值,而是通过其他方法间接确定一个区域中包含目标对象的图像信息的概率是否高于预设阈值。例如,当车载设备所获取的图像包含具有如下任一种特征的图像区域时,确定该图像区域是待检测图像区域:The image area to be detected refers to an area in the image that contains the image information of the target object with a high probability (for example, the probability is greater than a preset threshold). This is the definition of the image area to be detected in the embodiments of this application. However, in actual implementation, when determining the image area to be detected in the image, it is usually not necessary to directly determine the probability that an image area contains the image information of the target object , And there is no need to set the preset threshold in the vehicle-mounted device, but indirectly determine whether the probability of containing the image information of the target object in an area is higher than the preset threshold through other methods. For example, when the image acquired by the vehicle-mounted device contains an image area with any of the following characteristics, it is determined that the image area is the image area to be detected:
特征1:置信度低于或等于第一阈值。Feature 1: The confidence level is lower than or equal to the first threshold.
特征2:对应于车辆可行驶区域中与该车辆之间的距离大于或等于第二阈值的空间区域。Feature 2: Corresponding to a space area where the distance between the vehicle and the vehicle in the travelable area is greater than or equal to the second threshold.
特征3:在其所属图像中的位置是预设位置。Feature 3: The position in the image to which it belongs is a preset position.
也就是说,待检测图像区域是置信度低于或等于第一阈值的区域,或者,是对应于该车辆的可行驶区域中与该车辆之间的距离大于或等于第二阈值的空间区域的图像区域,或者,是图像中预设位置的区域。当然,在不冲突的情况下,上述特征1~3中的任意多个可以组合,作为待检测图像区域的特征。关于这3种特征的相关说明,以及具体示例可以参考下文。That is to say, the image area to be detected is an area with a confidence level lower than or equal to the first threshold, or it corresponds to a space area in the drivable area of the vehicle whose distance from the vehicle is greater than or equal to the second threshold. The image area, or, is the area at the preset position in the image. Of course, if there is no conflict, any of the above features 1 to 3 can be combined as the feature of the image area to be detected. For related descriptions of these three characteristics, and specific examples, please refer to the following.
待检测图像区域可以包含一个或多个对象的图像信息。待检测图像区域可能包含一个或多个目标对象的图像信息,也可能不包含目标对象的图像信息。待检测图像区 域可以是一帧图像中的部分或全部区域。The image area to be detected may contain image information of one or more objects. The image area to be detected may contain the image information of one or more target objects, or may not contain the image information of the target object. The image area to be detected can be part or all of the area in a frame of image.
一帧图像可以包含一个或多个待检测图像区域。本申请实施例中所描述的第一图像区域和第二图像区域均为待检测图像区域。A frame of image can contain one or more image regions to be detected. The first image area and the second image area described in the embodiments of the present application are both image areas to be detected.
6)、其他术语6), other terms
在本申请实施例中,“示例性的”或者“例如”等词用于表示作例子、例证或说明。本申请实施例中被描述为“示例性的”或者“例如”的任何实施例或设计方案不应被解释为比其它实施例或设计方案更优选或更具优势。确切而言,使用“示例性的”或者“例如”等词旨在以具体方式呈现相关概念。In the embodiments of the present application, words such as "exemplary" or "for example" are used as examples, illustrations, or illustrations. Any embodiment or design solution described as "exemplary" or "for example" in the embodiments of the present application should not be construed as being more preferable or advantageous than other embodiments or design solutions. To be precise, words such as "exemplary" or "for example" are used to present related concepts in a specific manner.
在本申请实施例中,“至少一个”是指一个或多个。“多个”是指两个或两个以上。In the embodiments of the present application, "at least one" refers to one or more. "Multiple" means two or more.
在本申请实施例中,“和/或”,仅仅是一种描述关联对象的关联关系,表示可以存在三种关系,例如,A和/或B,可以表示:单独存在A,同时存在A和B,单独存在B这三种情况。另外,本文中字符“/”,一般表示前后关联对象是一种“或”的关系。In the embodiments of the present application, "and/or" is merely an association relationship describing associated objects, indicating that there can be three types of relationships, for example, A and/or B, which can indicate that A exists alone, and both A and A B, there are three cases of B alone. In addition, the character "/" in this text generally indicates that the associated objects before and after are in an "or" relationship.
以下,结合附图对本申请实施例提供的图像处理方法进行说明。该方法可以应用于上文中所描述的车载设备101中。Hereinafter, the image processing method provided by the embodiments of the present application will be described with reference to the accompanying drawings. This method can be applied to the vehicle-mounted device 101 described above.
如图4所示,为本申请实施例提供的一种图像处理方法的流程示意图。该方法可以包括:As shown in FIG. 4, it is a schematic flowchart of an image processing method provided by an embodiment of this application. The method can include:
S101:车载设备获取N帧图像,该N帧图像包括该车载设备所在车辆的周边道路的图像信息。N是大于等于2的整数。S101: The vehicle-mounted device acquires N frames of images, where the N-frame images include image information of the surrounding roads of the vehicle where the vehicle-mounted device is located. N is an integer greater than or equal to 2.
该N帧图像可以是该车辆中安装的摄像头(如上述摄像头155)所拍摄的图像。车辆的周边道路可以包括:车辆的前方道路、后方道路和侧方道路等中的一种或多种。The N frames of images may be images taken by a camera (such as the aforementioned camera 155) installed in the vehicle. The surrounding roads of the vehicle may include one or more of the front road, the rear road, and the side road of the vehicle.
该N帧图像中的任意一帧图像可以是该车辆处于静止状态时所拍摄的图像,或者可以是该车辆处于运动状态(如直行、变道或转弯等)时所拍摄的图像。Any one of the N frames of images may be an image taken when the vehicle is in a stationary state, or may be an image taken when the vehicle is in a moving state (such as going straight, changing lanes, or turning).
可选地,该N帧图像是时序上连续的N帧图像。换句话说,该N帧图像是该车辆中安装的摄像头连续拍摄的N帧图像。Optionally, the N frames of images are consecutive N frames of images in time series. In other words, the N frames of images are N frames of images continuously captured by the camera installed in the vehicle.
基于该可选地实现方式,在多次执行图4所示的方法时,车载设备可以基于滑动窗口N,在摄像头所拍摄的图像中确定S101中的“N帧图像”。本申请实施例对N的取值不进行限定。在任意两次执行图4所示的方法时,N的取值可以相同,也可以不同。相邻两次执行图4所示的方法时,S101中的N帧图像之间可以有重叠,也可以没有重叠。例如,假设按照拍摄时间的先后顺序对摄像头所拍摄的图像进行排序后得到如下序列:图像1、图像2、图像3……图像n。n是整数,n的取值随着摄像头的拍摄次数的增加而增加;并且,N=3,那么,第一次执行图4所示的方法时,S101中的N帧图像可以是图像1~3,第二次执行图4所示的方法时,S101中的N帧图像可以是图像2~4,第三次执行图4所示的方法时,S101中的N帧图像可以是图像3~6,依此类推,从而执行本申请实施例提供的图像处理方法。Based on this optional implementation manner, when the method shown in FIG. 4 is executed multiple times, the on-vehicle device may determine the "N frame images" in S101 in the images captured by the camera based on the sliding window N. The embodiment of the present application does not limit the value of N. When the method shown in FIG. 4 is executed twice, the value of N may be the same or different. When the method shown in FIG. 4 is executed twice adjacently, the N frames of images in S101 may or may not overlap. For example, suppose that the images taken by the camera are sorted according to the order of shooting time to obtain the following sequence: image 1, image 2, image 3...image n. n is an integer, and the value of n increases as the number of shots of the camera increases; and, N=3, then, when the method shown in FIG. 4 is executed for the first time, the N frames of images in S101 can be images 1~ 3. When the method shown in FIG. 4 is executed for the second time, the N frames of images in S101 may be images 2 to 4, and when the method shown in FIG. 4 is executed for the third time, the N frames of images in S101 may be images 3 to 4 6. By analogy, the image processing method provided in the embodiment of the present application is executed.
S102:车载设备获取该N帧图像中N1帧图像的每帧图像中的第一图像区域。该N1帧图像的N1个第一图像区域对应于第一场景。其中,2≤N1≤N,N1是整数。不同第一图像区域是不同图像中的图像区域。S102: The vehicle-mounted device acquires the first image area in each frame of the N1 frame of images in the N frame of images. The N1 first image regions of the N1 frame image correspond to the first scene. Among them, 2≤N1≤N, N1 is an integer. Different first image areas are image areas in different images.
第一图像区域,是一个待检测图像区域。关于待检测图像区域的相关解释可以参考上文。The first image area is an image area to be detected. For the relevant explanation of the image area to be detected, please refer to the above.
第一场景,为该车辆周边的部分路况或者部分行驶视野。第一场景可以理解为该车辆周边的路况或者行驶视野中的一个或多个对象所在的空间区域。一个对象所在的空间区域,是指包含该对象的区域。例如,假设该N1帧图像中均包含同一个交通灯的图像信息,则第一场景可以为该交通灯所在的空间区域,第一图像区域可以为该N1帧图像的每帧图像中该交通灯的图像所在的图像区域。又如,假设该N1帧图像中均包含同一车辆的图像信息,则第一场景可以为该车辆所在的空间区域,第一图像区域可以为该N1帧图像的每帧图像中该车辆的图像所在的图像区域。The first scenario is part of the road conditions or part of the driving field of view around the vehicle. The first scene can be understood as the road conditions around the vehicle or the spatial area where one or more objects in the driving field of view are located. The spatial area where an object is located refers to the area containing the object. For example, assuming that the N1 frames of images all contain the image information of the same traffic light, the first scene may be the space area where the traffic light is located, and the first image area may be the traffic light in each frame of the N1 frame image. The image area where the image is located. For another example, assuming that the N1 frames of images all contain the image information of the same vehicle, the first scene may be the space area where the vehicle is located, and the first image area may be the image of the vehicle in each frame of the N1 frame image. Image area.
可以理解的是,随着车辆的移动,或者随着车辆周边道路上可移动的对象(如人或车辆等)的移动,会导致该车辆上安装的摄像头所拍摄的不同图像所包含的信息不同,这就可能导致,在S101中所获取的N帧图像中,并非每帧图像中均包含同一场景所对应的图像区域,当然也可能存在每帧图像中均包含同一场景所对应的图像区域,因此,在S102中,将车载设备所获取到的是第一图像区域的个数标记为N1,而非N。It is understandable that as the vehicle moves, or as the movable objects (such as people or vehicles, etc.) on the road around the vehicle move, different images captured by the camera installed on the vehicle will contain different information. , This may cause that in the N frames of images obtained in S101, not every frame of image contains the image area corresponding to the same scene, of course, there may also be each frame of image containing the image area corresponding to the same scene. Therefore, in S102, the number of first image regions acquired by the vehicle-mounted device is marked as N1 instead of N.
在一种实现方式中,对于该N1帧图像中的每帧图像来说,车载设备可以独立确定该图像中的第一图像区域。例如,根据上述特征1~3中的至少一种确定该图像中的第一图像区域。In an implementation manner, for each frame of the N1 frame of image, the vehicle-mounted device may independently determine the first image area in the image. For example, the first image area in the image is determined according to at least one of the aforementioned features 1 to 3.
在另一种实现方式中,对于该N帧图像中的部分图像来说,车载设备可以先确定该图像中的第一图像区域,例如,根据上述特征1~3中的至少一种确定该部分图像中的第一图像区域;再基于此推理得到该N帧图像中的其他图像中的第一图像区域。具体示例可以参考下文。In another implementation manner, for a part of the image in the N frames, the vehicle-mounted device may first determine the first image area in the image, for example, determine the part according to at least one of the above-mentioned features 1 to 3 The first image area in the image; based on this inference, the first image area in the other images in the N frames of images is obtained. For specific examples, please refer to the following.
可选地,该N帧图像中相邻两帧图像的拍摄时间间隔小于或等于第三阈值。其中,两帧图像的拍摄时间间隔,是指拍摄这两帧的时刻之间的时间段。本申请实施例对第三阈值的具体取值及取值方式不进行限定。这样,当车速较快时,有助于提高多帧超分的图像补偿效果。Optionally, the shooting time interval of two adjacent frames of images in the N frames of images is less than or equal to a third threshold. Among them, the shooting time interval of two frames of images refers to the time period between the moments when the two frames are shot. The embodiment of the present application does not limit the specific value and value method of the third threshold. In this way, when the vehicle speed is faster, it helps to improve the image compensation effect of multi-frame super-division.
可以理解的是,当车速较快时,如果相邻两帧图像的拍摄时间间隔较大,则该两帧图像中可能不存在对应于同一场景的图像区域,例如,如果车速较快,则可能出现一帧图像中包括:交通灯1的图像信息和车辆1~3的图像信息,且不包括其他对象的图像信息;而该图像的下一帧图像中包括:交通灯2的图像信息和车辆4~5的图像信息,且不包括其他对象的图像信息;那么,这两帧图像中没有对应同一场景的图像区域。这可能导致不能利用其它帧图像中的图像信息,对该场景所对应的图像区域中的图像信息进行补偿,从而导致多帧超分的图像补偿效果较差。因此,该可选地实现方式,有助于提高多帧超分的图像补偿效果。It can be understood that when the vehicle speed is fast, if the time interval between two adjacent frames of images is relatively large, there may not be an image area corresponding to the same scene in the two frames of images. For example, if the vehicle speed is relatively fast, it may A frame of image that appears includes: image information of traffic light 1 and image information of vehicles 1 to 3, and does not include image information of other objects; and the next frame of the image includes: image information of traffic light 2 and vehicle The image information of 4 to 5 does not include the image information of other objects; then, there is no image area corresponding to the same scene in the two frames. This may lead to the inability to use the image information in other frame images to compensate the image information in the image area corresponding to the scene, resulting in poor image compensation effect for multi-frame superdivision. Therefore, this optional implementation manner helps to improve the image compensation effect of multi-frame superdivision.
可选地,该N帧图像中的首帧图像与末帧图像的拍摄时间间隔小于或等于一阈值。这样,当车速较快时,有助于提高多帧超分的图像补偿效果。其具体分析过程可以参考上文。Optionally, the shooting time interval between the first frame image and the last frame image in the N frames of images is less than or equal to a threshold. In this way, when the vehicle speed is faster, it helps to improve the image compensation effect of multi-frame super-division. The specific analysis process can refer to the above.
S103:车载设备执行该N1个第一图像区域的景物对齐。S103: The vehicle-mounted device executes scene alignment of the N1 first image regions.
当车辆处于静止状态时,该车辆上安装的摄像头所拍摄的不同图像的背景不变,前景可能改变。因此,当车辆处于静止状态时,对N1帧图像中的第一图像区域进行对齐的步骤(即S103)可以是可选地步骤。当车辆处于运动状态时,该车辆上安装的摄像头所拍摄的不同图像的背景和前景均可能改变。该情况下,可以在执行S104的超 分运算之前,进行景物对齐。When the vehicle is at a standstill, the background of different images taken by the camera installed on the vehicle remains unchanged, and the foreground may change. Therefore, when the vehicle is in a stationary state, the step of aligning the first image area in the N1 frame of image (ie S103) may be an optional step. When the vehicle is in motion, the background and foreground of different images taken by the camera installed on the vehicle may change. In this case, the scene alignment can be performed before the superdivision calculation in S104 is executed.
景物对齐,是保证多帧图像中对应于同一场景的图像区域内包含统一或近似的前景和背景。例如,可以将该多帧图像中的对应于同一场景的图像区域缩放到统一尺寸,并通过角度等参数将该多帧图像中的部分或全部景象进行旋转,从而保证该多帧中对应于该场景的图像区域内包含统一或近似的前景和背景。Scene alignment is to ensure that the same or similar foreground and background are included in the image area corresponding to the same scene in multiple frames of images. For example, the image area corresponding to the same scene in the multi-frame image can be scaled to a uniform size, and some or all of the scene in the multi-frame image can be rotated by parameters such as angle, so as to ensure that the multi-frame corresponds to the The image area of the scene contains uniform or similar foreground and background.
S104:车载设备对景物对齐后的该N1个第一图像区域进行超分运算,得到第一目标图像区域。该超分运算也可称作多帧超分运算。S104: The in-vehicle device performs a hyperdivision operation on the N1 first image regions after scene alignment to obtain the first target image region. This super-division operation can also be called a multi-frame super-division operation.
具体的,车载设备根据景物对齐后的非参考图像中的第一图像区域,对景物对齐后的参考图像中的第一图像区域进行处理(如补偿边缘特征、锐化特征等),得到第一目标图像区域。其中,第一目标图像区域的分辨率高于参考图像中的第一图像区域的分辨率。具体的处理过程可以参考现有技术。Specifically, the in-vehicle device processes the first image area in the reference image after the scene is aligned (such as compensating for edge features, sharpening features, etc.) according to the first image area in the non-reference image after the scene is aligned to obtain the first image area. The target image area. Wherein, the resolution of the first target image area is higher than the resolution of the first image area in the reference image. The specific process can refer to the prior art.
本申请实施例对第一目标图像区域的应用场景不进行限定,例如可以应用于目标对象检测场景中。当应用于目标对象检测场景中时,上述方法还可以包括以下步骤S105:The embodiment of the present application does not limit the application scenario of the first target image area, for example, it may be applied to a target object detection scenario. When applied to a target object detection scene, the above method may further include the following step S105:
S105:车载设备确定第一目标图像区域中是否存在目标对象。若存在,则确定目标对象与该车辆的相对位置。若不存在,则结束。S105: The vehicle-mounted device determines whether there is a target object in the first target image area. If it exists, the relative position of the target object and the vehicle is determined. If it does not exist, it ends.
步骤S105的具体实现方式可以参考现有技术。例如,车载设备对第一目标图像区域进行目标检测或图像分割等,以确定第一目标图像区域中是否包含目标对象,并在存在的情况下,确定目标对象与该车辆的相对位置。The specific implementation of step S105 can refer to the prior art. For example, the in-vehicle device performs target detection or image segmentation on the first target image area to determine whether the target object is contained in the first target image area, and if it exists, determines the relative position of the target object and the vehicle.
本申请实施例提供的图像处理方法中,超分运算具体是多帧超分,这样可以结合多帧图像中的图像信息的特征信息进行图像补偿,与现有技术中使用单帧超分进行图像补偿的技术方案相比,有助于提高图像补偿效果。另外,由于本技术方案中,对该多帧图像中对应于同一场景的多个图像区域进行多帧超分,而非对该多帧图像本身进行超分运算,因此,有助于降低超分运算的复杂度,从而加快超分处理速率。进一步地,当本申请实施例提供的图像处理方法的处理结果应用于辅助自动驾驶路径规划时,有助于提高自动驾驶路径规划的精确度。In the image processing method provided by the embodiments of the present application, the super-division operation is specifically multi-frame super-division, which can combine the characteristic information of the image information in the multi-frame image to perform image compensation, which is similar to the prior art using single-frame super-division to perform image compensation. Compared with the compensation technical solution, it helps to improve the image compensation effect. In addition, since in the present technical solution, multiple image areas corresponding to the same scene in the multi-frame image are subjected to multi-frame super-division, instead of performing super-division operations on the multi-frame image itself, it is helpful to reduce the super-division. The complexity of the operation, thereby speeding up the processing rate of super-division. Further, when the processing result of the image processing method provided by the embodiment of the present application is applied to assist automatic driving path planning, it helps to improve the accuracy of automatic driving path planning.
可选地,在执行S101之后,该方法还可以包括:车载设备获取该N帧图像中N2帧图像的每帧图像中的第二图像区域。该N2帧图像的N2个第二图像区域对应于第二场景。2≤N2≤N,N2是整数。不同第二图像区域是不同图像中的图像区域。基于此,该方法还包括:步骤S102'~S105'。S102'~S105'是将S102~S105中的“第一图像区域”替换为“第二图像区域”,“N1”替换为“N2”,以及“第一目标图像区域”替换为“第二目标图像区域”后得到的。Optionally, after performing S101, the method may further include: the vehicle-mounted device acquires the second image area in each frame of the N2 frames of the N frame of images. The N2 second image regions of the N2 frame image correspond to the second scene. 2≤N2≤N, N2 is an integer. Different second image areas are image areas in different images. Based on this, the method further includes: steps S102' to S105'. S102'~S105' replace "first image area" in S102~S105 with "second image area", "N1" with "N2", and "first target image area" with "second target" Image area".
其中,第二图像区域是区别于第一图像区域的待检测图像区域。第一图像区域与第二图像区域之间可能部分重叠,也可能不重叠。Wherein, the second image area is an image area to be detected that is different from the first image area. The first image area and the second image area may or may not overlap partially.
其中,本申请实施例对N1与N2之间的大小关系不进行限定。并且,该N1帧图像和该N2帧图像可以包含相同的图像,也可以不包含相同的图像。Among them, the embodiment of the present application does not limit the magnitude relationship between N1 and N2. In addition, the N1 frame image and the N2 frame image may or may not include the same image.
可选地,该N1帧图像和该N2帧图像中均包含参考图像(即超分运算时所采用的参考图像)。例如,假设该N帧图像为第1~10帧图像,且第1~5帧图像中均包含交通灯1的图像信息,第1~7帧图像中均包含车辆1的图像信息,第1帧图像是参考图 像;那么,该N1帧图像可以是第1~5帧图像,第一图像区域可以是交通灯1的图像信息所在的区域,该N2帧图像可以是第1~7帧图像,第二图像区域可以是车辆1的图像信息所在的区域。Optionally, both the N1 frame image and the N2 frame image include a reference image (that is, a reference image used in the super-division operation). For example, suppose that the N frames of images are the first to tenth frames, and the first to fifth frames all contain the image information of traffic light 1, and the first to seventh frames all contain the image information of vehicle 1, and the first frame The image is a reference image; then, the N1 frame image can be the first to fifth frame images, the first image area can be the area where the image information of the traffic light 1 is located, and the N2 frame image can be the first to seventh frame images, The second image area may be the area where the image information of the vehicle 1 is located.
以下,通过具体示例对上文中提供的图像处理方法进行说明:Hereinafter, the image processing method provided above will be explained through specific examples:
实施例1Example 1
如图5所示,为本申请实施例提供的一种图像处理方法的流程示意图。该方法可以包括:As shown in FIG. 5, it is a schematic flowchart of an image processing method provided by an embodiment of this application. The method can include:
S201:可以参考上述S101,当然本申请实施例不限于此。S201: Refer to the above S101, of course, the embodiment of the present application is not limited to this.
S202:车载设备根据候选类型集合,对该N帧图像中的第一图像进行预处理(如图像分割或目标检测等),得到第一图像包括的至少一个候选图像区域、每个候选图像区域的识别结果以及每个识别结果的置信度。第一图像可以是该N帧图像中的任意一帧图像。可选的,第一图像是参考图像,该参考图像是指多帧超分中的参考图像。S202: The vehicle-mounted device performs preprocessing (such as image segmentation or target detection, etc.) on the first image in the N frames of images according to the candidate type set to obtain at least one candidate image area included in the first image, and each candidate image area Recognition results and the confidence level of each recognition result. The first image may be any one of the N frames of images. Optionally, the first image is a reference image, and the reference image refers to a reference image in a multi-frame super division.
候选类型集合,是至少一个候选类型构成的集合。候选类型是车载设备需要识别的目标对象的类型(即车载设备感兴趣的对象的类型)。对象的类型可以理解为该对象是什么。例如,如果一个对象是人,则该对象的类型是人;如果一个对象是交通灯,则该对象的类型是交通灯。示例的,在自动驾驶场景中,该候选类型集合可以包括:人、车、交通灯等。The candidate type set is a set composed of at least one candidate type. The candidate type is the type of target object that the vehicle-mounted device needs to recognize (that is, the type of object that the vehicle-mounted device is interested in). The type of object can be understood as what the object is. For example, if an object is a person, the type of the object is a person; if an object is a traffic light, the type of the object is a traffic light. For example, in an autonomous driving scenario, the candidate type set may include: people, cars, traffic lights, and so on.
候选图像区域,是包含具有候选类型的对象的图像信息的区域。示例的,如果候选类型集合是人、车和交通灯构成的集合,则候选图像区域包括:人的图像信息所在的图像区域、车的图像信息所在的图像区域和交通灯的图像信息所在的图像区域。例如,当“预处理”是目标检测时,以第一图像是图2所示的图像为例,图2中的矩形框所限定的区域可以作为一个候选图像区域。又如,当“预处理”是图像分割时,以第一图像是图3所示的图像为例,图3中所示的每个连通区域可以作为一个候选图像区域。S202中所描述的至少一个候选图像区域所占的区域可以是第一图像中的部分或全部区域。The candidate image area is an area that contains image information of objects of candidate types. For example, if the candidate type set is a set composed of people, cars, and traffic lights, the candidate image area includes: the image area where the image information of the person is located, the image area where the image information of the car is located, and the image where the image information of the traffic light is located. area. For example, when "preprocessing" is target detection, taking the first image as the image shown in FIG. 2 as an example, the area defined by the rectangular frame in FIG. 2 can be used as a candidate image area. For another example, when the "preprocessing" is image segmentation, taking the first image as the image shown in FIG. 3 as an example, each connected region shown in FIG. 3 can be used as a candidate image region. The area occupied by the at least one candidate image area described in S202 may be part or all of the area in the first image.
候选图像区域的识别结果,可以包括:该候选图像区域中的图像信息是哪一种目标对象的图像信息。可选地,还可以包括该目标对象与该车辆之间的相对位置。The recognition result of the candidate image area may include: which kind of target object the image information in the candidate image area is. Optionally, it may also include the relative position between the target object and the vehicle.
一个候选图像区域的识别结果的置信度,可以被称作该候选图像区域的置信度,可以理解为该候选图像区域的识别结果的准确度。The confidence of the recognition result of a candidate image area can be referred to as the confidence of the candidate image area, which can be understood as the accuracy of the recognition result of the candidate image area.
可以理解的是,对于同一对象来说,如果该对象距离车辆较近,则该车辆中的摄像头所拍摄的图像中该对象所占的图像区域较大,且较清晰;如果该对象距离该车辆较远,则该摄像头所拍摄的图像中该对象所占的图像区域较小,且较模糊。因此,在同一图像中,第一候选图像区域(即包含与车辆之间的距离较近的对象的图像信息的候选图像区域)的识别结果的置信度,一般高于,第二候选图像区域(即包含与车辆之间的距离较远的对象的图像信息的候选图像区域)的识别结果的置信度。例如,一个候选图像区域的识别结果为:该候选图像区域中包括的对象的图像信息是人的图像信息时,如果该对象距离车辆较近,则该对象实际是“人”的概率会较高,即识别结果的置信度较高;如果该对象距离车辆较远,则该对象实际是“人”的概率会较低,即识别结果的置信度较低,例如该对象实际可能是一个电线杆等。It is understandable that for the same object, if the object is closer to the vehicle, the image area taken by the object in the image taken by the camera in the vehicle is larger and clearer; if the object is closer to the vehicle Farther, the image area occupied by the object in the image taken by the camera is smaller and blurry. Therefore, in the same image, the confidence of the recognition result of the first candidate image area (that is, the candidate image area containing the image information of the object that is close to the vehicle) is generally higher than that of the second candidate image area ( That is, the confidence level of the recognition result of the candidate image region containing the image information of the object that is far away from the vehicle. For example, the recognition result of a candidate image area is: when the image information of the object included in the candidate image area is image information of a person, if the object is closer to the vehicle, the probability that the object is actually a "person" will be higher , That is, the confidence of the recognition result is high; if the object is far from the vehicle, the probability that the object is actually a "person" will be low, that is, the confidence of the recognition result is low, for example, the object may actually be a telephone pole Wait.
S203:车载设备获取第一图像中的至少一个待检测区域。其中,待检测区域是置信度小于或等于第一阈值的候选图像区域。该至少一个待检测区域包括第一图像区域。S203: The vehicle-mounted device acquires at least one area to be detected in the first image. Wherein, the area to be detected is a candidate image area whose confidence is less than or equal to the first threshold. The at least one area to be detected includes a first image area.
其中,获取第一图像中的第一图像区域,可以包括:获取第一图像区域在第一图像中的位置和第一图像区域的大小。Wherein, acquiring the first image area in the first image may include: acquiring the position of the first image area in the first image and the size of the first image area.
以下步骤中以该至少一个待检测区域包括第一图像区域为例进行说明。可选地,该至少一个待检测区域还可以包括第二图像区域等。In the following steps, the at least one area to be detected includes the first image area as an example for description. Optionally, the at least one area to be detected may further include a second image area and the like.
需要说明的是,对于置信度高于第一阈值的候选图像区域来说,车载设备在S202中所获取到的该候选图像区域的识别结果的准确度较高。因此,在后续步骤中,可以车载设备可以不再对这些候选图像区域中所包括的图像信息进行重复检测(或识别),也就是说,在后续步骤中,仅对置信度低于或等于第一阈值的候选图像区域所包括的图像信息进行检测(或识别),有助于降低检测复杂度,从而提高检测效率。It should be noted that, for a candidate image area with a confidence level higher than the first threshold, the recognition result of the candidate image area obtained by the vehicle-mounted device in S202 has a higher accuracy. Therefore, in the subsequent steps, the vehicle-mounted device can no longer repeatedly detect (or identify) the image information included in these candidate image regions, that is, in the subsequent steps, only the confidence level is lower than or equal to the first The detection (or recognition) of the image information included in the candidate image area with a threshold helps reduce the detection complexity, thereby improving the detection efficiency.
S204:车载设备获取该车辆的速度v、第一图像与第二图像的拍摄时间间隔T、拍摄第一图像时车辆与第一图像区域所对应的空间区域之间的相对角度θ1、拍摄第二图像时车辆与第一图像区域所对应的空间区域之间的相对角度θ2、以及在时间间隔T内车辆的朝向之间的相对角度(即车辆转向角度)α。S204: The vehicle-mounted device acquires the speed v of the vehicle, the shooting time interval T between the first image and the second image, the relative angle θ between the vehicle and the space area corresponding to the first image area when the first image is taken, and the second image is taken. The relative angle θ2 between the vehicle and the space area corresponding to the first image area at the time of the image, and the relative angle between the direction of the vehicle in the time interval T (ie, the vehicle steering angle) α.
车辆的速度可以是变化的,也可以是不变的。The speed of the vehicle can be variable or constant.
v、T、θ1、θ2和α均可以通过车辆中的相应传感器测量得到,或者用于获得这些参数中的部分或全部参数的原始信息均可以通过车辆中的相应传感器测量得到。例如,对于参数T来说,原始信息可以是第一图像的拍摄时刻t1和第二图像的拍摄时刻t2,其中,T=t2-t1。又如,对于参数α来说,原始信息可以是t1时刻车辆的朝向和t2时刻车辆的朝向。这些传感器中的部分或全部可以与车载设备集成在一起,也可以是独立设置的。v, T, θ1, θ2, and α can all be measured by corresponding sensors in the vehicle, or the original information used to obtain some or all of these parameters can be measured by corresponding sensors in the vehicle. For example, for the parameter T, the original information may be the shooting time t1 of the first image and the shooting time t2 of the second image, where T=t2-t1. For another example, for the parameter α, the original information may be the heading of the vehicle at time t1 and the heading of the vehicle at time t2. Some or all of these sensors can be integrated with on-board equipment, or they can be set independently.
S205:车载设备根据v、T、θ1和θ2,确定拍摄第一图像时车辆与第一图像区域所对应的空间区域之间的相对距离R1,以及拍摄第二图像时车辆与第一图像区域所对应的空间区域之间的相对距离R2。S205: The in-vehicle device determines, according to v, T, θ1, and θ2, the relative distance R1 between the vehicle and the space area corresponding to the first image area when the first image is taken, and the distance between the vehicle and the first image area when the second image is taken. The relative distance between the corresponding spatial regions R2.
如图6所示,假设:t1时刻,车体坐标系为X1轴和Y1轴构成X1-Y1坐标系,车辆所在的位置为X1-Y1坐标系的原点,且Y1轴为车辆行驶正前方向,X1轴为车辆切向右方向,摄像头拍摄的图像为第一图像;在t2时刻,车体坐标系为X2轴和Y2轴构成X2-Y2坐标系,车辆所在的位置为X2-Y2坐标系的原点,且Y2轴为车辆行驶正前方向,X2轴为车辆切向右方向,摄像头拍摄的图像为第二图像。那么:θ1为第一图像中的第一图像区域与Y1轴之间的夹角,θ2为第二图像中的第二图像区域与Y2轴之间的夹角。α、R1和R2在相应车体坐标系中的位置可以如图6所示。图6中的点划线区域为假设的车辆前视摄像头成像区域,其中,成像区域1为t1时刻的成像区域(即第一图像),成像区域2为t2时刻的成像区域(即第二图像)。当然在实际场景中,由于摄像头有视角等参数,因此,成像区域不会像图6中这样横平竖直。在本例中为了方便解释,假定摄像头的成像区域为车辆前方的一片矩形区域。As shown in Figure 6, suppose: at t1, the vehicle body coordinate system is X1 axis and Y1 axis to form the X1-Y1 coordinate system, the position of the vehicle is the origin of the X1-Y1 coordinate system, and the Y1 axis is the forward direction of the vehicle. , The X1 axis is the vehicle tangential to the right direction, and the image taken by the camera is the first image; at t2, the vehicle body coordinate system is the X2 axis and the Y2 axis to form the X2-Y2 coordinate system, and the position of the vehicle is the X2-Y2 coordinate system The origin of, and the Y2 axis is the forward direction of the vehicle, the X2 axis is the tangential right direction of the vehicle, and the image taken by the camera is the second image. Then: θ1 is the angle between the first image area in the first image and the Y1 axis, and θ2 is the angle between the second image area in the second image and the Y2 axis. The positions of α, R1 and R2 in the corresponding vehicle body coordinate system can be as shown in FIG. 6. The dashed-dotted area in Figure 6 is the assumed imaging area of the front-view camera of the vehicle, where imaging area 1 is the imaging area at time t1 (ie the first image), and imaging area 2 is the imaging area at time t2 (ie the second image) ). Of course, in the actual scene, since the camera has parameters such as the angle of view, the imaging area will not be horizontal and vertical as shown in Figure 6. For the convenience of explanation in this example, it is assumed that the imaging area of the camera is a rectangular area in front of the vehicle.
需要说明的是,图6中是基于车体坐标系确定R1和R2的,实际实现时,还可以基于其他坐标系(如世界坐标系、相机坐标系等)来确定R1和R2,本申请实施例对此不进行限定。It should be noted that R1 and R2 are determined based on the vehicle body coordinate system in FIG. 6. In actual implementation, R1 and R2 can also be determined based on other coordinate systems (such as the world coordinate system, camera coordinate system, etc.). This application is implemented The example does not limit this.
S206:车载设备根据第一图像中的第一图像区域、R1、R2和α,确定第二图像中的第一图像区域。S206: The vehicle-mounted device determines the first image area in the second image according to the first image area, R1, R2, and α in the first image.
第一图像和第二图像可以是S201中所描述的N帧图像中的任意两帧图像。在一种实现方式中,第一图像是该N帧图像中的参考图像(如第一帧图像),第二图像是该N帧图像中的任意一帧非参考图像。也就是说,本申请实施例支持“根据参考图像中的第一图像区域,确定其他任意一帧非参考图像中的第一图像区域”的技术方案。在另一种实现方式中,第一图像和第二图像是时序上相邻的图像。也就是说,本申请实施例支持“根据一帧图像的前一帧图像中的第一图像区域,确定该帧图像中的第一图像区域”的技术方案。The first image and the second image may be any two frames of the N frames of images described in S201. In an implementation manner, the first image is a reference image (such as the first image) in the N frames of images, and the second image is any non-reference image of the N frames of images. That is to say, the embodiments of the present application support the technical solution of "determining the first image area in any other frame of non-reference image according to the first image area in the reference image". In another implementation, the first image and the second image are images that are adjacent in time series. That is to say, the embodiment of the present application supports the technical solution of "determining the first image area in a frame image according to the first image area in the previous frame image of the frame image".
在图6中,ROI1为第一图像中的第一图像区域。当前示例中为一个矩形,其中心点已标粗处理。ROI2为第二图像中的第一图像区域,S206具体为:根据ROI1在第一图像中的位置和大小、R1、R2和α,确定ROI2在第二图像中的位置和大小。下面说明ROI2在第二图像中的位置和大小是如何确定的。In Fig. 6, ROI1 is the first image area in the first image. In the current example, it is a rectangle whose center point has been roughed. ROI2 is the first image region in the second image, and S206 is specifically: determining the position and size of ROI2 in the second image according to the position and size of ROI1 in the first image, R1, R2, and α. The following describes how the position and size of ROI2 in the second image are determined.
由于α是t1时刻至t2时刻这段时间内的车辆转向角度,因此,可以确定ROI2相对于ROI1的旋转角度也是α。假设ROI1和ROI2的中心点在t1、t2时刻没有发生变化,也就是说,在世界坐标系下,ROI1中心点和ROI2中心点的坐标相同。因此,可以将ROI1的中心点坐标从X1-Y1坐标系转换到世界坐标系下,再转换到X2-Y2坐标系下,从而可以得到ROI2中心点在X2-Y2坐标系下的坐标。至此,可以得到ROI2在第二图像中的位置。Since α is the steering angle of the vehicle from time t1 to time t2, it can be determined that the rotation angle of ROI2 relative to ROI1 is also α. It is assumed that the center points of ROI1 and ROI2 do not change at t1 and t2, that is, in the world coordinate system, the coordinates of the center point of ROI1 and the center point of ROI2 are the same. Therefore, the coordinates of the center point of ROI1 can be converted from the X1-Y1 coordinate system to the world coordinate system, and then to the X2-Y2 coordinate system, so that the coordinates of the center point of ROI2 in the X2-Y2 coordinate system can be obtained. So far, the position of ROI2 in the second image can be obtained.
将ROI1旋转α之后,得到ROI1’,如图7所示。由于在本示例中,相比t1时刻,当车辆行驶到t2时刻时,距离第一图像区域所对应的空间区域更近了,因此,根据成像原理,理论上t2时刻要用更大的框才能框住和t1时刻同样信息量的区域。基于此,在ROI1’的基础上,可以根据R1和R2的比值(可选地,还可以根据一定的权重),将ROI1’的长进行缩放(本示例中是放大),得到ROI2的长;同理可得ROI2的宽。至此,可以得到ROI2的大小。After ROI1 is rotated by α, ROI1' is obtained, as shown in Figure 7. Since in this example, compared to t1, when the vehicle travels to t2, it is closer to the spatial area corresponding to the first image area. Therefore, according to the imaging principle, theoretically a larger frame is required at t2. Frame the area with the same amount of information as at t1. Based on this, on the basis of ROI1', the length of ROI1' can be scaled (enlarged in this example) according to the ratio of R1 and R2 (optionally, according to a certain weight) to obtain the length of ROI2; Similarly, the width of ROI2 can be obtained. So far, the size of ROI2 can be obtained.
这里以一个示例,对t2时刻要用更大的框才能框住和t1时刻同样信息量的区域进行说明:假设第一图像为图8中的(a)图所示,第一图像中的第一图像区域为(a)图中矩形框所示的区域;第二图像为图8中的(b)图所示,第二图像中的第一图像区域为(b)图中矩形框所示的区域。基于此,可知,相比在t1时刻,第一图像区域所对应的车辆在t2时刻与本实施例的执行主体所在的车辆的距离较近,因此,第二图像中的第一图像区域的大小必须大于第一图像中的第一图像区域(如(b)图中的矩形框所示的区域)的大小,才可以使得第二图像中的第一图像区域包含该骑着摩托车的人的全部图像信息。Here is an example to illustrate that a larger frame is needed at t2 to frame the area with the same amount of information as at t1: suppose the first image is shown in Figure 8 (a), the first image in the first image An image area is the area shown by the rectangular box in (a); the second image is shown in (b) in FIG. 8, and the first image area in the second image is shown in the rectangular box in (b) Area. Based on this, it can be seen that the vehicle corresponding to the first image area is closer to the vehicle where the execution subject of this embodiment is located at time t2 than at time t1. Therefore, the size of the first image area in the second image is It must be larger than the size of the first image area in the first image (such as the area shown by the rectangular box in figure (b)) in order to make the first image area in the second image contain the person riding a motorcycle All image information.
上述S204~S206是以根据第一图像中的第一图像区域,推理得到第二图像中的第一图像区域为例进行说明的。据此,针对不同的第二图像,执行一次或多次执行S204~S206,可以获得N帧图像中的部分或全部图像中所包括的第一图像区域。The above S204 to S206 are described by taking the first image area in the second image inferred from the first image area in the first image as an example. Accordingly, for different second images, by executing S204-S206 one or more times, the first image area included in part or all of the N frames of images can be obtained.
需要说明的是,上述参数v、T、θ1、θ2、α、R1和R2统称为车体信息。其中,车体信息可以是直接通过车辆中安装的传感器等设备检测得到的该车辆的信息(如上述v、T、θ1、θ2和α等),也可以是对这些传感器等设备检测到的信息进行处理得 到的该车辆的信息(如R1和R2等)。上述S204~S206仅为本申请实施例提供的“一种基于车体信息推理得到第二图像中的第一图像区域”的示例,其不对可适用本申请实施例的“基于车体信息推理得到第二图像中的第一图像区域”的具体实现方式构成限定。具体实现时,可以基于比上述所列举的车体信息更多或更少的车体信息,来实现“推理得到第二图像中的第一图像区域”。It should be noted that the above parameters v, T, θ1, θ2, α, R1, and R2 are collectively referred to as vehicle body information. Among them, the vehicle body information can be the information of the vehicle (such as the above v, T, θ1, θ2, and α, etc.) detected directly by the sensors and other equipment installed in the vehicle, or it can be the information detected by these sensors and other equipment The information of the vehicle (such as R1 and R2, etc.) obtained by processing. The above S204 to S206 are only examples of "a first image area in the second image obtained by reasoning based on vehicle body information" provided by the embodiment of this application, which is not applicable to the "reduction based on vehicle body information in the embodiment of this application". The specific implementation of "the first image area in the second image" constitutes a limitation. In specific implementation, it is possible to implement "the first image area in the second image is obtained by reasoning" based on more or less vehicle body information than the vehicle body information listed above.
S207:车载设备执行该N1个第一图像区域的景物对齐。S207: The vehicle-mounted device executes scene alignment of the N1 first image regions.
当车辆处于运动状态时,由于车辆运动状态不定,因此,对齐过程涉及比较多的场景,不同场景可以使用不同的方案进行景物对齐,这里根据车辆运动轨迹举两个示例。以下两个示例均以摄像头是前视摄像头且安装在车辆内部的正前方为例,对两帧图像中的第一图像区域对齐过程进行说明。When the vehicle is in motion, because the vehicle motion state is uncertain, the alignment process involves more scenes, and different scenarios can use different solutions for scene alignment. Here are two examples based on the vehicle motion trajectory. In the following two examples, the camera is a front-view camera and is installed directly in front of the vehicle as an example to describe the alignment process of the first image area in the two frames of images.
示例1:车辆笔直向前方运动场景。Example 1: The scene of the vehicle moving straight ahead.
在车辆笔直向前运动的过程中,前视摄像头采集不同时刻的第一图像和第二图像,第一图像和第二图像包括的第一图像区域分别为ROI1和ROI2。若ROI1和ROI2的中心点恰好处于车辆正前方,则由于在获取第一图像区域时,已经确保了ROI2包含了和ROI1相同的景物(包括前景和背景)的信息,因此景物对齐,具体可以是将ROI1缩放(本例中是放大)到ROI2的大小即可。During the straight forward movement of the vehicle, the front-view camera collects the first image and the second image at different moments. The first image and the second image include the first image regions ROI1 and ROI2, respectively. If the center points of ROI1 and ROI2 are exactly in front of the vehicle, it has been ensured that ROI2 contains the same information of the scene (including foreground and background) as ROI1 when acquiring the first image area, so the scenes are aligned, which can be specifically Scale ROI1 (in this case, zoom in) to the size of ROI2.
示例2:车辆向右前方转弯的场景。Example 2: A scene where the vehicle turns to the front right.
在车辆向右前方转弯的过程中,前视摄像头采集不同时刻(即t1、t2时刻)的第一图像和第二图像,第一图像和第二图像包括的第一图像区域分别为ROI1和ROI2。车辆转弯会导致ROI1和ROI2除了拥有由于车辆位移带来的比例关系外,还拥有车辆方向偏移所带来的视角变换。若ROI1在车辆右前方,则车辆向右转弯(即X2-Y2坐标系的原点在X1-Y1坐标系的右上方)时,视角必然会更多的包含ROI左侧的信息,如图9所示。其中,图9中是以第一图像区域所包括的对象的图像信息是一个矩形为例进行说明的。在t1时刻,车体坐标系是X1-Y1坐标系。在t2时刻,车体坐标系是X2-Y2坐标系。When the vehicle turns to the front right, the front-view camera collects the first image and the second image at different moments (ie t1 and t2). The first image and the second image include the first image regions ROI1 and ROI2, respectively . Turning a vehicle will cause ROI1 and ROI2 to not only have the proportional relationship due to vehicle displacement, but also have the perspective change caused by the vehicle direction offset. If ROI1 is at the front right of the vehicle, when the vehicle turns to the right (that is, the origin of the X2-Y2 coordinate system is at the upper right of the X1-Y1 coordinate system), the angle of view will inevitably include more information on the left side of the ROI, as shown in Figure 9. Show. Wherein, in FIG. 9, the image information of the object included in the first image area is a rectangle as an example for description. At t1, the vehicle body coordinate system is the X1-Y1 coordinate system. At t2, the vehicle body coordinate system is the X2-Y2 coordinate system.
根据图9,可以很明显的看到,摄像头在t1时刻位置采集的第一图像中,该对象的正面信息更多,摄像头在t2时刻位置采集第二图像中,该对象的左侧信息更多。其中,摄像头所在的位置是坐标系的原点。这样,会使得第一图像、第二图像中的第一图像区域中的景物包含的信息量不同,从而除了ROI1、ROI2大小要对齐,其视野中的景物方向也要做一次对齐,也就是说,需要获取第一图像中的第一图像区域与第二图像中的第一图像区域之间的景物对齐角度。示例的,景物对齐角度可以为图10中β。图10是基于图6进行绘制的。如图10所示,可知,β=180°-γ-θ1获取;而γ=180°-α-θ2。其中,γ是为了方便计算而引入的一个中间量。由此可知,基于α、θ1和θ2可以得到景物对齐角度β。According to Figure 9, it can be clearly seen that in the first image collected by the camera at time t1, the front information of the object is more, and in the second image collected by the camera at time t2, the left side of the object has more information. . Among them, the position of the camera is the origin of the coordinate system. In this way, the amount of information contained in the scene in the first image area in the first image and the second image will be different, so that in addition to the ROI1 and ROI2 sizes, the direction of the scene in the field of view must also be aligned once, that is to say , It is necessary to obtain the scene alignment angle between the first image area in the first image and the first image area in the second image. For example, the scene alignment angle may be β in FIG. 10. Figure 10 is drawn based on Figure 6. As shown in Fig. 10, it can be seen that β=180°-γ-θ1 is obtained; and γ=180°-α-θ2. Among them, γ is an intermediate quantity introduced for the convenience of calculation. It can be seen that the scene alignment angle β can be obtained based on α, θ1, and θ2.
在景物方向对齐之后,需要按像素进行景物对齐。在一个示例中,基于如图10中的示例,可以按照景物对齐角度β将ROI2中的景物进行映射。图11为将ROI2以β为角度,以像素点为单位向ROI1所在的平面进行映射,得到ROI2’,图11中ROI2’所在的平面即为ROI1所在的平面。然后,将ROI1对照ROI2’进行缩放(本例是放大),以使得ROI1与ROI2’的大小相同。至此,完成了景物对齐。After the scene direction is aligned, the scene needs to be aligned by pixel. In one example, based on the example in FIG. 10, the scene in ROI2 can be mapped according to the scene alignment angle β. Fig. 11 shows ROI2 is mapped to the plane where ROI1 is located by taking β as the angle and pixels as the unit to obtain ROI2’. In Fig. 11, the plane where ROI2’ is located is the plane where ROI1 is located. Then, compare ROI1 to ROI2' for zooming (in this example, zooming in), so that ROI1 and ROI2' have the same size. At this point, the scene alignment is completed.
需要说明的是,为了便于描述,S207中的示例中是以第一图像和第二图像中的第一图像区域对齐为例进行说明的。这里的第一图像和第二图像仅仅是为了区分N1图像中的任意两帧图像。在实际实现时,对齐过程中的第一图像和第二图像,与上述确定第一图像区域的过程中的第一图像和第二图像可以对应相同,也可以不同。It should be noted that, for ease of description, the example in S207 takes the alignment of the first image area in the first image and the second image as an example for description. The first image and the second image here are only for distinguishing any two images in the N1 image. In actual implementation, the first image and the second image in the alignment process may correspond to the same or different from the first image and the second image in the process of determining the first image area.
需要说明的是,上述示例2仅为本申请实施例提供的“一种基于车体信息实现景物对齐”的示例,其不对可适用本申请实施例的“基于车体信息实现景物对齐”的实现方式构成限定。具体实现时,可以基于比示例2所列举的车体信息更多或更少的车体信息,来实现景物对齐。It should be noted that the above example 2 is only an example of "a realization of scene alignment based on vehicle body information" provided by the embodiment of this application, which is not applicable to the realization of "realization of scene alignment based on vehicle body information" in the embodiment of this application. The way constitutes limitation. In specific implementation, it is possible to achieve scene alignment based on more or less vehicle body information than the vehicle body information listed in Example 2.
S208~S209:可以参考上述S104~S105,当然本申请实施例不限于此。S208 to S209: The above S104 to S105 can be referred to, of course, the embodiment of the present application is not limited thereto.
本实施例中相关内容的解释,以及能够达到的有益效果均可以参考上述图4所示的实施例。除此之外,一方面,本实施例中,选择一帧图像中置信度低于第一阈值的候选图像区域作为第一图像区域,并对多帧图像中的第一图像区域进行超分处理。这样,不需要对置信度高于第一阈值的候选图像区域进行超分处理,有助于降低超分处理的复杂度,从而提高超分处理的效率。另一方面,本实施例中,依据一帧图像中的第一图像区域以及车体信息,确定另一帧图像中的第一图像区域,这样,在执行超分运算时,有助于车载设备获取到更多地空间信息,从而提高超分运算的精确度。For the explanation of related content in this embodiment and the beneficial effects that can be achieved, reference may be made to the embodiment shown in FIG. 4. In addition, on the one hand, in this embodiment, a candidate image area with a confidence level lower than the first threshold in one frame of image is selected as the first image area, and the first image area in the multi-frame image is super-divided. . In this way, there is no need to perform super-division processing on candidate image regions with a confidence level higher than the first threshold, which helps reduce the complexity of super-division processing, thereby improving the efficiency of super-division processing. On the other hand, in this embodiment, according to the first image area in one frame of image and the vehicle body information, the first image area in another frame of image is determined, so that when performing super-division operations, it is helpful for vehicle equipment Get more spatial information, thereby improving the accuracy of super-division operations.
实施例2Example 2
如图12所示,为本申请实施例提供的一种图像处理方法的流程示意图。该方法可以包括:As shown in FIG. 12, it is a schematic flowchart of an image processing method provided by an embodiment of this application. The method can include:
S301:可以参考上述S101,当然本申请实施例不限于此。S301: Refer to the above S101, of course, the embodiment of the present application is not limited to this.
S302:车载设备对该N帧图像中的第一图像进行预处理(如图像分割等),得到第一图像中的可行驶区域的图像信息。S302: The vehicle-mounted device performs preprocessing (such as image segmentation, etc.) on the first image in the N frames of images to obtain image information of the drivable area in the first image.
可行驶区域,是视野范围内各个方向上距离车辆的第一个对象之间的区域。例如,图13中的黑色线框住的区域表示一种可行驶区域的示意图。The drivable area is the area between the first objects that are away from the vehicle in all directions in the field of view. For example, the area enclosed by the black line in FIG. 13 represents a schematic diagram of a drivable area.
可以理解的是,由于图像分割技术或其他得到第一图像的可行区域的图像信息的技术,均不能百分之百确定所获得的可行区域中没有对象,因此,可以基于本申请实施例提供的技术方案进一步确定是否包含其他对象,也就是说,可以接着执行以下S303。It is understandable that, because image segmentation technology or other technologies for obtaining image information of the feasible area of the first image cannot be 100% sure that there is no object in the obtained feasible area, it can be further based on the technical solution provided in the embodiments of the present application. It is determined whether other objects are included, that is, the following S303 can be executed next.
关于第一图像的相关描述可以参考上文,此处不再赘述。For the relevant description of the first image, please refer to the above, which will not be repeated here.
S303:车载设备根据第一图像中的可行驶区域的图像信息,将可行驶区域中的与该车辆之间的距离大于或等于第二阈值的空间区域所对应的图像信息,作为第二图像中的第一图像区域。S303: According to the image information of the drivable area in the first image, the in-vehicle device uses the image information corresponding to the space area in the drivable area whose distance to the vehicle is greater than or equal to the second threshold as the image information in the second image The first image area.
可选地,车载设备可以先确定其所在车辆对应在摄像头所拍摄的图像中的位置,然后将第二图像中的且与该位置的距离大于或等于一个阈值的图像区域,作为第二图像中的第一图像区域。其中,第二阈值与该阈值之间的比例,等于空间距离与图像距离(即该空间距离映射到图像中的图像距离)之间的比例。Optionally, the in-vehicle device may first determine the position of the vehicle in which it is located in the image captured by the camera, and then use the image area in the second image whose distance from the position is greater than or equal to a threshold as the second image The first image area. Wherein, the ratio between the second threshold and the threshold is equal to the ratio between the spatial distance and the image distance (that is, the image distance mapped from the spatial distance to the image).
需要说明的是,虽然摄像头所拍摄的图像中没有本车辆(即本实施例中车载设备所在的车辆)的图像信息,但是,可以基于摄像头在该车辆中的位置信息,可选地还以基于该车辆的运动状态(如转弯或直行等),确定出该车辆对应在该图像中的位置。It should be noted that although the image captured by the camera does not contain the image information of the vehicle (that is, the vehicle in which the on-board equipment is located in this embodiment), it can be based on the location information of the camera in the vehicle, or optionally based on The motion state of the vehicle (such as turning or going straight) determines the position of the vehicle in the image.
例如,以摄像头位于车辆内部的前方正中间,且车辆与摄像头的相对位置不变为例,该车辆对应在图像中的位置可以是该图片的下边界正中间,如图14所示。图14中还示意出了该情况下的第一图像区域。For example, if the camera is located in the middle of the front of the vehicle and the relative position of the vehicle and the camera remains unchanged, the corresponding position of the vehicle in the image may be the middle of the lower boundary of the picture, as shown in FIG. 14. Figure 14 also illustrates the first image area in this case.
又如,以摄像头位于车辆内部的前方正中间,且摄像头相对于车辆向左旋转为例,该车辆对应在摄像头所拍摄的图像中的位置可以是该图片的右下角。For another example, if the camera is located in the middle of the front of the vehicle and the camera rotates to the left relative to the vehicle as an example, the corresponding position of the vehicle in the image captured by the camera may be the lower right corner of the picture.
确定车辆对应在摄像头所拍摄的图像中的位置的方法不限于此,如可参考现有技术。The method for determining the corresponding position of the vehicle in the image captured by the camera is not limited to this, for example, reference may be made to the prior art.
S304~S309:可以参考上述S204~S209,当然本申请实施例不限于此。S304-S309: The foregoing S204-S209 can be referred to, of course, the embodiment of the present application is not limited thereto.
依据S302~S306,车载设备可以获得N帧图像的N1帧图像中每帧图像所包括的第一图像区域。在本示例中,由于通常每帧图像中均具有可行驶区域,因此,通常N=N1。According to S302 to S306, the vehicle-mounted device can obtain the first image area included in each frame of the N1 frames of the N frames of images. In this example, since there is usually a drivable area in each frame of image, usually N=N1.
本实施例中相关内容的解释,以及能够达到的有益效果均可以参考上述图4所示的实施例。除此之外,一方面,本实施例中,将一帧图像中对应于车辆的可行驶区域中与该车辆之间的距离大于或等于第二阈值的空间区域的图像区域,作为第一图像区域,并对多帧图像中的第一图像区域进行超分处理。考虑到:虽然可行驶区域被定义为“视野范围内各个方向上距离车辆的第一个对象之间的区域”,但是,不能百分百保证所确定的可行驶区域中不包含对象(或目标对象),而对于所确定的可行驶区域而言,距离车辆越远的区域中包含目标对象的概率就越高,基于此,提出本实施例。这样,有助于在现有技术的基础上,提高检测目标对象的精确度。另一方面,本实施例中,依据一帧图像中的第一图像区域以及车体信息,确定另一帧图像中的第一图像区域,这样,在执行超分运算时,有助于车载设备获取到更多地空间信息,从而提高超分运算的精确度。For the explanation of related content in this embodiment and the beneficial effects that can be achieved, reference may be made to the embodiment shown in FIG. 4. In addition, on the one hand, in this embodiment, an image area corresponding to a space area in the drivable area of the vehicle whose distance from the vehicle is greater than or equal to the second threshold is taken as the first image Region, and super-divide the first image region in the multi-frame image. Considering: Although the drivable area is defined as the "area between the first object of the vehicle in various directions in the field of view", there is no guarantee that the determined drivable area does not contain objects (or targets). Object), and for the determined drivable area, the greater the probability that the target object is contained in the area farther from the vehicle, this embodiment is proposed based on this. In this way, it is helpful to improve the accuracy of detecting the target object on the basis of the existing technology. On the other hand, in this embodiment, according to the first image area in one frame of image and the vehicle body information, the first image area in another frame of image is determined, so that when performing super-division operations, it is helpful for vehicle equipment Get more spatial information, thereby improving the accuracy of super-division operations.
实施例3Example 3
如图15所示,为本申请实施例提供的一种图像处理方法的流程示意图。该方法可以包括:As shown in FIG. 15, it is a schematic flowchart of an image processing method provided by an embodiment of this application. The method can include:
S401:可以参考上述S101,当然本申请实施例不限于此。S401: Refer to the above S101, of course, the embodiment of the application is not limited to this.
S402:车载设备将该N帧图像中的第一图像中的预设位置的区域作为第一图像区域。S402: The in-vehicle device uses the area at the preset position in the first image in the N frames of images as the first image area.
在目标对象不同场景中,第一图像区域在第一图像中的位置可能相同,也可能不同。In different scenes of the target object, the position of the first image area in the first image may be the same or different.
例如,如果目标对象是交通灯,由于交通灯的图像信息通常在一帧图像中的上方,因此,可以将一帧图像内的上方预设位置的区域(如上方五分之二的区域)作为第一图像区域。图16中给出了一种第一图像区域的示意图。For example, if the target object is a traffic light, since the image information of the traffic light is usually above in a frame of image, the area (such as the upper two-fifths of the area) at the upper preset position in one frame of image can be taken as The first image area. A schematic diagram of the first image area is shown in FIG. 16.
又如,如果目标对象是高速公路上的限速指示牌,由于高速公路上的限速指示牌的图像信息通常在一帧图像中的右侧,因此,可以将一帧图像内的右侧预设位置的区域(如右侧四分之一的区域)作为第一图像区域。For another example, if the target object is a speed limit sign on an expressway, since the image information of the speed limit sign on an expressway is usually on the right side of an image, the right side of an image can be pre-defined. Set the location area (such as the area on the right side) as the first image area.
S403~S408:可以参考上述S204~S209,当然本申请实施例不限于此。S403 to S408: The foregoing S204 to S209 can be referred to, of course, the embodiment of the present application is not limited thereto.
依据S402~S405,车载设备可以获得N帧图像的N1帧图像中每帧图像所包括的第一图像区域。According to S402 to S405, the vehicle-mounted device can obtain the first image area included in each frame of the N1 frames of the N frames of images.
本实施例中相关内容的解释,以及能够达到的有益效果均可以参考上述图4所示 的实施例。除此之外,一方面,本实施例中,将一帧图像中预设位置的区域作为第一图像区域,并对多帧图像中的第一图像区域进行超分处理。该方法实现较简单,方便。另一方面,本实施例中,依据一帧图像中的第一图像区域以及车体信息,确定另一帧图像中的第一图像区域,这样,在执行超分运算时,有助于车载设备获取到更多地空间信息,从而提高超分运算的精确度。For the explanation of related content in this embodiment and the beneficial effects that can be achieved, reference may be made to the embodiment shown in FIG. 4 above. In addition, on the one hand, in this embodiment, an area at a preset position in one frame of image is used as the first image area, and super-division processing is performed on the first image area in multiple frames of images. This method is relatively simple and convenient to implement. On the other hand, in this embodiment, according to the first image area in one frame of image and the vehicle body information, the first image area in another frame of image is determined, so that when performing super-division operations, it is helpful for vehicle equipment Get more spatial information, thereby improving the accuracy of super-division operations.
上述主要从方法的角度对本申请实施例提供的方案进行了介绍。为了实现上述功能,其包含了执行各个功能相应的硬件结构和/或软件模块。本领域技术人员应该很容易意识到,结合本文中所公开的实施例描述的各示例的单元及算法步骤,本申请能够以硬件或硬件和计算机软件的结合形式来实现。某个功能究竟以硬件还是计算机软件驱动硬件的方式来执行,取决于技术方案的特定应用和设计约束条件。专业技术人员可以对每个特定的应用来使用不同方法来实现所描述的功能,但是这种实现不应认为超出本申请的范围。The foregoing mainly introduces the solutions provided in the embodiments of the present application from the perspective of methods. In order to realize the above-mentioned functions, it includes hardware structures and/or software modules corresponding to each function. Those skilled in the art should easily realize that in combination with the units and algorithm steps of the examples described in the embodiments disclosed herein, the present application can be implemented in the form of hardware or a combination of hardware and computer software. Whether a certain function is executed by hardware or computer software-driven hardware depends on the specific application and design constraint conditions of the technical solution. Professionals and technicians can use different methods for each specific application to implement the described functions, but such implementation should not be considered beyond the scope of this application.
本申请实施例可以根据上述方法示例对车载设备进行功能模块的划分,例如可以对应各个功能划分各个功能模块,也可以将两个或两个以上的功能集成在一个处理模块中。上述集成的模块既可以采用硬件的形式实现,也可以采用软件功能模块的形式实现。需要说明的是,本申请实施例中对模块的划分是示意性的,仅仅为一种逻辑功能划分,实际实现时可以有另外的划分方式。The embodiments of the present application may divide the in-vehicle equipment into functional modules according to the foregoing method examples. For example, each functional module may be divided corresponding to each function, or two or more functions may be integrated into one processing module. The above-mentioned integrated modules can be implemented in the form of hardware or software functional modules. It should be noted that the division of modules in the embodiments of the present application is illustrative, and is only a logical function division, and there may be other division methods in actual implementation.
如图17所示,为本申请实施例提供的一种设备,具体可以为车载设备170的结构示意图。作为设备的一个示例,车载设备170可以用于执行图4、图5、图12或图15所示的方法中车载设备所执行的步骤。As shown in FIG. 17, a device provided by an embodiment of this application may specifically be a schematic structural diagram of a vehicle-mounted device 170. As an example of the device, the vehicle-mounted device 170 may be used to execute the steps performed by the vehicle-mounted device in the method shown in FIG. 4, FIG. 5, FIG. 12, or FIG.
车载设备170可以包括:第一获取模块1701、第二获取模块1702和超分模块1703。其中,第一获取模块1701,用于获取多帧图像,该多帧图像包括该车载设备所在车辆的周边道路的图像信息。第二获取模块1702,用于获取该多帧图像的每帧图像中的第一图像区域;其中,该多帧图像的多个第一图像区域对应于第一场景。超分模块1703,用于对该多个第一图像区域进行超分运算。例如,结合图4,第一获取模块1701可以用于执行S101,第二获取模块1702可以用于执行S102,超分模块1703可以用于执行S104。The in-vehicle device 170 may include: a first acquisition module 1701, a second acquisition module 1702, and a super-division module 1703. Wherein, the first acquisition module 1701 is configured to acquire a multi-frame image, and the multi-frame image includes image information of the surrounding road of the vehicle where the on-board device is located. The second acquisition module 1702 is configured to acquire the first image area in each frame of the multi-frame image; wherein, the multiple first image areas of the multi-frame image correspond to the first scene. The super-division module 1703 is configured to perform super-division operations on the multiple first image regions. For example, with reference to FIG. 4, the first acquisition module 1701 may be used to perform S101, the second acquisition module 1702 may be used to perform S102, and the super-division module 1703 may be used to perform S104.
可选地,车载设备170还包括:确定模块1704,用于确定超分运算得到的图像区域中存在目标对象的图像信息。例如,结合图4,确定模块1704可以用于执行S105。Optionally, the in-vehicle device 170 further includes: a determining module 1704, configured to determine that the image information of the target object exists in the image area obtained by the super-division operation. For example, in conjunction with FIG. 4, the determining module 1704 may be used to perform S105.
可选地,对于该多帧图像的每帧图像:第一图像区域是置信度低于或等于第一阈值的区域;或者,第一图像区域对应于该车辆的可行驶区域中与该车辆之间的距离大于或等于第二阈值的空间区域;或者,第一图像区域是预设位置的区域。Optionally, for each image of the multi-frame image: the first image area is an area with a confidence level lower than or equal to a first threshold; or, the first image area corresponds to the difference between the drivable area of the vehicle and the vehicle. A space area where the distance between the two is greater than or equal to the second threshold; or, the first image area is an area at a preset position.
可选地,该多帧图像包括第一图像和第二图像。第二获取模块1702具体用于:获取第二图像中的第一图像区域。获取第二图像中的第一图像区域,包括:根据第一图像中的第一图像区域和该车辆的第一车体信息获取第二图像中的第一图像区域。例如,第一车体信息可以包括第一相对距离、第二相对距离和第一车辆转向角度中的至少一种。其中,第一相对距离是拍摄第一图像时该车辆与第一图像区域所对应的空间区域之间的相对距离,第二相对距离是拍摄第二图像时该车辆与第一图像区域所对应的空间区域之间的相对距离,第一车辆转向角度是在第一图像和第二图像的拍摄时间间隔 内,该车辆的朝向之间的夹角。例如,结合图5,第二获取模块1702可以用于执行S204和S205。可选的,所述第一车体信息还可以包括车身高度参数,例如摄像头离地高度和/或车体高度等。Optionally, the multi-frame image includes a first image and a second image. The second acquiring module 1702 is specifically configured to acquire the first image area in the second image. Obtaining the first image area in the second image includes: obtaining the first image area in the second image according to the first image area in the first image and the first body information of the vehicle. For example, the first vehicle body information may include at least one of a first relative distance, a second relative distance, and a first vehicle steering angle. Wherein, the first relative distance is the relative distance between the vehicle and the space area corresponding to the first image area when the first image is taken, and the second relative distance is the distance between the vehicle and the first image area when the second image is taken. The relative distance between the spatial regions, the first vehicle steering angle is the angle between the direction of the vehicle in the time interval of shooting the first image and the second image. For example, in conjunction with FIG. 5, the second acquisition module 1702 may be used to perform S204 and S205. Optionally, the first vehicle body information may also include vehicle height parameters, such as the height of the camera from the ground and/or the vehicle body height.
可选地,超分模块1703具体用于:对景物对齐后的多个第一图像区域进行超分运算。例如,超分模块1703可以用于执行图4中的S104,图5中的S208,图12中的S308或图15中的S407。Optionally, the super-division module 1703 is specifically configured to perform super-division operations on multiple first image regions after scene alignment. For example, the super-division module 1703 may be used to execute S104 in FIG. 4, S208 in FIG. 5, S308 in FIG. 12, or S407 in FIG. 15.
可选地,该多帧图像包括第三图像和第四图像。车载设备170还包括:对齐模块1705,用于根据该车辆的第二车体信息,执行该多个第一图像区域的景物对齐。例如,第二车体信息可以包括第一相对角度、第二相对角度和第二车辆转向角度中的至少一种。其中,第一相对角度是拍摄第三图像时该车辆与第一图像区域所对应的空间区域之间的相对角度,第二相对角度是拍摄所述第四图像时该车辆与第一图像区域所对应的空间区域之间的相对角度,第二车辆转向角度是在第三图像和第四图像的拍摄时间间隔内,该车辆的朝向之间的夹角。可选的,所述第二车体信息还可以包括车身高度参数,例如摄像头离地高度和/或车体高度等。Optionally, the multi-frame image includes a third image and a fourth image. The in-vehicle device 170 further includes an alignment module 1705, configured to perform scene alignment of the plurality of first image regions according to the second body information of the vehicle. For example, the second vehicle body information may include at least one of a first relative angle, a second relative angle, and a second vehicle steering angle. Wherein, the first relative angle is the relative angle between the vehicle and the space area corresponding to the first image area when the third image is taken, and the second relative angle is the relative angle between the vehicle and the first image area when the fourth image is taken. Corresponding to the relative angle between the spatial regions, the second vehicle steering angle is the angle between the direction of the vehicle in the time interval of shooting the third image and the fourth image. Optionally, the second vehicle body information may also include vehicle height parameters, such as the height of the camera from the ground and/or the vehicle body height.
可选地,该多帧图像是时序上连续的多帧图像。Optionally, the multi-frame images are consecutive multi-frame images in time series.
可选地,该多帧图像中的首帧图像的拍摄时刻与末帧图像的拍摄时刻之间的时间间隔小于或等于第三阈值。Optionally, the time interval between the shooting moment of the first frame image and the shooting moment of the last frame image in the multi-frame image is less than or equal to a third threshold.
可选地,第二获取模块1702还用于:获取该多帧图像的每帧图像中的第二图像区域;其中,该多帧图像的多个第二图像区域对应于第二场景。超分模块1703还用于:对该多个第二图像区域进行超分运算。Optionally, the second obtaining module 1702 is further configured to: obtain a second image area in each frame of the multi-frame image; wherein, the multiple second image areas of the multi-frame image correspond to the second scene. The super-division module 1703 is further configured to perform super-division operations on the multiple second image regions.
在一个示例中,参见图1,上述第一获取模块1701可以通过图1中的I/O接口115实现,上述第二获取模块1702、超分模块1703、确定模块1704和对齐模块1705中的至少一种可以由图1中的处理器103调用应用程序143实现。In an example, referring to FIG. 1, the above-mentioned first acquisition module 1701 may be implemented through the I/O interface 115 in FIG. 1. At least one of the above-mentioned second acquisition module 1702, super-division module 1703, determination module 1704, and alignment module 1705 One can be implemented by the processor 103 in FIG. 1 calling the application program 143.
关于上述可选方式的具体描述参见前述的方法实施例,此处不再赘述。上述提供的任一种车载设备170的解释以及有益效果的描述均可参考上述对应的方法实施例,不予赘述。For specific descriptions of the foregoing optional manners, refer to the foregoing method embodiments, which are not repeated here. For the explanation and the description of the beneficial effects of any of the on-board equipment 170 provided above, reference may be made to the corresponding method embodiment above, and details are not repeated.
需要说明的是,上述各个单元对应执行的动作仅是具体举例,各个单元实际执行的动作参照上述基于行图4、图5、图12或图15所述的实施例的描述中提及的动作或步骤。It should be noted that the actions corresponding to each unit above are only specific examples, and the actions actually performed by each unit refer to the actions mentioned in the above description based on the embodiment described in Figure 4, Figure 5, Figure 12 or Figure 15. Or steps.
本领域普通技术人员可以理解实现上述实施例的全部或部分步骤可通过程序来指令相关的硬件完成。所述的程序可以存储于一种计算机可读存储介质中。上述提到的存储介质可以是只读存储器,随机接入存储器等。上述处理单元或处理器可以是中央处理器,通用处理器、特定集成电路(application specific integrated circuit,ASIC)、微处理器(digital signal processor,DSP),现场可编程门阵列(field programmable gate array,FPGA)或者其他可编程逻辑器件、晶体管逻辑器件、硬件部件或者其任意组合。Those of ordinary skill in the art can understand that all or part of the steps for implementing the above-mentioned embodiments can be completed by a program instructing related hardware. The program can be stored in a computer-readable storage medium. The aforementioned storage medium may be a read-only memory, a random access memory, and the like. The above-mentioned processing unit or processor may be a central processing unit, a general-purpose processor, an application specific integrated circuit (ASIC), a microprocessor (digital signal processor, DSP), a field programmable gate array (field programmable gate array, FPGA) or other programmable logic devices, transistor logic devices, hardware components, or any combination thereof.
本申请实施例还提供了一种包含指令的计算机程序产品,当该指令在计算机上运行时,使得计算机执行上述实施例中的任意一种方法。该计算机程序产品包括一个或多个计算机指令。在计算机上加载和执行计算机程序指令时,全部或部分地产生按照本申请实施例所述的流程或功能。计算机可以是通用计算机、专用计算机、计算机网 络、或者其他可编程装置。计算机指令可以存储在计算机可读存储介质中,或者从一个计算机可读存储介质向另一个计算机可读存储介质传输,例如,计算机指令可以从一个网站站点、计算机、服务器或者数据中心通过有线(例如同轴电缆、光纤、数字用户线(digital subscriber line,DSL))或无线(例如红外、无线、微波等)方式向另一个网站站点、计算机、服务器或数据中心进行传输。计算机可读存储介质可以是计算机能够存取的任何可用介质或者是包含一个或多个可以用介质集成的服务器、数据中心等数据存储设备。可用介质可以是磁性介质(例如,软盘、硬盘、磁带),光介质(例如,DVD)、或者半导体介质(例如固态硬盘(solid state disk,SSD))等。The embodiments of the present application also provide a computer program product containing instructions, which when the instructions are run on a computer, cause the computer to execute any one of the methods in the foregoing embodiments. The computer program product includes one or more computer instructions. When the computer program instructions are loaded and executed on the computer, the processes or functions described in the embodiments of the present application are generated in whole or in part. The computer may be a general-purpose computer, a special-purpose computer, a computer network, or other programmable devices. Computer instructions may be stored in a computer-readable storage medium, or transmitted from one computer-readable storage medium to another computer-readable storage medium. For example, computer instructions may be transmitted from a website, computer, server, or data center through a cable (such as Coaxial cable, optical fiber, digital subscriber line (digital subscriber line, DSL) or wireless (such as infrared, wireless, microwave, etc.) transmission to another website site, computer, server, or data center. The computer-readable storage medium may be any available medium that can be accessed by a computer or may include one or more data storage devices such as a server or a data center that can be integrated with the medium. The usable medium may be a magnetic medium (for example, a floppy disk, a hard disk, and a magnetic tape), an optical medium (for example, a DVD), or a semiconductor medium (for example, a solid state disk (SSD)).
应注意,本申请实施例提供的上述用于存储计算机指令或者计算机程序的器件,例如但不限于,上述存储器、计算机可读存储介质和通信芯片等,均具有非易失性(non-transitory)。It should be noted that the foregoing devices for storing computer instructions or computer programs provided in the embodiments of the present application, such as but not limited to, the foregoing memory, computer-readable storage medium, and communication chip, are non-transitory. .
在实施所要求保护的本申请过程中,本领域技术人员通过查看附图、公开内容、以及所附权利要求书,可理解并实现公开实施例的其他变化。在权利要求中,“包括”(comprising)一词不排除其他组成部分或步骤,“一”或“一个”不排除多个的情况。单个处理器或其他单元可以实现权利要求中列举的若干项功能。相互不同的从属权利要求中记载了某些措施,但这并不表示这些措施不能组合起来产生良好的效果。In the process of implementing the claimed application, those skilled in the art can understand and implement other changes in the disclosed embodiments by viewing the drawings, the disclosure, and the appended claims. In the claims, the word "comprising" does not exclude other components or steps, and "a" or "one" does not exclude multiple. A single processor or other unit may implement several functions listed in the claims. Certain measures are described in mutually different dependent claims, but this does not mean that these measures cannot be combined to produce good results.
尽管结合具体特征及其实施例对本申请进行了描述,在不脱离本申请的精神和范围的情况下,可对其进行各种修改和组合。相应地,本说明书和附图仅仅是所附权利要求所界定的本申请的示例性说明,且视为已覆盖本申请范围内的任意和所有修改、变化、组合或等同物。Although the present application has been described with reference to specific features and embodiments, various modifications and combinations can be made without departing from the spirit and scope of the present application. Accordingly, this specification and drawings are merely exemplary descriptions of the application defined by the appended claims, and are deemed to have covered any and all modifications, changes, combinations or equivalents within the scope of the application.

Claims (20)

  1. 一种图像处理方法,应用于车载设备,其特征在于,所述方法包括:An image processing method applied to vehicle-mounted equipment, characterized in that the method includes:
    获取多帧图像,所述多帧图像包括所述车载设备所在车辆的周边道路的图像信息;Acquiring a multi-frame image, the multi-frame image including image information of a surrounding road of the vehicle where the on-board device is located;
    获取所述多帧图像的每帧图像中的第一图像区域;其中,所述多帧图像的多个第一图像区域对应于第一场景;Acquiring a first image area in each frame of the multi-frame image; wherein the multiple first image areas of the multi-frame image correspond to the first scene;
    对所述多个第一图像区域进行超分运算。Performing a hyperdivision operation on the plurality of first image regions.
  2. 根据权利要求1所述的方法,其特征在于,所述方法还包括:The method of claim 1, wherein the method further comprises:
    确定超分运算得到的图像区域中存在目标对象的图像信息。It is determined that the image information of the target object exists in the image area obtained by the super-division operation.
  3. 根据权利要求1或2所述的方法,其特征在于,对于所述多帧图像的每帧图像:The method according to claim 1 or 2, characterized in that, for each frame of the multi-frame image:
    所述第一图像区域是置信度低于或等于第一阈值的区域;The first image area is an area with a confidence level lower than or equal to a first threshold;
    或者,所述第一图像区域对应于所述车辆的可行驶区域中与所述车辆之间的距离大于或等于第二阈值的空间区域;Or, the first image area corresponds to a space area in the drivable area of the vehicle whose distance from the vehicle is greater than or equal to a second threshold;
    或者,所述第一图像区域是预设位置的区域。Alternatively, the first image area is an area of a preset position.
  4. 根据权利要求1至3任一项所述的方法,其特征在于,所述多帧图像包括第一图像和第二图像;The method according to any one of claims 1 to 3, wherein the multi-frame image includes a first image and a second image;
    所述获取所述多帧图像的每帧图像中的第一图像区域,包括:获取所述第二图像中的第一图像区域;The acquiring the first image area in each frame of the multi-frame image includes: acquiring the first image area in the second image;
    所述获取所述第二图像中的第一图像区域,包括:The acquiring the first image area in the second image includes:
    根据所述第一图像中的第一图像区域和所述车辆的第一车体信息,确定所述第二图像中的第一图像区域;其中,所述第一车体信息包括第一相对距离、第二相对距离和第一车辆转向角度中的至少一种,所述第一相对距离是拍摄所述第一图像时所述车辆与所述第一图像区域所对应的空间区域之间的相对距离,所述第二相对距离是拍摄所述第二图像时所述车辆与所述第一图像区域所对应的空间区域之间的相对距离,所述第一车辆转向角度是在所述第一图像和所述第二图像的拍摄时间间隔内,所述车辆的朝向之间的夹角。Determine the first image area in the second image according to the first image area in the first image and the first body information of the vehicle; wherein, the first vehicle body information includes a first relative distance , At least one of a second relative distance and a first vehicle steering angle, the first relative distance is the relative distance between the vehicle and the space area corresponding to the first image area when the first image is taken The second relative distance is the relative distance between the vehicle and the space area corresponding to the first image area when the second image is taken, and the first vehicle steering angle is the The angle between the direction of the vehicle in the time interval between the image and the second image.
  5. 根据权利要求1至4任一项所述的方法,其特征在于,所述对所述多个第一图像区域进行超分运算,包括:The method according to any one of claims 1 to 4, wherein the performing a hyperdivision operation on the plurality of first image regions comprises:
    对景物对齐后的所述多个第一图像区域进行超分运算。Performing a hyperdivision operation on the plurality of first image regions after scene alignment.
  6. 根据权利要求5所述的方法,其特征在于,所述多帧图像包括第三图像和第四图像;在所述对景物对齐后的所述多个第一图像区域进行超分运算之前,所述方法还包括:The method according to claim 5, wherein the multi-frame image includes a third image and a fourth image; before the superdivision operation is performed on the plurality of first image regions after the scene is aligned, the The method also includes:
    根据所述车辆的第二车体信息,执行所述多个第一图像区域的景物对齐;其中,所述第二车体信息包括第一相对角度、第二相对角度和第二车辆转向角度中的至少一种;其中,所述第一相对角度是拍摄所述第三图像时所述车辆与所述第一图像区域所对应的空间区域之间的相对角度,所述第二相对角度是拍摄所述第四图像时所述车辆与所述第一图像区域所对应的空间区域之间的相对角度,所述第二车辆转向角度是在所述第三图像和所述第四图像的拍摄时间间隔内,所述车辆的朝向之间的夹角。According to the second vehicle body information of the vehicle, perform scene alignment of the multiple first image areas; wherein the second vehicle body information includes the first relative angle, the second relative angle, and the second vehicle steering angle. At least one of; wherein the first relative angle is the relative angle between the vehicle and the space area corresponding to the first image area when the third image is taken, and the second relative angle is the The fourth image is the relative angle between the vehicle and the space area corresponding to the first image area, and the second vehicle steering angle is at the shooting time of the third image and the fourth image In the interval, the angle between the directions of the vehicles.
  7. 根据权利要求1至6任一项所述的方法,其特征在于,所述多帧图像是时序上连续的多帧图像。The method according to any one of claims 1 to 6, wherein the multi-frame images are consecutive multi-frame images in time series.
  8. 根据权利要求7所述的方法,其特征在于,所述多帧图像中的首帧图像的拍摄时刻与末帧图像的拍摄时刻之间的时间间隔小于或等于第三阈值。8. The method according to claim 7, wherein the time interval between the shooting moment of the first frame image and the shooting moment of the last frame image in the multi-frame images is less than or equal to a third threshold.
  9. 根据权利要求1至8任一项所述的方法,其特征在于,所述方法还包括:The method according to any one of claims 1 to 8, wherein the method further comprises:
    获取所述多帧图像的每帧图像中的第二图像区域;其中,所述多帧图像的多个第二图像区域对应于第二场景;Acquiring a second image area in each frame of the multi-frame image; wherein the multiple second image areas of the multi-frame image correspond to the second scene;
    对所述多个第二图像区域进行超分运算。Performing a hyperdivision operation on the plurality of second image regions.
  10. 一种车载设备,其特征在于,所述车载设备包括:A vehicle-mounted device, characterized in that, the vehicle-mounted device includes:
    第一获取模块,用于获取多帧图像,所述多帧图像包括所述车载设备所在车辆的周边道路的图像信息;The first acquisition module is configured to acquire multi-frame images, the multi-frame images including image information of the surrounding roads of the vehicle where the on-board equipment is located;
    第二获取模块,用于获取所述多帧图像的每帧图像中的第一图像区域;其中,所述多帧图像的多个第一图像区域对应于第一场景;The second acquisition module is configured to acquire the first image area in each frame of the multi-frame image; wherein the multiple first image areas of the multi-frame image correspond to the first scene;
    超分模块,用于对所述多个第一图像区域进行超分运算。The super-division module is used to perform super-division operations on the multiple first image regions.
  11. 根据权利要求10所述的车载设备,其特征在于,所述车载设备还包括:The vehicle-mounted device according to claim 10, wherein the vehicle-mounted device further comprises:
    确定模块,用于确定超分运算得到的图像区域中存在目标对象的图像信息。The determining module is used to determine that the image information of the target object exists in the image area obtained by the hyperdivision operation.
  12. 根据权利要求10或11所述的车载设备,其特征在于,对于所述多帧图像的每帧图像:The vehicle-mounted device according to claim 10 or 11, wherein, for each frame of the multi-frame image:
    所述第一图像区域是置信度低于或等于第一阈值的区域;The first image area is an area with a confidence level lower than or equal to a first threshold;
    或者,所述第一图像区域对应于所述车辆的可行驶区域中与所述车辆之间的距离大于或等于第二阈值的空间区域;Or, the first image area corresponds to a space area in the drivable area of the vehicle whose distance from the vehicle is greater than or equal to a second threshold;
    或者,所述第一图像区域是预设位置的区域。Alternatively, the first image area is an area of a preset position.
  13. 根据权利要求10至12任一项所述的车载设备,其特征在于,所述多帧图像包括第一图像和第二图像;The vehicle-mounted device according to any one of claims 10 to 12, wherein the multi-frame image includes a first image and a second image;
    所述第二获取模块具体用于:获取所述第二图像中的第一图像区域;所述获取所述第二图像中的第一图像区域,包括:The second acquiring module is specifically configured to: acquire the first image area in the second image; the acquiring the first image area in the second image includes:
    根据所述第一图像中的第一图像区域和所述车辆的第一车体信息,确定所述第二图像中的第一图像区域;其中,所述第一车体信息包括第一相对距离、第二相对距离和第一车辆转向角度中的至少一种,所述第一相对距离是拍摄所述第一图像时所述车辆与所述第一图像区域所对应的空间区域之间的相对距离,所述第二相对距离是拍摄所述第二图像时所述车辆与所述第一图像区域所对应的空间区域之间的相对距离,所述第一车辆转向角度是在所述第一图像和所述第二图像的拍摄时间间隔内,所述车辆的朝向之间的夹角。Determine the first image area in the second image according to the first image area in the first image and the first body information of the vehicle; wherein, the first vehicle body information includes a first relative distance , At least one of a second relative distance and a first vehicle steering angle, the first relative distance being the relative distance between the vehicle and the space area corresponding to the first image area when the first image is taken The second relative distance is the relative distance between the vehicle and the space area corresponding to the first image area when the second image is taken, and the first vehicle steering angle is the The angle between the direction of the vehicle in the time interval between the image and the second image.
  14. 根据权利要求10至13任一项所述的车载设备,其特征在于,The vehicle-mounted equipment according to any one of claims 10 to 13, wherein:
    所述超分模块具体用于:对景物对齐后的所述多个第一图像区域进行超分运算。The super-division module is specifically configured to perform a super-division operation on the plurality of first image regions after scene alignment.
  15. 根据权利要求14所述的车载设备,其特征在于,所述多帧图像包括第三图像和第四图像;所述车载设备还包括:The vehicle-mounted device according to claim 14, wherein the multi-frame image includes a third image and a fourth image; the vehicle-mounted device further includes:
    对齐模块,用于根据所述车辆的第二车体信息,执行所述多个第一图像区域的景物对齐;其中,所述第二车体信息包括第一相对角度、第二相对角度和第二车辆转向角度中的至少一种;其中,所述第一相对角度是拍摄所述第三图像时所述车辆与所述第一图像区域所对应的空间区域之间的相对角度,所述第二相对角度是拍摄所述第四 图像时所述车辆与所述第一图像区域所对应的空间区域之间的相对角度,所述第二车辆转向角度是在所述第三图像和所述第四图像的拍摄时间间隔内,所述车辆的朝向之间的夹角。The alignment module is configured to perform scene alignment of the multiple first image areas according to the second body information of the vehicle; wherein the second body information includes a first relative angle, a second relative angle, and a first relative angle. At least one of two vehicle steering angles; wherein, the first relative angle is the relative angle between the vehicle and the space area corresponding to the first image area when the third image is taken, and the first The second relative angle is the relative angle between the vehicle and the space area corresponding to the first image area when the fourth image is taken, and the second vehicle steering angle is between the third image and the first image area. The angle between the directions of the vehicles in the shooting time interval of the four images.
  16. 根据权利要求10至15任一项所述的车载设备,其特征在于,所述多帧图像是时序上连续的多帧图像。The vehicle-mounted device according to any one of claims 10 to 15, wherein the multi-frame image is a continuous multi-frame image in time series.
  17. 根据权利要求16所述的车载设备,其特征在于,所述多帧图像中的首帧图像的拍摄时刻与末帧图像的拍摄时刻之间的时间间隔小于或等于第三阈值。The vehicle-mounted device according to claim 16, wherein the time interval between the shooting moment of the first frame image and the shooting moment of the last frame image in the multi-frame images is less than or equal to a third threshold.
  18. 根据权利要求10至17任一项所述的车载设备,其特征在于,The in-vehicle device according to any one of claims 10 to 17, wherein:
    所述第二获取模块还用于:获取所述多帧图像的每帧图像中的第二图像区域;其中,所述多帧图像的多个第二图像区域对应于第二场景;The second acquisition module is further configured to: acquire a second image area in each frame of the multi-frame image; wherein the multiple second image areas of the multi-frame image correspond to a second scene;
    所述超分模块还用于:对所述多个第二图像区域进行超分运算。The super-division module is further configured to perform super-division operations on the plurality of second image regions.
  19. 一种图像处理装置,其特征在于,包括:存储器和处理器;所述存储器用于存储计算机程序,所述处理器用于调用所述计算机程序,以执行权利要求1至9任一项所述的方法。An image processing device, characterized by comprising: a memory and a processor; the memory is used to store a computer program, and the processor is used to call the computer program to execute any one of claims 1 to 9 method.
  20. 一种计算机可读存储介质,其特征在于,所述计算机可读存储介质上存储有计算机程序,当所述计算机程序在计算机上运行时,使得所述计算机执行权利要求1至9任一项所述的方法。A computer-readable storage medium, characterized in that a computer program is stored on the computer-readable storage medium, and when the computer program runs on a computer, the computer executes any one of claims 1 to 9 The method described.
PCT/CN2020/101717 2019-07-12 2020-07-13 Image processing method and apparatus WO2021008500A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201910633070.3A CN112215748A (en) 2019-07-12 2019-07-12 Image processing method and device
CN201910633070.3 2019-07-12

Publications (1)

Publication Number Publication Date
WO2021008500A1 true WO2021008500A1 (en) 2021-01-21

Family

ID=74047262

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2020/101717 WO2021008500A1 (en) 2019-07-12 2020-07-13 Image processing method and apparatus

Country Status (2)

Country Link
CN (1) CN112215748A (en)
WO (1) WO2021008500A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113378969A (en) * 2021-06-28 2021-09-10 北京百度网讯科技有限公司 Fusion method, device, equipment and medium of target detection results

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104219488A (en) * 2013-05-31 2014-12-17 索尼公司 Method and device of generating target image as well as video monitoring system
CN106600536A (en) * 2016-12-14 2017-04-26 同观科技(深圳)有限公司 Video imager super-resolution reconstruction method and apparatus
US9691133B1 (en) * 2013-12-16 2017-06-27 Pixelworks, Inc. Noise reduction with multi-frame super resolution
CN107480772A (en) * 2017-08-08 2017-12-15 浙江大学 A kind of car plate super-resolution processing method and system based on deep learning
CN109118430A (en) * 2018-08-24 2019-01-01 深圳市商汤科技有限公司 Super-resolution image reconstruction method and device, electronic equipment and storage medium

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8755636B2 (en) * 2011-09-14 2014-06-17 Mediatek Inc. Method and apparatus of high-resolution image reconstruction based on multi-frame low-resolution images
CN106845478B (en) * 2016-12-30 2019-09-10 同观科技(深圳)有限公司 A kind of secondary licence plate recognition method and device of character confidence level

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104219488A (en) * 2013-05-31 2014-12-17 索尼公司 Method and device of generating target image as well as video monitoring system
US9691133B1 (en) * 2013-12-16 2017-06-27 Pixelworks, Inc. Noise reduction with multi-frame super resolution
CN106600536A (en) * 2016-12-14 2017-04-26 同观科技(深圳)有限公司 Video imager super-resolution reconstruction method and apparatus
CN107480772A (en) * 2017-08-08 2017-12-15 浙江大学 A kind of car plate super-resolution processing method and system based on deep learning
CN109118430A (en) * 2018-08-24 2019-01-01 深圳市商汤科技有限公司 Super-resolution image reconstruction method and device, electronic equipment and storage medium

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113378969A (en) * 2021-06-28 2021-09-10 北京百度网讯科技有限公司 Fusion method, device, equipment and medium of target detection results
CN113378969B (en) * 2021-06-28 2023-08-08 北京百度网讯科技有限公司 Fusion method, device, equipment and medium of target detection results

Also Published As

Publication number Publication date
CN112215748A (en) 2021-01-12

Similar Documents

Publication Publication Date Title
US10964054B2 (en) Method and device for positioning
KR102523886B1 (en) A method and a device for detecting small target
US10990836B2 (en) Method and apparatus for recognizing object, device, vehicle and medium
JP6900081B2 (en) Vehicle travel route planning programs, devices, systems, media and devices
US11430228B2 (en) Dynamic driving metric output generation using computer vision methods
WO2020119567A1 (en) Data processing method, apparatus, device and machine readable medium
CN110386142A (en) Pitch angle calibration method for automatic driving vehicle
CN112740268B (en) Target detection method and device
CN112419385B (en) 3D depth information estimation method and device and computer equipment
US11004233B1 (en) Intelligent vision-based detection and ranging system and method
WO2021253245A1 (en) Method and device for identifying vehicle lane changing tendency
US11443627B2 (en) Navigation system with parking space identification mechanism and method of operation thereof
Kocur et al. Detection of 3D bounding boxes of vehicles using perspective transformation for accurate speed measurement
US10936920B2 (en) Determining geographical map features with multi-sensor input
US20160335775A1 (en) Visual navigation method, visual navigation device and robot
US11482007B2 (en) Event-based vehicle pose estimation using monochromatic imaging
CN111967396A (en) Processing method, device and equipment for obstacle detection and storage medium
WO2023088486A1 (en) Lane line extraction method and apparatus, vehicle and storage medium
WO2023179027A1 (en) Road obstacle detection method and apparatus, and device and storage medium
CN115147328A (en) Three-dimensional target detection method and device
WO2021008500A1 (en) Image processing method and apparatus
WO2022166606A1 (en) Target detection method and apparatus
CN115187941A (en) Target detection positioning method, system, equipment and storage medium
WO2023184869A1 (en) Semantic map construction and localization method and apparatus for indoor parking lot
CN116452911A (en) Target detection model training method and device, target detection method and device

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20840183

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20840183

Country of ref document: EP

Kind code of ref document: A1