WO2023155324A1 - Image fusion method and apparatus, device and storage medium - Google Patents

Image fusion method and apparatus, device and storage medium Download PDF

Info

Publication number
WO2023155324A1
WO2023155324A1 PCT/CN2022/094865 CN2022094865W WO2023155324A1 WO 2023155324 A1 WO2023155324 A1 WO 2023155324A1 CN 2022094865 W CN2022094865 W CN 2022094865W WO 2023155324 A1 WO2023155324 A1 WO 2023155324A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
fusion
infrared
visible light
infrared image
Prior art date
Application number
PCT/CN2022/094865
Other languages
French (fr)
Chinese (zh)
Inventor
赵尧
于洪英
闫奇
顾建超
Original Assignee
烟台艾睿光电科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 烟台艾睿光电科技有限公司 filed Critical 烟台艾睿光电科技有限公司
Publication of WO2023155324A1 publication Critical patent/WO2023155324A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques

Definitions

  • the present invention relates to the technical field of image processing, in particular to an image fusion method and device, image processing equipment, and a computer-readable storage medium.
  • Images are mainly divided into visible light images and infrared images.
  • Visible light images are rich in high-frequency details, which can better reflect the overall details of the shooting scene.
  • the background of the environment becomes blurred; while the principle of infrared imaging is mainly to display the shape and outline of the object through the thermal radiation intensity of the object, which has a good adaptability to weather and light, especially for hidden heat source targets, such as camouflaged enemies, Military targets such as weapons have good detectability, but infrared imaging has problems such as blurred image details, insufficient texture, less high-frequency information of the scene, poor contrast, and low definition.
  • the visible light images and infrared images are fused in the application scene to obtain a comprehensive and accurate image description of the shooting scene, to achieve full use of information, and to improve system analysis and decision-making accuracy and reliability.
  • Image fusion methods that fuse visible light images and infrared images are mainly divided into three categories according to the complexity of information processing in the fusion process: pixel-level fusion, feature-level fusion, and decision-level fusion.
  • Pixel-level fusion is the process of operating the pixels of an image to obtain a fused image.
  • the advantage is that it retains more information contained in the original image.
  • the disadvantage is that it needs to traverse, analyze and calculate the image pixel information, and the amount of data calculation and complexity is large. , The real-time performance of the system is low.
  • Feature-level image fusion is to extract edge, shape, texture, pixel density and other feature information from the image to be fused, and then form a multi-dimensional vector space according to these extracted features, and then analyze and process the feature vector in the vector space to form The feature set of the image is then trained and the image to be fused is fused according to the training result.
  • the algorithm of artificial neural network is mostly used in feature-level image fusion, which has the advantages of fast processing speed and small amount of calculation; the disadvantage is that there is more information loss and higher requirements for the operating system.
  • Decision-level image fusion is to first perform feature extraction, target feature recognition, and decision-making classification on the image to be fused, establish a preliminary judgment for the same target, and then fuse the decision information of visible light images and infrared images according to the fusion rules in terms of credibility. Get the result of a joint judgment.
  • decision-level fusion methods mainly include fusion algorithms based on support vector machines, neural networks, evidential reasoning, Bayesian reasoning, and fuzzy integrals, which are complex and require higher operating systems.
  • the embodiment of the present invention provides an image fusion method, device, image processing equipment, and computer-based image fusion method that can reduce invalid information in the image, reduce the amount of calculation and complexity, and improve the real-time performance of the system.
  • Read storage media Read storage media.
  • an image fusion method is provided, which is applied to an image processing device, including:
  • the infrared image is fused with the visible light image based on the target fusion area to obtain a fused image.
  • the fusion of the infrared image based on the target fusion area and the visible light image to obtain a fusion image includes:
  • the fused image of the brightness channel and the visible light image are fused to obtain a fused image.
  • the binarization of the infrared image to obtain a mask image, and determining the target fusion area according to the mask image includes:
  • the gray value of the pixel whose gray value is less than the binarization threshold is set to the first set value, and the gray value is greater than or
  • the grayscale value of the pixel point equal to the binarization threshold is set to a second set value to obtain a mask image;
  • the binarization threshold is determined according to the binarization strategy.
  • the matching binarization strategy is determined, including:
  • the gray histogram of the infrared image According to the distribution characteristics of the gray histogram of the infrared image, it is judged whether the gray histogram is in a unimodal distribution;
  • the matching binarization strategy is the Gaussian method or the Otsu method.
  • the matching binarization strategy is determined, including:
  • the determination of the binarization threshold according to the binarization strategy is specifically:
  • the maximum straight-line distance is determined through the triangle, and the binarization threshold is determined according to the gray level of the histogram corresponding to the maximum straight-line distance.
  • said according to the distribution characteristics of the grayscale histogram and average gradient of said infrared image, determine the binarization strategy of matching also include:
  • the matching binarization strategy is the Gaussian method
  • the determining the binarization threshold according to the binarization strategy specifically includes: calculating a Gaussian mean of the grayscale values of the infrared image within the target window function, and determining the binarization threshold according to the Gaussian mean.
  • said according to the distribution characteristics of the grayscale histogram and average gradient of said infrared image, determine the binarization strategy of matching also include:
  • the matching binarization strategy is the Otsu method
  • the determination of the binarization threshold according to the binarization strategy is specifically:
  • a binarization threshold is determined according to the inter-class variance value of the foreground image and the background image.
  • the channel separation of the infrared image and the visible light image is performed, and the separated brightness channel component representing the brightness of the image is fused according to the target fusion area to obtain a brightness channel fusion image, including:
  • the infrared image and the visible light image are respectively subjected to HSI channel separation, and the two separated I channel components are fused according to the Poisson image editing principle according to the target fusion area to obtain an I channel fusion image;
  • the merging of the luminance channel fused image and the visible light image to obtain a fused image includes:
  • an image fusion device including:
  • An acquisition module configured to acquire visible light images and infrared images synchronously collected for the target field of view
  • a fusion area determination module configured to binarize the infrared image to obtain a mask image, and determine a target fusion area according to the mask image
  • a fusion module configured to fuse the infrared image with the visible light image based on the target fusion area to obtain a fusion image.
  • the fusion module is specifically used to separate the channels of the infrared image and the visible light image, and fuse the separated brightness channel component representing the brightness of the image according to the target fusion area to obtain a brightness channel fusion image;
  • the fused image of the brightness channel and the visible light image are fused to obtain a fused image.
  • the fusion area determination module is specifically used to compare the gray value of each pixel in the infrared image with the binarization threshold, and the gray value of the pixel whose gray value is smaller than the binarization threshold Set the first set value, and set the gray value of the pixel point whose gray value is greater than or equal to the binarization threshold to the second set value to obtain a mask image; select the second set value in the mask image At least a part of the fixed-value pixel point distribution area is used as the target fusion area.
  • the fusion region determination module is further configured to determine a matching binarization strategy according to the gray histogram of the infrared image and the distribution characteristics of the average gradient; determine the binarization threshold according to the binarization strategy.
  • the fusion area determination module is also used to judge whether the gray histogram is a unimodal distribution according to the distribution characteristics of the gray histogram of the infrared image; if so, determine that the matching binarization strategy is a triangle If not, according to the comparison result of the average gradient of the infrared image and the average gradient of the visible light image, it is determined that the matching binarization strategy is the Gaussian method or the Otsu method.
  • the fusion region determination module is also used to determine the difference between the mode of the gray value of the infrared image and the average value of the gray value, if the difference is less than or equal to a preset value, determine the infrared
  • the gray histogram of the image is a unimodal distribution; the largest peak in the gray histogram is used as the apex to determine the triangle; the maximum straight-line distance is determined by the triangle, and the two are determined according to the gray level of the histogram corresponding to the maximum straight-line distance. Value threshold.
  • the fusion region determination module is further configured to determine the first average gradient of the infrared image and the second average gradient of the visible light image if the difference is greater than the preset value; if the second The average gradient is greater than or equal to the first average gradient, the Gaussian mean value of the gray value of the infrared image within the target window function is calculated, and the binarization threshold is determined according to the Gaussian mean value.
  • the fusion region determination module is further configured to segment the infrared image into a foreground image and a background image if the second average gradient is smaller than the first average gradient; according to the foreground image and the background image The between-class variance value of , and determine the binarization threshold.
  • the fusion module is also used to separately perform HSI channel separation on the infrared image and the visible light image, and fuse the separated two I-channel components according to the Poisson image editing principle according to the target fusion area, Obtain the I channel fused image; merge the I channel component of the I channel fused image with the H and S channel components separated from the visible light image to obtain a fused reference image; convert the fused reference image to RGB color space to obtain Blend images.
  • an image processing device including a processor, a memory connected to the processor, and a computer program stored in the memory and executable by the processor, the computer program being executed by the When executed by the processor, the steps of the image fusion method described in any embodiment of the present application are realized.
  • a computer-readable storage medium is provided, and a computer program is stored on the computer-readable storage medium.
  • the steps of the image fusion method described in any embodiment of the present application are implemented. .
  • the mask image is obtained by binarizing the infrared image
  • the target fusion area is determined according to the mask image
  • the infrared image is fused with the visible light image based on the target fusion area to obtain a fusion image
  • the part of the effective information contained in the infrared image can be extracted, and the extracted effective information can be fused with the visible light image, which can effectively prevent the fused image from containing useless information and reduce the image quality. , reducing invalid information in the image, reducing the amount of calculation and complexity, and improving the real-time performance of the system.
  • the image fusion device, image processing equipment, and computer-readable storage medium belong to the same concept as the corresponding image fusion method embodiments, and thus have the same technical effects as the corresponding image fusion method embodiments, and are not repeated here. repeat.
  • Fig. 1 is a schematic diagram of an application scene of an image fusion method in an embodiment
  • Fig. 2 is a flowchart of an image fusion method in an embodiment
  • Fig. 3 is the flowchart of image fusion method in another embodiment
  • Fig. 4 is the flowchart of image fusion method in another embodiment
  • Fig. 5 is a schematic diagram of a grayscale histogram of an example mid-infrared image
  • Fig. 6 is a schematic diagram of grayscale histogram data in an example in a unimodal distribution
  • Figure 7 is a schematic diagram of the comparison of the gray histogram of the infrared image showing a unimodal distribution, using the triangle method, the Gaussian method and the Otsu method for fusion;
  • Figure 8 is a schematic diagram of the comparison of the gray histogram of the infrared image, which is roughly uniformly distributed, and which is fused using the triangle method, the Gaussian method, and the Otsu method;
  • Figure 9 is a schematic diagram of the comparison of the gray histogram of the infrared image showing a bimodal distribution, using the triangle method, the Gaussian method and the Otsu method for fusion;
  • FIG. 10 is a flowchart of an image fusion method in an optional specific example
  • Fig. 11 is a schematic diagram of an infrared image used in the embodiment shown in Fig. 10;
  • Fig. 12 is a schematic diagram of a visible light image used in the embodiment shown in Fig. 10;
  • FIG. 13 is a schematic diagram of a fusion image obtained after fusion of an infrared image and a visible light image using the image fusion method described in the present application;
  • FIG. 14 is a schematic diagram of a fusion image obtained by fusing an infrared image and a visible light image using a known low-rank representation principle
  • FIG. 15 is a schematic diagram of a fusion image obtained by fusing an infrared image and a visible light image using a known non-subsampling shearlet transform principle
  • FIG. 16 is a schematic diagram of a fusion image obtained by fusing an infrared image and a visible light image using a known non-subsampling contourlet transform principle
  • Fig. 17 is a schematic diagram of a fusion image obtained by fusing an infrared image and a visible light image using a known Poisson image editing principle
  • Fig. 18 is a schematic diagram of an image fusion device in an embodiment
  • Fig. 19 is a schematic structural diagram of an image processing device in an embodiment.
  • first, second, and third are only used to distinguish similar objects, and do not represent a specific ordering of objects. Understandably, “first, second, and third” are used in Where permitted, the specific order or sequence may be interchanged such that the embodiments of the application described herein can be practiced in other sequences than illustrated or described herein.
  • FIG. 1 is a schematic diagram of an optional application scenario of the image processing method provided by the embodiment of the present application, wherein the image processing device 11 includes a processor 12, a memory 13 connected to the processor 12, and a visible light shooting module 14 And infrared camera module 15.
  • the image processing device 11 includes a processor 12, a memory 13 connected to the processor 12, and a visible light shooting module 14 And infrared camera module 15.
  • the image processing device 11 collects visible light images and infrared images synchronously and in real time through the visible light shooting module 14 and the infrared shooting module 14 and sends them to the processor 12, and the memory 13 stores the images provided by the embodiments of the present application
  • the computer program of the fusion method the processor 12 executes the computer program, binarizes the infrared image to obtain a mask image, determines the target fusion area through the mask image, and performs the infrared image based on the target fusion area and the visible light image. fusion to obtain a fusion image.
  • the image processing device 11 can be various types of intelligent terminals integrated with the visible light shooting module 14 and the infrared shooting module 15, and have storage and processing functions, such as security monitoring equipment, vehicle-mounted equipment, etc.; the image processing device 11 is also It can be a computer device connected to the visible light shooting module 14 and the infrared shooting module 15; the image processing device 11 can also be a dual-light fusion aiming device of white light and red light.
  • the image fusion method provided by an embodiment of the present application can be applied to the image processing device shown in Fig. 1 .
  • the image processing method includes the following steps:
  • the visible light image and the infrared image are acquired synchronously for the target field of view, so that the visible light image and the infrared image include the imaging of objects in the same target field of view.
  • the image processing device includes a visible light shooting module and an infrared shooting module, and the acquiring the visible light image and the infrared image synchronously collected for the target field of view includes: the image processing device simultaneously collects the visible light image and the infrared image through the visible light shooting module and the infrared shooting module image, and send the collected visible light image and infrared image to the processor.
  • the image processing device does not include an image capturing module, and the acquiring the visible light image and the infrared image synchronously collected for the target field of view includes: the image processing device acquires other images with the functions of capturing visible light images and infrared images Visible light images and infrared images synchronously collected for the target field of view sent by the smart device.
  • other smart devices may include infrared detectors, mobile terminals, and the cloud.
  • the binarization of the infrared image refers to assigning the gray value of each pixel on the infrared image to obtain a binarized image that can reflect the overall and local features of the image.
  • the step S103 is to binarize the infrared image to obtain a mask image, and determine the target fusion area according to the mask image, including:
  • the gray value of the pixel whose gray value is less than the binarization threshold is set to the first set value, and the gray value is greater than or
  • the grayscale value of the pixel point equal to the binarization threshold is set to a second set value to obtain a mask image;
  • the binarization threshold can be preset, or can be calculated according to the distribution characteristics of the gray value of pixels in the infrared image.
  • the first set value and the second set value can be respectively selected from the maximum value and the minimum value of the gray value interval, or can also be two gray values close to the maximum value and the minimum value respectively in the gray value interval.
  • the first set value is 0, and the second set value is 255, so that the entire image presents a black and white image effect.
  • the said binarization of the infrared image to obtain the mask image includes: performing binarization on the grayscale images of 256 brightness levels in the infrared image through the binarization threshold, and converting the grayscale value of each pixel in the infrared image to Compared with the binarization threshold, the gray value of pixels whose gray value is smaller than the binarization threshold is set to 0, and the gray value of pixels whose gray value is greater than the binarization threshold is set to 255, thus obtaining A binary image that reflects the overall and local features of the image, that is, the mask image.
  • determining the target fusion area according to the mask image may be based on the pixel point distribution area of the second set value in the mask image, that is, the white part determines the target fusion area, such as the white part in the mask image may be selected All as the target fusion area, or select a certain part of the white part in the mask image as the target fusion area.
  • Fusing the infrared image based on the target fusion area with the visible light image to obtain a fusion image may refer to merging the infrared image and the visible light image with image parts corresponding to the target fusion area, respectively, Other parts retain the image part of the visible light image to obtain a fused image; or, extract the target fusion area of the infrared image to form an image to be fused, and fuse the image to be fused with the visible light image, etc.
  • the mask image is obtained by binarizing the infrared image
  • the target fusion area is determined according to the mask image
  • the infrared image is fused with the visible light image based on the target fusion area to obtain a fusion image.
  • the determination of the area can extract the effective information contained in the infrared image, and fuse the extracted effective information with the visible light image, which can effectively avoid the image quality degradation caused by the useless information contained in the fused image, and reduce the invalid information in the image , reduce the amount of calculation and complexity, and can improve the real-time performance of the system.
  • the infrared image is fused based on the target fusion area with the visible light image to obtain a fused image, including:
  • a digital image it is a picture observed by the human eye, but from the perspective of a computer, a digital image is a bunch of points with different brightness.
  • a digital image with a size of M ⁇ N can be represented by a M ⁇ N matrix, the values of the elements in the matrix respectively represent the brightness of the corresponding pixel at this position, and the larger the pixel value, the brighter the pixel.
  • the grayscale image can be represented by a two-dimensional matrix
  • the color image can be represented by a three-dimensional matrix (M ⁇ N ⁇ 3), that is, a multi-channel image.
  • the hue and color of the image can be changed through the channel. For example, if only the red channel is saved, the image itself only retains red elements and information.
  • For each single channel it can be displayed as a pair of grayscale images (it should be noted that the grayscale image is not a black and white image), and the lightness and darkness in the grayscale image of a single channel correspond to the lightness and darkness of the color of the single channel, correspondingly representing The distribution of the color/light of the single channel on the image.
  • Channel-separating the infrared image and the visible light image, and merging the separated luminance channel component representing the brightness of the image according to the target fusion area to obtain the luminance channel fusion image can refer to channel separation of the infrared image and the visible light image, and the infrared image
  • the part corresponding to the target fusion area in the luminance channel component representing the image luminance separated in is fused with the part corresponding to the target fusion area in the luminance channel component representing image luminance separated from the visible light image, and the luminance channel fusion image is obtained .
  • the luminance channel components separated from the infrared image and the visible light image are fused according to the target fusion area, which can reduce the amount of calculation required for fusion, and can retain effective information in the target fusion area.
  • Image fusion refers to the image data of the same target collected by multi-source channels through image processing technology to maximize the extraction of beneficial information in each channel, and finally synthesize high-quality images to improve image information. Improve the utilization rate of computer, improve the accuracy and reliability of computer interpretation, and improve the spatial resolution and spectral resolution of the original image, which is beneficial to monitoring.
  • the brightness channel fusion image contains the effective information in the brightness channel component of the visible light image and the brightness channel component of the infrared image, and fuses the brightness channel fusion image and the visible light image, so that the brightness channel component in the brightness channel fusion image and other channels of the visible light image The components are combined to obtain a fused image.
  • the mask image is obtained by binarizing the infrared image
  • the target fusion area is determined according to the mask image
  • the channels of the infrared image and the visible light image are separated
  • the separated luminance channel component representing the brightness of the image is fused according to the target
  • the region is fused to obtain a luminance channel fusion image
  • the luminance channel fusion image and the visible light image are fused to obtain a fusion image.
  • the luminance channel components separated from the infrared image and the visible light image are fused, and then fused with the visible light image, which can effectively prevent the fused image from containing useless information and cause image quality degradation, reduce invalid information in the image, and reduce the amount of calculation and complexity.
  • the fused image can retain the respective advantages of visible light images and infrared images at the same time, whether it is for the fused image obtained after imaging under sufficient lighting conditions or for imaging under poor lighting conditions
  • the target can be better highlighted, ensuring that the target in the image is more clearly presented, and it is easier for the human eye to observe and identify.
  • the S103 binarize the infrared image to obtain a mask image, and in the step of determining the target fusion area according to the mask image, in the step of converting the infrared image Before comparing the gray value of each pixel with the binarization threshold, it includes:
  • the principles of binarization methods adopted by different binarization strategies are different, and the applicable target images are also different.
  • the distribution characteristics of the gray histogram and average gradient of the infrared image can be judged that the gray histogram of the infrared image is distributed with a single peak, a bimodal distribution or a roughly uniform distribution, so as to determine the matching binarization strategy.
  • the binarization strategy includes triangle method, Gaussian method, and Otsu method. If the gray histogram of the infrared image is distributed with a single peak, the triangle method is applied; if the gray histogram of the infrared image is more uniformly distributed, the Gaussian method is applied.
  • the mask image is obtained by binarizing the infrared image through the corresponding binarization strategy, and the area of the white part in the mask image is determined as the target fusion area.
  • the binarization strategy adapted to it, to ensure that the target in the image can be more accurately located after binarizing the infrared image.
  • the image area is binarized into the white part, so as to determine the target fusion area based on the mask image, and the effective information in the image can be more completely and comprehensively preserved after the brightness channel components separated from the visible light image and the infrared image are fused according to the target fusion area .
  • the S103 according to the gray histogram of the infrared image and the distribution characteristics of the average gradient, determine a matching binarization strategy, including:
  • the gray histogram of the infrared image According to the distribution characteristics of the gray histogram of the infrared image, it is judged whether the gray histogram is in a unimodal distribution;
  • the matching binarization strategy is the Gaussian method or the Otsu method.
  • the Gaussian method or Otsu method is selected according to the average gradient Law. Judging whether the binarization strategy for binarizing the current infrared image is suitable for the triangle method is to use the gray histogram data of the infrared image to judge whether it is a unimodal distribution, assuming the maximum peak of the gray histogram of the infrared image On the brightest side, use this to find the best threshold for infrared image binarization.
  • the comparison between the average gradient of the infrared image and the average gradient of the visible light image can be used to determine whether the distribution is relatively uniform or bimodal, so as to determine the applicable than Gauss method or Otsu method.
  • the binarization strategy is set to include triangle method, Gaussian method and Otsu method, and the distribution characteristics of the gray histogram and average gradient of the infrared image are analyzed to determine the binarization strategy adapted to it as triangular method, Gaussian method, and Gaussian method.
  • method or Otsu method to ensure that after binarizing the infrared image, the image area where the target in the image is located can be more accurately binarized into a white part, so that after the target fusion area is determined based on the mask image, the target fusion area can be aligned according to the target fusion area.
  • the luminance component fusion image obtained by fusing the luminance channel components separated from the visible light image and the infrared image, the effective information in the image can be more completely and comprehensively preserved.
  • said S1031 according to the distribution characteristics of the grayscale histogram and the average gradient of the infrared image, determine a matching binarization strategy, including:
  • the maximum straight-line distance is determined through the triangle, and the binarization threshold is determined according to the gray level of the histogram corresponding to the maximum straight-line distance.
  • the difference between the grayscale value mode and the grayscale average value corresponding to the infrared image is relative to The gray histogram is much smaller than other forms of infrared images.
  • the absolute value of the difference between the mode of the gray value and the average value of the gray value can be recorded as A-M (average-mode).
  • the difference A-M is less than the preset value, it means that the corresponding infrared image
  • the gray histogram of the infrared image shows a unimodal distribution, otherwise, it means that the gray histogram corresponding to the infrared image does not meet the characteristics of a unimodal distribution.
  • use the gray histogram data of the infrared image to find the optimal binarization threshold based on a pure geometric method, assuming that the largest peak in the gray histogram is on the brightest side, and find the largest straight line through the triangle Determine the gray level of the histogram corresponding to the maximum straight-line distance as the segmentation threshold.
  • the mode and average value of the gray value in the gray histogram corresponding to the infrared image are 121 and 127.734 respectively, where the mode of the gray value is the gray value with the most repetitions among the gray values , the calculation formula of the gray value AG is as follows: Formula 1:
  • H(i, j) represents the gray value of the pixel point whose coordinates are (i, j), M represents the maximum value of the abscissa, and N represents the maximum value of the ordinate.
  • the preset value is 10
  • the difference A-M between the mode of the gray value and the average value of the gray value is less than 10
  • the infrared image will be binarized using the triangle method, with the largest peak in the gray histogram as the apex A triangle is determined, a maximum straight-line distance is determined through the triangle, and a binarization threshold is determined according to the gray level of the histogram corresponding to the maximum straight-line distance.
  • the gray histogram of the infrared image shows a unimodal distribution.
  • the triangle method, Gaussian method and Otsu method are used as the corresponding binarization strategies to binarize the infrared image to obtain the mask image, and the visible light image Schematic diagram of fusion comparison, in which, after using the triangle method as the binarization strategy to determine the binarization threshold, the mask image obtained by binarizing the infrared image can highlight the image target more comprehensively and completely, and finally with The loss of effective image information in the triangular method fused image obtained after visible light image fusion is the smallest.
  • the step S1031 according to the distribution characteristics of the grayscale histogram and the average gradient of the infrared image, determining a matching binarization strategy, further includes:
  • the second average gradient is greater than or equal to the first average gradient, calculate the Gaussian mean value of the gray value of the infrared image within the target window function, and determine the binarization threshold according to the Gaussian mean value.
  • the difference A-M between the mode of the gray value and the average value of the gray value is greater than the preset value, it means that the gray histogram corresponding to the infrared image does not meet the characteristics of a unimodal distribution.
  • the relative size of the average gradient of the image is used to determine whether the gray histogram corresponding to the infrared image is more uniformly distributed or bimodal.
  • the average gradient of the infrared image is referred to as the first average gradient
  • the average gradient of the visible light image is referred to as the second average gradient.
  • the calculation formula of the average gradient of the image can be as follows formula two:
  • H(i, j) represents the gray value of the pixel point whose coordinates are (i, j), M represents the maximum value of the abscissa, and N represents the maximum value of the ordinate.
  • M represents the maximum value of the abscissa
  • N represents the maximum value of the ordinate.
  • the part to be fused and the part that does not need to be fused in the corresponding window function are determined, so that the gray value in the gray histogram is not concentrated in
  • the infrared image of a certain indicator that is, the infrared image whose gray value is concentrated on multiple indicators, can binarize the image areas where the targets corresponding to the multiple indicators in the infrared image are respectively located into white in the mask image.
  • the luminance component fusion image obtained by fusing the luminance channel components separated from the visible light image and the infrared image according to the target fusion area can retain the image more completely and comprehensively.
  • the effective information in the fused image is strengthened and the edges are highlighted.
  • the relative size of the average gradient of the visible light image and the average gradient of the infrared image is used to determine whether the gray value is distributed approximately uniformly, so as to quickly and accurately determine whether the binarization strategy of the infrared image conforms to Gaussian
  • the visible light image is clear and contains most of the effective information, so the Gaussian method is used to obtain the binarized image to ensure that the infrared image can be binarized more accurately
  • the image area where each target in the image is located is binarized into a white part, and after the target fusion area is determined based on the mask image, the luminance component fusion image obtained by fusing the luminance channel components separated from the visible light image and the infrared image according to the target fusion area
  • the effective information in the image can be preserved more completely and comprehensively.
  • the gray histogram of the infrared image is roughly evenly distributed.
  • the triangle method, Gaussian method, and Otsu method are used as the corresponding binarization strategies to binarize the infrared image to obtain the mask image.
  • Schematic diagram of the comparison of image fusion in which, the mask image obtained by binarizing the infrared image after using the Gaussian method as the binarization strategy to determine the binarization threshold, the target outline is clearer and more prominent, and finally fused with the visible light image
  • the Gaussian method fusion image obtained after the loss of image effective information is the smallest.
  • the determining a matching binarization strategy according to the distribution characteristics of the grayscale histogram and the average gradient of the infrared image further includes:
  • a binarization threshold is determined according to the inter-class variance value of the foreground image and the background image.
  • the difference A-M between the mode of the gray value and the average value of the gray value is greater than the preset value, it means that the gray histogram corresponding to the infrared image does not meet the characteristics of a unimodal distribution.
  • the relative size of the average gradient of the image is used to determine whether the gray histogram corresponding to the infrared image is more evenly distributed or bimodal.
  • the average gradient of the infrared image is called the first average gradient
  • the average gradient of the visible light image is called the second average gradient.
  • the binarization strategy applicable to the current infrared image is the Otsu method.
  • the principle of determining the binarization threshold by the Otsu method is to divide the image into two parts, the background and the target, according to the grayscale characteristics of the image. The greater the inter-class variance between the background and the target, the greater the difference between the two parts that make up the image. When part of the target is misclassified into the background or part of the background is misclassified into the target, the difference between the two parts will become smaller.
  • the binarization threshold segmentation based on the maximum variance between classes can minimize the probability of misclassification.
  • the relative size of the average gradient of the visible light image and the average gradient of the infrared image is used to judge whether the gray histogram is bimodal, so as to quickly and accurately judge whether the binarization strategy of the infrared image conforms to the Otsu method , the infrared image is divided into a foreground image and a background image by the Otsu method, wherein the foreground image contains main information, such as the energy radiated outward by other heat-generating target objects such as people, and the pixels in the area including the main information in the foreground image and the The difference between the gray values of adjacent pixels is large, so for the case where the average gradient of the infrared image is greater than the average gradient of the visible light image, it is more suitable to determine the area containing effective information according to the foreground image of the infrared image, and obtain the infrared image by the Otsu method.
  • the visible light is processed according to the target fusion area.
  • the luminance component fusion image obtained by fusing the luminance channel components separated from the image and the infrared image, the effective information in the image can be retained more completely and comprehensively.
  • the gray histogram of the infrared image shows a bimodal distribution.
  • the triangle method, Gaussian method, and Otsu method are used as the corresponding binarization strategies to binarize the infrared image to obtain the mask image, and then compare it with the visible light image. Schematic diagram of the fusion comparison.
  • the Otsu method is used as the binarization strategy to determine the binarization threshold
  • the mask image obtained by binarizing the infrared image has clearer and more prominent outlines of multiple targets, and is finally compared with the visible light image.
  • the loss of image effective information is the smallest in the fused image obtained by Otsu method after fusion.
  • the channel separation of the infrared image and the visible light image is carried out, and the separated brightness channel component representing the brightness of the image is fused according to the target fusion area to obtain a brightness channel fusion image, including:
  • the infrared image and the visible light image are respectively subjected to HSI channel separation, and the two separated I channel components are fused according to the Poisson image editing principle according to the target fusion area to obtain an I channel fusion image;
  • the merging of the luminance channel fused image and the visible light image to obtain a fused image includes:
  • HSI Human-Saturation-Intensity (Lightness)
  • H the color characteristics of the image with H, S, I three parameters, H defines the frequency of the color, called hue; S Indicates the depth of color, known as saturation; I indicates intensity or brightness.
  • HSI channel separation is performed on the infrared image and the visible light image respectively, and the two I-channel components separated from the target fusion regions corresponding to the infrared image and the visible light image are separated according to the Poisson image
  • the editing principle is used for fusion to obtain the I-channel fused image, which includes: separating the visible light image from the HSI channel, and separating the H-channel component, S-channel component, and I-channel component of the visible light image; according to the gray histogram and average gradient of the infrared image distribution characteristics, determine the matching binarization strategy, binarize the infrared image according to the binarization strategy to obtain a mask image, and determine the target fusion area according to the mask image; perform HSI channel separation on the infrared image , separate the H channel component, S channel component, and I channel component of the infrared image; fuse the I channel component of the infrared image and the I channel component of the visible light image according to the target fusion area determined by the mask image according
  • the infrared image and the visible light image are non-HSI format images
  • it before performing HSI channel separation on the infrared image and the visible light image, it also includes converting the infrared image and the visible light image into an infrared image and a visible light image in HSI format.
  • the obtained I channel fusion image and the visible light image are separated into H After the channel component and the S channel component are combined, they are converted to the RGB color space to obtain the fused image.
  • FIG. 10 Please refer to FIG. 12 , and use an optional example as an example to describe the image fusion method.
  • the adaptive threshold binarization strategy for infrared image matching to binarize the infrared image;
  • the adaptive threshold binarization strategy includes the triangle method, the Gaussian method, and the Otsu method;
  • the methods for selecting the adaptive binarization strategy include:
  • the region to be fused is determined according to the binarized image; by locking the region to be fused, the fusion calculation amount is reduced and the fusion processing efficiency is improved, and the Effective information in the image;
  • Figure 13 is the fused image obtained after using the image fusion method described in this application to fuse infrared images and visible light images
  • Figure 13 14 Fusion comparison effect of infrared image and visible light image fusion based on known low-rank representation principle
  • Figure 15 is fusion comparison of infrared image and visible light image fusion based on known non-subsampling shearlet transform principle Effect
  • Figure 16 shows the fusion and comparison effect of infrared images and visible light images based on the known principle of non-subsampling contourlet transformation. Fusion contrast effect for direct fusion.
  • IE Information Entropy
  • SF spatial Frequency
  • RMSE Root Mean Sqaured Error
  • SSIM Structuretural Similarity Index
  • TIME It refers to the fusion processing time.
  • NSST Non-subsampled Shearlet Transform
  • NSCT Nonsubsampled contourlet transform
  • an image fusion device including: an acquisition module 131, used to acquire visible light images and infrared images synchronously collected for the target field of view; a fusion area determination module 132, used for all Binarize the infrared image to obtain a mask image, and determine the target fusion area according to the mask image; the fusion module 134 is used to fuse the infrared image based on the target fusion area and the visible light image to obtain a fusion image .
  • the fusion module 134 is specifically configured to separate the channels of the infrared image and the visible light image, and fuse the separated brightness channel component representing the brightness of the image according to the target fusion area to obtain the brightness channel Fusing images: fusing the luminance channel fusion image and the visible light image to obtain a fusion image.
  • the fusion region determination module 132 is specifically configured to compare the grayscale value of each pixel in the infrared image with a binarization threshold, and the grayscale value of a pixel with a grayscale value smaller than the binarization threshold is The grayscale value is set to the first set value, and the grayscale value of the pixel point whose grayscale value is greater than or equal to the binarization threshold is set to the second set value to obtain a mask image; select the mask image described in At least a part of the pixel point distribution area of the second set value is used as the target fusion area.
  • the fusion region determination module 132 is further configured to determine a matching binarization strategy according to the distribution characteristics of the gray histogram and the average gradient of the infrared image; determine the binarization strategy according to the binarization strategy threshold.
  • the fusion region determination module 132 is further configured to judge whether the gray histogram is a unimodal distribution according to the distribution characteristics of the gray histogram of the infrared image; if so, determine the matching binarization The strategy is the triangle method; if not, according to the comparison result of the average gradient of the infrared image and the average gradient of the visible light image, it is determined that the matching binarization strategy is the Gaussian method or the Otsu method.
  • the fusion region determination module 132 is also used to determine the difference between the mode of the gray value of the infrared image and the average value of the gray value, if the difference is less than or equal to a preset value, determine The grayscale histogram of the infrared image is in a unimodal distribution; a triangle is determined with the largest peak in the grayscale histogram as the apex; the maximum straight-line distance is determined through the triangle, and the grayscale of the histogram corresponding to the maximum straight-line distance Level determines the binarization threshold.
  • the fusion region determining module 132 is further configured to determine the first average gradient of the infrared image and the second average gradient of the visible light image if the difference is greater than the preset value; The second average gradient is greater than or equal to the first average gradient, the Gaussian mean value of the gray value of the infrared image in the target window function is calculated, and the binarization threshold is determined according to the Gaussian mean value.
  • the fusion region determination module 132 is further configured to segment the infrared image into a foreground image and a background image if the second average gradient is smaller than the first average gradient; The inter-class variance value of the background image is used to determine the binarization threshold.
  • the fusion module 134 is further configured to separately perform HSI channel separation on the infrared image and the visible light image, and edit the two separated I-channel components according to the Poisson image editing principle according to the target fusion area.
  • Carrying out fusion to obtain an I channel fusion image merging the I channel component of the I channel fusion image with the H and S channel components separated from the visible light image to obtain a fusion reference image; converting the fusion reference image to RGB color space to obtain a fused image.
  • the division of the above-mentioned program modules is used as an example for illustration. In practical applications, the above-mentioned processing can be allocated according to needs. Completed by different program modules, that is, the internal structure of the device can be divided into different program modules to complete all or part of the method steps described above.
  • the image fusion device and the image fusion method embodiments provided in the above embodiments belong to the same idea, and the specific implementation process thereof is detailed in the method embodiments, and will not be repeated here.
  • FIG. 19 is a schematic diagram of an optional hardware structure of the image processing device provided by the embodiment of the present application.
  • the image processing device includes a processor 111, and the processing
  • the memory 112 connected to the device 111, the memory 112 is used to store various types of data to support the operation of the image processing equipment, and stores a computer program for realizing the image processing method provided by any embodiment of the present application, the computer program When executed by the processor, the steps of the image processing method provided by any embodiment of the present application can be realized, and the same technical effect can be achieved. To avoid repetition, details are not repeated here.
  • the image processing device further includes an infrared shooting module and a visible light shooting module connected to the processor 111, and the infrared shooting module and the visible light shooting module are used for synchronously shooting infrared images and visible light images for the same target field of view It is sent to the processor 111 as an image to be fused.
  • the embodiment of the present application also provides a computer-readable storage medium, where a computer program is stored on the computer-readable storage medium.
  • a computer program is stored on the computer-readable storage medium.
  • the computer-readable storage medium is, for example, a read-only memory (Read-Only Memory, ROM for short), a random access memory (Random Access Memory, RAM for short), a magnetic disk or an optical disk, and the like.
  • Nonvolatile memory can include read only memory (ROM), programmable ROM (PROM), electrically programmable ROM (EPROM), electrically erasable programmable ROM (EEPROM), or flash memory.
  • ROM read only memory
  • PROM programmable ROM
  • EPROM electrically programmable ROM
  • EEPROM electrically erasable programmable ROM
  • Volatile memory can include random access memory (RAM) or external cache memory.
  • RAM random access memory
  • RAM is available in many forms such as Static RAM (SRAM), Dynamic RAM (DRAM), Synchronous DRAM (SDRAM), Double Data Rate SDRAM (DDRSDRAM), Enhanced SDRAM (ESDRAM), Synchronous Chain Synchlink DRAM (SLDRAM), memory bus (Rambus) direct RAM (RDRAM), direct memory bus dynamic RAM (DRDRAM), and memory bus dynamic RAM (RDRAM), etc.

Landscapes

  • Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Image Processing (AREA)
  • Studio Devices (AREA)

Abstract

Disclosed in embodiments of the present invention are an image fusion method and apparatus, an image processing device and a computer-readable storage medium. The image fusion method comprises: acquiring a visible light image and an infrared image synchronously captured for a target field of view; binarizing the infrared image to obtain a mask image, and determining a target fusion area according to the mask image; and fusing the infrared image and the visible light image on the basis of the target fusion area to obtain a fused image. By means of the determination of the target fusion area, the valid information part contained in the infrared image an be extracted, and the extracted valid information and the visible light image are fused, so that the situation that image quality is reduced because the fused image contains useless information can be effectively avoided, invalid information in the image is reduced, the amount of calculation and complexity are reduced, and the real-time performance of a system can be improved.

Description

图像融合方法、装置及设备、存储介质Image fusion method, device and equipment, storage medium 技术领域technical field
本发明涉及图像处理技术领域,特别涉及一种图像融合方法、装置及图像处理设备、计算机可读存储介质。The present invention relates to the technical field of image processing, in particular to an image fusion method and device, image processing equipment, and a computer-readable storage medium.
背景技术Background technique
图像主要分为可见光图像和红外图像,可见光图像的高频细节丰富,能够较好地反映拍摄场景整体的细节特点,但在光照条件较差的情况下,图像质量降低,图像中需要检测的目标与环境背景变得模糊;而红外成像原理主要是通过物体的热辐射强度来显示物体的形状轮廓,对于天气、光照有很好的适应能力,特别是对于隐蔽的热源目标,如伪装的敌人、武器等军事目标具有很好的探测性,但红外成像存在图像细节模糊,纹理不够清晰,场景的高频信息比较少,对比度差,清晰度低等问题。Images are mainly divided into visible light images and infrared images. Visible light images are rich in high-frequency details, which can better reflect the overall details of the shooting scene. The background of the environment becomes blurred; while the principle of infrared imaging is mainly to display the shape and outline of the object through the thermal radiation intensity of the object, which has a good adaptability to weather and light, especially for hidden heat source targets, such as camouflaged enemies, Military targets such as weapons have good detectability, but infrared imaging has problems such as blurred image details, insufficient texture, less high-frequency information of the scene, poor contrast, and low definition.
为了能够兼顾可见光图像和红外图像的各自优点,在应用场景中将可见光图像和红外图像进行融合,以获得对拍摄场景全面准确的图像描述,达到对于信息的充分利用,同时也可以提高系统分析决策的准确性和可靠性。In order to take into account the respective advantages of visible light images and infrared images, the visible light images and infrared images are fused in the application scene to obtain a comprehensive and accurate image description of the shooting scene, to achieve full use of information, and to improve system analysis and decision-making accuracy and reliability.
将可见光图像和红外图像进行融合的图像融合方法按照融合过程中信息处理的复杂程度,主要分为三类:像素级融合、特征级融合、决策级融合。像素级融合是对图像的像素进行操作得到融合图像的过程,优点是对原始图像中所包含的信息保留较多,缺点是需要对图像像素信息进行遍历分析计算,数据计算量和复杂度较大,系统实时性低。特征级图像融合是先从待融合图像中提取边缘、形状、纹理、像素密度等特征信息,再根据提取到的这些特征构成多维向量空间,然后对向量空间中的特征向量进行分析和处理,形成图像的特征集合,然后进行训练并根据训练的结果对待融合图像进行融合。目前特征级图像融合多采用人工神经网络的算法,优点是处理速度快,计算量较小;缺点是信息丢失较多,对操作系统要求较高。决策级图像融合是先对待融合图像进行特征提取、目标特征识别和决策分类,建立对同一目标的初步判决,然后根据融合规则对可见光图像和红外图像的决策信息进行可信度上的融合,最终得到一个联合判决的结果。目前,决策级融合方法主要包括基于支持向量机、神经网络、证据推理、Bayes推理和模糊积分等融合算法,复杂度大,对操作系统的要求更高。Image fusion methods that fuse visible light images and infrared images are mainly divided into three categories according to the complexity of information processing in the fusion process: pixel-level fusion, feature-level fusion, and decision-level fusion. Pixel-level fusion is the process of operating the pixels of an image to obtain a fused image. The advantage is that it retains more information contained in the original image. The disadvantage is that it needs to traverse, analyze and calculate the image pixel information, and the amount of data calculation and complexity is large. , The real-time performance of the system is low. Feature-level image fusion is to extract edge, shape, texture, pixel density and other feature information from the image to be fused, and then form a multi-dimensional vector space according to these extracted features, and then analyze and process the feature vector in the vector space to form The feature set of the image is then trained and the image to be fused is fused according to the training result. At present, the algorithm of artificial neural network is mostly used in feature-level image fusion, which has the advantages of fast processing speed and small amount of calculation; the disadvantage is that there is more information loss and higher requirements for the operating system. Decision-level image fusion is to first perform feature extraction, target feature recognition, and decision-making classification on the image to be fused, establish a preliminary judgment for the same target, and then fuse the decision information of visible light images and infrared images according to the fusion rules in terms of credibility. Get the result of a joint judgment. At present, decision-level fusion methods mainly include fusion algorithms based on support vector machines, neural networks, evidential reasoning, Bayesian reasoning, and fuzzy integrals, which are complex and require higher operating systems.
技术问题technical problem
为了解决现有存在的技术问题,本发明实施例提供一种可减少图像中无效信息、减小计算量和复杂度、且可提升系统实时性的图像融合方法、装置、图像处理设备及计算机可读存储介质。In order to solve the existing technical problems, the embodiment of the present invention provides an image fusion method, device, image processing equipment, and computer-based image fusion method that can reduce invalid information in the image, reduce the amount of calculation and complexity, and improve the real-time performance of the system. Read storage media.
技术解决方案technical solution
本发明实施例第一方面,提供一种图像融合方法,应用于图像处理设备,包括:In the first aspect of the embodiments of the present invention, an image fusion method is provided, which is applied to an image processing device, including:
获取针对目标视场同步采集的可见光图像和红外图像;Acquire visible light images and infrared images synchronously collected for the target field of view;
对所述红外图像二值化得到掩膜图像,根据所述掩膜图像确定目标融合区域;Binarizing the infrared image to obtain a mask image, and determining a target fusion area according to the mask image;
将所述红外图像基于所述目标融合区域与所述可见光图像进行融合,得到融合图像。The infrared image is fused with the visible light image based on the target fusion area to obtain a fused image.
其中,所述将所述红外图像基于所述目标融合区域与所述可见光图像进行融合,得到融合图像,包括:Wherein, the fusion of the infrared image based on the target fusion area and the visible light image to obtain a fusion image includes:
分别将所述红外图像和所述可见光图像进行通道分离,对分离出的表征图像亮度的两个亮度通道分量根据所述目标融合区域进行融合,得到亮度通道融合图像;Channel-separating the infrared image and the visible light image respectively, and fusing the separated two brightness channel components representing image brightness according to the target fusion area to obtain a brightness channel fusion image;
将所述亮度通道融合图像和所述可见光图像进行融合,得到融合图像。The fused image of the brightness channel and the visible light image are fused to obtain a fused image.
其中,所述对所述红外图像二值化得到掩膜图像,根据所述掩膜图像确定目标融合区域,包括:Wherein, the binarization of the infrared image to obtain a mask image, and determining the target fusion area according to the mask image includes:
将所述红外图像中各像素点的灰度值与二值化阈值进行比较,灰度值小于所述二值化阈值的像素点的灰度值置第一设定值,灰度值大于或等于所述二值化阈值的像素点的灰度值置第二设定值,得到掩膜图像;Comparing the gray value of each pixel in the infrared image with the binarization threshold, the gray value of the pixel whose gray value is less than the binarization threshold is set to the first set value, and the gray value is greater than or The grayscale value of the pixel point equal to the binarization threshold is set to a second set value to obtain a mask image;
选择所述掩膜图像中所述第二设定值的像素点分布区域的至少一部分作为目标融合区域。Selecting at least a part of the pixel point distribution area of the second set value in the mask image as a target fusion area.
其中,所述将所述红外图像中各像素点的灰度值与二值化阈值进行比较之前,包括:Wherein, before comparing the gray value of each pixel in the infrared image with the binarization threshold, it includes:
根据所述红外图像的灰度直方图和平均梯度的分布特性,确定匹配的二值化策略;According to the gray histogram of the infrared image and the distribution characteristics of the average gradient, determine a matching binarization strategy;
按照所述二值化策略确定所述二值化阈值。The binarization threshold is determined according to the binarization strategy.
其中,所述根据所述红外图像的灰度直方图和平均梯度的分布特性,确定匹配的二值化策略,包括:Wherein, according to the distribution characteristics of the grayscale histogram and the average gradient of the infrared image, the matching binarization strategy is determined, including:
根据所述红外图像的灰度直方图的分布特性,判断所述灰度直方图是否呈单峰分布;According to the distribution characteristics of the gray histogram of the infrared image, it is judged whether the gray histogram is in a unimodal distribution;
若是,确定匹配的二值化策略为三角形法;If so, determine that the matching binarization strategy is the triangle method;
若否,根据所述红外图像的平均梯度与所述可见光图像的平均梯度的对比结果,确定匹配的二值化策略为高斯法或大津法。If not, according to the comparison result of the average gradient of the infrared image and the average gradient of the visible light image, it is determined that the matching binarization strategy is the Gaussian method or the Otsu method.
其中,所述根据所述红外图像的灰度直方图和平均梯度的分布特性,确定匹配的二值化策略,包括:Wherein, according to the distribution characteristics of the grayscale histogram and the average gradient of the infrared image, the matching binarization strategy is determined, including:
确定所述红外图像的灰度值众数与灰度平均值之间的差值,若所述差值小于或等于预设值,确定所述红外图像的灰度直方图呈单峰分布,则确定匹配的二值化策略为三角形法;Determining the difference between the mode of the gray value of the infrared image and the average value of the gray value, if the difference is less than or equal to a preset value, it is determined that the gray histogram of the infrared image is in a unimodal distribution, then Determine that the matching binarization strategy is the triangle method;
此时,所述按照所述二值化策略确定所述二值化阈值具体为:At this time, the determination of the binarization threshold according to the binarization strategy is specifically:
以所述灰度直方图中最大波峰为顶点确定三角形;Determining a triangle with the largest peak in the grayscale histogram as the apex;
通过所述三角形确定最大直线距离,根据所述最大直线距离对应的直方图灰度等级确定二值化阈值。The maximum straight-line distance is determined through the triangle, and the binarization threshold is determined according to the gray level of the histogram corresponding to the maximum straight-line distance.
其中,所述根据所述红外图像的灰度直方图和平均梯度的分布特性,确定匹配的二值化策略,还包括:Wherein, said according to the distribution characteristics of the grayscale histogram and average gradient of said infrared image, determine the binarization strategy of matching, also include:
若所述差值大于所述预设值,确定所述红外图像的第一平均梯度和所述可见光图像的第二平均梯度;If the difference is greater than the preset value, determine the first average gradient of the infrared image and the second average gradient of the visible light image;
若所述第二平均梯度大于或等于所述第一平均梯度,则确定匹配的二值化策略为高斯法;If the second average gradient is greater than or equal to the first average gradient, then determine that the matching binarization strategy is the Gaussian method;
此时,所述按照所述二值化策略确定所述二值化阈值具体为:计算目标窗函数内所述红外图像的灰度值的高斯均值,根据所述高斯均值确定二值化阈值。In this case, the determining the binarization threshold according to the binarization strategy specifically includes: calculating a Gaussian mean of the grayscale values of the infrared image within the target window function, and determining the binarization threshold according to the Gaussian mean.
其中,所述根据所述红外图像的灰度直方图和平均梯度的分布特性,确定匹配的二值化策略,还包括:Wherein, said according to the distribution characteristics of the grayscale histogram and average gradient of said infrared image, determine the binarization strategy of matching, also include:
若所述第二平均梯度小于所述第一平均梯度,则确定匹配的二值化策略为大津法;If the second average gradient is smaller than the first average gradient, it is determined that the matching binarization strategy is the Otsu method;
此时,所述按照所述二值化策略确定所述二值化阈值具体为:At this time, the determination of the binarization threshold according to the binarization strategy is specifically:
将所述红外图像分割为前景图像和背景图像;Segmenting the infrared image into a foreground image and a background image;
根据所述前景图像和所述背景图像的类间方差值,确定二值化阈值。A binarization threshold is determined according to the inter-class variance value of the foreground image and the background image.
其中,所述将所述红外图像和所述可见光图像进行通道分离,对分离出的表征图像亮度的亮度通道分量根据所述目标融合区域进行融合,得到亮度通道融合图像,包括:Wherein, the channel separation of the infrared image and the visible light image is performed, and the separated brightness channel component representing the brightness of the image is fused according to the target fusion area to obtain a brightness channel fusion image, including:
分别将所述红外图像和所述可见光图像进行HSI通道分离,根据所述目标融合区域对分离出的两个I通道分量按泊松图像编辑原理进行融合,得到I通道融合图像;The infrared image and the visible light image are respectively subjected to HSI channel separation, and the two separated I channel components are fused according to the Poisson image editing principle according to the target fusion area to obtain an I channel fusion image;
所述将所述亮度通道融合图像和所述可见光图像进行融合,得到融合图像,包括:The merging of the luminance channel fused image and the visible light image to obtain a fused image includes:
将所述I通道融合图像的I通道分量与所述可见光图像分离出的H通道分量、S通道分量合并,得到融合参考图像;Merging the I channel component of the I channel fused image with the H channel component and the S channel component separated from the visible light image to obtain a fused reference image;
将所述融合参考图像转换到RGB色彩空间,得到融合图像。Convert the fused reference image to RGB color space to obtain a fused image.
第二方面,还提供一种图像融合装置,包括:In a second aspect, an image fusion device is also provided, including:
获取模块,用于获取针对目标视场同步采集的可见光图像和红外图像;An acquisition module, configured to acquire visible light images and infrared images synchronously collected for the target field of view;
融合区域确定模块,用于对所述红外图像二值化得到掩膜图像,根据所述掩膜图像确定目标融合区域;A fusion area determination module, configured to binarize the infrared image to obtain a mask image, and determine a target fusion area according to the mask image;
融合模块,用于将所述红外图像基于所述目标融合区域与所述可见光图像进行融合,得到融合图像。A fusion module, configured to fuse the infrared image with the visible light image based on the target fusion area to obtain a fusion image.
其中,所述融合模块,具体用于将所述红外图像和所述可见光图像进行通道分离,对分离出的表征图像亮度的亮度通道分量根据所述目标融合区域进行融合,得到亮度通道融合图像;将所述亮度通道融合图像和所述可见光图像进行融合,得到融合图像。Wherein, the fusion module is specifically used to separate the channels of the infrared image and the visible light image, and fuse the separated brightness channel component representing the brightness of the image according to the target fusion area to obtain a brightness channel fusion image; The fused image of the brightness channel and the visible light image are fused to obtain a fused image.
其中,所述融合区域确定模块,具体用于将所述红外图像中各像素点的灰度值与二值化阈值进行比较,灰度值小于所述二值化阈值的像素点的灰度值置第一设定值,灰度值大于或等于所述二值化阈值的像素点的灰度值置第二设定值,得到掩膜图像;选择所述掩膜图像中所述第二设定值的像素点分布区域的至少一部分作为目标融合区域。Wherein, the fusion area determination module is specifically used to compare the gray value of each pixel in the infrared image with the binarization threshold, and the gray value of the pixel whose gray value is smaller than the binarization threshold Set the first set value, and set the gray value of the pixel point whose gray value is greater than or equal to the binarization threshold to the second set value to obtain a mask image; select the second set value in the mask image At least a part of the fixed-value pixel point distribution area is used as the target fusion area.
其中,所述融合区域确定模块,还用于根据所述红外图像的灰度直方图和平均梯度的分布特性,确定匹配的二值化策略;按照所述二值化策略确定二值化阈值。Wherein, the fusion region determination module is further configured to determine a matching binarization strategy according to the gray histogram of the infrared image and the distribution characteristics of the average gradient; determine the binarization threshold according to the binarization strategy.
其中,所述融合区域确定模块,还用于根据所述红外图像的灰度直方图的分布特性,判断所述灰度直方图是否呈单峰分布;若是,确定匹配的二值化策略为三角形法;若否,根据所述红外图像的平均梯度与所述可见光图像的平均梯度的对比结果,确定匹配的二值化策略为高斯法或大津法。Wherein, the fusion area determination module is also used to judge whether the gray histogram is a unimodal distribution according to the distribution characteristics of the gray histogram of the infrared image; if so, determine that the matching binarization strategy is a triangle If not, according to the comparison result of the average gradient of the infrared image and the average gradient of the visible light image, it is determined that the matching binarization strategy is the Gaussian method or the Otsu method.
其中,所述融合区域确定模块,还用于确定所述红外图像的灰度值众数与灰度平均值之间的差值,若所述差值小于或等于预设值,确定所述红外图像的灰度直方图呈单峰分布;以所述灰度直方图中最大波峰为顶点确定三角形;通过所述三角形确定最大直线距离,根据所述最大直线距离对应的直方图灰度等级确定二值化阈值。Wherein, the fusion region determination module is also used to determine the difference between the mode of the gray value of the infrared image and the average value of the gray value, if the difference is less than or equal to a preset value, determine the infrared The gray histogram of the image is a unimodal distribution; the largest peak in the gray histogram is used as the apex to determine the triangle; the maximum straight-line distance is determined by the triangle, and the two are determined according to the gray level of the histogram corresponding to the maximum straight-line distance. Value threshold.
其中,所述融合区域确定模块,还用于若所述差值大于所述预设值,确定所述红外图像的第一平均梯度和所述可见光图像的第二平均梯度;若所述第二平均梯度大于或等于所述第一平均梯度,计算目标窗函数内所述红外图像的灰度值的高斯均值,根据所述高斯均值确定二值化阈值。Wherein, the fusion region determination module is further configured to determine the first average gradient of the infrared image and the second average gradient of the visible light image if the difference is greater than the preset value; if the second The average gradient is greater than or equal to the first average gradient, the Gaussian mean value of the gray value of the infrared image within the target window function is calculated, and the binarization threshold is determined according to the Gaussian mean value.
其中,所述融合区域确定模块,还用于若所述第二平均梯度小于所述第一平均梯度,将所述红外图像分割为前景图像和背景图像;根据所述前景图像和所述背景图像的类间方差值,确定二值化阈值。Wherein, the fusion region determination module is further configured to segment the infrared image into a foreground image and a background image if the second average gradient is smaller than the first average gradient; according to the foreground image and the background image The between-class variance value of , and determine the binarization threshold.
其中,所述融合模块,还用于分别将所述红外图像和所述可见光图像进行HSI通道分离,根据所述目标融合区域对分离出的两个I通道分量按泊松图像编辑原理进行融合,得到I通道融合图像;将所述I通道融合图像的I通道分量与所述可见光图像分离出的H、S通道分量合并,得到融合参考图像;将所述融合参考图像转换到RGB色彩空间,得到融合图像。Wherein, the fusion module is also used to separately perform HSI channel separation on the infrared image and the visible light image, and fuse the separated two I-channel components according to the Poisson image editing principle according to the target fusion area, Obtain the I channel fused image; merge the I channel component of the I channel fused image with the H and S channel components separated from the visible light image to obtain a fused reference image; convert the fused reference image to RGB color space to obtain Blend images.
第三方面,还提供一种图像处理设备,包括处理器、与所述处理器连接的存储器及存储在所述存储器上并可被所述处理器执行的计算机程序,所述计算机程序被所述处理器执行时,实现本申请任一实施例所述的图像融合方法的步骤。In the third aspect, there is also provided an image processing device, including a processor, a memory connected to the processor, and a computer program stored in the memory and executable by the processor, the computer program being executed by the When executed by the processor, the steps of the image fusion method described in any embodiment of the present application are realized.
第四方面,提供一种计算机可读存储介质,所述计算机可读存储介质上存储有计算机程序,所述计算机程序被处理器执行时实现本申请任一实施例所述的图像融合方法的步骤。In a fourth aspect, a computer-readable storage medium is provided, and a computer program is stored on the computer-readable storage medium. When the computer program is executed by a processor, the steps of the image fusion method described in any embodiment of the present application are implemented. .
有益效果Beneficial effect
上述实施例所提供的图像融合方法,通过对红外图像进行二值化得到掩膜图像,根据掩膜图像确定目标融合区域,将红外图像基于所述目标融合区域与可见光图像进行融合,得到融合图像,如此,通过目标融合区域的确定,可以对红外图像中包含的有效信息的部分进行提取,将提取到的有效信息与可见光图像进行融合,可有效避免融合后图像包含无用信息而使得图像质量下降,减少图像中无效信息、减小计算量和复杂度、且可提升系统实时性。In the image fusion method provided by the above embodiment, the mask image is obtained by binarizing the infrared image, the target fusion area is determined according to the mask image, and the infrared image is fused with the visible light image based on the target fusion area to obtain a fusion image In this way, through the determination of the target fusion area, the part of the effective information contained in the infrared image can be extracted, and the extracted effective information can be fused with the visible light image, which can effectively prevent the fused image from containing useless information and reduce the image quality. , reducing invalid information in the image, reducing the amount of calculation and complexity, and improving the real-time performance of the system.
上述实施例中,图像融合装置、图像处理设备及计算机可读存储介质与对应的图像融合方法实施例属于同一构思,从而分别与对应的图像融合方法实施例具有相同的技术效果,在此不再赘述。In the above-mentioned embodiments, the image fusion device, image processing equipment, and computer-readable storage medium belong to the same concept as the corresponding image fusion method embodiments, and thus have the same technical effects as the corresponding image fusion method embodiments, and are not repeated here. repeat.
附图说明Description of drawings
图1为一实施例中图像融合方法的应用场景示意图;Fig. 1 is a schematic diagram of an application scene of an image fusion method in an embodiment;
图2为一实施例中图像融合方法的流程图;Fig. 2 is a flowchart of an image fusion method in an embodiment;
图3为另一实施例中图像融合方法的流程图;Fig. 3 is the flowchart of image fusion method in another embodiment;
图4为又一实施例中图像融合方法的流程图;Fig. 4 is the flowchart of image fusion method in another embodiment;
图5为一示例中红外图像的灰度直方图的示意图;Fig. 5 is a schematic diagram of a grayscale histogram of an example mid-infrared image;
图6为一示例中灰度直方图数据呈单峰分布的示意图;Fig. 6 is a schematic diagram of grayscale histogram data in an example in a unimodal distribution;
图7为红外图像的灰度直方图呈单峰分布,采用三角形法、高斯法和大津法进行融合的对比示意图;Figure 7 is a schematic diagram of the comparison of the gray histogram of the infrared image showing a unimodal distribution, using the triangle method, the Gaussian method and the Otsu method for fusion;
图8为红外图像的灰度直方图呈大致均匀分布,采用三角形法、高斯法和大津法进行融合的对比示意图;Figure 8 is a schematic diagram of the comparison of the gray histogram of the infrared image, which is roughly uniformly distributed, and which is fused using the triangle method, the Gaussian method, and the Otsu method;
图9为红外图像的灰度直方图呈双峰分布,采用三角形法、高斯法和大津法进行融合的对比示意图;Figure 9 is a schematic diagram of the comparison of the gray histogram of the infrared image showing a bimodal distribution, using the triangle method, the Gaussian method and the Otsu method for fusion;
图10为一可选的具体示例中图像融合方法的流程图;FIG. 10 is a flowchart of an image fusion method in an optional specific example;
图11为图10所示实施例中采用的红外图像的示意图;Fig. 11 is a schematic diagram of an infrared image used in the embodiment shown in Fig. 10;
图12为图10所示实施例中采用的可见光图像的示意图;Fig. 12 is a schematic diagram of a visible light image used in the embodiment shown in Fig. 10;
图13为采用本申请所述的图像融合方法对红外图像和可见光图像进行融合后得到的融合图像的示意图;13 is a schematic diagram of a fusion image obtained after fusion of an infrared image and a visible light image using the image fusion method described in the present application;
图14为采用已知的基于低秩表示原理对红外图像和可见光图像进行融合的融合图像的示意图;FIG. 14 is a schematic diagram of a fusion image obtained by fusing an infrared image and a visible light image using a known low-rank representation principle;
图15为采用已知的基于非下采样剪切波变换原理对红外图像和可见光图像进行融合的融合图像的示意图;FIG. 15 is a schematic diagram of a fusion image obtained by fusing an infrared image and a visible light image using a known non-subsampling shearlet transform principle;
图16为采用已知的基于非下采样轮廓波变换原理对红外图像和可见光图像进行融合的融合图像的示意图;FIG. 16 is a schematic diagram of a fusion image obtained by fusing an infrared image and a visible light image using a known non-subsampling contourlet transform principle;
图17采用已知的基于泊松图像编辑原理对红外图像和可见光图像进行融合的融合图像的示意图;Fig. 17 is a schematic diagram of a fusion image obtained by fusing an infrared image and a visible light image using a known Poisson image editing principle;
图18为一实施例中图像融合装置的示意图;Fig. 18 is a schematic diagram of an image fusion device in an embodiment;
图19为一实施例中图像处理设备的结构示意图。Fig. 19 is a schematic structural diagram of an image processing device in an embodiment.
本发明的实施方式Embodiments of the present invention
以下结合说明书附图及具体实施例对本发明技术方案做进一步的详细阐述。The technical solutions of the present invention will be further described in detail below in conjunction with the drawings and specific embodiments of the description.
除非另有定义,本文所使用的所有的技术和科学术语与属于本发明的技术领域的技术人员通常理解的含义相同。本文中在本发明的说明书中所使用的术语只是为了描述具体的实施例的目的,不是旨在于限制本发明的保护范围。本文所使用的术语“和/或”包括一个或多个相关的所列项目的任意的和所有的组合。Unless otherwise defined, all technical and scientific terms used herein have the same meaning as commonly understood by one of ordinary skill in the technical field of the invention. The terms used herein in the description of the present invention are only for the purpose of describing specific embodiments, and are not intended to limit the protection scope of the present invention. As used herein, the term "and/or" includes any and all combinations of one or more of the associated listed items.
在以下的描述中,涉及到“一些实施例”的表述,其描述了所有可能实施例的子集,但是应当理解,“一些实施例”可以是所有可能实施例的相同子集或不同子集,并且可以在不冲突的情况下相互结合。In the following description, the expression "some embodiments" refers to a subset of all possible embodiments, but it should be understood that "some embodiments" may be the same subset or different subsets of all possible embodiments , and can be combined with each other without conflict.
在以下的描述中,所涉及的术语“第一、第二、第三”仅仅是区别类似的对象,不代表针对对象的特定排序,可以理解地,“第一、第二、第三”在允许的情况下可以互换特定的顺序或先后次序,以使这里描述的本申请实施例能够以除了在这里图示或描述的以外的顺序实施。In the following description, the terms "first, second, and third" are only used to distinguish similar objects, and do not represent a specific ordering of objects. Understandably, "first, second, and third" are used in Where permitted, the specific order or sequence may be interchanged such that the embodiments of the application described herein can be practiced in other sequences than illustrated or described herein.
请参阅图1,为本申请实施例提供的图像处理方法的一可选应用场景的示意图,其中,图像处理设备11包括处理器12、与所述处理器12连接的存储器13、可见光拍摄模块14和红外拍摄模块15。所述图像处理设备11通过所述可见光拍摄模块14和所述红外拍摄模块14同步实时采集可见光图像和红外图像发送给处理器12,所述存储器13内存储有实施本申请实施例所提供的图像融合方法的计算机程序,处理器12通过执行所述计算机程序,对所述红外图像二值化得到掩膜图像,通过掩膜图像确定出目标融合区域,将红外图像基于目标融合区域与可见光图像进行融合,得到融合图像。其中,所述图像处理设备11可以是集成有可见光拍摄模块14和红外拍摄模块15、且具备存储和处理功能的各类智能终端,如安防监控设备、车载设备等;所述图像处理设备11也可以是与可见光拍摄模块14和红外拍摄模块15连接的计算机设备;所述图像处理设备11也可以是白光和红光的双光融合瞄准类设备。Please refer to FIG. 1, which is a schematic diagram of an optional application scenario of the image processing method provided by the embodiment of the present application, wherein the image processing device 11 includes a processor 12, a memory 13 connected to the processor 12, and a visible light shooting module 14 And infrared camera module 15. The image processing device 11 collects visible light images and infrared images synchronously and in real time through the visible light shooting module 14 and the infrared shooting module 14 and sends them to the processor 12, and the memory 13 stores the images provided by the embodiments of the present application The computer program of the fusion method, the processor 12 executes the computer program, binarizes the infrared image to obtain a mask image, determines the target fusion area through the mask image, and performs the infrared image based on the target fusion area and the visible light image. fusion to obtain a fusion image. Wherein, the image processing device 11 can be various types of intelligent terminals integrated with the visible light shooting module 14 and the infrared shooting module 15, and have storage and processing functions, such as security monitoring equipment, vehicle-mounted equipment, etc.; the image processing device 11 is also It can be a computer device connected to the visible light shooting module 14 and the infrared shooting module 15; the image processing device 11 can also be a dual-light fusion aiming device of white light and red light.
请参阅图2,为本申请一实施例提供的图像融合方法,可以应用于图1所 示的图像处理设备。其中,图像处理方法包括如下步骤:Referring to Fig. 2, the image fusion method provided by an embodiment of the present application can be applied to the image processing device shown in Fig. 1 . Wherein, the image processing method includes the following steps:
S101,获取针对目标视场同步采集的可见光图像和红外图像。S101. Acquire a visible light image and an infrared image that are synchronously collected for a target field of view.
可见光图像和红外图像是针对目标视场同步采集得到的,如此,可见光图像和红外图像中包含同一目标视场内的物体的成像。可选的,图像处理设备包括可见光拍摄模块和红外拍摄模块,所述获取针对目标视场同步采集的可见光图像和红外图像包括:图像处理设备通过可见光拍摄模块和红外拍摄模块同时采集可见光图像和红外图像,并将采集到的可见光图像和红外图像发送给处理器。在另一些可选的实施例中,图像处理设备不包括图像拍摄模块,所述获取针对目标视场同步采集的可见光图像和红外图像包括:图像处理设备获取具备可见光图像和红外图像拍摄功能的其它智能设备发送的针对目标视场同步采集的可见光图像和红外图像,这里,其它智能设备可以包括红外探测器、手机终端、云端等。The visible light image and the infrared image are acquired synchronously for the target field of view, so that the visible light image and the infrared image include the imaging of objects in the same target field of view. Optionally, the image processing device includes a visible light shooting module and an infrared shooting module, and the acquiring the visible light image and the infrared image synchronously collected for the target field of view includes: the image processing device simultaneously collects the visible light image and the infrared image through the visible light shooting module and the infrared shooting module image, and send the collected visible light image and infrared image to the processor. In some other optional embodiments, the image processing device does not include an image capturing module, and the acquiring the visible light image and the infrared image synchronously collected for the target field of view includes: the image processing device acquires other images with the functions of capturing visible light images and infrared images Visible light images and infrared images synchronously collected for the target field of view sent by the smart device. Here, other smart devices may include infrared detectors, mobile terminals, and the cloud.
S103,对所述红外图像二值化得到掩膜图像,根据所述掩膜图像确定目标融合区域。S103. Binarize the infrared image to obtain a mask image, and determine a target fusion area according to the mask image.
对红外图像二值化是指,将所述红外图像上的各像素点的灰度值分别进行赋值,得到可反映图像整体和局部特征的二值化图像。The binarization of the infrared image refers to assigning the gray value of each pixel on the infrared image to obtain a binarized image that can reflect the overall and local features of the image.
在可选的实施例中,所述S103,对所述红外图像二值化得到掩膜图像,根据所述掩膜图像确定目标融合区域,包括:In an optional embodiment, the step S103 is to binarize the infrared image to obtain a mask image, and determine the target fusion area according to the mask image, including:
将所述红外图像中各像素点的灰度值与二值化阈值进行比较,灰度值小于所述二值化阈值的像素点的灰度值置第一设定值,灰度值大于或等于所述二值化阈值的像素点的灰度值置第二设定值,得到掩膜图像;Comparing the gray value of each pixel in the infrared image with the binarization threshold, the gray value of the pixel whose gray value is less than the binarization threshold is set to the first set value, and the gray value is greater than or The grayscale value of the pixel point equal to the binarization threshold is set to a second set value to obtain a mask image;
选择所述掩膜图像中所述第二设定值的像素点分布区域的至少一部分作为目标融合区域。Selecting at least a part of the pixel point distribution area of the second set value in the mask image as a target fusion area.
二值化阈值可以是预先设定的,也可以是根据对红外图像中像素点灰度值的分布特征计算得到。第一设定值和第二设定值可以分别选自灰度值区间的最大值和最小值、或者也可以是选择灰度值区间中分别靠近最大值和最小值的两个灰度值。在一个具体示例中,第一设定值为0,第二设定值为255,使得整个图像呈现出黑白图像效果。所述对所述红外图像二值化得到掩膜图像包括:将红外图像中256个亮度等级的灰度图像通过二值化阈值进行二值化处理,将红外图像中各像素点的灰度值与二值化阈值进行比较,灰度值小于二值化阈值的像素点将其灰度值均置0,灰度值大于二值化阈值的像素点的灰度值均置255,从而得到可反映图像整体和局部特征的二值化图像,即掩膜图像。相应的,根据掩膜图像确定目标融合区域,可以是根据掩膜图像中所述第二设定值的像素点分布区域,也即白色部分确定目标融合区域,如可以选择掩膜图像中白色部分全部作为目标融合区域,或选择掩膜图像中白色部分中的某一部分作为目标融合区域。The binarization threshold can be preset, or can be calculated according to the distribution characteristics of the gray value of pixels in the infrared image. The first set value and the second set value can be respectively selected from the maximum value and the minimum value of the gray value interval, or can also be two gray values close to the maximum value and the minimum value respectively in the gray value interval. In a specific example, the first set value is 0, and the second set value is 255, so that the entire image presents a black and white image effect. The said binarization of the infrared image to obtain the mask image includes: performing binarization on the grayscale images of 256 brightness levels in the infrared image through the binarization threshold, and converting the grayscale value of each pixel in the infrared image to Compared with the binarization threshold, the gray value of pixels whose gray value is smaller than the binarization threshold is set to 0, and the gray value of pixels whose gray value is greater than the binarization threshold is set to 255, thus obtaining A binary image that reflects the overall and local features of the image, that is, the mask image. Correspondingly, determining the target fusion area according to the mask image may be based on the pixel point distribution area of the second set value in the mask image, that is, the white part determines the target fusion area, such as the white part in the mask image may be selected All as the target fusion area, or select a certain part of the white part in the mask image as the target fusion area.
S105,将所述红外图像基于所述目标融合区域与所述可见光图像进行融合,得到融合图像。S105. Fusion the infrared image with the visible light image based on the target fusion area to obtain a fusion image.
将所述红外图像基于所述目标融合区域与所述可见光图像进行融合,得到融合图像可以是指,将所述红外图像和所述可见光图像分别与所述目标融合区域对应的图像部位进行融合,其它部位保留可见光图像的图像部分,得到融合图像;或,将所述红外图像的所述目标融合区域提取出来形成待融合图像,将所述待融合图像与所述可见光图像进行融合等。Fusing the infrared image based on the target fusion area with the visible light image to obtain a fusion image may refer to merging the infrared image and the visible light image with image parts corresponding to the target fusion area, respectively, Other parts retain the image part of the visible light image to obtain a fused image; or, extract the target fusion area of the infrared image to form an image to be fused, and fuse the image to be fused with the visible light image, etc.
上述实施例中,通过对红外图像二值化得到掩膜图像,根据掩膜图像确定目标融合区域,将红外图像基于所述目标融合区域与可见光图像进行融合,得到融合图像,如此,通过目标融合区域的确定,可以对红外图像中包含的有效信息的部分进行提取,将提取到的有效信息与可见光图像进行融合,可有效避免融合后图像包含无用信息而使得图像质量下降,减少图像中无效信息、减小计算量和复杂度、且可提升系统实时性。In the above embodiment, the mask image is obtained by binarizing the infrared image, the target fusion area is determined according to the mask image, and the infrared image is fused with the visible light image based on the target fusion area to obtain a fusion image. In this way, through target fusion The determination of the area can extract the effective information contained in the infrared image, and fuse the extracted effective information with the visible light image, which can effectively avoid the image quality degradation caused by the useless information contained in the fused image, and reduce the invalid information in the image , reduce the amount of calculation and complexity, and can improve the real-time performance of the system.
可选的,请参阅图3,S105,将所述红外图像基于所述目标融合区域与所述可见光图像进行融合,得到融合图像,包括:Optionally, please refer to FIG. 3, S105, the infrared image is fused based on the target fusion area with the visible light image to obtain a fused image, including:
S1051,分别将所述红外图像和所述可见光图像进行通道分离,对分离出的表征图像亮度的两个亮度通道分量根据所述目标融合区域进行融合,得到亮度通道融合图像。S1051. Channel-separate the infrared image and the visible light image respectively, and fuse the separated two brightness channel components representing image brightness according to the target fusion area to obtain a brightness channel fusion image.
对于一幅数字图像,人眼观察到的一幅图片,但是从计算机来看,一副数字图像是一堆亮度各异的点,如,一副尺寸为M×N的数字图像可以用一个M×N的矩阵来表示,矩阵中元素的值分别表示这个位置上对应像素点的亮度,像素值越大表示该像素点越亮。通常,灰度图可用二维矩阵表示,彩色图像可用三维矩阵(M×N×3)表示,也即多通道图像。For a digital image, it is a picture observed by the human eye, but from the perspective of a computer, a digital image is a bunch of points with different brightness. For example, a digital image with a size of M×N can be represented by a M ×N matrix, the values of the elements in the matrix respectively represent the brightness of the corresponding pixel at this position, and the larger the pixel value, the brighter the pixel. Usually, the grayscale image can be represented by a two-dimensional matrix, and the color image can be represented by a three-dimensional matrix (M×N×3), that is, a multi-channel image.
通过通道可以改变图像的色相和颜色,例如,如仅保存红色通道,则图像本身就仅保留红色的元素和信息。针对每个单一通道,可分别显示为一副灰度图像(需要说明的是,该灰度图像非黑白图像),单一通道的灰度图像中的明暗对应所述单一通道色的明暗,相应表示所述单一通道色/光在图像上的分布情况。The hue and color of the image can be changed through the channel. For example, if only the red channel is saved, the image itself only retains red elements and information. For each single channel, it can be displayed as a pair of grayscale images (it should be noted that the grayscale image is not a black and white image), and the lightness and darkness in the grayscale image of a single channel correspond to the lightness and darkness of the color of the single channel, correspondingly representing The distribution of the color/light of the single channel on the image.
将红外图像和可见光图像进行通道分离,对分离出的表征图像亮度的亮度通道分量根据目标融合区域进行融合,得到亮度通道融合图像可以是指,将红外图像和可见光图像进行通道分离,将红外图像中分离出的表征图像亮度的亮度通道分量中与目标融合区域对应的部分、与可见光图像中分离出的表征图像亮度的亮度通道分量中与目标融合区域对应的部分进行融合,得到亮度通道融合图像。其中,对红外图像和可见光图像分离出的亮度通道分量根据目标融合区域进行融合,可以减少融合所需的运算量,且可以保留目标融合区域内的有效信息。Channel-separating the infrared image and the visible light image, and merging the separated luminance channel component representing the brightness of the image according to the target fusion area to obtain the luminance channel fusion image can refer to channel separation of the infrared image and the visible light image, and the infrared image The part corresponding to the target fusion area in the luminance channel component representing the image luminance separated in , is fused with the part corresponding to the target fusion area in the luminance channel component representing image luminance separated from the visible light image, and the luminance channel fusion image is obtained . Wherein, the luminance channel components separated from the infrared image and the visible light image are fused according to the target fusion area, which can reduce the amount of calculation required for fusion, and can retain effective information in the target fusion area.
S1052,将所述亮度通道融合图像和所述可见光图像进行融合,得到融合图像。S1052. Fusion the brightness channel fusion image and the visible light image to obtain a fusion image.
图像融合(Image Fusion)是指将多源信道所采集到的关于同一目标的图像数据经过图像处理技术,最大限度的提取各自信道中的有利信息,最后综合成 高质量的图像,以提高图像信息的利用率、改善计算机解译精度和可靠性、提升原始图像的空间分辨率和光谱分辨率,利于监测。亮度通道融合图像包含可见光图像的亮度通道分量和红外图像的亮度通道分量中的有效信息,将亮度通道融合图像和可见光图像进行融合,使得亮度通道融合图像中的亮度通道分量与可见光图像的其它通道分量合并,得到融合图像。Image fusion (Image Fusion) refers to the image data of the same target collected by multi-source channels through image processing technology to maximize the extraction of beneficial information in each channel, and finally synthesize high-quality images to improve image information. Improve the utilization rate of computer, improve the accuracy and reliability of computer interpretation, and improve the spatial resolution and spectral resolution of the original image, which is beneficial to monitoring. The brightness channel fusion image contains the effective information in the brightness channel component of the visible light image and the brightness channel component of the infrared image, and fuses the brightness channel fusion image and the visible light image, so that the brightness channel component in the brightness channel fusion image and other channels of the visible light image The components are combined to obtain a fused image.
上述实施例中,通过对红外图像二值化得到掩膜图像,根据掩膜图像确定目标融合区域,将红外图像和可见光图像进行通道分离,对分离出的表征图像亮度的亮度通道分量根据目标融合区域进行融合,得到亮度通道融合图像,将亮度通道融合图像和可见光图像进行融合,得到融合图像,如此,通过目标融合区域的确定,可以确保对红外图像中包含的有效信息的部分进行提取,将红外图像和可见光图像各自分离出的亮度通道分量进行融合,再与可见光图像进行融合,可有效避免融合后图像包含无用信息而使得图像质量下降,减少图像中无效信息、减小计算量和复杂度、且可提升系统实时性;融合后的图像可以同时保留可见光图像和红外图像各自的优点,无论是针对光照条件充分的条件下成像后获得的融合图像、还是针对光照条件不佳的情况下成像后获得融合图像,都可以让目标更好地凸显,确保对图像中关注的目标更加清晰地呈现,更便于人眼观察和识别。In the above embodiment, the mask image is obtained by binarizing the infrared image, the target fusion area is determined according to the mask image, the channels of the infrared image and the visible light image are separated, and the separated luminance channel component representing the brightness of the image is fused according to the target The region is fused to obtain a luminance channel fusion image, and the luminance channel fusion image and the visible light image are fused to obtain a fusion image. In this way, through the determination of the target fusion area, it is possible to ensure that the part of the effective information contained in the infrared image is extracted. The luminance channel components separated from the infrared image and the visible light image are fused, and then fused with the visible light image, which can effectively prevent the fused image from containing useless information and cause image quality degradation, reduce invalid information in the image, and reduce the amount of calculation and complexity. , and can improve the real-time performance of the system; the fused image can retain the respective advantages of visible light images and infrared images at the same time, whether it is for the fused image obtained after imaging under sufficient lighting conditions or for imaging under poor lighting conditions After the fused image is obtained, the target can be better highlighted, ensuring that the target in the image is more clearly presented, and it is easier for the human eye to observe and identify.
在一些实施例中,请参阅图4,所述S103,对所述红外图像二值化得到掩膜图像,根据所述掩膜图像确定目标融合区域的步骤中,在所述将所述红外图像中各像素点的灰度值与二值化阈值进行比较之前,包括:In some embodiments, please refer to FIG. 4, the S103, binarize the infrared image to obtain a mask image, and in the step of determining the target fusion area according to the mask image, in the step of converting the infrared image Before comparing the gray value of each pixel with the binarization threshold, it includes:
S1031,根据所述红外图像的灰度直方图和平均梯度的分布特性,确定匹配的二值化策略;S1031. Determine a matching binarization strategy according to the gray histogram of the infrared image and the distribution characteristics of the average gradient;
S1032,按照所述二值化策略确定所述二值化阈值。S1032. Determine the binarization threshold according to the binarization strategy.
根据所述二值化策略确定的二值化阈值对所述红外图像二值化得到掩膜图像,根据所述掩膜图像确定目标融合区域。Binarizing the infrared image according to the binarization threshold determined by the binarization strategy to obtain a mask image, and determining a target fusion area according to the mask image.
不同的二值化策略所采用的二值化方法的原理不同,适用的目标图像也不同。红外图像的灰度直方图和平均梯度的分布特性,可判断红外图像的灰度直方图呈单峰分布、双峰分布或大致较均匀地分布,以确定与其匹配的二值化策略。如二值化策略包括三角形法、高斯法和大津法,若红外图像的灰度直方图呈单峰分布,则适用三角形法;若红外图像的灰度直方图呈较均匀地分布,则适用高斯法;若红外图像的灰度直方图呈双峰分布,则适用大津法。通过确定与红外图像适用的二值化策略,通过对应二值化策略对红外图像进行二值化处理得到掩膜图像,将掩膜图像中的白色部分的区域确定为目标融合区域。The principles of binarization methods adopted by different binarization strategies are different, and the applicable target images are also different. The distribution characteristics of the gray histogram and average gradient of the infrared image can be judged that the gray histogram of the infrared image is distributed with a single peak, a bimodal distribution or a roughly uniform distribution, so as to determine the matching binarization strategy. For example, the binarization strategy includes triangle method, Gaussian method, and Otsu method. If the gray histogram of the infrared image is distributed with a single peak, the triangle method is applied; if the gray histogram of the infrared image is more uniformly distributed, the Gaussian method is applied. method; if the gray histogram of the infrared image shows a bimodal distribution, the Otsu method is applied. By determining the binarization strategy applicable to the infrared image, the mask image is obtained by binarizing the infrared image through the corresponding binarization strategy, and the area of the white part in the mask image is determined as the target fusion area.
上述实施例中,通过分析红外图像的灰度直方图和平均梯度的分布特性以确定与其适配的二值化策略,以确保对红外图像二值化后可更准确地将图像中的目标所在图像区域二值化为白色部分,以便于基于掩膜图像确定目标融合区域,按照目标融合区域对可见光图像和红外图像分离出的亮度通道分量融合后 可更完整、全面地保留图像中的有效信息。In the above-mentioned embodiment, by analyzing the gray histogram of the infrared image and the distribution characteristics of the average gradient to determine the binarization strategy adapted to it, to ensure that the target in the image can be more accurately located after binarizing the infrared image. The image area is binarized into the white part, so as to determine the target fusion area based on the mask image, and the effective information in the image can be more completely and comprehensively preserved after the brightness channel components separated from the visible light image and the infrared image are fused according to the target fusion area .
所述S1031,根据所述红外图像的灰度直方图和平均梯度的分布特性,确定匹配的二值化策略,包括:The S1031, according to the gray histogram of the infrared image and the distribution characteristics of the average gradient, determine a matching binarization strategy, including:
根据所述红外图像的灰度直方图的分布特性,判断所述灰度直方图是否呈单峰分布;According to the distribution characteristics of the gray histogram of the infrared image, it is judged whether the gray histogram is in a unimodal distribution;
若是,确定匹配的二值化策略为三角形法;If so, determine that the matching binarization strategy is the triangle method;
若否,根据所述红外图像的平均梯度与所述可见光图像的平均梯度的对比结果,确定匹配的二值化策略为高斯法或大津法。If not, according to the comparison result of the average gradient of the infrared image and the average gradient of the visible light image, it is determined that the matching binarization strategy is the Gaussian method or the Otsu method.
在确定与红外图像相适配的二值化策略过程中,首先根据红外图像直方图的分布特性,判断该红外图像是否适用于三角形法,若不符合,再根据平均梯度选择使用高斯法或大津法。判断对当前的红外图像进行二值化处理的二值化策略是否适用于三角形法,是使用红外图像的灰度直方图数据判断是否呈单峰分布,假设红外图像的灰度直方图的最大波峰在靠近最亮的一侧,以此寻找红外图像二值化的最佳阈值。使用红外图像的灰度直方图数据判断不满足单峰分布的情况下,再通过红外图像的平均梯度和可见光图像的平均梯度的对比结果判断呈较均匀地分布或呈双峰分布,以确定适用于高斯法或大津法。In the process of determining the binarization strategy suitable for the infrared image, firstly, according to the distribution characteristics of the infrared image histogram, it is judged whether the infrared image is suitable for the triangle method, if not, then the Gaussian method or Otsu method is selected according to the average gradient Law. Judging whether the binarization strategy for binarizing the current infrared image is suitable for the triangle method is to use the gray histogram data of the infrared image to judge whether it is a unimodal distribution, assuming the maximum peak of the gray histogram of the infrared image On the brightest side, use this to find the best threshold for infrared image binarization. If the gray histogram data of the infrared image is used to judge that the unimodal distribution is not satisfied, then the comparison between the average gradient of the infrared image and the average gradient of the visible light image can be used to determine whether the distribution is relatively uniform or bimodal, so as to determine the applicable than Gauss method or Otsu method.
上述实施例中,设定二值化策略包括三角形法、高斯法和大津法,分析红外图像的灰度直方图和平均梯度的分布特性以确定与其适配的二值化策略为三角形法、高斯法或者大津法,以确保对红外图像二值化后可更准确地将图像中的目标所在图像区域二值化为白色部分,以便于基于掩膜图像确定目标融合区域后,按照目标融合区域对可见光图像和红外图像分离出的亮度通道分量融合得到的亮度分量融合图像中,可更完整、全面地保留图像中的有效信息。In the above embodiment, the binarization strategy is set to include triangle method, Gaussian method and Otsu method, and the distribution characteristics of the gray histogram and average gradient of the infrared image are analyzed to determine the binarization strategy adapted to it as triangular method, Gaussian method, and Gaussian method. method or Otsu method to ensure that after binarizing the infrared image, the image area where the target in the image is located can be more accurately binarized into a white part, so that after the target fusion area is determined based on the mask image, the target fusion area can be aligned according to the target fusion area. In the luminance component fusion image obtained by fusing the luminance channel components separated from the visible light image and the infrared image, the effective information in the image can be more completely and comprehensively preserved.
在一些实施例中,所述S1031,根据所述红外图像的灰度直方图和平均梯度的分布特性,确定匹配的二值化策略,包括:In some embodiments, said S1031, according to the distribution characteristics of the grayscale histogram and the average gradient of the infrared image, determine a matching binarization strategy, including:
确定所述红外图像的灰度值众数与灰度平均值之间的差值,若所述差值小于或等于预设值,确定所述红外图像的灰度直方图呈单峰分布;Determine the difference between the mode of the gray value of the infrared image and the average value of the gray value, if the difference is less than or equal to a preset value, determine that the gray histogram of the infrared image is a unimodal distribution;
以所述灰度直方图中最大波峰为顶点确定三角形;Determining a triangle with the largest peak in the grayscale histogram as the apex;
通过所述三角形确定最大直线距离,根据所述最大直线距离对应的直方图灰度等级确定二值化阈值。The maximum straight-line distance is determined through the triangle, and the binarization threshold is determined according to the gray level of the histogram corresponding to the maximum straight-line distance.
根据红外图像的灰度直方图数据,若灰度直方图中灰度值集中分布于某一数值,则对应所述红外图像的灰度值众数与灰度平均值之间的差值相对于灰度直方图呈其它形式的红外图像而言要小很多,如,可以将灰度值众数与灰度平均值之间的差值的绝对值记为A-M(average-mode),通过判断灰度值众数与灰度平均值之间的差值A-M是否小于预设值,以判断灰度直方图呈单峰分布的程度,若差值A-M小于预设值,则表示对应所述红外图像的灰度直方图呈单峰分 布,反之,则表示对应红外图像的灰度直方图不满足呈单峰分布的特征。如图5所示,使用红外图像的灰度直方图数据,基于纯几何方法来寻找最佳二值化阈值,假设灰度直方图中最大波峰在靠近最亮的一侧,通过三角形求最大直线距离,确定最大直线距离对应的直方图灰度等级为分割阈值。According to the grayscale histogram data of the infrared image, if the grayscale values in the grayscale histogram are concentrated on a certain value, then the difference between the grayscale value mode and the grayscale average value corresponding to the infrared image is relative to The gray histogram is much smaller than other forms of infrared images. For example, the absolute value of the difference between the mode of the gray value and the average value of the gray value can be recorded as A-M (average-mode). By judging the gray Whether the difference A-M between the mode of the degree value and the gray-scale average value is less than the preset value is used to judge the degree of the gray histogram showing a unimodal distribution. If the difference A-M is less than the preset value, it means that the corresponding infrared image The gray histogram of the infrared image shows a unimodal distribution, otherwise, it means that the gray histogram corresponding to the infrared image does not meet the characteristics of a unimodal distribution. As shown in Figure 5, use the gray histogram data of the infrared image to find the optimal binarization threshold based on a pure geometric method, assuming that the largest peak in the gray histogram is on the brightest side, and find the largest straight line through the triangle Determine the gray level of the histogram corresponding to the maximum straight-line distance as the segmentation threshold.
如图6所示,对应红外图像的灰度直方图中灰度值众数与灰度平均值分别为121、127.734,其中,灰度值众数为灰度值中重复次数最多的灰度值,灰度平均值AG的计算公式如下公式一:As shown in Figure 6, the mode and average value of the gray value in the gray histogram corresponding to the infrared image are 121 and 127.734 respectively, where the mode of the gray value is the gray value with the most repetitions among the gray values , the calculation formula of the gray value AG is as follows: Formula 1:
Figure PCTCN2022094865-appb-000001
Figure PCTCN2022094865-appb-000001
H(i,j)表示坐标为(i,j)的像素点的灰度值,M表示横坐标最大值,N表示纵坐标最大值。假设预设值为10,若灰度值众数与灰度平均值的差值A-M小于10,则对红外图像采用三角形法对其进行二值化处理,以灰度直方图中最大波峰为顶点确定三角形,通过所述三角形确定最大直线距离,根据所述最大直线距离对应的直方图灰度等级确定二值化阈值。H(i, j) represents the gray value of the pixel point whose coordinates are (i, j), M represents the maximum value of the abscissa, and N represents the maximum value of the ordinate. Assuming that the preset value is 10, if the difference A-M between the mode of the gray value and the average value of the gray value is less than 10, the infrared image will be binarized using the triangle method, with the largest peak in the gray histogram as the apex A triangle is determined, a maximum straight-line distance is determined through the triangle, and a binarization threshold is determined according to the gray level of the histogram corresponding to the maximum straight-line distance.
上述实施例中,通过计算灰度直方图中灰度值众数与灰度平均值之间的差值,基于差值与预设值的相对大小,衡量灰度值是否集中于某一个指标,以判断对应红外图像的灰度直方图是否呈单峰分布,以实现快速且准确地判断红外图像的灰度直方图的分布特性。In the above embodiment, by calculating the difference between the mode of the gray value in the gray histogram and the average value of the gray value, based on the relative size of the difference and the preset value, it is measured whether the gray value is concentrated in a certain index, To judge whether the grayscale histogram corresponding to the infrared image has a unimodal distribution, so as to realize the fast and accurate judgment of the distribution characteristics of the grayscale histogram of the infrared image.
请参阅图7,为红外图像的灰度直方图呈单峰分布,采用三角形法、高斯法和大津法作为对应的二值化策略对红外图像进行二值化得到掩膜图像后,与可见光图像进行融合的对比示意图,其中,采用三角形法作为二值化策略确定二值化阈值后,对红外图像进行二值化处理得到的掩膜图像,可以更全面地和完整地突出图像目标,最终与可见光图像融合后得到的三角形法融合图像中图像有效信息损失最小。Please refer to Figure 7. The gray histogram of the infrared image shows a unimodal distribution. The triangle method, Gaussian method and Otsu method are used as the corresponding binarization strategies to binarize the infrared image to obtain the mask image, and the visible light image Schematic diagram of fusion comparison, in which, after using the triangle method as the binarization strategy to determine the binarization threshold, the mask image obtained by binarizing the infrared image can highlight the image target more comprehensively and completely, and finally with The loss of effective image information in the triangular method fused image obtained after visible light image fusion is the smallest.
在一些实施例中,所述S1031,根据所述红外图像的灰度直方图和平均梯度的分布特性,确定匹配的二值化策略,还包括:In some embodiments, the step S1031, according to the distribution characteristics of the grayscale histogram and the average gradient of the infrared image, determining a matching binarization strategy, further includes:
若所述差值大于所述预设值,确定所述红外图像的第一平均梯度和所述可见光图像的第二平均梯度;If the difference is greater than the preset value, determine the first average gradient of the infrared image and the second average gradient of the visible light image;
若所述第二平均梯度大于或等于所述第一平均梯度,计算目标窗函数内所述红外图像的灰度值的高斯均值,根据所述高斯均值确定二值化阈值。If the second average gradient is greater than or equal to the first average gradient, calculate the Gaussian mean value of the gray value of the infrared image within the target window function, and determine the binarization threshold according to the Gaussian mean value.
若灰度值众数与灰度平均值之间的差值A-M大于预设值,则表示对应红外图像的灰度直方图不满足呈单峰分布的特征,通过确定红外图像的平均梯度和可见光图像的平均梯度的相对大小,以确定对应红外图像的灰度直方图是否呈较均匀地分布或呈双峰分布。为了便于区分,将红外图像的平均梯度称为第一平均梯度,将可见光图像的平均梯度称为第二平均梯度。其中,图像的平均梯度的计算公式可如下公式二:If the difference A-M between the mode of the gray value and the average value of the gray value is greater than the preset value, it means that the gray histogram corresponding to the infrared image does not meet the characteristics of a unimodal distribution. By determining the average gradient of the infrared image and the visible light The relative size of the average gradient of the image is used to determine whether the gray histogram corresponding to the infrared image is more uniformly distributed or bimodal. For ease of distinction, the average gradient of the infrared image is referred to as the first average gradient, and the average gradient of the visible light image is referred to as the second average gradient. Wherein, the calculation formula of the average gradient of the image can be as follows formula two:
Figure PCTCN2022094865-appb-000002
Figure PCTCN2022094865-appb-000002
H(i,j)表示坐标为(i,j)的像素点的灰度值,M表示横坐标最大值,N表示纵坐标最大值。若第二平均梯度大于第一平均梯度,则表示灰度直方图呈较为均匀地分布,与当前红外图像适用的二值化策略为高斯法。通过高斯法确定二值化阈值的原理是,通过计算窗函数内图像灰度的高斯均值,将所述高斯均值作为二值化阈值对该部分图像进行二值化操作。高斯法是通过获取局部阈值的方法实现二值化,通过优化窗函数的尺度,确定对应窗函数内要融合的部分和不需要融合的部分,以对灰度直方图中灰度值并非集中于某一个指标的红外图像,也即灰度值分别集中于多个指标的红外图像,可将所述红外图像中多个指标各自对应的目标分别所在图像区域二值化为掩膜图像中的白色部分,以便于基于掩膜图像中白色部分确定目标融合区域后,按照目标融合区域对可见光图像和红外图像分离出的亮度通道分量融合得到的亮度分量融合图像中,可更完整、全面地保留图像中的有效信息,融合后图像中的细节信息加强、边缘突出。H(i, j) represents the gray value of the pixel point whose coordinates are (i, j), M represents the maximum value of the abscissa, and N represents the maximum value of the ordinate. If the second average gradient is greater than the first average gradient, it means that the gray histogram is relatively uniformly distributed, and the binarization strategy applicable to the current infrared image is the Gaussian method. The principle of determining the binarization threshold by the Gaussian method is to calculate the Gaussian mean value of the grayscale of the image within the window function, and use the Gaussian mean value as the binarization threshold value to perform binarization operation on the part of the image. The Gaussian method achieves binarization by obtaining local thresholds. By optimizing the scale of the window function, the part to be fused and the part that does not need to be fused in the corresponding window function are determined, so that the gray value in the gray histogram is not concentrated in The infrared image of a certain indicator, that is, the infrared image whose gray value is concentrated on multiple indicators, can binarize the image areas where the targets corresponding to the multiple indicators in the infrared image are respectively located into white in the mask image. Part, so that after the target fusion area is determined based on the white part of the mask image, the luminance component fusion image obtained by fusing the luminance channel components separated from the visible light image and the infrared image according to the target fusion area can retain the image more completely and comprehensively. The effective information in the fused image is strengthened and the edges are highlighted.
上述实施例中,通过可见光图像的平均梯度和红外图像的平均梯度的相对大小,以判断灰度值是否呈大致均匀地分布,以实现快速且准确地判断红外图像的二值化策略是否符合高斯法,对于可见光图像的平均梯度大于红外图像的平均梯度的情况,可见光图像清晰且包含大部分有效信息,从而采用高斯法获得二值化图像,以确保对红外图像二值化后可更准确地将图像中的各个目标所在图像区域均二值化为白色部分,基于掩膜图像确定目标融合区域后,按照目标融合区域对可见光图像和红外图像分离出的亮度通道分量融合得到的亮度分量融合图像中,可更完整、全面地保留图像中的有效信息。In the above-mentioned embodiment, the relative size of the average gradient of the visible light image and the average gradient of the infrared image is used to determine whether the gray value is distributed approximately uniformly, so as to quickly and accurately determine whether the binarization strategy of the infrared image conforms to Gaussian For the case where the average gradient of the visible light image is greater than the average gradient of the infrared image, the visible light image is clear and contains most of the effective information, so the Gaussian method is used to obtain the binarized image to ensure that the infrared image can be binarized more accurately The image area where each target in the image is located is binarized into a white part, and after the target fusion area is determined based on the mask image, the luminance component fusion image obtained by fusing the luminance channel components separated from the visible light image and the infrared image according to the target fusion area In , the effective information in the image can be preserved more completely and comprehensively.
请参阅图8,为红外图像的灰度直方图呈大致均匀分布,分别采用三角形法、高斯法和大津法作为对应的二值化策略对红外图像进行二值化得到掩膜图像后,与可见光图像进行融合的对比示意图,其中,采用高斯法作为二值化策略确定二值化阈值后,对红外图像进行二值化处理得到的掩膜图像,目标轮廓更加清晰和突出,最终与可见光图像融合后得到的高斯法融合图像中图像有效信息损失最小。Please refer to Figure 8. The gray histogram of the infrared image is roughly evenly distributed. The triangle method, Gaussian method, and Otsu method are used as the corresponding binarization strategies to binarize the infrared image to obtain the mask image. Schematic diagram of the comparison of image fusion, in which, the mask image obtained by binarizing the infrared image after using the Gaussian method as the binarization strategy to determine the binarization threshold, the target outline is clearer and more prominent, and finally fused with the visible light image The Gaussian method fusion image obtained after the loss of image effective information is the smallest.
在一些实施例中,所述根据所述红外图像的灰度直方图和平均梯度的分布特性,确定匹配的二值化策略,还包括:In some embodiments, the determining a matching binarization strategy according to the distribution characteristics of the grayscale histogram and the average gradient of the infrared image further includes:
若所述第二平均梯度小于所述第一平均梯度,将所述红外图像分割为前景图像和背景图像;if the second average gradient is smaller than the first average gradient, segmenting the infrared image into a foreground image and a background image;
根据所述前景图像和所述背景图像的类间方差值,确定二值化阈值。A binarization threshold is determined according to the inter-class variance value of the foreground image and the background image.
若灰度值众数与灰度平均值之间的差值A-M大于预设值,则表示对应红外图像的灰度直方图不满足呈单峰分布的特征,通过确定红外图像的平均梯度和可见光图像的平均梯度得相对大小,以确定对应红外图像的灰度直方图是否呈较均匀地分布或呈双峰分布。为了便于区分,将红外图像的平均梯度称为第一 平均梯度,将可见光图像的平均梯度称为第二平均梯度。若第二平均梯度小于第一平均梯度,则表示灰度直方图呈双峰分布,与当前红外图像适用的二值化策略为大津法。通过大津法确定二值化阈值的原理是,按图像的灰度特性,将图像分成背景和目标两部分。背景和目标之间的类间方差越大,说明构成图像的两部分的差别越大,当部分目标错分为背景或部分背景错分为目标都会导致两部分之间的差别变小,如此,基于类间方差最大的二值化阈值分割可最大程度地减小错分概率。If the difference A-M between the mode of the gray value and the average value of the gray value is greater than the preset value, it means that the gray histogram corresponding to the infrared image does not meet the characteristics of a unimodal distribution. By determining the average gradient of the infrared image and the visible light The relative size of the average gradient of the image is used to determine whether the gray histogram corresponding to the infrared image is more evenly distributed or bimodal. In order to facilitate the distinction, the average gradient of the infrared image is called the first average gradient, and the average gradient of the visible light image is called the second average gradient. If the second average gradient is smaller than the first average gradient, it means that the gray histogram has a bimodal distribution, and the binarization strategy applicable to the current infrared image is the Otsu method. The principle of determining the binarization threshold by the Otsu method is to divide the image into two parts, the background and the target, according to the grayscale characteristics of the image. The greater the inter-class variance between the background and the target, the greater the difference between the two parts that make up the image. When part of the target is misclassified into the background or part of the background is misclassified into the target, the difference between the two parts will become smaller. Thus, The binarization threshold segmentation based on the maximum variance between classes can minimize the probability of misclassification.
上述实施例中,通过可见光图像的平均梯度和红外图像的平均梯度的相对大小,以判断灰度直方图是否呈双峰分布,实现快速且准确地判断红外图像的二值化策略是否符合大津法,通过大津法将所述红外图像分割为前景图像和背景图像,其中前景图像中包含主要信息,如人等其它发热目标物体向外辐射的能量,前景图像中包括主要信息的区域内像素和与之相邻像素的灰度值相差较大,从而对于红外图像的平均梯度大于可见光图像的平均梯度的情况,更适合根据红外图像的前景图像确定包含有效信息的区域,通过大津法获得红外图像的二值化图像,以确保对红外图像二值化后可更准确地将图像中的目标所在图像区域均二值化为白色部分,基于掩膜图像确定目标融合区域后,按照目标融合区域对可见光图像和红外图像分离出的亮度通道分量融合得到的亮度分量融合图像中,可更完整、全面地保留图像中的有效信息。In the above-mentioned embodiment, the relative size of the average gradient of the visible light image and the average gradient of the infrared image is used to judge whether the gray histogram is bimodal, so as to quickly and accurately judge whether the binarization strategy of the infrared image conforms to the Otsu method , the infrared image is divided into a foreground image and a background image by the Otsu method, wherein the foreground image contains main information, such as the energy radiated outward by other heat-generating target objects such as people, and the pixels in the area including the main information in the foreground image and the The difference between the gray values of adjacent pixels is large, so for the case where the average gradient of the infrared image is greater than the average gradient of the visible light image, it is more suitable to determine the area containing effective information according to the foreground image of the infrared image, and obtain the infrared image by the Otsu method. Binarize the image to ensure that the image area where the target in the image is located can be more accurately binarized into a white part after binarizing the infrared image. After determining the target fusion area based on the mask image, the visible light is processed according to the target fusion area. In the luminance component fusion image obtained by fusing the luminance channel components separated from the image and the infrared image, the effective information in the image can be retained more completely and comprehensively.
请参阅图9,为红外图像的灰度直方图呈双峰分布,采用三角形法、高斯法和大津法作为对应的二值化策略对红外图像进行二值化得到掩膜图像后,与可见光图像进行融合的对比示意图,其中,采用大津法作为二值化策略确定二值化阈值后,对红外图像进行二值化处理得到的掩膜图像,多目标的轮廓更加清晰和突出,最终与可见光图像融合后得到的大津法融合图像中图像有效信息损失最小。Please refer to Figure 9. The gray histogram of the infrared image shows a bimodal distribution. The triangle method, Gaussian method, and Otsu method are used as the corresponding binarization strategies to binarize the infrared image to obtain the mask image, and then compare it with the visible light image. Schematic diagram of the fusion comparison. After the Otsu method is used as the binarization strategy to determine the binarization threshold, the mask image obtained by binarizing the infrared image has clearer and more prominent outlines of multiple targets, and is finally compared with the visible light image. The loss of image effective information is the smallest in the fused image obtained by Otsu method after fusion.
在一些实施例中,所述将所述红外图像和所述可见光图像进行通道分离,对分离出的表征图像亮度的亮度通道分量根据所述目标融合区域进行融合,得到亮度通道融合图像,包括:In some embodiments, the channel separation of the infrared image and the visible light image is carried out, and the separated brightness channel component representing the brightness of the image is fused according to the target fusion area to obtain a brightness channel fusion image, including:
分别将所述红外图像和所述可见光图像进行HSI通道分离,根据所述目标融合区域对分离出的两个I通道分量按泊松图像编辑原理进行融合,得到I通道融合图像;The infrared image and the visible light image are respectively subjected to HSI channel separation, and the two separated I channel components are fused according to the Poisson image editing principle according to the target fusion area to obtain an I channel fusion image;
所述将所述亮度通道融合图像和所述可见光图像进行融合,得到融合图像,包括:The merging of the luminance channel fused image and the visible light image to obtain a fused image includes:
将所述I通道融合图像的I通道分量与所述可见光图像分离出的H通道分量、S通道分量合并,得到融合参考图像;Merging the I channel component of the I channel fused image with the H channel component and the S channel component separated from the visible light image to obtain a fused reference image;
将所述融合参考图像转换到RGB色彩空间,得到融合图像。Convert the fused reference image to RGB color space to obtain a fused image.
其中,HSI(Hue-Saturation-Intensity(Lightness))是指一个数字图像的颜色模型,HSI颜色模型用H、S、I三参数描述图像的颜色特性,H定义颜色的频率, 称为色调;S表示颜色的深浅程度,称为饱和度;I表示强度或亮度。Among them, HSI (Hue-Saturation-Intensity (Lightness)) refers to the color model of a digital image, the HSI color model describes the color characteristics of the image with H, S, I three parameters, H defines the frequency of the color, called hue; S Indicates the depth of color, known as saturation; I indicates intensity or brightness.
可选的,将所述红外图像和所述可见光图像分别进行HSI通道分离,对所述红外图像和所述可见光图像分别对应的所述目标融合区域分离出的两个I通道分量按泊松图像编辑原理进行融合,得到I通道融合图像包括:将可见光图像进行HSI通道分离,分离出可见光图像的H通道分量、S通道分量、I通道分量;根据所述红外图像的灰度直方图和平均梯度的分布特性,确定匹配的二值化策略,按照所述二值化策略对所述红外图像二值化得到掩膜图像,根据所述掩膜图像确定目标融合区域;将红外图像进行HSI通道分离,分离出红外图像的H通道分量、S通道分量、I通道分量;将红外图像的I通道分量和可见光图像的I通道分量按照掩膜图像所确定的目标融合区域按泊松图像编辑原理进行融合,得到I通道融合图像。其中,若红外图像和可见光图像为非HSI格式图像,则将所述红外图像和所述可见光图像进行HSI通道分离之前还包括,将红外图像和可见光图像转换为HSI格式的红外图像和可见光图像。Optionally, HSI channel separation is performed on the infrared image and the visible light image respectively, and the two I-channel components separated from the target fusion regions corresponding to the infrared image and the visible light image are separated according to the Poisson image The editing principle is used for fusion to obtain the I-channel fused image, which includes: separating the visible light image from the HSI channel, and separating the H-channel component, S-channel component, and I-channel component of the visible light image; according to the gray histogram and average gradient of the infrared image distribution characteristics, determine the matching binarization strategy, binarize the infrared image according to the binarization strategy to obtain a mask image, and determine the target fusion area according to the mask image; perform HSI channel separation on the infrared image , separate the H channel component, S channel component, and I channel component of the infrared image; fuse the I channel component of the infrared image and the I channel component of the visible light image according to the target fusion area determined by the mask image according to the Poisson image editing principle , to get the I channel fusion image. Wherein, if the infrared image and the visible light image are non-HSI format images, before performing HSI channel separation on the infrared image and the visible light image, it also includes converting the infrared image and the visible light image into an infrared image and a visible light image in HSI format.
上述实施例中,通过提取待融合的可见光图像的I通道分量和红外图像的I通道分量按掩膜图像所规定的目标融合区域进行融合,将得到的I通道融合图像与可见光图像分离出的H通道分量、S通道分量合并后,再转换到RGB色彩空间得到融合图像,如此,通过分别提取出可见光图像和红外图像的I通道分量、以及确定出目标融合区域内的I通道分量进行融合,既能确保得到的融合图像可保留可见光图像和红外图像中的有效信息,保留红外图像和可见光图像所包含的细节信息,同时可提高融合图像的客观评价指标,减小运算量,提高融合图像的处理效率。In the above-mentioned embodiment, by extracting the I channel component of the visible light image to be fused and the I channel component of the infrared image and performing fusion according to the target fusion area specified by the mask image, the obtained I channel fusion image and the visible light image are separated into H After the channel component and the S channel component are combined, they are converted to the RGB color space to obtain the fused image. In this way, by extracting the I channel component of the visible light image and the infrared image respectively, and determining the I channel component in the target fusion area for fusion, both It can ensure that the obtained fusion image can retain the effective information in the visible light image and the infrared image, retain the detailed information contained in the infrared image and the visible light image, and at the same time improve the objective evaluation index of the fusion image, reduce the amount of calculation, and improve the processing of the fusion image efficiency.
为了能够对本申请实施例提供的图像融合方法具有更加整体的理解,请结合参阅图10至图12,以一可选的示例为例,对所述图像融合方法进行说明。In order to have a more overall understanding of the image fusion method provided by the embodiment of the present application, please refer to FIG. 10 to FIG. 12 , and use an optional example as an example to describe the image fusion method.
S11,读取红外图像和可见光图像;如图11和图12所示,红外图像IR_1和可见光图像VIS_2;S11, read the infrared image and the visible light image; as shown in Figure 11 and Figure 12, the infrared image IR_1 and the visible light image VIS_2;
选择红外图像匹配的自适应阈值二值化策略,将红外图像二值化;自适应阈值二值化策略包括三角形法、高斯法和大津法;选择自适应二值化策略的方法包括:Select the adaptive threshold binarization strategy for infrared image matching to binarize the infrared image; the adaptive threshold binarization strategy includes the triangle method, the Gaussian method, and the Otsu method; the methods for selecting the adaptive binarization strategy include:
S121,根据红外图像的灰度直方图二值化分布特性,计算图像灰度值的灰度值众数与灰度平均值的差A-M值;S121, according to the grayscale histogram binarization distribution characteristics of the infrared image, calculate the difference A-M value between the grayscale value mode and the grayscale average value of the image grayscale value;
S122,判断A-M值是否大于预设值;S122, judging whether the A-M value is greater than a preset value;
S123,若A-M值小于或等于预设值,表示灰度直方图呈单峰分布,采用三角形法对红外图像进行二值化处理;S123, if the A-M value is less than or equal to the preset value, it means that the gray histogram is in a unimodal distribution, and the infrared image is binarized using the triangle method;
S124,若A-M值大于预设值,计算红外图像的平均梯度AG1和可见光图像的平均梯度AG2;S124, if the A-M value is greater than the preset value, calculate the average gradient AG1 of the infrared image and the average gradient AG2 of the visible light image;
S125,判断AG1与AG2的差是否大于0;S125, judging whether the difference between AG1 and AG2 is greater than 0;
S126,若AG1与AG2的差小于或等于0,表示灰度直方图呈较均匀地分布,采用高斯法对红外图像进行二值化处理;S126, if the difference between AG1 and AG2 is less than or equal to 0, it means that the gray histogram is more evenly distributed, and the infrared image is binarized by using the Gaussian method;
S127,若AG1与AG2的差大于0,表示灰度直方图呈双峰分布,采用大津法对红外图像进行二值化处理;S127, if the difference between AG1 and AG2 is greater than 0, it means that the gray histogram shows a bimodal distribution, and the infrared image is binarized by the Otsu method;
S13,根据二值化处理后的二值化图像生成掩膜图像,根据掩膜图像规定要融合区域;S13, generating a mask image according to the binarized image after binarization processing, and specifying an area to be fused according to the mask image;
S14,对红外图像进行HSI通道分离,得到红外图像的H通道分量、S通道分量和I通道分量;S14, performing HSI channel separation on the infrared image to obtain the H channel component, the S channel component and the I channel component of the infrared image;
S15,对可见光图像进行HSI通道分离,得到可见光图像的H通道分量、S通道分量和I通道分量;S15, performing HSI channel separation on the visible light image to obtain the H channel component, the S channel component and the I channel component of the visible light image;
S16,将红外图像和可见光图像的I通道分量根据掩膜图像所规定的要融合区域依泊松原理进行融合;S16, merging the I channel components of the infrared image and the visible light image according to the Poisson principle in the area to be fused according to the mask image;
S17,将融合后图像的I通道与可见光图像的H、S通道合并;S17, merging the I channel of the fused image with the H and S channels of the visible light image;
S18,将合并后得到的图像转换到RGB色彩空间,得到融合图像。S18, converting the merged image into an RGB color space to obtain a fusion image.
上述实施例所提供的图像融合方法,至少具备如下特点:The image fusion method provided by the foregoing embodiments has at least the following characteristics:
第一、通过对红外图像选用匹配的二值化策略进行二值化处理,根据二值化图像确定要融合区域;通过锁定要融合区域以减小融合计算量和提高融合处理效率,且有效保留图像中有效信息;First, by selecting a matching binarization strategy for the infrared image for binarization processing, the region to be fused is determined according to the binarized image; by locking the region to be fused, the fusion calculation amount is reduced and the fusion processing efficiency is improved, and the Effective information in the image;
第二、提供了如何根据红外图像的灰度直方图和平均梯度的分布特性,确定灰度直方图是否呈单峰分布、较为均匀地分布及双峰分布来选定匹配的二值化策略的方法,由此确保基于二值化结果确定要融合区域后,可保留红外图像和可见光图像所包含的有效信息和细节信息;Second, it provides instructions on how to determine whether the gray histogram is unimodal, relatively uniform, or bimodal according to the distribution characteristics of the gray histogram and average gradient of the infrared image to select a matching binarization strategy. method, thereby ensuring that the effective information and detailed information contained in the infrared image and the visible light image can be retained after the region to be fused is determined based on the binarization result;
第三、将红外图像和可见光图像的I通道分量进行融合后,与可见光图像的H、S通道分量进行合并来获得融合图像,可有效避免融合后图像包含无用信息而使得图像质量下降,减少图像中无效信息、减小计算量和复杂度、且可提升系统实时性,如图13所示,为采用本申请所述的图像融合方法对红外图像和可见光图像进行融合后得到的融合图像,图14采用已知的基于低秩表示原理对红外图像和可见光图像进行融合的融合对比效果、图15为采用已知的基于非下采样剪切波变换原理对红外图像和可见光图像进行融合的融合对比效果、图16为采用已知的基于非下采样轮廓波变换原理对红外图像和可见光图像进行融合的融合对比效果、图17为采用已知的基于泊松图像编辑原理对红外图像和可见光图像整体直接进行融合的融合对比效果。Third, after merging the I channel component of the infrared image and the visible light image, they are merged with the H and S channel components of the visible light image to obtain a fused image, which can effectively prevent the fused image from containing useless information and reduce the image quality and reduce the image quality. Invalid information, reducing the amount of calculation and complexity, and can improve the real-time performance of the system, as shown in Figure 13, which is the fused image obtained after using the image fusion method described in this application to fuse infrared images and visible light images, Figure 13 14 Fusion comparison effect of infrared image and visible light image fusion based on known low-rank representation principle, Figure 15 is fusion comparison of infrared image and visible light image fusion based on known non-subsampling shearlet transform principle Effect. Figure 16 shows the fusion and comparison effect of infrared images and visible light images based on the known principle of non-subsampling contourlet transformation. Fusion contrast effect for direct fusion.
图13至图17对应的融合图像的评价指标对比如下表一所示:The evaluation index comparison of the fused images corresponding to Figure 13 to Figure 17 is shown in Table 1 below:
Figure PCTCN2022094865-appb-000003
Figure PCTCN2022094865-appb-000003
Figure PCTCN2022094865-appb-000004
Figure PCTCN2022094865-appb-000004
其中,IE(Information Entropy)是指信息熵;SF(Spatial Frequency)是指空间频率;RMSE(Root Mean Sqaured Error)是指均方根误差;SSIM(Structural Similarity Index)是指结构相似性指数;TIME是指融合处理时长。NSST(Non-subsampled Shearlet Transform)是指非下采样剪切波变换;NSCT(Nonsubsampledcontourlet transform)是指非下采样轮廓波变换原理。结合图示及表一可知,采用本申请所述的图像融合方法对红外图像和可见光图像进行融合后得到的融合图像,均方根误差值最小、融合处理时长明显大幅减小、结构相似性指数接近于1、且信息熵和空间频率仍保持相对较大值,图像性能综合表现明显优于其它融合方法得到的融合图像。Among them, IE (Information Entropy) refers to information entropy; SF (Spatial Frequency) refers to spatial frequency; RMSE (Root Mean Sqaured Error) refers to root mean square error; SSIM (Structural Similarity Index) refers to structural similarity index; TIME It refers to the fusion processing time. NSST (Non-subsampled Shearlet Transform) refers to non-subsampled shearlet transform; NSCT (Nonsubsampled contourlet transform) refers to the principle of non-subsampled contourlet transform. Combining the illustrations and Table 1, it can be seen that the fused image obtained by using the image fusion method described in this application to fuse the infrared image and the visible light image has the smallest root mean square error value, the fusion processing time is significantly reduced, and the structural similarity index It is close to 1, and the information entropy and spatial frequency still maintain a relatively large value, and the overall performance of the image performance is obviously better than the fusion image obtained by other fusion methods.
请参阅图18,本申请另一方面,提供一种图像融合装置,包括:获取模块131,用于获取针对目标视场同步采集的可见光图像和红外图像;融合区域确定模块132,用于对所述红外图像二值化得到掩膜图像,根据所述掩膜图像确定目标融合区域;融合模块134,用于将所述红外图像基于所述目标融合区域与所述可见光图像进行融合,得到融合图像。Please refer to FIG. 18 , another aspect of the present application provides an image fusion device, including: an acquisition module 131, used to acquire visible light images and infrared images synchronously collected for the target field of view; a fusion area determination module 132, used for all Binarize the infrared image to obtain a mask image, and determine the target fusion area according to the mask image; the fusion module 134 is used to fuse the infrared image based on the target fusion area and the visible light image to obtain a fusion image .
可选的,所述融合模块134,具体用于将所述红外图像和所述可见光图像进行通道分离,对分离出的表征图像亮度的亮度通道分量根据所述目标融合区域进行融合,得到亮度通道融合图像;将所述亮度通道融合图像和所述可见光图像进行融合,得到融合图像。Optionally, the fusion module 134 is specifically configured to separate the channels of the infrared image and the visible light image, and fuse the separated brightness channel component representing the brightness of the image according to the target fusion area to obtain the brightness channel Fusing images: fusing the luminance channel fusion image and the visible light image to obtain a fusion image.
可选的,所述融合区域确定模块132,具体用于将所述红外图像中各像素点的灰度值与二值化阈值进行比较,灰度值小于所述二值化阈值的像素点的灰度值置第一设定值,灰度值大于或等于所述二值化阈值的像素点的灰度值置第二设定值,得到掩膜图像;选择所述掩膜图像中所述第二设定值的像素点分布区域的至少一部分作为目标融合区域。Optionally, the fusion region determination module 132 is specifically configured to compare the grayscale value of each pixel in the infrared image with a binarization threshold, and the grayscale value of a pixel with a grayscale value smaller than the binarization threshold is The grayscale value is set to the first set value, and the grayscale value of the pixel point whose grayscale value is greater than or equal to the binarization threshold is set to the second set value to obtain a mask image; select the mask image described in At least a part of the pixel point distribution area of the second set value is used as the target fusion area.
可选的,所述融合区域确定模块132,还用于根据所述红外图像的灰度直方图和平均梯度的分布特性,确定匹配的二值化策略;按照所述二值化策略确定二值化阈值。Optionally, the fusion region determination module 132 is further configured to determine a matching binarization strategy according to the distribution characteristics of the gray histogram and the average gradient of the infrared image; determine the binarization strategy according to the binarization strategy threshold.
可选的,所述融合区域确定模块132,还用于根据所述红外图像的灰度直方图的分布特性,判断所述灰度直方图是否呈单峰分布;若是,确定匹配的二值化策略为三角形法;若否,根据所述红外图像的平均梯度与所述可见光图像的平均梯度的对比结果,确定匹配的二值化策略为高斯法或大津法。Optionally, the fusion region determination module 132 is further configured to judge whether the gray histogram is a unimodal distribution according to the distribution characteristics of the gray histogram of the infrared image; if so, determine the matching binarization The strategy is the triangle method; if not, according to the comparison result of the average gradient of the infrared image and the average gradient of the visible light image, it is determined that the matching binarization strategy is the Gaussian method or the Otsu method.
可选的,所述融合区域确定模块132,还用于确定所述红外图像的灰度值众数与灰度平均值之间的差值,若所述差值小于或等于预设值,确定所述红外 图像的灰度直方图呈单峰分布;以所述灰度直方图中最大波峰为顶点确定三角形;通过所述三角形确定最大直线距离,根据所述最大直线距离对应的直方图灰度等级确定二值化阈值。Optionally, the fusion region determination module 132 is also used to determine the difference between the mode of the gray value of the infrared image and the average value of the gray value, if the difference is less than or equal to a preset value, determine The grayscale histogram of the infrared image is in a unimodal distribution; a triangle is determined with the largest peak in the grayscale histogram as the apex; the maximum straight-line distance is determined through the triangle, and the grayscale of the histogram corresponding to the maximum straight-line distance Level determines the binarization threshold.
可选的,所述融合区域确定模块132,还用于若所述差值大于所述预设值,确定所述红外图像的第一平均梯度和所述可见光图像的第二平均梯度;若所述第二平均梯度大于或等于所述第一平均梯度,计算目标窗函数内所述红外图像的灰度值的高斯均值,根据所述高斯均值确定二值化阈值。Optionally, the fusion region determining module 132 is further configured to determine the first average gradient of the infrared image and the second average gradient of the visible light image if the difference is greater than the preset value; The second average gradient is greater than or equal to the first average gradient, the Gaussian mean value of the gray value of the infrared image in the target window function is calculated, and the binarization threshold is determined according to the Gaussian mean value.
可选的,所述融合区域确定模块132,还用于若所述第二平均梯度小于所述第一平均梯度,将所述红外图像分割为前景图像和背景图像;根据所述前景图像和所述背景图像的类间方差值,确定二值化阈值。Optionally, the fusion region determination module 132 is further configured to segment the infrared image into a foreground image and a background image if the second average gradient is smaller than the first average gradient; The inter-class variance value of the background image is used to determine the binarization threshold.
可选的,所述融合模块134,还用于分别将所述红外图像和所述可见光图像进行HSI通道分离,根据所述目标融合区域对分离出的两个I通道分量按泊松图像编辑原理进行融合,得到I通道融合图像;将所述I通道融合图像的I通道分量与所述可见光图像分离出的H、S通道分量合并,得到融合参考图像;将所述融合参考图像转换到RGB色彩空间,得到融合图像。Optionally, the fusion module 134 is further configured to separately perform HSI channel separation on the infrared image and the visible light image, and edit the two separated I-channel components according to the Poisson image editing principle according to the target fusion area. Carrying out fusion to obtain an I channel fusion image; merging the I channel component of the I channel fusion image with the H and S channel components separated from the visible light image to obtain a fusion reference image; converting the fusion reference image to RGB color space to obtain a fused image.
需要说明的是:上述实施例提供的图像融合装置在实现可见光图像和红外图像融合处理过程中,仅以上述各程序模块的划分进行举例说明,在实际应用中,可以根据需要而将上述处理分配由不同的程序模块完成,即可将装置的内部结构划分成不同的程序模块,以完成以上描述的全部或者部分方法步骤。另外,上述实施例提供的图像融合装置与图像融合方法实施例属于同一构思,其具体实现过程详见方法实施例,这里不再赘述。It should be noted that: in the image fusion device provided by the above-mentioned embodiments, in the process of realizing fusion processing of visible light images and infrared images, the division of the above-mentioned program modules is used as an example for illustration. In practical applications, the above-mentioned processing can be allocated according to needs. Completed by different program modules, that is, the internal structure of the device can be divided into different program modules to complete all or part of the method steps described above. In addition, the image fusion device and the image fusion method embodiments provided in the above embodiments belong to the same idea, and the specific implementation process thereof is detailed in the method embodiments, and will not be repeated here.
本申请另一方面提供一种图像处理设备,请参阅图19,为本申请实施例提供的图像处理设备的一个可选的硬件结构示意图,所述图像处理设备包括处理器111、与所述处理器111连接的存储器112,存储器112内用于存储各种类别的数据以支持图像处理设备的操作,且存储有用于实现本申请任一实施例提供的图像处理方法的计算机程序,所述计算机程序被所述处理器执行时,实现本申请任一实施例提供的图像处理方法的步骤,且能达到相同的技术效果,为避免重复,这里不再赘述。Another aspect of the present application provides an image processing device. Please refer to FIG. 19 , which is a schematic diagram of an optional hardware structure of the image processing device provided by the embodiment of the present application. The image processing device includes a processor 111, and the processing The memory 112 connected to the device 111, the memory 112 is used to store various types of data to support the operation of the image processing equipment, and stores a computer program for realizing the image processing method provided by any embodiment of the present application, the computer program When executed by the processor, the steps of the image processing method provided by any embodiment of the present application can be realized, and the same technical effect can be achieved. To avoid repetition, details are not repeated here.
可选的,所述图像处理设备还包括与所述处理器111连接的红外拍摄模块和可见光拍摄模块,所述红外拍摄模块和可见光拍摄模块用于同步针对同一目标视场拍摄红外图像和可见光图像作为待融合图像发送给所述处理器111。Optionally, the image processing device further includes an infrared shooting module and a visible light shooting module connected to the processor 111, and the infrared shooting module and the visible light shooting module are used for synchronously shooting infrared images and visible light images for the same target field of view It is sent to the processor 111 as an image to be fused.
本申请实施例还提供一种计算机可读存储介质,所述计算机可读存储介质上存储有计算机程序,该计算机程序被处理器执行时实现上述图像处理方法实施例的各个过程,且能达到相同的技术效果,为避免重复,这里不再赘述。其中,所述的计算机可读存储介质,如只读存储器(Read-OnlyMemory,简称ROM)、随机存取存储器(RandomAccessMemory,简称RAM)、磁碟或者光盘等。The embodiment of the present application also provides a computer-readable storage medium, where a computer program is stored on the computer-readable storage medium. When the computer program is executed by a processor, each process of the above-mentioned image processing method embodiment is realized, and the same To avoid repetition, the technical effects will not be repeated here. Wherein, the computer-readable storage medium is, for example, a read-only memory (Read-Only Memory, ROM for short), a random access memory (Random Access Memory, RAM for short), a magnetic disk or an optical disk, and the like.
需要说明的是,在本文中,术语“包括”、“包含”或者其任何其他变体意在 涵盖非排他性的包含,从而使得包括一系列要素的过程、方法、物品或者装置不仅包括那些要素,而且还包括没有明确列出的其他要素,或者是还包括为这种过程、方法、物品或者装置所固有的要素。在没有更多限制的情况下,由语句“包括一个……”限定的要素,并不排除在包括该要素的过程、方法、物品或者装置中还存在另外的相同要素。It should be noted that, in this document, the term "comprising", "comprising" or any other variation thereof is intended to cover a non-exclusive inclusion such that a process, method, article or apparatus comprising a set of elements includes not only those elements, It also includes other elements not expressly listed, or elements inherent in the process, method, article, or device. Without further limitations, an element defined by the phrase "comprising a ..." does not preclude the presence of additional identical elements in the process, method, article, or apparatus comprising that element.
通过以上的实施方式的描述,本领域的技术人员可以清楚地了解到上述实施例方法可借助软件加必需的通用硬件平台的方式来实现,当然也可以通过硬件,但很多情况下前者是更佳的实施方式。基于这样的理解,本发明的技术方案本质上或者说对现有技术做出贡献的部分可以软件产品的形式体现出来,该计算机软件产品存储在一个存储介质(如ROM/RAM、磁碟、光盘)中,包括若干指令用以使得一台终端(可以是手机,计算机,服务器,或网络设备等)执行本发明各个实施例所述的方法。Through the description of the above embodiments, those skilled in the art can clearly understand that the methods of the above embodiments can be implemented by means of software plus a necessary general-purpose hardware platform, and of course also by hardware, but in many cases the former is better implementation. Based on such an understanding, the essence of the technical solution of the present invention or the part that contributes to the prior art can be embodied in the form of software products, and the computer software products are stored in a storage medium (such as ROM/RAM, magnetic disk, optical disk, etc.) ) includes several instructions to make a terminal (which may be a mobile phone, a computer, a server, or a network device, etc.) execute the methods described in various embodiments of the present invention.
本领域普通技术人员可以理解的,实现上述实施例方法中的全部或部分流程,是可以通过计算机程序来指令相关的硬件来完成,所述的程序可存储于一非易失性计算机可读取存储介质中,该程序在执行时,可包括如上述各方法的实施例的流程。其中,本申请各实施例中所使用的对存储器、存储、数据库或其它介质的任何引用,均可包括非易失性和/或易失性存储器。非易失性存储器可包括只读存储器(ROM)、可编程ROM(PROM)、电可编程ROM(EPROM)、电可擦除可编程ROM(EEPROM)或闪存。易失性存储器可包括随机存取存储器(RAM)或者外部高速缓冲存储器。作为说明而非局限,RAM以多种形式可得,诸如静态RAM(SRAM)、动态RAM(DRAM)、同步DRAM(SDRAM)、双数据率SDRAM(DDRSDRAM)、增强型SDRAM(ESDRAM)、同步链路(Synchlink)DRAM(SLDRAM)、存储器总线(Rambus)直接RAM(RDRAM)、直接存储器总线动态RAM(DRDRAM)、以及存储器总线动态RAM(RDRAM)等。Those of ordinary skill in the art can understand that all or part of the processes in the methods of the above-mentioned embodiments can be completed by instructing related hardware through computer programs, and the programs can be stored in a non-volatile computer-readable In the storage medium, when the program is executed, it may include the processes of the embodiments of the above-mentioned methods. Wherein, any reference to memory, storage, database or other media used in various embodiments of the present application may include non-volatile and/or volatile memory. Nonvolatile memory can include read only memory (ROM), programmable ROM (PROM), electrically programmable ROM (EPROM), electrically erasable programmable ROM (EEPROM), or flash memory. Volatile memory can include random access memory (RAM) or external cache memory. By way of illustration and not limitation, RAM is available in many forms such as Static RAM (SRAM), Dynamic RAM (DRAM), Synchronous DRAM (SDRAM), Double Data Rate SDRAM (DDRSDRAM), Enhanced SDRAM (ESDRAM), Synchronous Chain Synchlink DRAM (SLDRAM), memory bus (Rambus) direct RAM (RDRAM), direct memory bus dynamic RAM (DRDRAM), and memory bus dynamic RAM (RDRAM), etc.
以上所述,仅为本发明的具体实施方式,但本发明的保护范围并不局限于此,任何熟悉本技术领域的技术人员在本发明揭露的技术范围内,可轻易想到变化或替换,都应涵盖在本发明的保护范围之内。本发明的保护范围应以所述权利要求的保护范围以准。The above is only a specific embodiment of the present invention, but the scope of protection of the present invention is not limited thereto. Anyone skilled in the art can easily think of changes or substitutions within the technical scope disclosed in the present invention. Should be covered within the protection scope of the present invention. The protection scope of the present invention shall be determined by the protection scope of the claims.

Claims (12)

  1. 一种图像融合方法,应用于图像处理设备,其特征在于,包括:An image fusion method applied to an image processing device, characterized in that it comprises:
    获取针对目标视场同步采集的可见光图像和红外图像;Acquire visible light images and infrared images synchronously collected for the target field of view;
    对所述红外图像二值化得到掩膜图像,根据所述掩膜图像确定目标融合区域;Binarizing the infrared image to obtain a mask image, and determining a target fusion area according to the mask image;
    将所述红外图像基于所述目标融合区域与所述可见光图像进行融合,得到融合图像。The infrared image is fused with the visible light image based on the target fusion area to obtain a fused image.
  2. 如权利要求1所述的图像融合方法,其特征在于,所述将所述红外图像基于所述目标融合区域与所述可见光图像进行融合,得到融合图像,包括:The image fusion method according to claim 1, wherein the fusion of the infrared image based on the target fusion area and the visible light image to obtain a fusion image comprises:
    分别将所述红外图像和所述可见光图像进行通道分离,对分离出的表征图像亮度的两个亮度通道分量根据所述目标融合区域进行融合,得到亮度通道融合图像;Channel-separating the infrared image and the visible light image respectively, and fusing the separated two brightness channel components representing image brightness according to the target fusion area to obtain a brightness channel fusion image;
    将所述亮度通道融合图像和所述可见光图像进行融合,得到融合图像。The fused image of the brightness channel and the visible light image are fused to obtain a fused image.
  3. 如权利要求1所述的图像融合方法,其特征在于,所述对所述红外图像二值化得到掩膜图像,根据所述掩膜图像确定目标融合区域,包括:The image fusion method according to claim 1, wherein said binarizing said infrared image to obtain a mask image, and determining a target fusion area according to said mask image comprises:
    将所述红外图像中各像素点的灰度值与二值化阈值进行比较,灰度值小于所述二值化阈值的像素点的灰度值置第一设定值,灰度值大于或等于所述二值化阈值的像素点的灰度值置第二设定值,得到掩膜图像;Comparing the gray value of each pixel in the infrared image with the binarization threshold, the gray value of the pixel whose gray value is less than the binarization threshold is set to the first set value, and the gray value is greater than or The grayscale value of the pixel point equal to the binarization threshold is set to a second set value to obtain a mask image;
    选择所述掩膜图像中所述第二设定值的像素点分布区域的至少一部分作为目标融合区域。Selecting at least a part of the pixel point distribution area of the second set value in the mask image as a target fusion area.
  4. 如权利要求3所述的图像融合方法,其特征在于,所述将所述红外图像中各像素点的灰度值与二值化阈值进行比较之前,包括:The image fusion method according to claim 3, wherein before comparing the gray value of each pixel in the infrared image with the binarization threshold, it includes:
    根据所述红外图像的灰度直方图和平均梯度的分布特性,确定匹配的二值化策略;According to the gray histogram of the infrared image and the distribution characteristics of the average gradient, determine a matching binarization strategy;
    按照所述二值化策略确定所述二值化阈值。The binarization threshold is determined according to the binarization strategy.
  5. 如权利要求3所述的图像融合方法,其特征在于,所述根据所述红外图像的灰度直方图和平均梯度的分布特性,确定匹配的二值化策略,包括:The image fusion method according to claim 3, wherein, determining a matching binarization strategy according to the distribution characteristics of the gray histogram and the average gradient of the infrared image includes:
    根据所述红外图像的灰度直方图的分布特性,判断所述灰度直方图是否呈单峰分布;According to the distribution characteristics of the gray histogram of the infrared image, it is judged whether the gray histogram is in a unimodal distribution;
    若是,确定匹配的二值化策略为三角形法;If so, determine that the matching binarization strategy is the triangle method;
    若否,根据所述红外图像的平均梯度与所述可见光图像的平均梯度的对比结果,确定匹配的二值化策略为高斯法或大津法。If not, according to the comparison result of the average gradient of the infrared image and the average gradient of the visible light image, it is determined that the matching binarization strategy is the Gaussian method or the Otsu method.
  6. 如权利要求4所述的图像融合方法,其特征在于,所述根据所述红外图像的灰度直方图和平均梯度的分布特性,确定匹配的二值 化策略,包括:The image fusion method according to claim 4, wherein, the gray histogram according to the infrared image and the distribution characteristics of the average gradient determine the matching binarization strategy, comprising:
    确定所述红外图像的灰度值众数与灰度平均值之间的差值,若所述差值小于或等于预设值,确定所述红外图像的灰度直方图呈单峰分布,则确定匹配的二值化策略为三角形法;Determining the difference between the mode of the gray value of the infrared image and the average value of the gray value, if the difference is less than or equal to a preset value, it is determined that the gray histogram of the infrared image is in a unimodal distribution, then Determine that the matching binarization strategy is the triangle method;
    此时,所述按照所述二值化策略确定所述二值化阈值具体为:At this time, the determination of the binarization threshold according to the binarization strategy is specifically:
    以所述灰度直方图中最大波峰为顶点确定三角形;Determining a triangle with the largest peak in the grayscale histogram as the apex;
    通过所述三角形确定最大直线距离,根据所述最大直线距离对应的直方图灰度等级确定二值化阈值。The maximum straight-line distance is determined through the triangle, and the binarization threshold is determined according to the gray level of the histogram corresponding to the maximum straight-line distance.
  7. 如权利要求6所述的图像融合方法,其特征在于,所述根据所述红外图像的灰度直方图和平均梯度的分布特性,确定匹配的二值化策略,还包括:The image fusion method according to claim 6, wherein, determining a matching binarization strategy according to the distribution characteristics of the gray histogram and the average gradient of the infrared image, further comprising:
    若所述差值大于所述预设值,确定所述红外图像的第一平均梯度和所述可见光图像的第二平均梯度;If the difference is greater than the preset value, determine the first average gradient of the infrared image and the second average gradient of the visible light image;
    若所述第二平均梯度大于或等于所述第一平均梯度,则确定匹配的二值化策略为高斯法;If the second average gradient is greater than or equal to the first average gradient, then determine that the matching binarization strategy is the Gaussian method;
    此时,所述按照所述二值化策略确定所述二值化阈值具体为:计算目标窗函数内所述红外图像的灰度值的高斯均值,根据所述高斯均值确定二值化阈值。In this case, the determining the binarization threshold according to the binarization strategy specifically includes: calculating a Gaussian mean of the grayscale values of the infrared image within the target window function, and determining the binarization threshold according to the Gaussian mean.
  8. 如权利要求7所述的图像融合方法,其特征在于,所述根据所述红外图像的灰度直方图和平均梯度的分布特性,确定匹配的二值化策略,还包括:The image fusion method according to claim 7, wherein, determining a matching binarization strategy according to the distribution characteristics of the gray histogram and the average gradient of the infrared image, further comprising:
    若所述第二平均梯度小于所述第一平均梯度,则确定匹配的二值化策略为大津法;If the second average gradient is smaller than the first average gradient, it is determined that the matching binarization strategy is the Otsu method;
    此时,所述按照所述二值化策略确定所述二值化阈值具体为:At this time, the determination of the binarization threshold according to the binarization strategy is specifically:
    将所述红外图像分割为前景图像和背景图像;Segmenting the infrared image into a foreground image and a background image;
    根据所述前景图像和所述背景图像的类间方差值,确定二值化阈值。A binarization threshold is determined according to the inter-class variance value of the foreground image and the background image.
  9. 如权利要求2所述的图像融合方法,其特征在于,所述将所述红外图像和所述可见光图像进行通道分离,对分离出的表征图像亮度的亮度通道分量根据所述目标融合区域进行融合,得到亮度通道融合图像,包括:The image fusion method according to claim 2, wherein the channel separation of the infrared image and the visible light image is carried out, and the separated brightness channel component representing the brightness of the image is fused according to the target fusion area , to obtain the luminance channel fusion image, including:
    分别将所述红外图像和所述可见光图像进行HSI通道分离,根据所述目标融合区域对分离出的两个I通道分量按泊松图像编辑原理进行融合,得到I通道融合图像;The infrared image and the visible light image are respectively subjected to HSI channel separation, and the two separated I channel components are fused according to the Poisson image editing principle according to the target fusion area to obtain an I channel fusion image;
    所述将所述亮度通道融合图像和所述可见光图像进行融合,得到融合图像,包括:The merging of the luminance channel fused image and the visible light image to obtain a fused image includes:
    将所述I通道融合图像的I通道分量与所述可见光图像分离出的H通道分量、S通道分量合并,得到融合参考图像;Merging the I channel component of the I channel fused image with the H channel component and the S channel component separated from the visible light image to obtain a fused reference image;
    将所述融合参考图像转换到RGB色彩空间,得到融合图像。Convert the fused reference image to RGB color space to obtain a fused image.
  10. 一种图像融合装置,其特征在于,包括:An image fusion device, characterized in that it comprises:
    获取模块,用于获取针对目标视场同步采集的可见光图像和红外 图像;Obtaining module, be used for obtaining the visible light image and the infrared image that are collected synchronously for target field of view;
    融合区域确定模块,用于对所述红外图像二值化得到掩膜图像,根据所述掩膜图像确定目标融合区域;A fusion area determination module, configured to binarize the infrared image to obtain a mask image, and determine a target fusion area according to the mask image;
    融合模块,用于将所述红外图像基于所述目标融合区域与所述可见光图像进行融合,得到融合图像。A fusion module, configured to fuse the infrared image with the visible light image based on the target fusion area to obtain a fusion image.
  11. 一种图像处理设备,其特征在于,包括处理器、与所述处理器连接的存储器及存储在所述存储器上并可被所述处理器执行的计算机程序,所述计算机程序被所述处理器执行时实现如权利要求1至9中任一项所述的图像融合方法。An image processing device, characterized by comprising a processor, a memory connected to the processor, and a computer program stored on the memory and executable by the processor, the computer program being executed by the processor When executed, the image fusion method according to any one of claims 1 to 9 is realized.
  12. 一种计算机可读存储介质,其特征在于,所述计算机可读存储介质上存储有计算机程序,所述计算机程序被处理器执行时实现如权利要求1至9中任一项所述的图像融合方法。A computer-readable storage medium, characterized in that a computer program is stored on the computer-readable storage medium, and when the computer program is executed by a processor, the image fusion according to any one of claims 1 to 9 is realized method.
PCT/CN2022/094865 2022-02-21 2022-05-25 Image fusion method and apparatus, device and storage medium WO2023155324A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202210157058.1A CN114519808A (en) 2022-02-21 2022-02-21 Image fusion method, device and equipment and storage medium
CN202210157058.1 2022-02-21

Publications (1)

Publication Number Publication Date
WO2023155324A1 true WO2023155324A1 (en) 2023-08-24

Family

ID=81598755

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/094865 WO2023155324A1 (en) 2022-02-21 2022-05-25 Image fusion method and apparatus, device and storage medium

Country Status (2)

Country Link
CN (1) CN114519808A (en)
WO (1) WO2023155324A1 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117773405A (en) * 2024-02-28 2024-03-29 茌平鲁环汽车散热器有限公司 Method for detecting brazing quality of automobile radiator
CN117876836A (en) * 2024-03-11 2024-04-12 齐鲁工业大学(山东省科学院) Image fusion method based on multi-scale feature extraction and target reconstruction
CN117893525A (en) * 2024-02-28 2024-04-16 广州威睛光学科技有限公司 Chip hot spot detection method based on infrared thermal imaging

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114519808A (en) * 2022-02-21 2022-05-20 烟台艾睿光电科技有限公司 Image fusion method, device and equipment and storage medium
CN115278016A (en) * 2022-07-25 2022-11-01 烟台艾睿光电科技有限公司 Infrared intelligent shooting method and device, infrared thermal imaging equipment and medium
CN115170810B (en) * 2022-09-08 2022-12-13 南京理工大学 Visible light infrared image fusion target detection example segmentation method
CN116433695B (en) * 2023-06-13 2023-08-22 天津市第五中心医院 Mammary gland region extraction method and system of mammary gland molybdenum target image
CN116977154B (en) * 2023-09-22 2024-03-19 南方电网数字电网研究院有限公司 Visible light image and infrared image fusion storage method, device, equipment and medium

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1545064A (en) * 2003-11-27 2004-11-10 上海交通大学 Infrared and visible light image merging method
CN103778618A (en) * 2013-11-04 2014-05-07 国家电网公司 Method for fusing visible image and infrared image
CN108665443A (en) * 2018-04-11 2018-10-16 中国石油大学(北京) A kind of the infrared image sensitizing range extracting method and device of mechanical equipment fault
KR20200102907A (en) * 2019-11-12 2020-09-01 써모아이 주식회사 Method and apparatus for object recognition based on visible light and infrared fusion image
CN112102340A (en) * 2020-09-25 2020-12-18 Oppo广东移动通信有限公司 Image processing method, image processing device, electronic equipment and computer readable storage medium
CN112767289A (en) * 2019-10-21 2021-05-07 浙江宇视科技有限公司 Image fusion method, device, medium and electronic equipment
CN114519808A (en) * 2022-02-21 2022-05-20 烟台艾睿光电科技有限公司 Image fusion method, device and equipment and storage medium

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1545064A (en) * 2003-11-27 2004-11-10 上海交通大学 Infrared and visible light image merging method
CN103778618A (en) * 2013-11-04 2014-05-07 国家电网公司 Method for fusing visible image and infrared image
CN108665443A (en) * 2018-04-11 2018-10-16 中国石油大学(北京) A kind of the infrared image sensitizing range extracting method and device of mechanical equipment fault
CN112767289A (en) * 2019-10-21 2021-05-07 浙江宇视科技有限公司 Image fusion method, device, medium and electronic equipment
KR20200102907A (en) * 2019-11-12 2020-09-01 써모아이 주식회사 Method and apparatus for object recognition based on visible light and infrared fusion image
CN112102340A (en) * 2020-09-25 2020-12-18 Oppo广东移动通信有限公司 Image processing method, image processing device, electronic equipment and computer readable storage medium
CN114519808A (en) * 2022-02-21 2022-05-20 烟台艾睿光电科技有限公司 Image fusion method, device and equipment and storage medium

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117773405A (en) * 2024-02-28 2024-03-29 茌平鲁环汽车散热器有限公司 Method for detecting brazing quality of automobile radiator
CN117893525A (en) * 2024-02-28 2024-04-16 广州威睛光学科技有限公司 Chip hot spot detection method based on infrared thermal imaging
CN117773405B (en) * 2024-02-28 2024-05-14 茌平鲁环汽车散热器有限公司 Method for detecting brazing quality of automobile radiator
CN117893525B (en) * 2024-02-28 2024-06-11 广州威睛光学科技有限公司 Chip hot spot detection method based on infrared thermal imaging
CN117876836A (en) * 2024-03-11 2024-04-12 齐鲁工业大学(山东省科学院) Image fusion method based on multi-scale feature extraction and target reconstruction
CN117876836B (en) * 2024-03-11 2024-05-24 齐鲁工业大学(山东省科学院) Image fusion method based on multi-scale feature extraction and target reconstruction

Also Published As

Publication number Publication date
CN114519808A (en) 2022-05-20

Similar Documents

Publication Publication Date Title
WO2023155324A1 (en) Image fusion method and apparatus, device and storage medium
CN107610114B (en) optical satellite remote sensing image cloud and snow fog detection method based on support vector machine
Kumar et al. Review on image segmentation techniques
CN111104943B (en) Color image region-of-interest extraction method based on decision-level fusion
Ajmal et al. A comparison of RGB and HSV colour spaces for visual attention models
CN108898132B (en) Terahertz image dangerous article identification method based on shape context description
US11238301B2 (en) Computer-implemented method of detecting foreign object on background object in an image, apparatus for detecting foreign object on background object in an image, and computer-program product
CN108320294B (en) Intelligent full-automatic portrait background replacement method for second-generation identity card photos
CN113792827B (en) Target object recognition method, electronic device, and computer-readable storage medium
CN111695373B (en) Zebra stripes positioning method, system, medium and equipment
CN108830857A (en) A kind of adaptive Chinese character rubbings image binaryzation partitioning algorithm
CN111489330A (en) Weak and small target detection method based on multi-source information fusion
CN107239761B (en) Fruit tree branch pulling effect evaluation method based on skeleton angular point detection
CN107704864B (en) Salient object detection method based on image object semantic detection
KR100488014B1 (en) YCrCb color based human face location detection method
JP2001167273A (en) Method and device for detecting face and computer readable medium
Ying et al. Region-aware RGB and near-infrared image fusion
Khan et al. Shadow removal from digital images using multi-channel binarization and shadow matting
Jeong et al. Homogeneity patch search method for voting-based efficient vehicle color classification using front-of-vehicle image
Storcz et al. Histogram based segmentation of shadowed leaf images
Liu et al. A Fusion-based Enhancement Method for Low-light UAV Images
Mubin et al. Identification of parking lot status using circle blob detection
Entuni et al. Severity estimation of plant leaf diseases using segmentation method
Long et al. An Efficient Method For Dark License Plate Detection
AU2021101444A4 (en) A fruit image identification system and a method thereof

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22926635

Country of ref document: EP

Kind code of ref document: A1