CN110991547A - Image significance detection method based on multi-feature optimal fusion - Google Patents

Image significance detection method based on multi-feature optimal fusion Download PDF

Info

Publication number
CN110991547A
CN110991547A CN201911276766.1A CN201911276766A CN110991547A CN 110991547 A CN110991547 A CN 110991547A CN 201911276766 A CN201911276766 A CN 201911276766A CN 110991547 A CN110991547 A CN 110991547A
Authority
CN
China
Prior art keywords
pixel
image
saliency
super
color
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201911276766.1A
Other languages
Chinese (zh)
Inventor
李建平
顾小丰
胡健
王青松
蒋涛
陈强强
贺喜
李天凯
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
University of Electronic Science and Technology of China
Original Assignee
University of Electronic Science and Technology of China
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by University of Electronic Science and Technology of China filed Critical University of Electronic Science and Technology of China
Priority to CN201911276766.1A priority Critical patent/CN110991547A/en
Publication of CN110991547A publication Critical patent/CN110991547A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/253Fusion techniques of extracted features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F17/00Digital computing or data processing equipment or methods, specially adapted for specific functions
    • G06F17/10Complex mathematical operations
    • G06F17/14Fourier, Walsh or analogous domain transformations, e.g. Laplace, Hilbert, Karhunen-Loeve, transforms
    • G06F17/141Discrete Fourier transforms
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2411Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on the proximity to a decision surface, e.g. support vector machines
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/50Extraction of image or video features by performing operations within image blocks; by using histograms, e.g. histogram of oriented gradients [HoG]; by summing image-intensity values; Projection analysis
    • G06V10/507Summing image-intensity values; Histogram projection analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/56Extraction of image or video features relating to colour

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Mathematical Physics (AREA)
  • General Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Computation (AREA)
  • Computational Mathematics (AREA)
  • Multimedia (AREA)
  • Mathematical Analysis (AREA)
  • Mathematical Optimization (AREA)
  • Pure & Applied Mathematics (AREA)
  • Discrete Mathematics (AREA)
  • Algebra (AREA)
  • Databases & Information Systems (AREA)
  • Software Systems (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses an image saliency detection method based on multi-feature optimal fusion, aiming at the problem that the existing saliency detection algorithm only selects the bottom layer features of color, texture, direction and the like of an image in a space domain or only analyzes the singleness of a problem from the angle of a frequency domain, and the saliency of an object is calculated by selecting the features which can show the saliency of the object through comparison between the space domain and the frequency domain; the method can set different weights for various characteristics according to the importance degree of the various characteristics of the image to the significance detection by learning the training data through the support vector machine; the invention provides a saliency detection algorithm capable of obtaining a more accurate and clear saliency map so as to allocate more computer resources to the extracted saliency areas in various computer vision field tasks and efficiently process various vision field tasks.

Description

Image significance detection method based on multi-feature optimal fusion
Technical Field
The invention belongs to the technical field of saliency detection, and particularly relates to an image saliency detection method based on multi-feature optimal fusion.
Background
In recent decades, the field of significance detection has been developed vigorously, and many significance detection models have been proposed by scholars. The main idea of these models is based on a feature fusion theory and a visual attention theory, and through a Center-periphery (Center-around) mechanism, color, brightness and direction features of an image are calculated to simulate a bottom-up visual attention mechanism of a human visual system, and a saliency detection model is established by calculating contrast of color features.
These algorithms generally work well by mimicking the mechanism by which the human visual system can quickly locate objects of interest, but there is still some room for improvement.
However, at present, the knowledge of the attention mechanism of the visual system is not perfect, and the diversity of images and the interference of noise cause that many methods can only start from the primary features without paying attention to the importance of advanced features such as image semantic information, and many methods have low calculation efficiency and the accuracy of the calculation result needs to be improved. And because many methods are proposed aiming at some specific tasks, the universality of the method is not strong, and some methods have good effect on a standard data set, but when the foreground and the background of the image are relatively close and the like, the deviation between the position of the obtained saliency map and the true value is large, and the saliency region is difficult to accurately extract.
Therefore, the significance detection algorithm capable of obtaining more accurate and fresh significance map is provided, and the significance detection algorithm has great significance for fully utilizing computer resources and efficiently processing various visual field tasks.
Disclosure of Invention
Aiming at the defects in the prior art, the image saliency detection method based on multi-feature optimal fusion provided by the invention solves the problems that the conventional saliency detection method is single in feature analysis angle and is difficult to accurately reflect the image saliency.
In order to achieve the purpose of the invention, the invention adopts the technical scheme that: an image saliency detection method based on multi-feature optimal fusion comprises the following steps:
s1, acquiring an image to be detected;
s2, preprocessing an image to be detected through a linear iterative clustering segmentation algorithm to obtain a plurality of super pixels formed by adjacent pixels;
s3, extracting color features and texture features of the super pixels, and extracting frequency domain features of the image to be detected;
and S4, inputting the extracted color features, texture features and frequency domain features into a trained support vector machine to obtain corresponding saliency values, forming a final saliency map, and realizing image saliency detection.
Further, in step S3, the method for extracting color features of the super-pixels specifically includes:
establishing a color histogram for each super pixel, taking each super pixel as a target super pixel in sequence, and performing super pixel color comparison on the target super pixel and all other super pixels to obtain a color comparison significant value of the target super pixel as a color feature of the super pixel.
Further, the color contrast saliency value S (r) of said target superpixelk) The calculation formula of (2) is as follows:
Figure RE-GDA0002367546370000021
in the formula, Ds(rk,ri) Is a target super pixel riAnd a super pixel rkThe spatial distance between;
w(ri) Is a target super pixel riThe spatial weight of (2);
δsfor the influencing variable of the spatial weighting for the calculation of the color contrast saliency value, said δsThe smaller the space weight value is, the greater the calculation effect on the color significant contrast is;
Dr(rk,ri) Is a target super pixel riAnd a super pixel rkThe calculation formula of the color space distance between the two is specifically as follows:
Figure RE-GDA0002367546370000031
in the formula, f (c)1I) is the i-th color in the super pixel c1The probability of occurrence of (a);
f(c2j) is the jth color in the super pixel c2The probability of occurrence of (a);
D(c1,i,c2,j) Is a super pixel c1And the ith color of (c) and the super pixel2The spatial distance of the jth color in (a);
n1,n2are respectively a super pixel c1And super pixel c1Total number of colors in (1).
Further, in step S3, the method for extracting texture features of the super pixels specifically includes:
a1 calculating the super pixel R by two-dimensional Gabor filterjIn each pixel IiFeature vector G (R)j,Ii);
A2, based on feature vector G (R)j,Ii) Calculating the superpixel RjEach pixel of (1)iTexture feature vector G (R)j);
A3, calculating all pixels IiTexture feature vector G (R)j) Average value to obtain super pixel RjTexture saliency value S oft(j) As a super pixel RjThe texture feature of (1).
Further, the expression of the two-dimensional Gabor filter in the step a1 is:
Figure RE-GDA0002367546370000032
wherein, the real number part of the two-dimensional Gabor filter is:
Figure RE-GDA0002367546370000033
the imaginary part of the two-dimensional Gabor filter is:
Figure RE-GDA0002367546370000034
wherein x '═ x cos θ + y sin θ, y' ═ x sin θ + y cos θ;
x is the value of a pixel on the x-axis in two-dimensional space;
y is the value of the pixel on the y-axis in two-dimensional space;
λ is the wavelength of the sine function;
θ represents the direction of the Gabor kernel function;
ψ denotes a corresponding phase shift amount;
σ represents the standard deviation of the Gaussian function;
γ represents the width to height ratio of the space;
the super-pixel R in the step A1jIn each pixel IiFeature vector G (R)j,Ii) Comprises the following steps:
Figure RE-GDA0002367546370000041
in the formula, Gi(s, o) is the pixel I at a certain scale and orientationiThe characteristic vector after being filtered by a two-dimensional Gabor filter, s is a first-dimensional characteristic vector, and o is a second-dimensional characteristic vector;
in the step A2, the super-pixel RjEach pixel of (1)iTexture feature vector G (R)j) Comprises the following steps:
Figure RE-GDA0002367546370000042
in the formula, NiIs a super pixel RjThe total number of pixels in;
in the step A3, the super-pixel RjTexture saliency value S oft(j) Comprises the following steps:
Figure RE-GDA0002367546370000043
in the formula, NtThe number of superpixels in the image to be detected is shown;
D(G(Ri),G(Rj) Is a super pixel RjAnd a super pixel RiThe euclidean distance between.
Further, in step S3, the method for extracting the frequency domain feature of the image to be detected i (x) specifically includes:
b1, converting the image I (x) to be detected in the original space domain into a frequency domain through Fourier transform, and calculating a phase spectrum and a magnitude spectrum of the image in the frequency domain;
b2, calculating a log spectrum of the amplitude spectrum, and filtering the log spectrum;
b3, calculating spectrum residual error information corresponding to the filtered log spectrum;
b4, performing inverse Fourier transform on the frequency spectrum residual error information and the magnitude spectrum, and performing Gaussian smoothing on the inverse Fourier transform result to obtain a saliency map Sf(x) And taking the frequency domain feature of the image to be detected as the frequency domain feature of the image to be detected.
Further, the calculation formula of the amplitude spectrum p (f) in the step B1 is:
P(f)=S(F[I(x)])
wherein S (-) is a function of amplitude as a function of angular frequency;
f [. cndot. ] is a Fourier transform;
phase spectrum a (f) is:
A(f)=R(F[I(x)])
wherein R (-) is a function of phase with angular frequency;
in step B2, the log spectrum l (f) of the amplitude spectrum p (f) is:
L(f)=log(A(f))
in the formula, log (-) is a logarithm operator;
in step B3, the spectrum residual information r (f) is:
R(f)=L(f)-hn(f)*L(f)
in the formula, hn(f) Is an average filter, hn(f) L (f) is a filtering operation performed on the log spectrum l (f) by a mean filter;
in the step B4, the saliency map Sf(x) Comprises the following steps:
Sf(x)=g(x)*F-1[exp(R(f)+P(f))]2
wherein g (x) is a Gaussian smoothing filter;
F-1[·]an inverse of a Fourier transform;
exp (. cndot.) is an exponential function.
Further, in step S4, the training data set used for training the support vector machine is a super-pixel set T { (x)1,y1),(x2,y2),.....(xN,yN) In which xiFor each super pixel a feature vector (c)i,ti,fi),ciTo correspond to the average color characteristic of a superpixel, tiAverage texture feature of corresponding superpixel fiFor the average frequency domain characteristics of the corresponding image, yiControl variable, y, for the class of the corresponding superpixeliWhen 1, yiIndicating that the corresponding superpixel is a salient target region, yiWhen equal to 0, yiRepresenting the corresponding superpixel as a background area;
in step S4, the method for training the support vector machine specifically includes:
c1, segmenting data in the training data set through a linear hyperplane to obtain a corresponding hyperplane;
and C2, determining the hyperplane with the maximum interval in all hyperplanes to obtain the trained support vector machine.
Further, in the step C1, the expression when the training data set is segmented by the linear hyperplane is as follows:
h(x)=ωTx+b
wherein h (x) is a hyperplane after division;
omega is a normal vector and is used for determining the direction of the hyperplane;
b is a displacement term used for the distance between the absolute hyperplane and the origin;
wherein (ω, b) denotes a hyperplane, and the hyperplane satisfies the condition:
h(xi)yi≥1。
further, the step S4 is specifically:
and learning input data through a trained support vector machine to respectively obtain optimal fusion coefficients of color features, texture features and frequency domain features, calculating a significant value of each super pixel based on the fusion coefficients, further obtaining a significant graph, and realizing image significance detection.
The invention has the beneficial effects that: the invention discloses an image saliency detection method based on multi-feature optimal fusion, aiming at the problem that the existing saliency detection algorithm only selects the bottom layer features of color, texture, direction and the like of an image in a space domain or only analyzes the singleness of a problem from the angle of a frequency domain, and the saliency of an object is calculated by selecting the features which can show the saliency of the object through comparison between the space domain and the frequency domain; the method can set different weights for various characteristics according to the importance degree of the various characteristics of the image to the significance detection by learning the training data through the support vector machine; the invention provides a saliency detection algorithm capable of obtaining a more accurate and clear saliency map so as to allocate more computer resources to the extracted saliency areas in various computer vision field tasks and efficiently process various vision field tasks.
Drawings
Fig. 1 is a flowchart of an image saliency detection method based on multi-feature fusion provided by the invention.
Fig. 2 is a schematic comparison diagram of a saliency map of each algorithm of the MSRA-1000 dataset according to an embodiment of the present invention.
Fig. 3 is a schematic diagram illustrating a saliency map comparison of each algorithm of the SED2 data set in an embodiment provided by the present invention.
FIG. 4 is a diagram illustrating PR curves of each algorithm on an SOD data set according to an embodiment of the present invention.
FIG. 5 is a diagram illustrating a comparison of PR curves of algorithms on MSRA-1000 in an embodiment of the present invention.
FIG. 6 is a diagram illustrating a comparison of PR curves of algorithms on the SED2 according to an embodiment of the present invention.
FIG. 7 is a diagram illustrating PR curves of algorithms on SOD according to an embodiment of the present invention.
Detailed Description
The following description of the embodiments of the present invention is provided to facilitate the understanding of the present invention by those skilled in the art, but it should be understood that the present invention is not limited to the scope of the embodiments, and it will be apparent to those skilled in the art that various changes may be made without departing from the spirit and scope of the invention as defined and defined in the appended claims, and all matters produced by the invention using the inventive concept are protected.
As shown in fig. 1, an image saliency detection method based on multi-feature optimal fusion includes the following steps:
s1, acquiring an image to be detected;
s2, preprocessing an image to be detected through a linear iterative clustering segmentation algorithm to obtain a plurality of super pixels formed by adjacent pixels;
s3, extracting color features and texture features of the super pixels, and extracting frequency domain features of the image to be detected;
and S4, inputting the extracted color features, texture features and frequency domain features into a trained support vector machine to obtain corresponding saliency values, forming a final saliency map, and realizing image saliency detection.
In the above step S2, in order to facilitate the extraction of the features of the subsequent image, it is first necessary to process the image by using a super-pixel segmentation technique of the image to obtain a super-pixel block for the subsequent processing, where the pixel is the minimum representation unit of the image, most of the previous image saliency detection techniques use the pixel as a basic processing unit for processing, but since the subsequent image size is larger and larger, the calculation complexity is higher, and the image processing using the pixel as a basic unit cannot well utilize the local structural features of the image, in this context, the proposed super-pixel concept is favored by researchers, the super-pixel refers to a process in which adjacent pixels cluster these pixels into pixel blocks having a certain visual meaning but irregular shape according to the similarity of the features such as color, texture, brightness, etc., the super-pixel segmentation aims to cluster the pixels into the super-pixel block according to the similarity of the features, due to the fact that the local structure information of the image can be well obtained through superpixel segmentation, the efficiency of subsequent calculation is greatly improved, and therefore the method becomes a preprocessing step of multiple processes.
At present, various superpixel segmentation algorithms have appeared, each algorithm has corresponding characteristics, and the superpixel segmentation algorithms can be divided into two types according to different principles of segmenting an image to generate superpixels: graph theory based algorithms and gradient based algorithms. The first class of algorithms generally converts the image to be processed into a weighted undirected graph structure, the relationship of any adjacent pixels is mapped to the edges of the undirected graph, the weights on the edges represent the similarity of the corresponding pixel characteristics, then the undirected graph is divided, and an objective cost function is constructed, the optimal value of which is calculated to obtain the superpixels. Based on the gradient descent segmentation algorithm, part of pixel points are usually selected as seed points, then clustering is carried out by taking the pixel points as centers, and then the corresponding clustering results are simultaneously modified in the continuous iteration process until the corresponding convergence conditions are reached, and the operation is stopped. The invention processes images using a SLIC superpixel segmentation algorithm with high accuracy.
In the step S3, when extracting the color feature, the contrast of the color generally refers to the degree of difference between the color of a certain region in the image and the color of the surrounding region, and generally, the larger the difference in color, the more noticeable the human is. Since the global contrast can better show the whole saliency region, the algorithm uses the global contrast to calculate the saliency value of the target region.
Therefore, the method for extracting the color feature of the super-pixel specifically comprises the following steps:
establishing a color histogram for each super pixel, taking each super pixel as a target super pixel in sequence, and performing super pixel color comparison on the target super pixel and all other super pixels to obtain a color comparison significant value of the target super pixel as a color feature of the super pixel.
Wherein the color contrast saliency value S (r) of the target superpixelk) The calculation formula of (2) is as follows:
Figure RE-GDA0002367546370000091
in the formula, Ds(rk,ri) Is a target super pixel riAnd a super pixel rkThe spatial distance therebetween;
w(ri) Is a target super pixel riThe spatial weight of (2);
δsfor the influencing variable of the spatial weighting for the calculation of the color contrast saliency value, said δsThe smaller the space weight value is, the greater the calculation effect on the color significant contrast is;
Dr(rk,ri) Is a target super pixel riAnd a super pixel rkThe calculation formula of the color space distance between the two is specifically as follows:
Figure RE-GDA0002367546370000092
in the formula, f (c)1I) is the i-th color in the super pixel c1The probability of occurrence of (a);
f(c2j) is the jth color in the super pixel c2The probability of occurrence of (a);
D(c1,i,c2,j) Is a super pixel c1And the ith color of (c) and the super pixel2The spatial distance of the jth color in (1)Separating;
n1,n2are respectively a super pixel c1And super pixel c1Total number of colors in (1).
Color contrast saliency value S (r) at target superpixelk) The calculation formula (2) takes into account the influence of the spatial distance between the super-pixel blocks on the contrast of the super-pixel blocks to enhance the enhancement effect of the closer distance on the contrast calculation and inhibit the inhibition effect of the farther distance on the contrast calculation.
In the step S3, the texture features can reflect the uniform or non-uniform variation degree of a certain structure in a certain image region, and can visually reflect the characteristics of the object, and the method for extracting the texture features of the image includes a model method, a statistical method, a spectrum method and a structure analysis method. The Gabor filter can analyze image information in a frequency domain and a space domain simultaneously, the expression of the frequency and the direction of the image information is very similar to the human visual system, and the texture information of the image can be well extracted. Therefore, the method for extracting the texture feature of the super pixel in step S3 specifically includes:
a1 calculating the super pixel R by two-dimensional Gabor filterjIn each pixel IiFeature vector G (R)j,Ii);
Specifically, a Gabor filter is used for carrying out filtering processing on the segmented super-pixel block, the filter is set to use 4 directions and 4 scale parameters to obtain 4 groups of 4 different scales and 16 different Gabor characteristic vectors in 4 different directions, and a pixel I under a certain scale and direction is setiThe Gabor filtered feature vector of (1) is Gi(s, o) to obtain a super pixel RjIn each pixel IiFeature vector G (R)j,Ii);
A2, based on feature vector G (R)j,Ii) Calculating the superpixel RjEach pixel of (1)iTexture feature vector G (R)j);
A3, calculating all pixels IiTexture feature vector G (R)j) Average value to obtain super pixel RjTexture saliency value S oft(j) As a super pixel RjThe texture feature of (1).
Wherein, the expression of the two-dimensional Gabor filter in the step a1 is:
Figure RE-GDA0002367546370000101
wherein, the real number part of the two-dimensional Gabor filter is:
Figure RE-GDA0002367546370000102
the imaginary part of the two-dimensional Gabor filter is:
Figure RE-GDA0002367546370000111
wherein x '═ x cos θ + y sin θ, y' ═ x sin θ + y cos θ;
x is the value of a pixel on the x-axis in two-dimensional space;
y is the value of the pixel on the y-axis in two-dimensional space;
λ is the wavelength of the sine function;
θ represents the direction of the Gabor kernel function;
ψ denotes a corresponding phase shift amount;
σ represents the standard deviation of the Gaussian function;
γ represents the width to height ratio of the space;
the super-pixel R in the step A1jIn each pixel IiFeature vector G (R)j,Ii) Comprises the following steps:
Figure RE-GDA0002367546370000112
in the formula, Gi(s, o) is the pixel I at a certain scale and orientationiThe feature vector after being filtered by a two-dimensional Gabor filter, s is a first-dimensional feature vector, and o is a second-dimensional feature vector;
In the step A2, the super-pixel RjEach pixel of (1)iTexture feature vector G (R)j) Comprises the following steps:
Figure RE-GDA0002367546370000113
in the formula, NiIs a super pixel RjThe total number of pixels in;
in the step A3, the super-pixel RjTexture saliency value S oft(j) Comprises the following steps:
Figure RE-GDA0002367546370000114
in the formula, NtThe number of superpixels in the image to be detected is shown;
D(G(Ri),G(Rj) Is a super pixel RjAnd a super pixel RiThe euclidean distance between.
In the step S3, when performing frequency domain feature extraction on the image i (x) to be detected, the saliency target of the image can be efficiently analyzed by converting the image information into the frequency domain and calculating the logarithmic frequency of the image. The main principle of image significance analysis from the angle of a frequency domain is that a human visual system is sensitive to comparative special information and can inhibit frequently-occurring information, so that image information can be divided into redundant information and special information from the angle of information theory, and the redundant information of the image is removed to obtain the significance information of the remaining image. Therefore, the method for extracting the frequency domain feature of the image i (x) to be detected in step S3 specifically includes:
b1, converting the image I (x) to be detected in the original space domain into a frequency domain through Fourier transform, and calculating a phase spectrum and a magnitude spectrum of the image in the frequency domain;
b2, calculating a log spectrum of the amplitude spectrum, and filtering the log spectrum;
b3, calculating spectrum residual error information corresponding to the filtered log spectrum;
b4, performing inverse Fourier transform on the frequency spectrum residual error information and the magnitude spectrum, and performing Gaussian smoothing on the inverse Fourier transform result to obtain a saliency map Sf(x) And taking the frequency domain feature of the image to be detected as the frequency domain feature of the image to be detected.
The calculation formula of the amplitude spectrum p (f) in the step B1 is:
P(f)=S(F[I(x)])
wherein S (-) is a function of amplitude as a function of angular frequency;
f [. cndot. ] is a Fourier transform;
phase spectrum a (f) is:
A(f)=R(F[I(x)])
wherein R (-) is a function of phase with angular frequency;
in step B2, the log spectrum l (f) of the amplitude spectrum p (f) is:
L(f)=log(A(f))
in the formula, log (-) is a logarithm operator;
in step B3, the spectrum residual information r (f) is:
R(f)=L(f)-hn(f)*L(f)
in the formula, hn(f) Is an average filter, hn(f) L (f) is a filtering operation performed on the log spectrum l (f) by a mean filter;
in the step B4, the saliency map Sf(x) Comprises the following steps:
Sf(x)=g(x)*F-1[exp(R(f)+P(f))]2
wherein g (x) is a Gaussian smoothing filter;
F-1[·]an inverse of a Fourier transform;
exp (. cndot.) is an exponential function.
Most of the traditional image saliency detection methods simply add or multiply the extracted feature maps linearly, so that different importance degrees of the extracted feature maps to the finally formed saliency map are not considered, and different contribution degrees of the extracted different features to the saliency map generation can be better distinguished by learning the super-pixel features of the target region and the background region by using a support vector machine, and meanwhile, in a classification model of an image, the saliency detection problem can be regarded as a problem of dividing the image region into a target and a background. The support vector machine is a model for processing classification and regression problems in a learning mode under supervision, is widely applied to the fields of pattern recognition, text classification and the like, and has high accuracy in small sample and high-dimensional data classification. The method has the advantages of high robustness and the like, can process significance detection tasks, and mainly comprises two stages by using a support vector machine to carry out significance detection; the first stage is to use an initial support vector machine to learn the characteristics of the super-pixels of the salient object or the background, and obtain a classifier through a continuous iterative optimization process; the second stage is to use the classifier to compute its saliency values for the features of the superpixels of the input image, forming the final saliency image. The basic idea of the support vector machine is that for given data, if the data are linearly separable, a hyperplane needs to be found to distinguish the data, but the hyperplane is often many, so that a separating hyperplane which maximizes two types of support vectors needs to be selected, and if the data are linearly inseparable, the current data needs to be mapped to a high-dimensional space for processing. Based on this, when the saliency of the image is detected by the support vector machine in step S4,
the training data set used to train the support vector machine is the superpixel set T { (x)1,y1),(x2,y2),.....(xN,yN) In which x1For each super pixel a feature vector (c)i,ti,fi),ciTo correspond to the average color characteristic of a superpixel, tiAverage texture feature of corresponding superpixel fiFor the average frequency domain characteristics of the corresponding image, yiControl variable, y, for the class of the corresponding superpixeliWhen 1, yiRepresenting corresponding superThe pixel is a salient target region, yiWhen equal to 0, yiRepresenting the corresponding superpixel as a background area;
the method for training the support vector machine specifically comprises the following steps:
c1, segmenting data in the training data set through a linear hyperplane to obtain a corresponding hyperplane;
and C2, determining the hyperplane with the maximum interval in all hyperplanes to obtain the trained support vector machine.
In step C1, the expression when the training data set is segmented by the linear hyperplane is:
h(x)=ωTx+b
wherein h (x) is a hyperplane after division;
omega is a normal vector and is used for determining the direction of the hyperplane;
b is a displacement term used for the distance between the absolute hyperplane and the origin;
(ω, b) to represent the hyperplane that can correctly classify the training data, then the hyperplane satisfies the condition:
h(xi)yi≥1;
meanwhile, in order to search for the hyperplane having the maximum interval, it is necessary to solve (ω, b) that the sum of the distances from the support vector to the hyperplane, where the equal sign of the above equation holds, is the maximum, and assuming that the sum of the distances from the support vector to the hyperplane is represented by γ, the calculation formula is as follows:
Figure RE-GDA0002367546370000151
the problem of finding the maximum spacing then translates into finding (ω, b) such that
Figure RE-GDA0002367546370000152
At the most, the specific requirements are as follows:
Figure RE-GDA0002367546370000153
meanwhile, the above formula can be transformed into the following problem, and the appropriate (ω, b) is obtained to minimize | | ω |, and the specific requirement is as follows:
Figure RE-GDA0002367546370000154
corresponding (omega) is obtained by solving the minimization of the above formula*,b*) The corresponding separation hyperplane is:
*)Tx+b*=0
after the trained support vector machine (classifier) is obtained, the feature vector x of each super pixel can be inputiAnd obtaining the significant value of the super pixel, and finally obtaining a corresponding significant image.
Therefore, the step S4 is specifically:
and learning input data through a trained support vector machine to respectively obtain optimal fusion coefficients of color features, texture features and frequency domain features, calculating a significant value of each super pixel based on the fusion coefficients, further obtaining a significant graph, and realizing image significance detection.
Example 2:
in one embodiment of the invention, the method of the invention is compared with the existing 8 classical algorithms, the algorithms are IT, LC, HC, SR, FT, AC, CA and GB respectively, the saliency maps generated by the classical algorithms are generated by the codes of the corresponding algorithms, and the validity of the algorithms is verified by combining the true value maps and the evaluated indexes;
(1) experimental Environment
The effectiveness of the multi-feature fusion based algorithm was verified experimentally, and the programming environment used was Matlab2014 a. The computer is configured as Windows7(64 bits), Intel i5-3470@3.2GHz, 8G memory. Experiments were performed in this section on the MSRA-1000 dataset, the SED2 dataset and the SOD dataset.
(2) Subjective effect comparison
In MSRA-1000, SED2 and SOD data set random drawing picture, using the invention proposed based on the multi-feature optimal fusion significance detection algorithm generated significance map, and the other mentioned in the above section of the significance map generated by some other classical algorithms to qualitatively compare, the classical algorithm obtained significance map and the algorithm obtained significance map is shown in the lower graph. In fig. 2 to 4, the original image, the saliency maps obtained by the IT, LC, HC, SR, FT, AC, CA, and GB algorithms, and the saliency map and the truth map obtained by the present algorithm are shown in order from left to right.
By comparing the significance detection effects of the algorithms on different data sets, the results of most algorithms on the MSRA-1000 data set are the best, the results on the SED2 data set are good, and the results on the SOD data set are poor, which is mainly because the images in the MSRA-1000 data set only have one target area, the difference between foreground and background characteristics is large, the targets are clear, most of the target areas are located in the center of the images, and the difficulty of detecting the target areas is low. The image in the SED2 dataset typically has two target regions and the target regions are typically located relatively far from the centre and are difficult to detect. However, the target area of the image in the SOD data set is usually not clear enough, and the background is complex, so that the interference to foreground detection is large, and the salient area is not easy to detect. The target area obtained by the IT algorithm is found to be incomplete and poor in effect by comparing the effect of the IT algorithm with that of other algorithms on a single data set; the HC algorithm can better resist the interference of background noise, but can not highlight a salient object; the SR algorithm and the GB algorithm can only detect the boundary of an object, and the comparison cavity inside the object is obvious; the CA algorithm can only detect the edge of an object and cannot highlight the inner area of a salient object; LC algorithms, AC algorithms and FT algorithms are often prone to false detection of the background as a target area when the foreground and target are relatively close. In general, a classical saliency detection algorithm can detect saliency objects, but the problems that a generated target region is incomplete or background noise is easy to detect as a saliency object exist, and the like, while the algorithm provided in this chapter can completely detect the saliency target region, obtain clearer object edges and highlight target regions, inhibit the interference of background noise, and show better effect in the detection of natural scene pictures.
(3) Objective effect comparison
Comparing the performance of the algorithm of the chapter with 8 classical algorithms on public data sets MSRA-1000, SED2 and SOD, and mainly using the following performance indexes: PR curve, F-measure value, MAE value, and AUC value. The PR curve is mainly used for describing the relation between the precision ratio and the recall ratio; the F-measure value comprehensively reflects the relationship between the accuracy and the recall rate; MAE represents the error degree between the saliency map obtained by the algorithm and the artificially labeled real map; AUC represents the probability that, for one positive and one negative case of randomness, the probability value predicted as a positive case by the algorithm is greater than the probability value predicted as a negative case. The PR curve image, F-measure image, MAE value image and AUC image at a fixed threshold for the algorithm of the invention and the other 8 classical algorithms were compared, respectively, mainly in three sets of data sets MSRA-1000, SED2 and SOD.
Fig. 5-7 show PR curve comparisons between the algorithm of the present invention and 8 other algorithms on different datasets at a fixed threshold, respectively.
As can be seen from fig. 5-7, the PR curves for most algorithms are higher than the IT algorithms proposed earlier, and the PR curves for the various algorithms are highest on the MSRA-1000 dataset, lower on the SED2 dataset, and lowest on the SOD dataset. The reason is that the picture in the MSRA-1000 generally has only one target area, the difference between the foreground characteristic and the background characteristic is large, the target is clear, most of the target area is located in the center of the image, and the difficulty of detecting the target area is low. It is for this reason that the image in the SED2 dataset typically has two target regions and the target is typically located relatively far off-center, making it difficult to detect the target regions, which results in the algorithm being less accurate on the SED2 dataset than it is on the MSRA-1000. The target area of the image in the SOD dataset is usually not clear enough, and the foreground and background image features are relatively close, and are relatively disturbed by background noise, so that the salient area is not easy to detect, and therefore the PR curve of an algorithm on the dataset is much lower than that of the other datasets. The PR curve of the algorithm is positioned above other PR curves on the MSRA-1000 data set, the accuracy rate of the algorithm under the same recall rate is almost higher than that of all other algorithms, and the algorithm can well detect a significant target; at the same accuracy, the recall rate of the algorithm in this chapter is almost close to or higher than that of all other algorithms. The comparison proves that the algorithm has better significance detection performance, and the obtained significance map is closer to the real map. Even in the SED2 data set and the SOD data set with high difficulty in detecting the obvious target, the PR curve of the algorithm is slightly higher than that of other classical algorithms, because most of the algorithms only simply fuse the feature maps of the images, and the algorithm fuses the image features by using the optimal weight by using a support vector machine, so that a more accurate target region can be obtained.
The invention has the beneficial effects that: the invention discloses an image saliency detection method based on multi-feature optimal fusion, aiming at the problem that the existing saliency detection algorithm only selects the bottom layer features of color, texture, direction and the like of an image in a space domain or only analyzes the singleness of a problem from the angle of a frequency domain, and the saliency of an object is calculated by selecting the features which can show the saliency of the object through comparison between the space domain and the frequency domain; the method can set different weights for various characteristics according to the importance degree of the various characteristics of the image to the significance detection by learning the training data through the support vector machine; the invention provides a saliency detection algorithm capable of obtaining a more accurate and clear saliency map so as to allocate more computer resources to the extracted saliency areas in various computer vision field tasks and efficiently process various vision field tasks.

Claims (10)

1. An image saliency detection method based on multi-feature optimal fusion is characterized by comprising the following steps:
s1, acquiring an image to be detected;
s2, preprocessing an image to be detected through a linear iterative clustering segmentation algorithm to obtain a plurality of super pixels formed by adjacent pixels;
s3, extracting color features and texture features of the super pixels, and extracting frequency domain features of the image to be detected;
and S4, inputting the extracted color features, texture features and frequency domain features into a trained support vector machine to obtain corresponding saliency values, forming a final saliency map, and realizing image saliency detection.
2. The method for detecting image saliency based on multi-feature optimal fusion according to claim 1, wherein in the step S3, the method for extracting color features of the super-pixels specifically comprises:
establishing a color histogram for each super pixel, taking each super pixel as a target super pixel in sequence, and performing super pixel color comparison on the target super pixel and all other super pixels to obtain a color comparison significant value of the target super pixel as a color feature of the super pixel.
3. The image saliency detection method based on multi-feature optimal fusion according to claim 2, characterized in that the color contrast saliency value S (r) of the target superpixelk) The calculation formula of (2) is as follows:
Figure FDA0002315770400000011
in the formula, Ds(rk,ri) Is a target super pixel riAnd a super pixel rkThe spatial distance therebetween;
w(ri) Is a target super pixel riThe spatial weight of (2);
δsfor the influencing variable of the spatial weighting for the calculation of the color contrast saliency value, said δsThe smaller the space weight value is, the greater the calculation effect on the color significant contrast is;
Dr(rk,ri) Is a target super pixel riAnd a super pixel rkThe calculation formula of the color space distance between the two is specifically as follows:
Figure FDA0002315770400000021
in the formula, f (c)1I) is the i-th color in the super pixel c1The probability of occurrence of (a);
f(c2j) is the jth color in the super pixel c2The probability of occurrence of (a);
D(c1,i,c2,j) Is a super pixel c1And the ith color of (c) and the super pixel2The spatial distance of the jth color in (a);
n1,n2are respectively a super pixel c1And super pixel c1Total number of colors in (1).
4. The method for detecting image saliency based on multi-feature optimal fusion according to claim 1, wherein in the step S3, the method for extracting texture features of the superpixels specifically comprises:
a1 calculating the super pixel R by two-dimensional Gabor filterjIn each pixel IiFeature vector G (R)j,Ii);
A2, based on feature vector G (R)j,Ii) Calculating the superpixel RjEach pixel of (1)iTexture feature vector G (R)j);
A3, calculating all pixels IiTexture feature vector G (R)j) Average value to obtain super pixel RjTexture saliency value S oft(j) As a super pixel RjThe texture feature of (1).
5. The method for detecting the image saliency based on multi-feature optimal fusion according to claim 4, wherein the expression of the two-dimensional Gabor filter in the step A1 is as follows:
Figure FDA0002315770400000022
wherein, the real number part of the two-dimensional Gabor filter is:
Figure FDA0002315770400000023
the imaginary part of the two-dimensional Gabor filter is:
Figure FDA0002315770400000024
wherein x '═ xcos θ + ysin θ, y' ═ xsin θ + ycos θ;
x is the value of a pixel on the x-axis in two-dimensional space;
y is the value of the pixel on the y-axis in two-dimensional space;
λ is the wavelength of the sine function;
θ represents the direction of the Gabor kernel function;
ψ denotes a corresponding phase shift amount;
σ represents the standard deviation of the Gaussian function;
γ represents the width to height ratio of the space;
the super-pixel R in the step A1jIn each pixel IiFeature vector G (R)j,Ii) Comprises the following steps:
Figure FDA0002315770400000031
in the formula, Gi(s, o) is the pixel I at a certain scale and orientationiThe characteristic vector after being filtered by a two-dimensional Gabor filter, s is a first-dimensional characteristic vector, and o is a second-dimensional characteristic vector;
in the step A2, the super-pixel RjEach pixel of (1)iTexture feature vector G (R)j) Comprises the following steps:
Figure FDA0002315770400000032
in the formula, NiIs a super pixel RjThe total number of pixels in;
in the step A3, the super-pixel RjTexture saliency value S oft(j) Comprises the following steps:
Figure FDA0002315770400000033
in the formula, NtThe number of superpixels in the image to be detected is shown;
D(G(Ri),G(Rj) Is a super pixel RjAnd a super pixel RiThe euclidean distance between.
6. The image saliency detection method based on multi-feature optimal fusion according to claim 1, wherein in the step S3, the method for extracting the frequency domain features of the image to be detected i (x) specifically comprises:
b1, converting the image I (x) to be detected in the original space domain into a frequency domain through Fourier transform, and calculating a phase spectrum and a magnitude spectrum of the image in the frequency domain;
b2, calculating a log spectrum of the amplitude spectrum, and filtering the log spectrum;
b3, calculating spectrum residual error information corresponding to the filtered log spectrum;
b4, performing inverse Fourier transform on the frequency spectrum residual error information and the magnitude spectrum, and performing Gaussian smoothing on the inverse Fourier transform result to obtain a saliency map Sf(x) And taking the frequency domain feature of the image to be detected as the frequency domain feature of the image to be detected.
7. The image saliency detection method based on multi-feature optimal fusion according to claim 6, characterized in that the calculation formula of the magnitude spectrum P (f) in the step B1 is as follows:
P(f)=S(F[I(x)])
wherein S (-) is a function of amplitude as a function of angular frequency;
f [. cndot. ] is a Fourier transform;
phase spectrum a (f) is:
A(f)=R(F[I(x)])
wherein R (-) is a function of phase with angular frequency;
in step B2, the log spectrum l (f) of the amplitude spectrum p (f) is:
L(f)=log(A(f))
in the formula, log (-) is a logarithm operator;
in step B3, the spectrum residual information r (f) is:
R(f)=L(f)-hn(f)*L(f)
in the formula, hn(f) Is an average filter, hn(f) L (f) is a filtering operation performed on the log spectrum l (f) by a mean filter;
in the step B4, the saliency map Sf(x) Comprises the following steps:
Sf(x)=g(x)*F-1[exp(R(f)+P(f))]2
wherein g (x) is a Gaussian smoothing filter;
F-1[·]an inverse of a Fourier transform;
exp (. cndot.) is an exponential function.
8. The method for detecting image saliency based on multi-feature optimal fusion of claim 1, characterized in that in step S4, the training data set for training support vector machine is a super-pixel set T { (x)1,y1),(x2,y2),.....(xN,yN) In which xiFor each super pixel a feature vector (c)i,ti,fi),ciTo correspond to the average color characteristic of a superpixel, tiAverage texture feature of corresponding superpixel fiFor the average frequency domain characteristics of the corresponding image, yiControl variable, y, for the class of the corresponding superpixeliWhen 1, yiIndicating that the corresponding superpixel is a salient target region, yiWhen equal to 0, yiRepresenting the corresponding superpixel as a background area;
in step S4, the method for training the support vector machine specifically includes:
c1, segmenting data in the training data set through a linear hyperplane to obtain a corresponding hyperplane;
and C2, determining the hyperplane with the maximum interval in all hyperplanes to obtain the trained support vector machine.
9. The method for detecting image saliency based on multi-feature optimal fusion according to claim 8, wherein in the step C1, the expression when the training data set is segmented by the linear hyperplane is as follows:
h(x)=ωTx+b
wherein h (x) is a hyperplane after division;
omega is a normal vector and is used for determining the direction of the hyperplane;
b is a displacement term used for the distance between the absolute hyperplane and the origin;
wherein (ω, b) denotes a hyperplane, and the hyperplane satisfies the condition:
h(xi)yi≥1。
10. the method for detecting image saliency based on multi-feature optimal fusion according to claim 1, wherein the step S4 specifically includes:
and learning input data through a trained support vector machine to respectively obtain optimal fusion coefficients of color features, texture features and frequency domain features, calculating a significant value of each super pixel based on the fusion coefficients, further obtaining a significant graph, and realizing image significance detection.
CN201911276766.1A 2019-12-12 2019-12-12 Image significance detection method based on multi-feature optimal fusion Pending CN110991547A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911276766.1A CN110991547A (en) 2019-12-12 2019-12-12 Image significance detection method based on multi-feature optimal fusion

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911276766.1A CN110991547A (en) 2019-12-12 2019-12-12 Image significance detection method based on multi-feature optimal fusion

Publications (1)

Publication Number Publication Date
CN110991547A true CN110991547A (en) 2020-04-10

Family

ID=70093043

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911276766.1A Pending CN110991547A (en) 2019-12-12 2019-12-12 Image significance detection method based on multi-feature optimal fusion

Country Status (1)

Country Link
CN (1) CN110991547A (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112734695A (en) * 2020-12-23 2021-04-30 中国海洋大学 SAR image change detection method based on regional enhancement convolutional neural network
CN113553966A (en) * 2021-07-28 2021-10-26 中国科学院微小卫星创新研究院 Method for extracting effective starry sky area of single star map
CN113808166A (en) * 2021-09-15 2021-12-17 西安电子科技大学 Single-target tracking method based on clustering difference and depth twin convolutional neural network
CN114782878A (en) * 2022-05-26 2022-07-22 广东南方电信规划咨询设计院有限公司 Video significance detection method
CN115953672A (en) * 2023-03-13 2023-04-11 南昌工程学院 Method for identifying surface cracks of underwater dam
CN117830322A (en) * 2024-03-06 2024-04-05 慧创科仪(北京)科技有限公司 Method and device for performing significance difference analysis on near infrared data

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103020965A (en) * 2012-11-29 2013-04-03 奇瑞汽车股份有限公司 Foreground segmentation method based on significance detection
CN105760886A (en) * 2016-02-23 2016-07-13 北京联合大学 Image scene multi-object segmentation method based on target identification and saliency detection
CN107256547A (en) * 2017-05-26 2017-10-17 浙江工业大学 A kind of face crack recognition methods detected based on conspicuousness
CN107665347A (en) * 2017-09-22 2018-02-06 中国科学院西安光学精密机械研究所 Visual saliency target detection method based on filtering optimization
CA3032487A1 (en) * 2016-08-03 2018-02-08 Jiangsu University Saliency-based method for extracting road target from night vision infrared image
CN107977660A (en) * 2017-10-13 2018-05-01 天津工业大学 Region of interest area detecting method based on background priori and foreground node
CN109242854A (en) * 2018-07-14 2019-01-18 西北工业大学 A kind of image significance detection method based on FLIC super-pixel segmentation
CN109522908A (en) * 2018-11-16 2019-03-26 董静 Image significance detection method based on area label fusion
CN109886267A (en) * 2019-01-29 2019-06-14 杭州电子科技大学 A kind of soft image conspicuousness detection method based on optimal feature selection
WO2020188121A1 (en) * 2019-03-21 2020-09-24 Five AI Limited Perception uncertainty

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103020965A (en) * 2012-11-29 2013-04-03 奇瑞汽车股份有限公司 Foreground segmentation method based on significance detection
CN105760886A (en) * 2016-02-23 2016-07-13 北京联合大学 Image scene multi-object segmentation method based on target identification and saliency detection
CA3032487A1 (en) * 2016-08-03 2018-02-08 Jiangsu University Saliency-based method for extracting road target from night vision infrared image
CN107256547A (en) * 2017-05-26 2017-10-17 浙江工业大学 A kind of face crack recognition methods detected based on conspicuousness
CN107665347A (en) * 2017-09-22 2018-02-06 中国科学院西安光学精密机械研究所 Visual saliency target detection method based on filtering optimization
CN107977660A (en) * 2017-10-13 2018-05-01 天津工业大学 Region of interest area detecting method based on background priori and foreground node
CN109242854A (en) * 2018-07-14 2019-01-18 西北工业大学 A kind of image significance detection method based on FLIC super-pixel segmentation
CN109522908A (en) * 2018-11-16 2019-03-26 董静 Image significance detection method based on area label fusion
CN109886267A (en) * 2019-01-29 2019-06-14 杭州电子科技大学 A kind of soft image conspicuousness detection method based on optimal feature selection
WO2020188121A1 (en) * 2019-03-21 2020-09-24 Five AI Limited Perception uncertainty

Non-Patent Citations (6)

* Cited by examiner, † Cited by third party
Title
HUA WANG等: "Hyperspectral Classification Based on Coupling Multiscale Super-Pixels and Spatial Spectral Features", 《IEEE GEOSCIENCE AND REMOTE SENSING LETTERS》 *
TYQ101010: "图像处理:显著性区域检测总结(二)", 《HTTPS://BLOG.CSDN.NET/TYQ101010/ARTICLE/DETAILS/48626995》 *
崔玲玲等: "融合双特征图信息的图像显著性检测方法", 《中国图象图形学报》 *
袁小艳等: "显著性目标检测中的视觉特征及融合", 《计算机应用与软件》 *
袁小艳等: "融入频域信息聚焦特征的显著性目标检测算法", 《计算机科学》 *
韩辰希等: "视觉显著性纹理—色彩特征融合的图像目标分类", 《电子测量技术》 *

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112734695A (en) * 2020-12-23 2021-04-30 中国海洋大学 SAR image change detection method based on regional enhancement convolutional neural network
CN112734695B (en) * 2020-12-23 2022-03-22 中国海洋大学 SAR image change detection method based on regional enhancement convolutional neural network
CN113553966A (en) * 2021-07-28 2021-10-26 中国科学院微小卫星创新研究院 Method for extracting effective starry sky area of single star map
CN113553966B (en) * 2021-07-28 2024-03-26 中国科学院微小卫星创新研究院 Method for extracting effective starry sky area of single star map
CN113808166A (en) * 2021-09-15 2021-12-17 西安电子科技大学 Single-target tracking method based on clustering difference and depth twin convolutional neural network
CN114782878A (en) * 2022-05-26 2022-07-22 广东南方电信规划咨询设计院有限公司 Video significance detection method
CN114782878B (en) * 2022-05-26 2024-04-30 广东南方电信规划咨询设计院有限公司 Video saliency detection method
CN115953672A (en) * 2023-03-13 2023-04-11 南昌工程学院 Method for identifying surface cracks of underwater dam
CN115953672B (en) * 2023-03-13 2024-02-27 南昌工程学院 Method for identifying surface cracks of underwater dam
CN117830322A (en) * 2024-03-06 2024-04-05 慧创科仪(北京)科技有限公司 Method and device for performing significance difference analysis on near infrared data

Similar Documents

Publication Publication Date Title
CN110991547A (en) Image significance detection method based on multi-feature optimal fusion
CN110334762B (en) Feature matching method based on quad tree combined with ORB and SIFT
Singh et al. Svm-bdt pnn and fourier moment technique for classification of leaf shape
WO2020107717A1 (en) Visual saliency region detection method and apparatus
CN104200495B (en) A kind of multi-object tracking method in video monitoring
CN104835175B (en) Object detection method in a kind of nuclear environment of view-based access control model attention mechanism
Davarzani et al. Scale-and rotation-invariant texture description with improved local binary pattern features
Wang et al. Recognition and localization of occluded apples using K-means clustering algorithm and convex hull theory: a comparison
CN110135438B (en) Improved SURF algorithm based on gradient amplitude precomputation
CN103984920B (en) Three-dimensional face identification method based on sparse representation and multiple feature points
CN106874942B (en) Regular expression semantic-based target model rapid construction method
CN111510792B (en) Video abstract generation method and system based on adaptive weighted graph difference analysis
CN107180436A (en) A kind of improved KAZE image matching algorithms
CN111091129B (en) Image salient region extraction method based on manifold ordering of multiple color features
Nguyen et al. Satellite image classification using convolutional learning
Zhang et al. Saliency-driven oil tank detection based on multidimensional feature vector clustering for SAR images
CN107301643A (en) Well-marked target detection method based on robust rarefaction representation Yu Laplce's regular terms
CN114821358A (en) Optical remote sensing image marine ship target extraction and identification method
CN107527348B (en) Significance detection method based on multi-scale segmentation
CN110633691A (en) Binocular in-vivo detection method based on visible light and near-infrared camera
CN111127407B (en) Fourier transform-based style migration forged image detection device and method
CN109902690A (en) Image recognition technology
Tao et al. Illumination-insensitive image representation via synergistic weighted center-surround receptive field model and weber law
Gao et al. A novel patterned fabric defect detection algorithm based on GHOG and low-rank recovery
CN116310416A (en) Deformable object similarity detection method based on Radon transformation and electronic equipment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
AD01 Patent right deemed abandoned

Effective date of abandoning: 20230317

AD01 Patent right deemed abandoned