CN108154147A - The region of interest area detecting method of view-based access control model attention model - Google Patents

The region of interest area detecting method of view-based access control model attention model Download PDF

Info

Publication number
CN108154147A
CN108154147A CN201810034712.3A CN201810034712A CN108154147A CN 108154147 A CN108154147 A CN 108154147A CN 201810034712 A CN201810034712 A CN 201810034712A CN 108154147 A CN108154147 A CN 108154147A
Authority
CN
China
Prior art keywords
image
scale
notable
gabor
feature
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201810034712.3A
Other languages
Chinese (zh)
Inventor
徐振辉
毛保全
朱守瑞
白向华
杨雨迎
韩小平
吴东亚
冯帅
李程
张天意
辛学敏
郑博文
王之千
李俊
朱锐
李晓刚
兰图
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Academy of Armored Forces of PLA
Original Assignee
Academy of Armored Forces of PLA
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Academy of Armored Forces of PLA filed Critical Academy of Armored Forces of PLA
Priority to CN201810034712.3A priority Critical patent/CN108154147A/en
Publication of CN108154147A publication Critical patent/CN108154147A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/25Determination of region of interest [ROI] or a volume of interest [VOI]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/46Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
    • G06V10/462Salient features, e.g. scale invariant feature transforms [SIFT]

Abstract

The present invention relates to a kind of region of interest area detecting methods of view-based access control model attention model, are related to technical field of image processing.The present invention is improved Itti vision modes, add the visual perception process of vision noticing mechanism simulation people, increase and play target movable information so that testing result more meets the physiological characteristic of people, is significantly improved to the detection result of Missile Body mark class Small object under complex background.

Description

The region of interest area detecting method of view-based access control model attention model
Technical field
The present invention relates to technical field of image processing, and in particular to a kind of area-of-interest inspection of view-based access control model attention model Survey method.
Background technology
People are often put to interested object more note that this is because vision attention is selective, and claim For visual selective or vision significance.When in face of complex scene, vision system can be primarily focused in scene rapidly Certain marking areas.
Region of interest (Regions Of Interest, ROI) is people's observation and generates interest, concern when understanding image Or the region paid attention to, i.e., it most can be intriguing in image, it can most show the region of picture material.In the application of its attention mechanism The key concept of proposition.Region of interest may be considered in image the most notable pixel set of (saliency), i.e., significant point or The set of point of interest.How region of interest is automatically extracted from piece image, be exactly region of interest detection technique.
Region of interest detect, i.e., salient region detection is using computer technology simulate human visual system, using regarding Feel attention model, extract some key messages of image as significant point, the appropriate area using centered on significant point is emerging as sense Interesting area.Area-of-interest selects not relying on the unique characteristics of scene areas, but relies on its generation compared with peripheral region Relative characteristic, i.e. vision significance.Conspicuousness is strong, is chosen as area-of-interest.How the region of interest inspection of image a kind of is designed Survey method so that the physiological characteristic that testing result more meets people becomes the technical issues of urgently to be resolved hurrily.
Invention content
(1) technical problems to be solved
The technical problem to be solved by the present invention is to:How the region of interest detection method of image a kind of is designed so that detection As a result more meet the physiological characteristic of people.
(2) technical solution
In order to solve the above technical problem, the present invention provides a kind of area-of-interest detections of view-based access control model attention model Method includes the following steps:
Step 1: establish multi-scale image structure
One width two dimensional image, scale space under different scale represent can by input picture I (x, y) and Gaussian kernel G (x, Y, σ) convolution obtains:
L (x, y, σ)=G (x, y, σ) * I (x, y) (2)
In formula (2), the location of pixels of (x, y) representative image, σ is known as the scale space factor, and value is smaller, characterizes the image quilt Smooth is fewer, and corresponding scale is also just smaller, and large scale corresponds to the general picture feature of image, and small scale corresponds to the thin of image Feature is saved, L represents the scale space of image;
Step 2: low-level visual features are extracted
1st, brightness extracts
If color video frequency image, r, g, b represent red, green and blue in image respectively, then brightness calculation is public Formula is:
I=(r+g+b)/3 (3)
If gray level image, then gray feature directly chooses the gray value of each pixel;2nd, color feature extracted
It is respectively red, green, blue, yellow 4 Color Channels to define R, G, B, Y, then:
R=r- (g+b)/2
G=g- (r+b)/2 (4)
B=b- (r+g)/2
Y=(r+g)/2- (r-g)/2-b
Then RG and BY channels are:RG=R-GB, Y=B- (R+G)/2;
3rd, Directional feature extraction
Direction character is extracted with Gabor filter:
One-dimensional Gabor functions, i.e. 1D-Gabor functions:
In formula, standard deviations of the σ for Gaussian function, w0For the spatial frequency of complex plane wave, x0Center for 1D-Gabor functions Point coordinates, the formula of the odd, even component of 1D-Gabor functions are respectively:
According to 3 σ principles, l=6 σ/w are taken0
Two-dimensional Gabor function, i.e. 2D-Gabor functions are one under the Gauss envelopes multiple change sine waves along x-axis, two Tieing up Gabor function expressions is:
The real and imaginary parts of two-dimensional Gabor function are respectively:
θ is directioin parameter, and σ is bigger, and energy more disperses, smaller more to concentrate, and has high frequency σh, low frequency σlTwo centre frequencies;
The selection of deflection θ:θ=0,30,60,90,120,150, filtering window 32*32 share 24 Gabor filters Wave device;
Input picture I (x, y) and the convolution of Gabor wavelet kernel function are:R (x, y)=∫ ∫ I (ε, η) g (x- ε, y- η) d ε d η (8)
Gabor wavelet transformation after the result is that plural number, take plural number mould | | r (x, y) | | mean value and variance as small echo change The result changed;
θ=0 is selected, the Gabor filter output of 30,60,90,120,150 six directions passes through as direction character (7) formula obtains the Gaobr wave filters of six direction, and then each tomographic image of gaussian pyramid structure is filtered with these wave filters Wave obtains the Feature Mapping figure on corresponding six direction;
4th, Motion feature extraction
Motion feature includes movement velocity and two category feature of the direction of motion;
1) motion vector extracts:The motion vector of target is extracted using background subtraction, by calculating each pixel in image The relevance in time domain or spatial domain between point using the background of former frame as reference, calculates the background of present frame, then by present frame Image and its background subtracting, obtain difference image, if t frames image is I (x, y, t), corresponding background image is B (x, y, t), Then difference image is:
2) Motion feature extraction:According to the difference image, obtain the movement of each pixel in adjacent two field pictures away from From DijThe move distance of each pixel on 0 °, 45 °, 90 ° and 135 ° four direction is projected, then obtains each pixel four by (x, y) Move distance in a direction of motion, and then the movement velocity for obtaining 0 °, 45 °, 90 ° and 135 ° four direction is followed successively by:
D(i, j)=pixel is in the component in x directions
D90°(i, j)=pixel is in the component in y directions
Step 3: characteristic pattern is calculated with notable figure
The center of visual attention model-periphery operation is to carry out interlayer phase reducing to central stratum and peripheral tier, that is, is calculated The difference of the corresponding large and small scale of different resolution, the mathematic(al) representation of characteristic pattern E are:
E (c, s)=| E (c) Θ E (s) | (11)
Wherein, Θ represents that the interpolation of two different scale images is subtracted each other, i.e., first to peripheral tier into row interpolation, pixel number is increased The difference operation of central stratum and peripheral tier is carried out after to corresponding central stratum pixel number, c represents center scale, behalf neighboring area Scale, c ∈ { 2,3,4 }, i.e., 2,3,4 layers are central stratum, s=c+ δ, scale difference δ ∈ { 3,4 }, i.e., each central stratum corresponding week Boundary layer adds 3 or 4 for the number of plies of central stratum;
The characteristic pattern of gray scale, color, direction and motion feature is acquired according to formula (11):
Gray feature figure:
I (c, s)=| I (c) Θ I (s) | (12)
Color characteristic figure:RG (c, s)=| (R (c)-G (c)) Θ (G (s)-R (s)) |
BY (c, s)=| (B (c)-Y (c)) Θ (Y (s)-B (s)) | (13)
Direction character figure:
O (c, s, θ)=| O (c, θ) Θ O (s, θ) | (14)
Motion feature figure:Ds(c, s)=| Ds(c)ΘDs(s)|
D(c, s)=| D(c)ΘD(s)|
D45°(c, s)=| D45°(c)ΘD45°(s)| (15)
D90°(c, s)=| D90°(c)ΘD90°(s)|
D135°(c, s)=| D135°(c)ΘD135°(s)|
All kinds of characteristic patterns are merged using default normalization operator N (), obtain the notable figure of each category feature.
Gray scale notable figure:
Color notable figure:
Direction notable figure:
Movement velocity notable figure:
Direction of motion notable figure:
⊕ is default blending algorithm, and characteristic remarkable picture further normalizes, and addition obtains comprehensive notable figure:
After obtaining the synthesis notable figure of image, each position competes with one another in comprehensive notable figure, and the position of triumph, which becomes, to be paid attention to Focus forms area-of-interest.
Preferably, it in step 3, will successively select in the victor is a king in comprehensive notable figure input Itti vision modes network The region of conspicuousness maximum in notable figure is selected as area-of-interest.
Preferably, it is further included after step 3 and focus-of-attention is realized by the inhibition of return mechanism in Itti vision modes The step of transfer.
Preferably, w0=1.
(3) advantageous effect
The present invention is improved Itti vision modes, adds the visual perception process of vision noticing mechanism simulation people, Increase and play target movable information so that testing result more meets the physiological characteristic of people, to the small mesh of Missile Body mark class under complex background Target detection result is significantly improved.
Description of the drawings
Fig. 1 is the complicated ground image detection result figure based on Itti visual attention models;Wherein (a) is artwork, and (b) is The synthesis notable figure of visual attention model, (c) are vision attention testing results;
Fig. 2 is the improvement visual attention model schematic diagram that the present invention uses;
Fig. 3 is the multi-scale image structure diagram that the present invention uses;
Fig. 4 is the one-dimensional Gabor functional digraphs of the present invention;
Fig. 5 is the two-dimensional Gabor functional digraph of the present invention;
Fig. 6 is that the size of the present invention is 4 × 6 Gabor filter group.
Specific embodiment
To make the purpose of the present invention, content and advantage clearer, with reference to the accompanying drawings and examples, to the present invention's Specific embodiment is described in further detail.
In numerous visual attention models, Itti proposes that visual attention model is of greatest concern.It belongs to bottom-up Attention mechanism carries out Multiresolution Decomposition after being filtered to linearity, extract brightness, color and the direction character of image as regarding Feel primary features, image office is calculated using center-periphery (Center-Surround, C-S) calculus of differences and non-linear normalizing The characteristic attribute (brightness, color and direction) of portion's range obtains the characteristic pattern under different scale, each characteristic pattern is carried out multiple dimensioned The notable figure of each feature dimensions is obtained after fusion and normalization.The mode of nonuniform sampling is then based on, is merged using multi-characteristic The notable figure that strategy ties up these different characteristics merges, and forms final synthesis notable figure.It is finally that comprehensive notable figure is defeated Enter in " the victor is a king " (winner-take-all WTA) network, select successively the region of conspicuousness maximum in notable figure as Area-of-interest.Attention mobility is carried out using the neural network and " inhibition of return " mechanism of " the victor is a king ".The model can utilize Human eye is to the attention mechanisms of the features such as color of image, brightness and direction, the adaptively area-of-interest in detection image, in mesh Have great advantage in mark detection.
A more complicated ground scene image is selected, as shown in Fig. 1 (a).When human eye watches the image attentively, can most draw Play human eye attention is a building building of image, it belongs to man-made target, is typical ground assault target.Due to ground Scape is complicated, is difficult accurately to detect the building with conventional dividing method.Utilize the region of interest based on Itti visual attention models Domain detection algorithm can then detect building position well, and (b) is comprehensive notable figure in Fig. 1, and (c) is testing result, wherein face Color is redder to represent that attention value is higher.
According to the imaging characteristic of tank target, in high s/n ratio, tank target has brightness, color and texture spy Sign, can directly be detected it using Itti visual attention models.
According to Missile Body target imaging characteristic, bullet mark has certain brightness, but color and textural characteristics unobvious, directly It connects and utilizes Itti visual attention models, detection result is undesirable.Below as the modeling approach of Itti vision modes, to model into Row improves, and increases Missile Body target kinetic characteristic, the improvement visual attention model for being suitble to the detection of Missile Body mark is established, such as Fig. 2 institutes Show.Specifically, detection method includes the following steps for the area-of-interest of the view-based access control model attention model of the present invention:
Step 1: establish multi-scale image structure
The sampling density and vision addressability of human visual system is as the distance to retinal centre increases and reduces, i.e., The process of human eye understanding things is ten split-phase of process of multi-resolution decomposition process and vision system understanding things from thick to thin Seemingly, thus can analog vision attention model this nonuniform sampling mechanism well.Nonuniform sampling is earliest by Kronauer& Zeevi (1985) is proposed, is then frequently used (Rybak 1998, Itti2005) in the model of simulation human vision.Its In most it is representative be Itti&Koch (2001) propose gaussian pyramid model, as shown in Figure 3.
Image Multiscale structure describes original image using a series of word image of different resolutions, wherein subgraph at different levels (or coefficient) represents the input picture that resolution ratio reduces step by step in succession.The formation of multi-resolution framework is to use to carry out certainly image Decomposition of the bottom to top, each tomographic image are that its previous tomographic image is formed by certain template convolution.Multi-resolution framework can To efficiently perform many basic image operations, one group of low pass or band logical image can be generated.It is more by the interconnection of grade and grade Resolution structural provides contacting between Local treatment and Global treatment.
Koendetink proves that Gaussian convolution core is the unique translation core for realizing change of scale, and Gaussian kernel is unique line Property core.
Two-dimensional Gaussian function is defined as follows:
σ represents the standard deviation of Gauss normal distribution, and x, y are independent variable.
One width two dimensional image, the scale space expression under different scale can be by input picture I (x, y) and Gauss nuclear convolution It obtains:
L (x, y, σ)=G (x, y, σ) * I (x, y)
(2) in formula (2), the location of pixels of (x, y) representative image, σ is known as the scale space factor, and the smaller then characterization of value should Image is smoothed fewer, and corresponding scale is also just smaller.Large scale corresponds to the general picture feature of image, and small scale corresponds to figure The minutia of picture.L represents the scale space of image.
Step 2: low-level visual features are extracted
Human visual system is classified and is identified by capturing clarification of objective.Therefore, the spy of target is extracted Sign, is the basic premise for establishing visual attention model.Clarification of objective generally comprises color, brightness, direction, size, shape, fortune Dynamic, solid degree, curvature, topological attribute, closure etc., wherein vision is most sensitive to motion feature.Therefore the present invention is in Itti On the basis of model, for Missile Body mark vision-based detection, motion feature is introduced, and special with color characteristic, brightness, direction Sign is together as low-level visual features.
1st, brightness extracts
If color video frequency image, r, g, b represent red, green and blue in image respectively.Then general brightness Calculation formula is:
I=(r+g+b)/3 (3)
If gray level image, then gray feature directly chooses the gray value of each pixel.2nd, color feature extracted
It is respectively red, green, blue, yellow 4 Color Channels to define R, G, B, Y, then:
R=r- (g+b)/2
G=g- (r+b)/2 (4)
B=b- (r+g)/2
Y=(r+g)/2- (r-g)/2-b
Then RG and BY channels are:RG=R-GB, Y=B- (R+G)/2.
3rd, Directional feature extraction
Direction character is extracted with Gabor filter.One key property of Gabor transformation is that Gabor becomes The regularity of distribution that coefficient discloses the local frequencies of a signal or piece image is changed, rather than Fourier transformation coefficients only Reflect the information of global frequencies.Gabor transformation is highly useful in peacekeeping 2D signal processing, and advantage is multiple Research direction is proved to, such as speech recognition, the detection of signal, compression of images, texture analysis, the segmentation of image and identification side Face.Another interesting characteristics of Gabor transformation are that Gabor basic functions have the property similar to human vision primitive, especially It is secondly this function waveform of Wiki, similar to the impression surface wave shape of mammalian visual systems, this is just mathematically dynamic for research The spatial character in object visual experience face provides a kind of effective mode.Pass through spectrum analysis, it is known that Gabor functions are on frequency domain It is a bandpass filter, by adjusting the direction of Gabor functions, target can be extracted different towards in upper assigned frequency band Information, and other unconcerned information are all filtered out, therefore, it is good property detector.
It is a kind of integrated approach based on gray scale and feature based that Gabor wavelet feature, which describes method,.It says it is based on spy Sign, be because the Gabor wavelet coefficient of extraction different directions and scale is as feature in being distributed from gradation of image, with one group Feature and its location expression target signature are obtained by optimization, do not need to priori.Say it is based on gray scale be because For using Gabor wavelet extract feature can complete reconstruction image, other than the mean value of image, do not lose position, shape Information.
One-dimensional Gabor functions, i.e. 1D-Gabor functions:
In formula, standard deviations of the σ for Gaussian function, w0For the spatial frequency of complex plane wave, x0Center for 1D-Gabor functions Point coordinates takes w0The formula of the odd, even component of=1,1D-Gabor function is respectively:
According to 3 σ principles, l=6 σ/w are taken0
Fig. 4 is the figure of the one-dimensional odd, even component of Gabor functions.
Two-dimensional Gabor function, i.e. 2D-Gabor functions are one under the Gauss envelopes multiple change sine waves along x-axis, such as Shown in Fig. 5.Two-dimensional Gabor function expression is:
The real and imaginary parts of two-dimensional Gabor function are respectively:
There are three the parameters for determining Gabor:f,θ,σ.F determines scale, and θ is directioin parameter, and σ is bigger, and energy more disperses, more It is small more to concentrate, usually there is high frequency σh, low frequency σlTwo centre frequencies.
The selection of deflection θ:The efficient transformation section that can learn θ is [0,2 π], since the Fourier transformation of image exists There is symmetry, so the value range of θ is [0, π] in frequency domain.
It is assumed that f=4, θ=0,30,60,90,120,150, filtering window 32*32, share 24 Gabor filters, Fig. 6 gives the spatial domain expression figure of this group of wave filter.In figure, line direction indicates 6 directions of wave filter from 0 to 150 Variation, column direction illustrate the variation of 4 frequencies of wave filter from high to low.
Input picture I (x, y) and the convolution of Gabor wavelet kernel function are:R (x, y)=∫ ∫ I (ε, η) g (x- ε, y- η) d ε d η (10)
Gabor wavelet transformation after the result is that plural number, take plural number mould | | r (x, y) | | mean value and variance as small echo change The result changed.
As can be seen that with the continuous raising of filter frequencies, the feature for the image that wave filter can characterize significantly increases, I.e. its response in Gabor domains also constantly becomes larger.And as can see from Figure 6 image it is filtered after its direction spy Sign is significantly characterized.Therefore, either from the angle of frequency still from the angle in direction, by Gabor filter After group filtering, apparent response of the feature in Gabor wavelet domain is all higher in image, and unconspicuous feature is filtered Its response is then correspondingly very low after wave.Simultaneously as the advantageous characteristic of Gabor filter in itself, such as scale invariability, rotation Invariance etc. but also target is passing through certain translation, rotates, after dimensional variation, response of the corresponding points in Gabor domains Value equally has constant characteristic, consequently facilitating the target with fixed characteristic point is described using these features.
The present invention selects θ=0, and the Gabor filter output of 30,60,90,120,150 six directions is used as direction character, The Gaobr wave filters of six direction can be obtained by (7) formula, then with these wave filters to each of gaussian pyramid structure Tomographic image filters, you can obtains the Feature Mapping figure on corresponding six direction.
4th, Motion feature extraction
Motion feature mainly includes movement velocity and two category feature of the direction of motion.
1) motion vector extracts.The motion vector of target is extracted using background subtraction.By calculating each pixel in image The relevance in time domain or spatial domain between point using the background of former frame as reference, calculates the background of present frame.Then by present frame Image and its background subtracting, obtain difference image.If t frames image is I (x, y, t), corresponding background image is B (x, y, t), Then difference image is:
2) Motion feature extraction.According to the difference image, the fortune of each pixel in adjacent two field pictures can be obtained Dynamic distance Dij(x, y) projects the move distance of each pixel on 0 °, 45 °, 90 ° and 135 ° four direction, then can obtain each Move distance of the pixel in four directions of motion, and then the movement velocity of 0 °, 45 °, 90 ° and 135 ° four direction can be obtained It is followed successively by:
D(i, j)=pixel is in the component in x directions
D90°(i, j)=pixel is in the component in y directions
Step 3: characteristic pattern is calculated with notable figure
The impression of retina also has center-surrounding features (Center-Surround) in human visual system.Center- Surrounding features refer to that vision system is most sensitive regarding signal to central area, and it is inhibited to regard signal to peripheral region.Vision attention The center of model-periphery operation is to carry out interlayer phase reducing to central stratum and peripheral tier, that is, it is corresponding to calculate different resolution The difference of large and small scale, the mathematic(al) representation of characteristic pattern E are:
E (c, s)=| E (c) Θ E (s) | (13)
Wherein, Θ represents that the interpolation of two different scale images is subtracted each other, i.e., first to peripheral tier into row interpolation, pixel number is increased The difference operation of central stratum and peripheral tier is carried out after to corresponding central stratum pixel number, c represents center scale, behalf neighboring area Scale, c ∈ { 2,3,4 }, i.e., 2,3,4 layers are central stratum, s=c+ δ, scale difference δ ∈ { 3,4 }, i.e., each central stratum corresponding week Boundary layer adds 3 or 4 for the number of plies of central stratum.
It can be in the hope of the characteristic pattern of gray scale, color, direction and motion feature according to above formula.
Gray feature figure:
I (c, s)=| I (c) Θ I (s) | (14)
Color characteristic figure:RG (c, s)=| (R (c)-G (c)) Θ (G (s)-R (s)) |
BY (c, s)=| (B (c)-Y (c)) Θ (Y (s)-B (s)) | (15)
Direction character figure:
O (c, s, θ)=| O (c, θ) Θ O (s, θ) | (16)
Motion feature figure:Ds(c, s)=| Ds(c)ΘDs(s)|
D(c, s)=| D(c)ΘD(s)|
D45°(c, s)=| D45°(c)ΘD45°(s)| (17)
D90°(c, s)=| D90°(c)ΘD90°(s)|
D135°(c, s)=| D135°(c)ΘD135°(s)|
All kinds of characteristic patterns are merged using default normalization operator N (), obtain the notable figure of each category feature.
Gray scale notable figure:
Color notable figure:
Direction notable figure:
Movement velocity notable figure:
Direction of motion notable figure:
⊕ is default blending algorithm, and characteristic remarkable picture further normalizes, and addition obtains comprehensive notable figure:
After obtaining the synthesis notable figure of image, each position competes with one another in comprehensive notable figure, using " the victor is a king " mechanism, The position of triumph becomes focus-of-attention, forms area-of-interest.The transfer of focus-of-attention is realized by " inhibition of return " mechanism.
The above is only the preferred embodiment of the present invention, it is noted that for the ordinary skill people of the art For member, without departing from the technical principles of the invention, several improvement and deformation can also be made, these are improved and deformation Also it should be regarded as protection scope of the present invention.

Claims (4)

1. a kind of region of interest area detecting method of view-based access control model attention model, which is characterized in that include the following steps:
Step 1: establish multi-scale image structure
One width two dimensional image, the scale space expression under different scale can be by input picture I (x, y) and Gaussian kernel G (x, y, σ) Convolution obtains:
L (x, y, σ)=G (x, y, σ) * I (x, y) (2)
In formula (2), the location of pixels of (x, y) representative image, σ is known as the scale space factor, and value is smaller, characterizes the image quilt Smooth is fewer, and corresponding scale is also just smaller, and large scale corresponds to the general picture feature of image, and small scale corresponds to the thin of image Feature is saved, L represents the scale space of image;
Step 2: low-level visual features are extracted
1st, brightness extracts
If color video frequency image, r, g, b represent red, green and blue in image respectively, then brightness calculation formula is:
I=(r+g+b)/3 (3)
If gray level image, then gray feature directly chooses the gray value of each pixel;
2nd, color feature extracted
It is respectively red, green, blue, yellow 4 Color Channels to define R, G, B, Y, then:
Y=(r+g)/2- (r-g)/2-b
Then RG and BY channels are:RG=R-GB, Y=B- (R+G)/2;
3rd, Directional feature extraction
Direction character is extracted with Gabor filter:
One-dimensional Gabor functions, i.e. 1D-Gabor functions:
In formula, standard deviations of the σ for Gaussian function, w0For the spatial frequency of complex plane wave, x0Central point for 1D-Gabor functions is sat Mark, the formula of the odd, even component of 1D-Gabor functions are respectively:
According to 3 σ principles, l=6 σ/w are taken0
Two-dimensional Gabor function, i.e. 2D-Gabor functions are one under the Gauss envelopes multiple change sine waves along x-axis, two dimension Gabor function expressions are:
The real and imaginary parts of two-dimensional Gabor function are respectively:
θ is directioin parameter, and σ is bigger, and energy more disperses, smaller more to concentrate, and has high frequency σh, low frequency σlTwo centre frequencies;
The selection of deflection θ:θ=0,30,60,90,120,150, filtering window 32*32 share 24 Gabor filtering Device;
Input picture I (x, y) and the convolution of Gabor wavelet kernel function are:
R (x, y)=∫ ∫ I (ε, η) g (x- ε, y- η) d ε d η (8)
Gabor wavelet transformation after the result is that plural number, take plural number mould | | r (x, y) | | mean value and variance as wavelet transformation As a result;
θ=0 is selected, the Gabor filter output of 30,60,90,120,150 six directions passes through (7) formula as direction character The Gaobr wave filters of six direction are obtained, then each tomographic image of gaussian pyramid structure is filtered with these wave filters, is obtained Feature Mapping figure on to corresponding six direction;
4th, Motion feature extraction
Motion feature includes movement velocity and two category feature of the direction of motion;
1) motion vector extracts:The motion vector of target is extracted using background subtraction, by calculating in image between each pixel Time domain or spatial domain on relevance, using the background of former frame as reference, the background of present frame is calculated, then by current frame image With its background subtracting, difference image is obtained, if t frames image is I (x, y, t), corresponding background image is B (x, y, t), then poor Partial image is:
2) Motion feature extraction:According to the difference image, the move distance D of each pixel in adjacent two field pictures is obtainedij (x, y) projects the move distance of each pixel on 0 °, 45 °, 90 ° and 135 ° four direction, then obtains each pixel in four fortune Move distance on dynamic direction, and then the movement velocity for obtaining 0 °, 45 °, 90 ° and 135 ° four direction is followed successively by:
Step 3: characteristic pattern is calculated with notable figure
The center of visual attention model-periphery operation is to carry out interlayer phase reducing to central stratum and peripheral tier, that is, is calculated different The difference of the corresponding large and small scale of resolution ratio, the mathematic(al) representation of characteristic pattern E are:
E (c, s)=| E (c) Θ E (s) | (11)
Wherein, Θ represents that the interpolation of two different scale images is subtracted each other, i.e., first to peripheral tier into row interpolation, pixel number is increased to pair Carrying out the difference operation of central stratum and peripheral tier after the central stratum pixel number answered, c represents center scale, behalf neighboring area scale, C ∈ { 2,3,4 }, i.e., 2,3,4 layers are central stratum, s=c+ δ, scale difference δ ∈ { 3,4 }, i.e., the corresponding peripheral tier of each central stratum is The number of plies of central stratum adds 3 or 4;
The characteristic pattern of gray scale, color, direction and motion feature is acquired according to formula (11):
Gray feature figure:
I (c, s)=| I (c) Θ I (s) | (12)
Color characteristic figure:RG (c, s)=| (R (c)-G (c)) Θ (G (s)-R (s)) |
BY (c, s)=| (B (c)-Y (c)) Θ (Y (s)-B (s)) | (13)
Direction character figure:
O (c, s, θ)=| O (c, θ) Θ O (s, θ) | (14)
Motion feature figure:
All kinds of characteristic patterns are merged using default normalization operator N (), obtain the notable figure of each category feature.
Gray scale notable figure:
Color notable figure:
Direction notable figure:
Movement velocity notable figure:
Direction of motion notable figure:
To preset blending algorithm, characteristic remarkable picture further normalizes, and addition obtains comprehensive notable figure:
After obtaining the synthesis notable figure of image, each position competes with one another in comprehensive notable figure, and the position of triumph becomes focus-of-attention, Form area-of-interest.
2. the method as described in claim 1, which is characterized in that in step 3, comprehensive notable figure is inputted into Itti vision modes In in the victor is a king network, select the region of conspicuousness maximum in notable figure successively as area-of-interest.
3. the method as described in claim 1, which is characterized in that further included after step 3 through returning in Itti vision modes Return the step of suppression mechanism realizes the transfer of focus-of-attention.
4. the method as described in claims 1 or 2 or 3, which is characterized in that w0=1.
CN201810034712.3A 2018-01-15 2018-01-15 The region of interest area detecting method of view-based access control model attention model Pending CN108154147A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810034712.3A CN108154147A (en) 2018-01-15 2018-01-15 The region of interest area detecting method of view-based access control model attention model

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810034712.3A CN108154147A (en) 2018-01-15 2018-01-15 The region of interest area detecting method of view-based access control model attention model

Publications (1)

Publication Number Publication Date
CN108154147A true CN108154147A (en) 2018-06-12

Family

ID=62461249

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810034712.3A Pending CN108154147A (en) 2018-01-15 2018-01-15 The region of interest area detecting method of view-based access control model attention model

Country Status (1)

Country Link
CN (1) CN108154147A (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109284670A (en) * 2018-08-01 2019-01-29 清华大学 A kind of pedestrian detection method and device based on multiple dimensioned attention mechanism
CN109670516A (en) * 2018-12-19 2019-04-23 广东工业大学 A kind of image characteristic extracting method, device, equipment and readable storage medium storing program for executing
CN110688783A (en) * 2019-11-07 2020-01-14 济南大学 Downhole pump target detection method based on MATLAB GUI
CN110956086A (en) * 2019-10-24 2020-04-03 国家电网有限公司 Machine room equipment online identification method based on image identification technology
CN111654647A (en) * 2020-06-02 2020-09-11 苏州睿新微系统技术有限公司 Induction imaging control method, device and control circuit of infrared detector
CN112699832A (en) * 2021-01-12 2021-04-23 腾讯科技(深圳)有限公司 Target detection method, device, equipment and storage medium
CN113780078A (en) * 2021-08-05 2021-12-10 广州西威科智能科技有限公司 Method for quickly and accurately identifying fault object in unmanned visual navigation

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100189354A1 (en) * 2009-01-28 2010-07-29 Xerox Corporation Modeling images as sets of weighted features
CN103106672A (en) * 2013-01-25 2013-05-15 西北工业大学 Method for detecting interested region of image based on color features
CN104835175A (en) * 2015-05-26 2015-08-12 西南科技大学 Visual attention mechanism-based method for detecting target in nuclear environment
CN105825238A (en) * 2016-03-30 2016-08-03 江苏大学 Visual saliency object detection method
CN106462771A (en) * 2016-08-05 2017-02-22 深圳大学 3D image significance detection method

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100189354A1 (en) * 2009-01-28 2010-07-29 Xerox Corporation Modeling images as sets of weighted features
CN103106672A (en) * 2013-01-25 2013-05-15 西北工业大学 Method for detecting interested region of image based on color features
CN104835175A (en) * 2015-05-26 2015-08-12 西南科技大学 Visual attention mechanism-based method for detecting target in nuclear environment
CN105825238A (en) * 2016-03-30 2016-08-03 江苏大学 Visual saliency object detection method
CN106462771A (en) * 2016-08-05 2017-02-22 深圳大学 3D image significance detection method

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
均陈晓: "基于尺度不变特征变换的图像匹配1", 《百度文库》 *
徐振辉: "一种基于改进视觉注意模型和局部自相似性的目标自动检测算法研究", 《科学技术与工程》 *
曾姝彦等: "基于Gabor滤波器的图像目标识别方法", 《北京航天航空大学学报》 *
韩成美: "图像感兴趣区域数字水印算法研究", 《中国优秀硕士学位论文全文数据库 信息科技辑》 *

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109284670A (en) * 2018-08-01 2019-01-29 清华大学 A kind of pedestrian detection method and device based on multiple dimensioned attention mechanism
CN109670516A (en) * 2018-12-19 2019-04-23 广东工业大学 A kind of image characteristic extracting method, device, equipment and readable storage medium storing program for executing
CN110956086A (en) * 2019-10-24 2020-04-03 国家电网有限公司 Machine room equipment online identification method based on image identification technology
CN110688783A (en) * 2019-11-07 2020-01-14 济南大学 Downhole pump target detection method based on MATLAB GUI
CN111654647A (en) * 2020-06-02 2020-09-11 苏州睿新微系统技术有限公司 Induction imaging control method, device and control circuit of infrared detector
CN111654647B (en) * 2020-06-02 2022-12-09 苏州睿新微系统技术有限公司 Induction imaging control method, device and control circuit of infrared detector
CN112699832A (en) * 2021-01-12 2021-04-23 腾讯科技(深圳)有限公司 Target detection method, device, equipment and storage medium
CN112699832B (en) * 2021-01-12 2023-07-04 腾讯科技(深圳)有限公司 Target detection method, device, equipment and storage medium
CN113780078A (en) * 2021-08-05 2021-12-10 广州西威科智能科技有限公司 Method for quickly and accurately identifying fault object in unmanned visual navigation
CN113780078B (en) * 2021-08-05 2024-03-19 广州西威科智能科技有限公司 Rapid and accurate fault object identification method in unmanned visual navigation

Similar Documents

Publication Publication Date Title
CN108154147A (en) The region of interest area detecting method of view-based access control model attention model
Zhang et al. A categorization of multiscale-decomposition-based image fusion schemes with a performance study for a digital camera application
Vincent et al. A descriptive algorithm for sobel image edge detection
Batool et al. Fast detection of facial wrinkles based on Gabor features using image morphology and geometric constraints
CN105550685B (en) The large format remote sensing image area-of-interest exacting method of view-based access control model attention mechanism
CN104463877B (en) A kind of water front method for registering based on radar image Yu electronic chart information
CN107066916B (en) Scene semantic segmentation method based on deconvolution neural network
WO2018145470A1 (en) Image detection method and device
CN104835175B (en) Object detection method in a kind of nuclear environment of view-based access control model attention mechanism
CN104217221A (en) Method for detecting calligraphy and paintings based on textural features
CN107330371A (en) Acquisition methods, device and the storage device of the countenance of 3D facial models
Draper et al. Evaluation of selective attention under similarity transformations
CN102844766A (en) Human eyes images based multi-feature fusion identification method
CN110163119A (en) A kind of finger vein identification method and system
CN105809173B (en) A kind of image RSTN invariable attribute feature extraction and recognition methods based on bionical object visual transform
CN101615245A (en) Expression recognition method based on AVR and enhancing LBP
CN107292907A (en) A kind of method to following target to be positioned and follow equipment
CN105023253A (en) Visual underlying feature-based image enhancement method
CN107945111A (en) A kind of image split-joint method based on SURF feature extraction combination CS LBP descriptors
CN110263605A (en) Pedestrian's dress ornament color identification method and device based on two-dimension human body guise estimation
CN115100719A (en) Face recognition method based on fusion of Gabor binary pattern and three-dimensional gradient histogram features
CN106599891A (en) Remote sensing image region-of-interest rapid extraction method based on scale phase spectrum saliency
Yang et al. Adversarial reconstruction CNN for illumination-robust frontal face image recovery and recognition
CN106296632B (en) A kind of well-marked target detection method based on amplitude spectrum analysis
Roy et al. LGVTON: a landmark guided approach for model to person virtual try-on

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20180612