CN113192092B - Contour detection method for simulating fusion of properties of receptor field of XYW cell - Google Patents

Contour detection method for simulating fusion of properties of receptor field of XYW cell Download PDF

Info

Publication number
CN113192092B
CN113192092B CN202110493954.0A CN202110493954A CN113192092B CN 113192092 B CN113192092 B CN 113192092B CN 202110493954 A CN202110493954 A CN 202110493954A CN 113192092 B CN113192092 B CN 113192092B
Authority
CN
China
Prior art keywords
response
channel
contour
image
cells
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202110493954.0A
Other languages
Chinese (zh)
Other versions
CN113192092A (en
Inventor
林川
张贞光
吴海晨
万术娟
李福章
文泽奇
潘勇才
韦艳霞
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangxi University of Science and Technology
Original Assignee
Guangxi University of Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangxi University of Science and Technology filed Critical Guangxi University of Science and Technology
Priority to CN202110493954.0A priority Critical patent/CN113192092B/en
Publication of CN113192092A publication Critical patent/CN113192092A/en
Application granted granted Critical
Publication of CN113192092B publication Critical patent/CN113192092B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/13Edge detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/181Segmentation; Edge detection involving edge growing; involving edge linking
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/44Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20024Filtering details
    • G06T2207/20028Bilateral filtering
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20212Image combination
    • G06T2207/20221Image fusion; Image merging

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

The invention provides a contour detection method for simulating the fusion of properties of a receptor field of XYW cells, which comprises the following steps: A. acquiring a gray image and a parallax image of an image to be detected; B. carrying out rapid bilateral filtering on the gray level image of the image to be detected; C. the filtered image is processed by X, Y and W channels respectively; D. inhibiting through sparse coding, and removing redundant textures; E. summing the suppressed X-channel contour response, Y-channel contour response and W-channel contour response with the parallax contour response of the corresponding channel of the parallax image respectively; F. fusing the X channel contour response, the Y channel contour response and the W channel contour response of the strengthened weak contour by using different watching planes to obtain a final contour and an optimal direction; G. and carrying out non-maximum suppression on the final contour response by using the optimal direction to obtain a suppressed final contour. The invention enhances the integrity and continuity of the contour and further improves the detection performance of the contour.

Description

Contour detection method for simulating fusion of properties of receptor field of XYW cell
Technical Field
The invention relates to the field of image processing, in particular to a contour detection method for simulating fusion of XYW cell receptive field characteristics.
Background
In real life, the target contour is often embedded in a complex background, and it is a difficult task to distinguish the target contour from the background texture well. The following three neuronal cells with different space summation characteristics exist in the biological visual nerve mechanism: x-type cells, Y-type cells, and W-type cells. However, the existing contour detection model based on biological visual mechanism only considers the action mechanism of X-type cells or Y-type cells singly, neglects the action of W-type cells in contour detection, and does not consider the effective fusion of the response characteristics of the three types of cells. And the lack of important characteristic clues of disparity information causes incomplete and discontinuous contour detection in previous work, thereby reducing the performance of a contour detection model.
Disclosure of Invention
The invention aims to provide a contour detection method for simulating fusion of receptive field characteristics of XYW cells, and provides a bionic contour detection model based on a unique response mechanism of X, Y and W neuron cells to different visual stimuli in a V1 region and an information fusion method thereof.
The technical scheme of the invention is as follows:
the contour detection method for simulating the fusion of the properties of the receptor field of the XYW cells comprises the following steps:
A. acquiring a gray image and a parallax image of an image to be detected;
B. carrying out rapid bilateral filtering on a gray level image of an image to be detected to obtain a filtered image;
C. processing the filtered image by X, Y and W channels respectively to obtain X channel contour response, Y channel contour response and W channel contour response;
D. respectively suppressing the X channel contour response, the Y channel contour response and the W channel contour response through sparse coding, removing redundant textures, and obtaining the suppressed X channel contour response, Y channel contour response and W channel contour response;
E. respectively summing the suppressed X-channel contour response, Y-channel contour response and W-channel contour response with the parallax contour response of the corresponding channel of the parallax image to obtain the X-channel contour response, Y-channel contour response and W-channel contour response with strengthened weak contours;
F. fusing the X channel contour response, the Y channel contour response and the W channel contour response which strengthen the weak contour by using different watching planes to obtain a final contour and an optimal direction;
G. and carrying out non-maximum suppression on the final contour response by using the optimal direction to obtain a suppressed final contour.
In step B, the formula of the fast bilateral filtering is:
f g (x,y)=exp(μ· - f b (x,y)+f d (x,y)) (1)
Figure GDA0004054055870000021
/>
f d (x,y)=log(I(x,y)-f b (x,y)) (3)
where I (x, y) represents a gray image, σ represents a standard deviation of a Gaussian function, and f b (x, y) represents an image of the gray-scale image after smoothing by a Gaussian function, f d (x, y) is represented by a gray image and f b (x, y) is a difference and logarithmic to obtain an image, and μ represents f b (x, y) and f d The (x, y) link strength, symbol · represents multiplication, and x represents convolution.
In the step C, the formula for obtaining the X channel contour response is:
FR X⌒C (x,y)=E X⌒C (x,y)-α X⌒C ·T X⌒C (x,y) (4)
wherein: e X⌒C (X, y) is the inhibitory response of the classical receptor field of the X channel, T X⌒C (X, y) is the inhibitory response of the X-channel nonclassical receptive field, α X⌒C The inhibition intensity is shown.
The inhibition response T for obtaining the X channel classical receptive field X⌒C The process of (x, y) is as follows:
a. the classical receptive field of X cells was simulated using the first derivative of gaussian with directional selectivity:
Figure GDA0004054055870000022
Figure GDA0004054055870000023
Figure GDA0004054055870000024
wherein, the rotation angle theta represents the direction parameter 0 epsilon [0, pi) of the receptive field, and gamma represents the constant gamma =0.5 of the ratio of the long axis and the short axis of the receptive field;
b. order to
Figure GDA0004054055870000025
Obtaining the energy response E of the classical receptive field of the X cells to the external stimulus X⌒C (x,y;θ i ):
E X⌒C (x,y;θ i )=f C (x,y)*RF(x,y;θ i ,σ XC ) (8)
Wherein C belongs to { g, d }, g represents a gray level image, and d represents a parallax image;
theta as above i Each pixel point is represented as N θ The direction angle in each direction is calculated as follows:
Figure GDA0004054055870000031
c. comparing the energy information E of each pixel point in different direction angles X⌒C (x,y;θ i ) (ii) a Obtaining the optimal response E of the current position X⌒C (x, y) obtaining an inhibitory response of the classical receptive field; simultaneously obtaining the corresponding optimal direction angle theta X⌒g (x,g):
E X⌒C (x,y)=max{E X⌒C (x,y;θ i )|i=1,2,…,N θ } (10)
Figure GDA0004054055870000034
The inhibition response E for acquiring the X channel non-classical receptive field X⌒C The process of (x, y) is as follows:
a. optimal energy response and suppression weight omega of each pixel point in classical receptive field u (x, y) convolution to obtain the inhibitory response T of the non-classical receptive field X⌒C (x,y):
T X⌒C (x,y)=E X⌒C (x,y)*ω u (x,y) (12)
Figure GDA0004054055870000032
Wherein | | xi | purple 1 Represents L 1 Norm, H (Z) = max (0, Z);
DOG is a two-dimensional Gaussian difference function, and the formula is as follows:
Figure GDA0004054055870000033
wherein σ u Standard deviation of the non-classical receptive field is indicated and k represents the scale ratio of the peripheral inhibitory region to the central excitatory region.
In the step C, the formula for obtaining the Y channel contour response is:
FR Y⌒C (x,y)=S Y⌒C (x,y)-α Y⌒C ·T Y⌒C (x,y) (15)
wherein S is Y⌒C (x, Y) is the Y-channel nonlinear quench response, T Y⌒C (x, Y) is the Y-channel linear suppression response, α Y⌒C The inhibition intensity is shown.
Obtaining Y-channel nonlinear suppression response S Y⌒C The process of (x, y) is as follows:
a. by sigma Y Substitution of σ in equations 1-3, 5, 6 X Obtaining the classical receptive field response E of the Y cell Y⌒C (x,y;θ i ) Then E is added Y⌒C (x,y;θ i ) Substituting into formula 9-11 to obtain the optimal response E of Y cell Y⌒C (x, y) and corresponding optimal direction angle θ y⌒g (x,y);
b. Setting a region S of finite size xy Each pixel point in the region corresponds to the central position of one subunit, the rectification property of the subunit is simulated by a Gaussian function, and then, each pixel point is subjected to rectificationSumming the contributions of the subunits results in a non-linear subunit response S Y⌒C (x, y) in the formula:
Figure GDA0004054055870000041
Figure GDA0004054055870000042
wherein the content of the first and second substances,
Figure GDA0004054055870000043
representing a window S centered on (x, y) xy Intrinsic response E Y_C Information in (x, y). Sigma s =σ YC Calculating the spatial standard deviation of the subunit size; s xy Radius of (d) size r s =12·σ YC (ii) a N is
Figure GDA0004054055870000044
The number of pixels of (a); v represents the ordinate of the Y-type cell subunit;
the acquisition of the linear suppressive response T of the Y channel Y⌒C The process of (x, y) is as follows: will S Y⌒C (x, y) instead of E X⌒C (x, y) is calculated by substituting (x, y) as an input into equation 12.
In the step C, the calculation process for obtaining the W channel contour response is as follows:
let σ = σ n - Will σ n Substituting into formula 5 to obtain Gaussian first derivative function RF (x, y; theta, sigma) with different scales and shapes n ) (ii) a Standard deviation sigma n The calculation formula of (a) is as follows:
Figure GDA0004054055870000045
σ WC is the standard deviation of CRF for W cells, since CRF is similar for W cells and X cells, here σ WC =σ XC (ii) a S represents the classical perception of W cellsThe size of the field scale change; n denotes the number of classical receptive field scale changes, where n =25/σ WC
b. From a Gaussian first derivative function RF (x, y; theta, sigma) n ) Selecting different micro-motion areas, wherein the selection formula of the micro-motion areas is as follows:
Figure GDA0004054055870000051
RF (x, y; theta, sigma) n ) Multiplying by equation 19 yields a gaussian first derivative function with fixation and fixation micromotion characteristics:
RM n (x,y;θ i ,σ n )=RF(x,y;θ i ,σ n )·M n (x,y) (20);
c. the response of the classical receptive field under the visual micromotion to the external stimulus is input from the characteristic diagram and RM n (x,y;θ i ,σ n ) Convolution yields an energy response E W⌒C (x,y;θ i ,σ n ):
E W⌒C (x,y;θ i ,σ n )=f C (x,y)*RM n (x,y;θ i ,σ n ) (21);
d. From E W⌒C (x,y;θ i ,σ n ) In place of E X⌒C (x,y;θ i ) Substituting into the formula 9-11 to calculate the optimal response E of W cells W⌒C (x, y) and corresponding optimal direction angle θ W⌒g (x,y);
From E W⌒C (x, y) instead of E X⌒C (x, y) as input into equation 12 to calculate the non-classical receptor suppression response T of W cells W⌒C (x,y);
e. W cell contour response FR W⌒C The formula for the calculation (x, y) is:
FR W⌒C (x,y)=E W⌒C (x,y)-α W⌒C ·T W⌒C (22)
wherein E is W⌒C (x, y) is the optimal response of W cells, T X⌒C (x, y) is WtongInhibitory response of the Dow nonclassical receptor field, alpha W⌒C The inhibition intensity is shown.
The sparse coding formula in the step D is as follows:
Figure GDA0004054055870000052
wherein e is M (x,y)∈{FR M⌒g (X, Y) }, namely M belongs to { X, Y, W } and represents the stimulation response under X, Y and W channels based on the gray level image information;
Figure GDA0004054055870000053
is a histogram of local gradient amplitude values of each information channel centered at (x, y), and t is
Figure GDA0004054055870000054
Dimension of (d); i 1 Is L 1 Norm, | | 2 Is L 2 A norm; />
Figure GDA0004054055870000055
Represents->
Figure GDA0004054055870000056
Min represents the minimum value of the two;
applying sparse coding to inhibit, and obtaining formulas of inhibited X channel contour response, Y channel contour response and W channel contour response as follows;
LR M⌒g (x,y)=sm M (x,y)·FR M⌒g (x,y) (24)。
in the step E, the suppressed X-channel contour response, Y-channel contour response and W-channel contour response are respectively compared with the parallax contour response FR of the corresponding channel of the parallax image M⌒d The formula for the sum (x, y) is:
R M =FR M⌒d (x,y)+LR M⌒g (x,y) (25)。
in the step F, the process of fusing the X channel contour response, the Y channel contour response and the W channel contour response which strengthen the weak contour by using different gazing planes is as follows:
a. the visual information of each watching plane is processed by an X channel, the visual information of a near plane is processed by a Y channel, the visual information of a far plane is processed by a W channel, and the confirmation formula of the watching planes is as follows:
Figure GDA0004054055870000061
b. normalizing the parallax image to [0, P]Parallax range of (1), using d (x, y) represents the parallax image after normalization; setting a parallax range on a watching plane as beta; when round (G (x, y), 1) =0.5, it represents a near plane; when round (G (x, y), 1) = -0.5 is set, a far plane is represented; setting G (x, y) to other values, indicating on the gaze plane;
c. and fusing visual information on X, Y and W channels through three gazing planes to obtain a final contour R (X, Y) and an optimal direction theta (X, Y):
Figure GDA0004054055870000062
Figure GDA0004054055870000063
the invention designs a unique simulation function, models the respective unique physiological characteristic response of X, Y and W cells, respectively and simultaneously processes the gray image and the parallax image, and finally effectively fuses the contour characteristics of X, Y and W channels according to the fixation surface determined by the parallax image. Experimental results show that the contour detection method based on the model can better preserve the integrity of the contour and inhibit background textures, and has good application prospect.
Drawings
FIG. 1 is a schematic diagram of the fusion process of three channels X, Y and W according to three fixation surfaces in example 1;
FIG. 2 is a graph showing the comparison between the contour detection method of example 1 and the contour detection method of reference 1
Detailed Description
Example 1
1. The contour detection method for simulating the fusion of the properties of the receptor field of the XYW cell provided by the embodiment comprises the following steps:
the contour detection method for simulating the fusion of the properties of the receptor field of the XYW cell comprises the following steps:
A. acquiring a gray image and a parallax image of an image to be detected;
B. carrying out rapid bilateral filtering on the gray level image of the image to be detected to obtain a filtered image; the formula for fast bilateral filtering is:
f g (x,y)=exp(μ· - f b (x,y)+f d (x,y)) (1)
Figure GDA0004054055870000071
f d (x,y)=log(I(x,y)-f b (x,y)) (3)
wherein I (x, y) represents a grayscale image; σ represents a standard deviation of the Gaussian function; f. of b (x, y) represents an image after the gray image is smoothed by a gaussian function; f. of d (x, y) is represented by a gray image and f b (x, y) subtracting and logarithmically obtaining an image; μ denotes f b (x, y) and f d The value of the connection strength of (x, y) in this embodiment is 0.25; symbol · denotes multiplication; * Represents a convolution;
C. processing the filtered image by X, Y and W channels respectively to obtain X channel contour response, Y channel contour response and W channel contour response;
the formula for obtaining the X channel contour response is as follows:
FR X⌒C (x,y)=E X⌒C (x,y)-α X⌒C ·T X⌒C (x,y) (4)
wherein E is X⌒C (X, y) is X channelSuppression of the classical receptive field, T X⌒C (X, y) is the inhibitory response of the X-channel nonclassical receptive field, α X⌒C The inhibition intensity is shown.
The inhibition response E for obtaining the X channel classical receptive field X⌒C The process of (x, y) is as follows:
a. the classical receptive field of X cells was simulated using the first derivative of gaussian with directional selectivity:
Figure GDA0004054055870000072
Figure GDA0004054055870000073
Figure GDA0004054055870000074
wherein, the rotation angle theta represents the direction parameter theta epsilon [0, pi ] of the receptive field, gamma represents the constant of the ratio of the long axis and the short axis of the receptive field, and gamma =0.5;
b. let σ = σ XC Obtaining the energy response E of the classical receptive field of the X cells to the external stimulus X⌒C (x,y;θ i ):
E X⌒C (x,y;θ i )=f C (x,y)*RF(x,y;θ i ,σ XC ) (8)
Wherein C belongs to { g, d }, g represents a gray level image, and d represents a parallax image;
theta as above i Each pixel point is represented as N θ The direction angle in each direction is calculated as follows:
Figure GDA0004054055870000081
c. comparing the energy information E of each pixel point in different direction angles X⌒C (x,y;θ i ) (ii) a Is obtained whenOptimal response E of the front position X⌒C (x, y) obtaining an inhibitory response of the classical receptive field; simultaneously acquiring corresponding optimal direction angle theta X⌒g (x,y):
E X⌒C (x,y)=max{E X⌒C (x,y;θ i )|i=1,2,…,N θ } (10)
Figure GDA0004054055870000082
Obtaining the inhibition response E of the X channel non-classical receptive field X⌒C The process of (x, y) is as follows:
a. optimal energy response and suppression weight omega of each pixel point in classical receptive field u (x, y) convolution to obtain the inhibitory response T of the non-classical receptive field X⌒C (x,y):
T X⌒C (x,y)=E X⌒C (x,y)*ω n (x,y) (12)
Figure GDA0004054055870000083
Wherein | | xi | purple 1 Represents L 1 Norm, H (Z) = max (0, Z);
DOG is a two-dimensional gaussian difference function, and its formula is:
Figure GDA0004054055870000084
wherein σ u Standard deviation of the non-classical receptive field is indicated, K represents the scale ratio of the peripheral inhibitory region to the central excitatory region, K =4 in this example.
In the step C, the formula for obtaining the Y channel contour response is:
FR Y⌒C (x,y)=S Y⌒C (x,y)-α Y⌒C ·T Y⌒C (x,y) (15)
wherein S is Y⌒C (x, Y) is the Y-channel nonlinear quench response, T Y⌒C (x, Y) is the Y-channel linear suppression response, α Y⌒C The inhibition intensity is shown.
The obtained Y channel nonlinear suppression response T X⌒C The process of (x, y) is as follows:
a. by sigma Y Substitution of σ in equations 1-3, 5, 6 X Obtaining the classical receptive field response E of the Y cell Y⌒C (x,y;θ i ) Then E is added Y⌒C (x,y;θ i ) Substituting into formula 9-11 to obtain the optimal response E of Y cell Y⌒C (x, y) and corresponding optimal direction angle θ y⌒g (x,y);
b. Setting a region S of finite size xy Each pixel point in the region corresponds to the central position of one subunit, the rectification property of the subunit is simulated by a Gaussian function, and then the nonlinear subunit response S is obtained by summing the contribution of each subunit Y⌒C (x, y) in the formula:
Figure GDA0004054055870000091
Figure GDA0004054055870000092
wherein the content of the first and second substances,
Figure GDA0004054055870000093
representing a window S centered at (x, y) xy Intrinsic response E Y_C Information in (x, y). Sigma s =σ YC Calculating the spatial standard deviation of the subunit size; s xy Radius of (r) s =12·σ YC (ii) a N is
Figure GDA0004054055870000094
The number of pixels of (a); v represents the ordinate of the Y-type cell subunit;
obtaining a Y-channel linear inhibitory response T Y⌒C The process of (x, y) is as follows: will S Y⌒C (x, y) instead of E X⌒C (x, y) as inputAnd substituting the formula 12 for calculation.
In the step C, the calculation process for obtaining the W channel contour response is as follows:
let σ = σ n - Will σ n Substituting into formula 5 to obtain first derivative function RF (x, y; theta, sigma) of Gaussian with different scales and shapes n ) (ii) a Standard deviation sigma n The calculation formula of (a) is as follows:
Figure GDA0004054055870000095
σ WC is the standard deviation of CRF for W cells, since W cells are similar to CRF for X cells, here σ WC =σ XC (ii) a S represents the size of the scale change of the classical receptive field of the W cell, and S is 0.2 in the embodiment; n denotes the number of classical receptive field scale changes, where n =25/σ WC
b. From a Gaussian first derivative function RF (x, y; theta, sigma) n ) Selecting different micro-motion areas, wherein the selection formula of the micro-motion areas is as follows:
Figure GDA0004054055870000101
RF (x, y; theta, sigma) n ) Multiplying by equation 19 yields a gaussian first derivative function with fixation and micromotion characteristics:
RM n (x,y;θ i ,σ n )=RF(x,y;θ i ,σ n )·M n (x,y) (20);
c. the response of the classical receptive field under the visual micromotion to the external stimulus is input from the characteristic diagram and RM n (x,y;θ i ,σ n ) Convolution yields an energy response E W⌒C (x,y;θ i ,σ n ):
E W⌒C (x,y;θ i ,σ n )=f C (x,y)*RM n (x,y;θ i ,σ n ) (21);
d. From E W⌒C (x,y;θ i ,σ n ) In place of E X⌒C (x,y;θ i ) Substituting into the formula 9-11 to calculate the optimal response E of W cells W⌒C (x, y) and corresponding optimal direction angle θ W⌒g (x,y);
From E W⌒C (x, y) instead of E X⌒C (x, y) as input into equation 12 to calculate the non-classical receptor suppression response T of W cells W⌒C (x,y);
e. W cell contour response FR W⌒C The formula for (x, y) is:
FR W⌒C (x,y)=E W⌒C (x,y)-α W⌒C ·T W⌒C (22)
wherein E is W⌒C (x, y) is the optimal response of W cells, T X⌒C (x, y) is the inhibitory response of the W channel non-classical receptive field, α W⌒C The inhibition intensity is shown.
D. Respectively suppressing the X channel contour response, the Y channel contour response and the W channel contour response through sparse coding, removing redundant textures, and obtaining suppressed X channel contour response, Y channel contour response and W channel contour response; the sparse coding formula is as follows:
Figure GDA0004054055870000102
wherein e is M (x,y)∈{FR M⌒g (X, Y) }, namely M belongs to { X, Y, W } and represents the stimulation response under X, Y and W channels based on the gray level image information;
Figure GDA0004054055870000111
the histogram of the local gradient amplitude values of each information channel with (x, y) as the center; t is
Figure GDA0004054055870000112
In this example, a 25 × 25 square window; i 1 Is L 1 Norm, | · | luminance 2 Is L 2 A norm;
Figure GDA0004054055870000113
represents->
Figure GDA0004054055870000114
Min represents the minimum value of the two;
applying sparse coding to inhibit, and obtaining formulas of inhibited X channel contour response, Y channel contour response and W channel contour response as follows;
LR M⌒g (x,y)=sm M (x,y)·FR M⌒g (x,y) (24)。
E. summing the suppressed X-channel contour response, Y-channel contour response and W-channel contour response with the parallax contour response of the corresponding channel of the parallax image respectively to obtain the X-channel contour response, Y-channel contour response and W-channel contour response with strengthened weak contours;
respectively matching the suppressed X-channel contour response, Y-channel contour response and W-channel contour response with the parallax contour response FR of the corresponding channel of the parallax image M⌒d The formula for the sum (x, y) is:
R M =FR M⌒d (x,y)+LR M⌒g (x,y) (25)。
F. as shown in fig. 1, fusing the X-channel contour response, the Y-channel contour response, and the W-channel contour response, which strengthen the weak contour, by using different gaze planes to obtain a final contour and an optimal direction;
the process of fusing the X channel contour response, the Y channel contour response and the W channel contour response which strengthen the weak contour by using different watching planes comprises the following steps:
a. the visual information of each watching plane is processed by an X channel, the visual information of a near plane is processed by a Y channel, the visual information of a far plane is processed by a W channel, and the confirmation formula of the watching planes is as follows:
Figure GDA0004054055870000115
b. normalizing the parallax image to [0, P]Parallax range of (1), using d (x, y) represents the parallax image after normalization; setting the parallax range on the gazing plane as beta, wherein beta is set as 2 in the embodiment; when round (G (x, y), 1) =0.5, it means a near plane; when round (G (x, y), 1) = -0.5 is set, a far plane is represented; setting G (x, y) to other values, indicating on the gaze plane;
c. and fusing visual information on X, Y and W channels through three gazing planes to obtain a final contour R (X, Y) and an optimal direction theta (X, Y):
Figure GDA0004054055870000116
Figure GDA0004054055870000121
G. and carrying out non-maximum suppression on the final contour response by using the optimal direction to obtain a suppressed final contour.
2. Comparing the contour identification performance tests based on the method:
1. the method of document 1 was used for comparison:
document 1: lin C, zhang Q, cao Y. Multi-scale control detection model based on quantitative on temporal movement mechanism [ J ]. Signal, image and Video Processing, 2019.
The parameters used in document 1 are the optimal parameters of the model, as in the original text.
2. For quantitative performance evaluation of the final profile, we used the same performance measurement criteria as in document 1, specifically evaluated as follows:
Figure GDA0004054055870000122
wherein P represents precision and R represents recall. The larger the value of F, the better the performance.
The parameters used in document 1 are the optimal parameters of the model, as in the original text.
The comparative test results are shown in FIG. 2:
FIG. 2 shows four natural images randomly selected by the NYUD data set, a real profile, an optimal profile detected by the method of document 1 and an optimal profile of the invention; among them, F-score is shown in the lower right corner of FIG. 2, and the present invention circled in a red oval in the figure is better in texture suppression than document 1 in this region.
From the experimental effect, the detection method of the invention is superior to the detection method of the document 1.

Claims (10)

1. A contour detection method for simulating the fusion of XYW cell receptive field characteristics is characterized by comprising the following steps:
A. acquiring a gray image and a parallax image of an image to be detected;
B. carrying out rapid bilateral filtering on the gray level image of the image to be detected to obtain a filtered image;
C. processing the filtered image by X, Y and W channels respectively to obtain X channel contour response, Y channel contour response and W channel contour response;
D. respectively suppressing the X channel contour response, the Y channel contour response and the W channel contour response through sparse coding, removing redundant textures, and obtaining suppressed X channel contour response, Y channel contour response and W channel contour response;
E. respectively summing the suppressed X-channel contour response, Y-channel contour response and W-channel contour response with the parallax contour response of the corresponding channel of the parallax image to obtain the X-channel contour response, Y-channel contour response and W-channel contour response with strengthened weak contours;
F. fusing the X channel contour response, the Y channel contour response and the W channel contour response of the strengthened weak contour by using different watching planes to obtain a final contour and an optimal direction;
G. and carrying out non-maximum suppression on the final contour response by using the optimal direction to obtain a suppressed final contour.
2. The method of claim 1 for contour detection that mimics the fusion of receptor field properties of XYW cells, comprising:
in step B, the formula of the fast bilateral filtering is:
f g (x,y)=exp(μ· - f b (x,y)+f d (x,y) (1)
Figure QLYQS_1
f d (x,y)=log(I(x,y)-f b (x,y)) (3)
where I (x, y) represents a gray image, σ represents a standard deviation of a Gaussian function, and f b (x, y) represents an image of the gray image after the Gaussian function smoothing, f d (x, y) is represented by a gray image and f b (x, y) is a difference and logarithmic, and μ represents f b (x, y) and f d The (x, y) link strength, sign · represents multiplication, and one represents convolution.
3. The method of claim 1 for contour detection that mimics the fusion of receptor field properties of a XYW cell, comprising:
in the step C, the formula for obtaining the X channel contour response is:
Figure QLYQS_2
wherein the content of the first and second substances,
Figure QLYQS_3
is a suppressive response of the X-channel classical receptive field, is->
Figure QLYQS_4
Is a suppressive response of the X-channel non-classical receptive field, is based on the presence of a receptor in the receptor>
Figure QLYQS_5
The inhibition intensity is shown.
4. The method of claim 3 for contour detection that mimics the fusion of receptor field properties of XYW cells, comprising:
obtaining the inhibition response of the classical receptive field of the X channel
Figure QLYQS_6
The calculation process comprises the following steps:
a. the classical receptive field of X cells was simulated using the first derivative of gaussian with directional selectivity:
Figure QLYQS_7
Figure QLYQS_8
/>
Figure QLYQS_9
wherein, the rotation angle theta represents the direction parameter theta epsilon [0, pi ] of the receptive field, gamma represents the constant of the ratio of the long axis and the short axis of the receptive field, and gamma =0.5;
b. let σ = σ XC Obtaining the energy response of the classical receptive field of the X cells to the external stimulus
Figure QLYQS_10
Figure QLYQS_11
C belongs to { g, d }, g represents a gray level image, and d represents a parallax image;
theta as above i Each pixel point is represented as N θ The direction angle in each direction is calculated as follows:
Figure QLYQS_12
c. comparing the energy information of each pixel point in different direction angles
Figure QLYQS_13
Obtaining an optimal response to the current location->
Figure QLYQS_14
Obtaining the inhibition response of the classical receptive field; simultaneously acquiring corresponding optimal direction angles
Figure QLYQS_15
Figure QLYQS_16
Figure QLYQS_17
Obtaining inhibition response of X-channel non-classical receptive field
Figure QLYQS_18
The calculation process comprises the following steps:
a. optimal energy response and suppression weight omega of each pixel point in classical receptive field u (x, y) convolution to obtain the inhibitory response of the non-classical receptive field
Figure QLYQS_19
Figure QLYQS_20
Figure QLYQS_21
Wherein | | xi | purple 1 Represents L 1 Norm, H (Z) = max (0, Z);
DOG is a two-dimensional gaussian difference function, and its formula is:
Figure QLYQS_22
wherein σ u Standard deviation of the non-classical receptive field is indicated and k represents the scale ratio of the peripheral inhibitory region to the central excitatory region.
5. The method of claim 4 for contour detection that mimics the fusion of receptor field properties of XYW cells, comprising:
in the step C, the formula for obtaining the Y channel contour response is:
Figure QLYQS_23
wherein the content of the first and second substances,
Figure QLYQS_24
for a Y-channel non-linear suppression response, a->
Figure QLYQS_25
For a Y-channel linear suppression response, a->
Figure QLYQS_26
The inhibition intensity is indicated.
6. The method of claim 5 for contour detection that mimics the fusion of receptor field properties of XYW cells, comprising:
obtaining Y-channel nonlinear rejection response
Figure QLYQS_27
The calculation process comprises the following steps:
a. by sigma Y Substitution of σ in equations 1-3, 5, 6 X Obtaining the classical receptor field response of Y cells
Figure QLYQS_28
Then will
Figure QLYQS_29
Substitution into equations 9-11 results in an optimal response for Y cells>
Figure QLYQS_30
And corresponding optimal direction angle
Figure QLYQS_31
b. Setting a region S of finite size xy Each pixel point in the region corresponds to the central position of one subunit, the rectification property of the subunit is simulated by a Gaussian function, and then the nonlinear subunit response is obtained by summing the contribution of each subunit
Figure QLYQS_32
The formula is as follows:
Figure QLYQS_33
Figure QLYQS_34
wherein the content of the first and second substances,
Figure QLYQS_35
representing a window S centered on (x, y) xy Intrinsic response E Y_C Information in (x, y), σ s =σ YC Calculating the spatial standard deviation of the size of the subunit; s xy Radius of (d) size r s =12·σ YC (ii) a N is->
Figure QLYQS_36
The number of pixels of (a); v represents the ordinate of the Y-cell subunit;
the linear suppressive response of the acquisition Y channel
Figure QLYQS_37
The calculation process comprises the following steps: will->
Figure QLYQS_38
Instead of the former
Figure QLYQS_39
And is calculated by substituting the formula 12 as an input.
7. The method of claim 4 for contour detection that mimics fusion of receptor field properties of a XYW cell, characterized in that:
in the step C, the calculation process of obtaining the W channel contour response is:
let σ n = σ n - Will σ n Substituting into formula 5 to obtain first derivative function RF (x, y; theta, sigma) of Gaussian with different scales and shapes n ) (ii) a Standard deviation sigma n The calculation formula of (a) is as follows:
Figure QLYQS_40
σ WC is the standard deviation of CRF for W cells, since W cells are similar to CRF for X cells, here σ WCXC (ii) a S represents the size of the scale change of the classical receptive field of the W cell; n denotes the number of classical receptive field scale changes, where n =25/σ WC
b. From a Gaussian first derivative function RF (x, y; θ i, σ) n ) Selecting different micro-motion areas, wherein the selection formula of the micro-motion areas is as follows:
Figure QLYQS_41
RF (x, y; theta, sigma) n ) Multiplying by equation 19 yields a gaussian first derivative function with fixation and fixation micromotion characteristics:
RM n (x,y;θ i ,σ n )=RF(x,y;θ i ,σ n )·M n (x,y) (20);
c. the response of the classical receptive field under the visual micromotion to the external stimulus is input from the characteristic diagram and RM n (x,y;θ i ,σ n ) Convolution yields an energy response
Figure QLYQS_42
Figure QLYQS_43
d. By
Figure QLYQS_44
In place of>
Figure QLYQS_45
Substituting into equation 9-11 to calculate the optimal response of W cells
Figure QLYQS_46
And the corresponding optimum direction angle->
Figure QLYQS_47
/>
By
Figure QLYQS_48
In place of>
Figure QLYQS_49
Substituting as input into equation 12 calculates the non-canonical receptive field suppression response +for W cells>
Figure QLYQS_50
e. W cell contour response
Figure QLYQS_51
The calculation formula is as follows:
Figure QLYQS_52
wherein the content of the first and second substances,
Figure QLYQS_53
for optimal response of W cells>
Figure QLYQS_54
In response to suppression of the W channel non-classical receptive field>
Figure QLYQS_55
The inhibition intensity is shown.
8. The method of claim 1 for contour detection that mimics the fusion of receptor field properties of XYW cells, comprising:
the sparse coding formula in the step D is as follows:
Figure QLYQS_56
wherein the content of the first and second substances,
Figure QLYQS_57
that is, M belongs to { X, Y, W } represents the stimulation response under X, Y and W channels based on the gray image information; />
Figure QLYQS_58
Is a histogram of local gradient amplitude values of each information channel centered at (x, y), t is
Figure QLYQS_59
Dimension of (d); i | · | live through 1 Is L 1 Norm, | · | luminance 2 Is L 2 A norm; />
Figure QLYQS_60
To represent
Figure QLYQS_61
Min represents the minimum value of the two;
applying sparse coding to inhibit, and obtaining formulas of inhibited X channel contour response, Y channel contour response and W channel contour response as follows;
Figure QLYQS_62
9. the method of claim 1 for contour detection that mimics the fusion of receptor field properties of a XYW cell, comprising:
in the step E, the suppressed X-channel contour response, Y-channel contour response and W-channel contour response are respectively compared with the parallax contour response of the corresponding channel of the parallax image
Figure QLYQS_63
The formula for the summation is:
Figure QLYQS_64
10. the method of claim 1 for contour detection that mimics the fusion of receptor field properties of XYW cells, comprising:
in the step F, the process of fusing the X channel contour response, the Y channel contour response and the W channel contour response which strengthen the weak contour by using different gazing planes is as follows:
a. the visual information of each watching plane is processed by an X channel, the visual information of a near plane is processed by a Y channel, the visual information of a far plane is processed by a W channel, and the confirmation formula of the watching planes is as follows:
Figure QLYQS_65
b. normalizing the parallax image to [0, P]Parallax range of (1), using d (x, y) represents the parallax image after normalization; setting the parallax range on the gazing plane as beta; when round (G (x, y), 1) =0.5, it represents a near plane; when round (G (x, y), 1) = -0.5 is set, a far plane is represented; setting G (x, y) to other values, indicating on the gaze plane;
c. and fusing visual information on X, Y and W channels through three gazing planes to obtain a final contour R (X, Y) and an optimal direction theta (X, Y):
Figure QLYQS_66
Figure QLYQS_67
/>
CN202110493954.0A 2021-05-07 2021-05-07 Contour detection method for simulating fusion of properties of receptor field of XYW cell Active CN113192092B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110493954.0A CN113192092B (en) 2021-05-07 2021-05-07 Contour detection method for simulating fusion of properties of receptor field of XYW cell

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110493954.0A CN113192092B (en) 2021-05-07 2021-05-07 Contour detection method for simulating fusion of properties of receptor field of XYW cell

Publications (2)

Publication Number Publication Date
CN113192092A CN113192092A (en) 2021-07-30
CN113192092B true CN113192092B (en) 2023-04-07

Family

ID=76983981

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110493954.0A Active CN113192092B (en) 2021-05-07 2021-05-07 Contour detection method for simulating fusion of properties of receptor field of XYW cell

Country Status (1)

Country Link
CN (1) CN113192092B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117079397B (en) * 2023-09-27 2024-03-26 青海民族大学 Wild human and animal safety early warning method based on video monitoring

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107067407A (en) * 2017-04-11 2017-08-18 广西科技大学 Profile testing method based on non-classical receptive field and linear non-linear modulation

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4803736A (en) * 1985-11-27 1989-02-07 The Trustees Of Boston University Neural networks for machine vision
US6941016B1 (en) * 2001-12-31 2005-09-06 Cognex Technology And Investment Method for finding contours in an image of an object
US9092692B2 (en) * 2012-09-13 2015-07-28 Los Alamos National Security, Llc Object detection approach using generative sparse, hierarchical networks with top-down and lateral connections for combining texture/color detection and shape/contour detection
CN109919945B (en) * 2019-02-01 2022-03-25 广西科技大学 Contour detection method based on non-classical receptive field non-linear two-side subunit response
CN111179294B (en) * 2019-12-30 2020-09-15 广西科技大学 Bionic type contour detection method based on X, Y parallel visual channel response
CN111968140B (en) * 2020-06-24 2023-07-14 广西科技大学 Contour detection method based on classical receptive field vision-strengthening micro-motion mechanism
CN111968141B (en) * 2020-06-30 2023-06-16 广西科技大学 Non-classical receptive field contour detection method based on multi-scale superposition

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107067407A (en) * 2017-04-11 2017-08-18 广西科技大学 Profile testing method based on non-classical receptive field and linear non-linear modulation

Also Published As

Publication number Publication date
CN113192092A (en) 2021-07-30

Similar Documents

Publication Publication Date Title
Sun et al. Low-illumination image enhancement algorithm based on improved multi-scale Retinex and ABC algorithm optimization
CN109117826B (en) Multi-feature fusion vehicle identification method
Zhou et al. Utilizing dictionary learning and machine learning for blind quality assessment of 3-D images
WO2015180527A1 (en) Image saliency detection method
CN111402285B (en) Contour detection method based on visual mechanism dark edge enhancement
CN114926459B (en) Image quality evaluation method, system and computer readable medium
CN110210493B (en) Contour detection method and system based on non-classical receptive field modulation neural network
Saif et al. Gradient based image edge detection
CN109788275A (en) Naturality, structure and binocular asymmetry are without reference stereo image quality evaluation method
CN105046202B (en) Adaptive recognition of face lighting process method
CN116385315B (en) Image enhancement method and system for simulated ablation of tumor therapeutic instrument
Dash et al. Illumination normalized based technique for retinal blood vessel segmentation
CN113192092B (en) Contour detection method for simulating fusion of properties of receptor field of XYW cell
Gu et al. Research on the improvement of image edge detection algorithm based on artificial neural network
CN110555820A (en) Image fusion method based on convolutional neural network and dynamic guide filtering
Madrid-Herrera et al. Human image complexity analysis using a fuzzy inference system
CN113129390B (en) Color blindness image re-coloring method and system based on joint significance
CN115187569A (en) Image quality evaluation model training method, device, equipment and medium
CN114066812B (en) No-reference image quality evaluation method based on spatial attention mechanism
CN109376782B (en) Support vector machine cataract classification method and device based on eye image features
CN114982846B (en) Control method and device suitable for coffee roasting and electronic equipment
CN115187519B (en) Image quality evaluation method, system and computer readable medium
CN107024480A (en) A kind of stereoscopic image acquisition device
CN111968140B (en) Contour detection method based on classical receptive field vision-strengthening micro-motion mechanism
Kumar et al. Signal and image processing in medical applications

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
EE01 Entry into force of recordation of patent licensing contract

Application publication date: 20210730

Assignee: HUALI FAMILY PRODUCTS CO.,LTD.

Assignor: GUANGXI University OF SCIENCE AND TECHNOLOGY

Contract record no.: X2023980054119

Denomination of invention: A contour detection method for simulating the fusion of receptive field characteristics of XYW cells

Granted publication date: 20230407

License type: Common License

Record date: 20231226

EE01 Entry into force of recordation of patent licensing contract