CN112613427B - Road obstacle detection method based on visual information flow partition projection coding model - Google Patents

Road obstacle detection method based on visual information flow partition projection coding model Download PDF

Info

Publication number
CN112613427B
CN112613427B CN202011578651.0A CN202011578651A CN112613427B CN 112613427 B CN112613427 B CN 112613427B CN 202011578651 A CN202011578651 A CN 202011578651A CN 112613427 B CN112613427 B CN 112613427B
Authority
CN
China
Prior art keywords
response
map
contour
formula
azimuth
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202011578651.0A
Other languages
Chinese (zh)
Other versions
CN112613427A (en
Inventor
范影乐
杨瑞
武薇
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hangzhou Dianzi University
Original Assignee
Hangzhou Dianzi University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hangzhou Dianzi University filed Critical Hangzhou Dianzi University
Priority to CN202011578651.0A priority Critical patent/CN112613427B/en
Publication of CN112613427A publication Critical patent/CN112613427A/en
Application granted granted Critical
Publication of CN112613427B publication Critical patent/CN112613427B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/56Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
    • G06V20/58Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/44Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/46Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/56Extraction of image or video features relating to colour
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/60Extraction of image or video features relating to illumination properties, e.g. using a reflectance or lighting model
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/52Surveillance or monitoring of activities, e.g. for recognising suspicious objects
    • G06V20/54Surveillance or monitoring of activities, e.g. for recognising suspicious objects of traffic, e.g. cars on the road, trains or boats
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/513Sparse representations
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Data Mining & Analysis (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Evolutionary Computation (AREA)
  • Evolutionary Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Software Systems (AREA)
  • Image Processing (AREA)
  • Image Analysis (AREA)

Abstract

The invention relates to a road obstacle detection method based on a visual information flow partition projection coding model. Firstly, constructing a parallel vision path model in a V1 topological projection profile global perception unit, and extracting the brightness edge and the color edge of a road traffic map in parallel to obtain a topological projection profile map representing the overall characteristics of an obstacle and an optimal azimuth index matrix; then, a visual information difference enhancement model is constructed in a V4 sparse coding fine feature extraction unit, and the position response difference and the azimuth response difference are utilized to carry out contrast enhancement on the topological projection profile to obtain a refined profile; thirdly, providing a self-adaptive size sparse coding model, realizing intelligent focusing on the contour features of the obstacle according to the sparseness of the thinned contour map, and obtaining a pooling map representing the local features of the obstacle; and finally, simulating feedback regulation and control mechanisms of different brain regions of the visual cortex, correcting the topological projection contour map of the cross-visual area by using the pooling map, and fusing to obtain a final obstacle contour appearance result.

Description

Road obstacle detection method based on visual information flow partition projection coding model
Technical Field
The invention belongs to the field of machine vision, and particularly relates to a road obstacle detection method based on a visual information flow partition projection coding model.
Background
Road obstacle detection tasks are one of the subtasks of intelligent traffic systems, and for vehicles traveling at high speeds, obstacles can be classified into various forms such as pedestrians, road traffic signs, mountain falling stones, vehicle spills, and the like. Road obstacle detection is of great importance for safe driving of vehicles and efficient management of traffic. Contours are an efficient representation of the target subject as low-dimensional spatial features of the image. The obstacle outline acquisition in the road traffic map is beneficial to improving the rapidity and accuracy of tasks such as subsequent obstacle classification, recognition and the like.
The contours detected by the traditional image processing technology adopting the mathematical differentiation operator often contain a large amount of texture information, and road traffic target barriers cannot be well distinguished from complex backgrounds. Along with the great emergence of research results in the field of biological vision and the rapid development of nerve computation, a target perception method based on a bionic mechanism is widely focused. The sensitivity characteristics of the classical receptive field and the edge line segments of the primary visual cortex are simulated by a two-dimensional Gabor energy model, the physiological structure of the non-classical receptive field is simulated by a Gaussian differential model, an isotropy and anisotropy inhibition method is provided, and texture information is effectively inhibited; in addition, the modeling of the color antagonism mechanism in the video path is studied, so that the effective extraction of the contour information of the color image is realized. And then further suppressing texture information by using an improved sparsity measurement method on the basis. The contour detection method models the processing process of the information flow in the visual path by utilizing different biological vision mechanisms, but simplifies the information flow projection process in the visual path, simplifies the processing effect of the advanced visual cortex on the visual information, performs black box simulation on the physiological characteristics of neuron information coding or feedback regulation and the like only on the basis of single visual path modeling, and is unfavorable for effectively extracting the obstacle target because the target usually does not have a structured contour shape for the road obstacle target under a complex background.
Disclosure of Invention
The invention provides a road obstacle detection method based on a visual information flow partition projection coding model.
The model provided by the invention is composed of a V1 topological projection contour global perception unit and a V4 sparse coding fine feature extraction unit, and respectively simulates the front-stage characteristic of a V1 layer and the coding characteristic of the V4 layer. Firstly, constructing a parallel view path model in a V1 topological projection profile global perception unit, extracting the brightness edge and the color edge of a road traffic map in parallel, and acquiring a topological projection profile E (x, y) representing the overall characteristics of an obstacle and an optimal azimuth index matrix Θ (x, y) in a V1 layer; then constructing a visual information difference enhancement model in a V4 sparse coding fine feature extraction unit, and respectively performing texture suppression and contour enhancement on the topological projection contour map E (x, y) by using the position response difference and the azimuth response difference to obtain a refined contour map E t (x, y); then put forward aSeed adaptive size sparse coding model according to refined contour map E t The sparseness degree of (x, y) realizes intelligent focusing on the outline features of the obstacle, and a pooling graph E for representing the local features of the obstacle is obtained at a V4 layer s (x, y); finally, simulating feedback regulation and control mechanisms of different brain regions of visual cortex, and utilizing pooling graph E s And (x, y) correcting the topological projection profile map E (x, y) of the cross-view area, and fusing to obtain a final barrier profile result.
Compared with the prior art, the invention has the following effects:
the invention constructs a parallel vision path model, and simulates the information flow partition projection characteristic to primarily sense the outline of the road obstacle. In consideration of the detail sensitivity difference of M-type ganglion cells and P-type ganglion cells to brightness information flow and color information flow and the partition projection characteristic of two types of visual information flow in a primary visual path, a parallel visual path model is constructed at the front stage of a V1 layer, the brightness edge and the color edge of a road traffic map are respectively extracted, primary contour response is obtained by fusion, and the model ensures the integrity of obstacle contour information perception in the road traffic map.
The invention provides a visual information difference enhancement model, which utilizes the physiological characteristics of a visual receptive field to carry out contrast enhancement on primary contour response. Considering the azimuth sensitivity characteristic of a classical receptive field, calculating azimuth response difference by utilizing the optimal azimuth information and orthogonal azimuth information corresponding to the primary profile response to enhance the profile pixel point; and calculating the position response difference to restrain the texture pixel point by using the response difference between the classical receptive field and the non-classical receptive field. The model refines the primary profile response by enhancing contrast.
The invention provides a novel sparse coding method for adaptively selecting sparse kernel sizes. In consideration of rapidity and accuracy of the visual nerve center in sensing the spatial position of the target main body, a sparse coding model with adaptively selected sparse kernel size is constructed, the adaptively selected sparse kernel size is carried out according to the sparse degree of pixel distribution in the refined contour map, and information redundancy is removed while intelligent focusing is completed on the contour features of the obstacle. Compared with the traditional method for sparse coding by adopting a fixed-size sparse kernel, the self-adaptive size sparse coding method provided by the invention is more in line with the dynamic focusing characteristic of a biological vision system on the target contour characteristic.
A road obstacle detection method based on an information flow partition projection coding model is provided. Firstly, simulating partition projection characteristics of different information flows, constructing a parallel vision path model, respectively extracting brightness edges and color edges of a road traffic map, and fusing in a V1 layer to obtain a primary contour response; then, constructing a visual information difference enhancement model by utilizing the physiological characteristics of the visual receptive field, and refining the primary contour response by using the position response difference and the azimuth response difference; then simulating dynamic focusing characteristics of the visual nerve center on the target contour features, constructing a self-adaptive size sparse coding model, and obtaining a pooling graph; and finally, simulating a feedback regulation mechanism between visual layers, correcting the primary contour response of the cross-visual area by using the pooling graph, and fusing to obtain a final obstacle contour result.
Drawings
Fig. 1 is a flowchart of a road obstacle detection method according to the present invention.
FIG. 2 is a schematic diagram of the optimal orientation and orthogonal orientation of the receptive field.
The specific embodiment is as follows:
note that: taking E (x, y) as an example, where (x, y) represents the two-dimensional coordinate position of the pixel in the image E, E (x, y) represents the pixel value at the coordinate position (x, y) in the image E, and the steps will not be described.
A specific embodiment of the present invention will be described with reference to fig. 1 and 2.
And (1) constructing a parallel vision path model, and extracting a topological projection profile E (x, y) and an optimal azimuth index matrix theta (x, y). For the road traffic diagram to be detected, the brightness component I (x, y) and the red, green and blue color components R (x, y), G (x, y) and B (x, y) are decomposed, and the number of rows and columns of each component are respectively m and n. The parallel view path model is composed of a brightness path and a color path, and extracts the brightness edge and the color edge of the road traffic map respectively.
Step 1.1 the mathematical model of the luminance path is shown in formula (1):
wherein sigma and gamma represent the size and ellipticity of classical receptive fields, and the default values are set to 2 and 0.5, respectively.
Representing selective orientations, defaulting to 8 orientations equally spaced, i.e. θ i = {0 °,45 °,90 °,135 °,180 °,225 °,270 °,315 ° }, which represents convolution operation, e (x, y; θ i ) Representative pixel coordinates are (x, y), selective orientation is θ i The corresponding luminance component edge response.
For each pixel, selecting the maximum value of the brightness component edge response corresponding to all the selective orientations, linearly normalizing the maximum value to be used as output, and simultaneously recording the corresponding optimal orientation index to obtain the contour response E of the brightness path L (x, y) and an optimal azimuth index matrix Θ L (x, y) as shown in formula (2):
where N (-) represents the linear normalization operation.
Step 1.2 the color channels are modeled using a color antagonistic mechanism, and are divided into four types of antagonistic channels, R-on/G-off, G-on/R-off, B-on/Y-off, Y-on/B-off, wherein the yellow component Y (x, Y) = (R (x, Y) +G (x, Y))/2. The R-on/G-off type antagonistic channel is exemplified. Firstly, simulating the action of cone cells, processing R (x, y) component and G (x, y) component by a Gaussian filter (default variance is set to be 1), and marking the result asThen calculating the edge response of the single antagonistic receptive field, and marking the result as S RG (x, y) as shown in formula (3):
then simulating the action of the double antagonistic receptive fields, calculating the position of pixel coordinates (x, y) and the selective azimuth is theta i Corresponding color component edge response d RG (x,y;θ i ) As shown in formula (4):
for each pixel, selecting the maximum value of the edge response of the color component corresponding to all the selective orientations, linearly normalizing the maximum value to be used as output, and simultaneously recording the corresponding optimal orientation index to obtain the edge response D of the R-on/G-off type antagonism channel RG (x, y) and an optimal azimuth index matrix Θ RG (x, y) as shown in formula (5):
similar to the R-on/G-off type antagonistic channels, the edge response D of the other three types of antagonistic channels is calculated GR 、D BY 、D YB And an optimal azimuth index matrix Θ GR 、Θ BY 、Θ YB Or by calculation. For each pixel coordinate position, taking the maximum value of edge responses in four antagonistic channels as output, and simultaneously recording the corresponding optimal azimuth index to obtain the contour response E of the color channel C (x, y) and an optimal azimuth index matrix Θ C (x, y) as shown in formula (6):
and 1.3, respectively fusing the contour response of the brightness path and the color path and the optimal azimuth index matrix, and simulating the front-stage characteristic of the visual cortex V1 functional layer to obtain a topological projection contour map E (x, y) and the optimal azimuth index matrix theta (x, y). As shown in formula (7):
step (2) constructing a visual information difference enhancement model, and carrying out contrast enhancement on the topological projection profile E (x, y) by utilizing the position response difference and the azimuth response difference to obtain a refined profile E t (x, y). First, for the topology projection profile E (x, y) obtained in step (1), a Gaussian function G (x, y; σ) and a Gaussian difference function DoG are used, respectively + (x, y; sigma) convolving with E (x, y) to obtain classical receptive field visual input L C (x, y; sigma) and non-classical receptive field visual input L N (x, y; sigma) as shown in formula (8):
wherein,
the default value of sigma in the formula (8) is the same as that in the formula (1), and L in the formula C (x, y; sigma) and L N (x, y; sigma) performing a difference operation to obtain a position response difference DeltaL (x, y; sigma) as shown in formula (10):
ΔL(x,y;σ)=max{L C (x,y;σ)-L N (x,y;σ),0} (10)
then, an orthogonal azimuth index matrix Θ of the topology projection profile E (x, y) is calculated + (x, y) as shown in formula (11):
Θ + (x,y)=(Θ(x,y)+2)mod 8 (11)
where mod represents the remainder taking operation; the theta (x, y) obtained in the step (1) stores the optimal azimuth information at the pixel coordinates (x, y) in the form of indexes, wherein the value range is {0,1,2,3,4,5,6,7}, {0,4} represents the horizontal optimal azimuth, {2,6 represents a vertical optimum orientation, {1,5} represents a forward optimum orientation, {3,7} represents a reverse optimum orientation; theta (theta) + (x, y) represents an orthogonal orientation complementary to the optimal orientation Θ (x, y), wherein the horizontal orientation is complementary to the vertical orientation and the left diagonal orientation is complementary to the right diagonal orientation. A schematic relationship of the optimal orientation to the orthogonal orientation is shown in fig. 2.
For each pixel coordinate (x, y) in the topology projection profile E, the sum of the two neighborhood pixel values is used as response information. Calculating optimal azimuth response O respectively Θ (x, y) and orthogonal azimuth response O Θ+ (x, y), and performing a difference operation on the two to obtain an azimuth response difference Δo (x, y), as shown in formula (12):
finally, contrast enhancement is carried out on E (x, y) to obtain a refined contour map E t (x, y) as shown in formula (13):
E t (x,y)=E(x,y)+E enha (x,y)-E inhi (x,y) (13)
wherein E is enha (x, y) represents the result of linear normalization of the azimuth response difference Δo (x, y) for enhancing the contour pixel point. E (E) inhi (x, y) represents the result of exponential normalization of the position response difference DeltaL (x, y; sigma) for suppressing background pixels.
Step (3) constructing an adaptive size sparse coding model, and refining a contour map E t (x, y) performing sparse coding to obtain a pooling graph E s (x, y). For the refined profile E obtained in step (2) t And (x, y), calculating an adaptive threshold value thresh according to a maximum inter-class variance method, binarizing and counting the proportion of the contour pixels, wherein the proportion is represented by a formula (14):
wherein,representing the binarized result of the refined contour map, the count (& gt) operation is used for counting binary imagesThe number of pixel values 1.
Then, calculating the size of a sparse kernel of the sparse coding model, selecting the shape of a sparse kernel window as a square window for simplifying calculation, and adaptively selecting the window size according to the outline pixel proportion description size, wherein the formula (15) is as follows:
wherein w is 1 、w 2 、w 3 Representing sparse kernel window sizes of different sizes. The invention is provided with w 1 =3、w 2 =5、w 3 =7. thresh1 and thresh2 are used to measure the refined profile E t Threshold parameters of the sparseness degree default to the value thresh1=0.2, thresh2=0.1. According to E t The relative sizes of the middle outline pixel proportion project and the threshold parameter are equal to the sparse kernel window size w s And performing self-adaptive selection. Adopting mirror symmetry method to refine contour map E before sparse coding t And filling boundary pixels. For E t Calculating an average value of all pixels in a window taking the average value as the center as sparse coding output of the pixel point, taking the window size as window moving step length, and realizing sparse expression of visual information in space, wherein the sparse expression is shown in a formula (16):
wherein w and h are respectively the lateral offset and the longitudinal offset, and w s The size of the sparse nuclear window is also the window moving step length; floor (·) represents a rounding down function, E s (x, y) is a pooled graph after the thinning process.
Step (4)And calculating an obstacle outline result T (x, y) after feedback correction. Pooling the FIG. E obtained in step (3) s And (x, y) feeding back to the V1 visual cortex in a cross-visual area mode, correcting the topological projection profile map E (x, y) obtained in the step (1) in a form of an adjustment coefficient, and obtaining a final barrier profile result T (x, y) through pixel-by-pixel multiplication and fusion, wherein the final barrier profile result T (x, y) is shown as a formula (17):
wherein, the size (·) represents a bilinear interpolation operation,the size of the sparse precise sketch enlarged by bilinear interpolation operation is the same as that of the topological projection outline.

Claims (8)

1. The road obstacle detection method based on the visual information flow partition projection coding model is characterized by comprising the following steps of:
step one: at V 1 Constructing a parallel vision path model in a topology projection profile global perception unit, and extracting the brightness edge and the color edge of a road traffic map in parallel so as to obtain a topology projection profile E (x, y) representing the overall characteristics of the obstacle and an optimal azimuth index matrix theta (x, y);
step two: at V 4 Constructing a visual information difference enhancement model in the sparse coding fine feature extraction unit, and respectively performing texture suppression and contour enhancement on the topological projection contour map E (x, y) by using the position response difference and the azimuth response difference to obtain a refined contour map E t (x,y);
Step three: an adaptive size sparse coding model is provided for refining the contour map E t (x, y) carrying out pooling operation, realizing intelligent focusing on the outline features of the obstacle according to the image characteristics, and obtaining a pooling graph E representing the local features of the obstacle s (x,y);
Step four: feedback regulation mechanism for simulating different brain regions of visual cortexUsing pooled graphs E s (x, y) correcting the topological projection profile map E (x, y) of the cross-view area, and fusing to obtain a final barrier profile result T (x, y);
at V 1 Constructing a parallel vision path model in a topology projection profile global perception unit, and extracting the brightness edge and the color edge of a road traffic map in parallel so as to obtain a topology projection profile E (x, y) representing the overall characteristics of the obstacle and an optimal azimuth index matrix theta (x, y); the method comprises the following steps:
for a road traffic diagram to be detected, decomposing a brightness component I (x, y) and red, green and blue color components R (x, y), G (x, y) and B (x, y); the number of rows and columns of each component are m and n respectively; the parallel vision path model consists of a brightness path and a color path, and extracts the brightness edge and the color edge of the road traffic map respectively;
step 1.1 the mathematical model of the luminance path is shown in formula (1):
wherein sigma and gamma represent the size and ellipticity of classical receptive field,representing selective orientations, taking equally spaced 8 orientations, i.e. θ i = {0 °,45 °,90 °,135 °,180 °,225 °,270 °,315 ° }, which represents convolution operation, e (x, y; θ i ) Representative pixel coordinates are (x, y), selective orientation is θ i The corresponding luminance component edge response;
for each pixel, selecting the maximum value of the brightness component edge response corresponding to all the selective orientations, linearly normalizing the maximum value to be used as output, and simultaneously recording the corresponding optimal orientation index to obtain the contour response E of the brightness path L (x, y) and an optimal azimuth index matrix Θ L (x, y) as shown in formula (2):
wherein, N (·) represents a linear normalization operation;
step 1.2, modeling a color channel by using a color antagonism mechanism, wherein the color channel is divided into four types of antagonism channels of R-on/G-off, G-on/R-off, B-on/Y-off and Y-on/B-off, and yellow component Y (x, Y) = (R (x, Y) +G (x, Y))/2; take the R-on/G-off type antagonistic channel as an example; firstly, simulating the action of cone cells, processing R (x, y) component and G (x, y) component by using Gaussian filter respectively, and recording the result as Then calculating the edge response of the single antagonistic receptive field, and marking the result as S RG (x, y) as shown in formula (3):
then simulating the action of the double antagonistic receptive fields, calculating the position of pixel coordinates (x, y) and the selective azimuth is theta i Corresponding color component edge response d RG (x,y;θ i ) As shown in formula (4):
for each pixel, selecting the maximum value of the edge response of the color component corresponding to all the selective orientations, linearly normalizing the maximum value to be used as output, and simultaneously recording the corresponding optimal orientation index to obtain the edge response D of the R-on/G-off type antagonism channel RG (x, y) and an optimal azimuth index matrix Θ RG (x, y) as shown in formula (5):
the same calculation mode as that of R-on/G-off type antagonistic channels is adopted to obtain the edge response D of other three antagonistic channels GR 、D BY 、D YB And an optimal azimuth index matrix Θ GR 、Θ BY 、Θ YB The method comprises the steps of carrying out a first treatment on the surface of the For each pixel coordinate position, taking the maximum value of edge responses in four antagonistic channels as output, and simultaneously recording the corresponding optimal azimuth index to obtain the contour response E of the color channel C (x, y) and an optimal azimuth index matrix Θ C (x, y) as shown in formula (6):
step 1.3, respectively fusing the contour response of the brightness path and the color path and an optimal azimuth index matrix, and simulating the front-stage characteristic of the visual cortex V1 functional layer to obtain a topological projection contour map E (x, y) and an optimal azimuth index matrix theta (x, y); as shown in formula (7):
2. the method for detecting the road obstacle based on the visual information flow partition projection coding model according to claim 1, wherein the method comprises the following steps: sigma, gamma are set to 2 and 0.5 in order.
3. The method for detecting the road obstacle based on the visual information flow partition projection coding model according to claim 1, wherein the method comprises the following steps: the second step is as follows: at V 4 Constructing a visual information difference enhancement model in the sparse coding fine feature extraction unit, and respectively performing texture suppression and contour enhancement on the topological projection contour map E (x, y) by using the position response difference and the azimuth response difference to obtain a refined contour map E t (x, y); the method comprises the following steps:
first, for the topology projection profile E (x, y) obtained in the first step, a Gaussian function G (x, y; sigma) and a Gaussian difference function DoG are used, respectively + (xY; sigma) is convolved with E (x, y) to obtain visual input L of classical receptive field C (x, y; sigma) and non-classical receptive field visual input L N (x, y; sigma) as shown in formula (8):
wherein,
in the pair (8), L C (x, y; sigma) and L N (x, y; sigma) performing a difference operation to obtain a position response difference DeltaL (x, y; sigma) as shown in formula (10):
△L(x,y;σ)=max{L C (x,y;σ)-L N (x,y;σ),0} (10)
then, an orthogonal azimuth index matrix Θ of the topology projection profile E (x, y) is calculated + (x, y) as shown in formula (11):
Θ + (x,y)=(Θ(x,y)+2)mod 8 (11)
where mod represents the remainder taking operation; the theta (x, y) obtained in the step (1) stores the optimal azimuth information at the pixel coordinates (x, y) in an index form, wherein the value range is {0,1,2,3,4,5,6,7}, {0,4} represents the horizontal optimal azimuth, {2,6} represents the vertical optimal azimuth, {1,5} represents the forward-inclined optimal azimuth, {3,7} represents the reverse-inclined optimal azimuth; theta (theta) + (x, y) represents an orthogonal orientation complementary to the optimal orientation Θ (x, y), wherein the horizontal orientation is complementary to the vertical orientation, and the left diagonal orientation is complementary to the right diagonal orientation;
for each pixel coordinate (x, y) in the topological projection profile diagram E, taking the sum of two neighborhood pixel values as response information; calculating optimal azimuth response O respectively Θ (x, y) and orthogonal azimuth responseAnd performing difference operation on the two to obtain azimuth response difference delta O (x, y), as shown in formula (12):
finally, contrast enhancement is carried out on E (x, y) to obtain a refined contour map E t (x, y) as shown in formula (13):
E t (x,y)=E(x,y)+E enha (x,y)-E inhi (x,y) (13)
wherein E is enha (x, y) represents the result of linear normalization of the azimuth response difference delta O (x, y) and is used for enhancing the contour pixel points; e (E) inhi (x, y) represents the result of exponential normalization of the position response difference DeltaL (x, y; sigma) for suppressing background pixels.
4. The method for detecting road obstacle based on visual information flow partition projection coding model according to claim 3, wherein: sigma is set to 2.
5. The method for detecting road obstacle based on visual information flow partition projection coding model according to claim 3, wherein: and step three: an adaptive size sparse coding model is provided for refining the contour map E t (x, y) carrying out pooling operation, realizing intelligent focusing on the outline features of the obstacle according to the image characteristics, and obtaining a pooling graph E representing the local features of the obstacle s (x, y); the method comprises the following steps:
for the refined profile E obtained in the step two t And (x, y), calculating an adaptive threshold value thresh according to a maximum inter-class variance method, binarizing and counting the proportion of the contour pixels, wherein the proportion is represented by a formula (14):
wherein,binarization of representative refined profileThe count (·) operation is used to count binary imagesThe number of middle pixel values 1;
then, calculating the size of a sparse kernel of the sparse coding model, selecting the shape of a sparse kernel window as a square window for simplifying calculation, and adaptively selecting the window size according to the outline pixel proportion description size, wherein the formula (15) is as follows:
wherein w is 1 、w 2 、w 3 Representing sparse kernel window sizes of different sizes; thresh1 and thresh2 are used to measure the refined profile E t Threshold parameters of sparseness according to E t The relative sizes of the middle outline pixel proportion project and the threshold parameter are equal to the sparse kernel window size w s Performing self-adaptive selection; adopting mirror symmetry method to refine contour map E before sparse coding t Filling boundary pixels; for E t Calculating an average value of all pixels in a window taking the average value as the center as sparse coding output of the pixel point, taking the window size as window moving step length, and realizing sparse expression of visual information in space, wherein the sparse expression is shown in a formula (16):
wherein w and h are respectively the lateral offset and the longitudinal offset, and w s The size of the sparse nuclear window is also the window moving step length; floor (·) represents a rounding down function, E s (x, y) is a pooled graph after the thinning process.
6. The method for detecting road obstacle based on the visual information flow partition projection coding model according to claim 5, wherein: w (w) 1 =3,w 2 =5,w 3 =7。
7. The method for detecting road obstacle based on the visual information flow partition projection coding model according to claim 5, wherein: thresh1=0.2, thresh2=0.1.
8. The method for detecting road obstacle based on the visual information flow partition projection coding model according to claim 5, wherein: and step four: feedback regulation mechanism simulating different brain regions of visual cortex and using pooling diagram E s (x, y) correcting the topological projection profile map E (x, y) of the cross-view area, and fusing to obtain a final barrier profile result T (x, y); the method comprises the following steps:
pooling graph E obtained in step three s And (x, y) feeding back to the V1 visual cortex in a cross-visual area mode, correcting the topological projection profile map E (x, y) obtained in the step one in a form of an adjustment coefficient, and obtaining a final barrier profile shape result T (x, y) through pixel-by-pixel multiplication and fusion, wherein the final barrier profile shape result T (x, y) is shown as a formula (17):
wherein, the size (·) represents a bilinear interpolation operation,the dimension of the pooled map amplified by bilinear interpolation operation is the same as that of the topological projection profile map.
CN202011578651.0A 2020-12-28 2020-12-28 Road obstacle detection method based on visual information flow partition projection coding model Active CN112613427B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011578651.0A CN112613427B (en) 2020-12-28 2020-12-28 Road obstacle detection method based on visual information flow partition projection coding model

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011578651.0A CN112613427B (en) 2020-12-28 2020-12-28 Road obstacle detection method based on visual information flow partition projection coding model

Publications (2)

Publication Number Publication Date
CN112613427A CN112613427A (en) 2021-04-06
CN112613427B true CN112613427B (en) 2024-02-27

Family

ID=75248224

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011578651.0A Active CN112613427B (en) 2020-12-28 2020-12-28 Road obstacle detection method based on visual information flow partition projection coding model

Country Status (1)

Country Link
CN (1) CN112613427B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113435455B (en) * 2021-05-12 2024-03-22 深圳灵图创新科技有限公司 Image contour extraction method based on space-time pulse coding
CN113391368B (en) * 2021-06-30 2022-10-21 山东国瑞新能源有限公司 Road exploration method and equipment based on virtual imaging technology

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104484667A (en) * 2014-12-30 2015-04-01 华中科技大学 Contour extraction method based on brightness characteristic and contour integrity
CN109489576A (en) * 2018-10-19 2019-03-19 杭州电子科技大学 A kind of profile testing method based on primary vision access computation model
CN111222518A (en) * 2020-01-16 2020-06-02 杭州电子科技大学 Contour feature extraction method based on frequency division visual mechanism
CN111402285A (en) * 2020-01-16 2020-07-10 杭州电子科技大学 Contour detection method based on visual mechanism dark edge enhancement

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104484667A (en) * 2014-12-30 2015-04-01 华中科技大学 Contour extraction method based on brightness characteristic and contour integrity
CN109489576A (en) * 2018-10-19 2019-03-19 杭州电子科技大学 A kind of profile testing method based on primary vision access computation model
CN111222518A (en) * 2020-01-16 2020-06-02 杭州电子科技大学 Contour feature extraction method based on frequency division visual mechanism
CN111402285A (en) * 2020-01-16 2020-07-10 杭州电子科技大学 Contour detection method based on visual mechanism dark edge enhancement

Also Published As

Publication number Publication date
CN112613427A (en) 2021-04-06

Similar Documents

Publication Publication Date Title
Han et al. Underwater image processing and object detection based on deep CNN method
CN111079685B (en) 3D target detection method
Fritsch et al. Monocular road terrain detection by combining visual and spatial information
CN103914699A (en) Automatic lip gloss image enhancement method based on color space
CN103996198B (en) The detection method of area-of-interest under Complex Natural Environment
CN106650640A (en) Negative obstacle detection method based on local structure feature of laser radar point cloud
CN108665463A (en) A kind of cervical cell image partition method generating network based on confrontation type
CN110619638A (en) Multi-mode fusion significance detection method based on convolution block attention module
CN104408711B (en) Multi-scale region fusion-based salient region detection method
CN107633513A (en) The measure of 3D rendering quality based on deep learning
CN112613427B (en) Road obstacle detection method based on visual information flow partition projection coding model
CN103699900B (en) Building horizontal vector profile automatic batch extracting method in satellite image
CN104766096B (en) A kind of image classification method based on multiple dimensioned global characteristics and local feature
CN111563447A (en) Crowd density analysis and detection positioning method based on density map
CN110414385B (en) Lane line detection method and system based on homography transformation and characteristic window
CN112488046B (en) Lane line extraction method based on high-resolution images of unmanned aerial vehicle
CN104616308A (en) Multiscale level set image segmenting method based on kernel fuzzy clustering
CN107315998A (en) Vehicle class division method and system based on lane line
CN103996185A (en) Image segmentation method based on attention TD-BU mechanism
CN102799646B (en) A kind of semantic object segmentation method towards multi-view point video
CN103679718A (en) Fast scenario analysis method based on saliency
CN111369617A (en) 3D target detection method of monocular view based on convolutional neural network
CN103955945A (en) Self-adaption color image segmentation method based on binocular parallax and movable outline
CN106529441B (en) Depth motion figure Human bodys' response method based on smeared out boundary fragment
Khan et al. Lrdnet: lightweight lidar aided cascaded feature pools for free road space detection

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant