CN113191281A - ORB feature extraction method based on region of interest and adaptive radius - Google Patents

ORB feature extraction method based on region of interest and adaptive radius Download PDF

Info

Publication number
CN113191281A
CN113191281A CN202110495753.4A CN202110495753A CN113191281A CN 113191281 A CN113191281 A CN 113191281A CN 202110495753 A CN202110495753 A CN 202110495753A CN 113191281 A CN113191281 A CN 113191281A
Authority
CN
China
Prior art keywords
image
region
sub
interest
gray
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202110495753.4A
Other languages
Chinese (zh)
Inventor
孙超
孙佳
董璐
王远大
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nanjing Yunzhikong Industrial Technology Research Institute Co ltd
Original Assignee
Nanjing Yunzhikong Industrial Technology Research Institute Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanjing Yunzhikong Industrial Technology Research Institute Co ltd filed Critical Nanjing Yunzhikong Industrial Technology Research Institute Co ltd
Priority to CN202110495753.4A priority Critical patent/CN113191281A/en
Publication of CN113191281A publication Critical patent/CN113191281A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/10Terrestrial scenes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2415Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on parametric or probabilistic models, e.g. based on likelihood ratio or false acceptance rate versus a false rejection rate
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/047Probabilistic or stochastic networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/25Determination of region of interest [ROI] or a volume of interest [VOI]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/44Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Computing Systems (AREA)
  • Software Systems (AREA)
  • Molecular Biology (AREA)
  • Computational Linguistics (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Mathematical Physics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Multimedia (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Evolutionary Biology (AREA)
  • Probability & Statistics with Applications (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses an ORB feature extraction method based on a region of interest and a self-adaptive radius. Secondly, a Region of Interest (RoI) marking technology is designed to mark any Region where ORB features may exist based on the gray variance in each layer of the laplacian of gaussian pyramid. And finally, adjusting the radius of the template according to the illumination brightness and the image contrast to extract the features by adopting a self-adaptive ORB feature extraction method. The method can obviously improve the extraction quantity and the uniform distribution of the ORB characteristic points under the complex illumination.

Description

ORB feature extraction method based on region of interest and adaptive radius
Technical Field
The invention relates to the field of image processing, in particular to an ORB (object-oriented bounding box) feature extraction method based on a region of interest and a self-adaptive radius.
Background
An image feature extraction algorithm is an essential key step in computer vision, and since a cheap camera can provide a large amount of environmental information such as color, shape and texture, a robot expert widely applies the computer vision technology to the fields of Automatic Guided Vehicle (AGV) visual navigation, target detection and tracking and the like. Some challenging problems for vision-based AGVs running in complex work areas involve large amounts of camera data, appearance feature variations due to non-uniform and dynamic lighting, and intensive computations in the feature extraction process.
Currently, in order to improve the intelligence and adaptability of automated guided vehicles and reduce the reliance on navigation infrastructure, robotics experts must develop computer vision techniques that can efficiently and stably identify natural landmarks in dynamic lighting environments. Since natural landmarks are not a navigation infrastructure that is deliberately placed in the environment, it is inconvenient to acquire a priori knowledge of the existing distribution and characteristic appearance of natural landmarks. Under the visual SLAM framework, the quality of feature extraction seriously affects the accuracy of feature matching, which is a key step for determining the accuracy and robustness of attitude estimation. In this sense, feature detection is a difficult and important process for visual navigation with natural landmarks.
Common natural landmark feature extraction algorithms SIFT and SURF cause a large amount of calculation due to high-dimensional feature description, which limits the real-time performance of robot navigation under the condition that software and hardware resources of a vehicle-mounted controller are limited. In comparison, the ORB algorithm improves the efficiency and robustness of feature detection. However, the existing ORB-based feature detection algorithm still has some problems that 1) feature points are concentrated in a region with complex texture, and other regions have fewer feature points. 2) For many real-time applications, the processing time for feature point detection is still long. 3) The number of ORB feature points decreases dramatically when the illumination intensity or image contrast conditions deteriorate.
Disclosure of Invention
In view of the above deficiencies of the prior art, the present invention aims to provide an ORB feature extraction method based on an interested region and an adaptive radius, so as to solve the problem that the ORB feature extraction method in the prior art is difficult to balance the advantages of robustness, distribution uniformity and time overhead in a complex illumination environment.
In order to achieve the purpose, the technical scheme adopted by the invention is as follows:
an ORB feature extraction method based on an interested region and an adaptive radius comprises the following steps:
1) collecting RGB images as the bottom layer image, namely a basic octave 0, by using an on-board RGB-D camera, performing convolution on the RGB images by using a Gaussian core, establishing a Gaussian pyramid, and generating a series of different octave images which gradually decrease from bottom to top;
2) aiming at each layer of image in the Gaussian pyramid, namely each octave, performing Laplace transform on the image by utilizing different convolution cores so as to generate a group of gradually blurred images with the same size and different intervals from bottom to top in each octave, and finally establishing the Gaussian Laplace pyramid;
3) combining Laplace transform and a Gaussian pyramid to construct a Gaussian Laplace pyramid, and realizing multi-resolution expression in a multi-scale space;
4) marking any area which is possible to have ORB characteristics according to the gray variance at each interval of each octave of the Gaussian Laplacian pyramid by adopting an interested area marking technology;
5) and aiming at the region where the ORB features possibly exist, namely the region of interest, adaptively adjusting the template radius according to the illumination brightness and the image contrast of the region, and extracting the potential ORB features.
Preferably, the RGB-D camera in step 1) is mounted on a mobile robot, and performs the natural environment image I in the surrounding environment without landmarksx,yCapture, image Ix,yHas a size of B1×B2Image Ix,yThe pixel coordinates of (2) are x, y, and the total frequency multiplication path number of the Gaussian pyramid is:
l=[log2(min(B1,B2))] (1)
preferably, the laplace transform in step 2) is implemented as follows: first, a gaussian function with variable scale is defined, as follows:
Figure BDA0003054193430000021
for the image octave at the l-rd layer of the gaussian pyramid, its kernel scale σ is calculated as σ (l) ═ σ 02l/2,σ0Is an initial scale;
input image I by using a variable scale Gaussian function G (x, y, σ)x,yAnd (3) performing convolution, wherein x and y are image pixel coordinates, and obtaining a multiresolution expression of the image octaves with the same size, wherein the convolution operation is as follows:
L(x,y,σ)=G(x,y,σ)*Ix,y (3)
preferably, the laplacian of gaussian pyramid in step 3) is constructed by combining laplacian transformation with a gaussian pyramid, so as to realize multi-resolution expression in a multi-scale space, in the multi-resolution expression in the multi-scale space, a group of images which gradually become more and more blurred at different intervals form an image octave, a group of gradually reduced image octaves form a gaussian pyramid at different layers, and the image kernel scale σ at the s-th interval in the l-th octave of the laplacian of pyramid is calculated as σ(s)l=σ(l)(21/s)n-sAnd n is the total number of image intervals in the l-th layer octave.
Preferably, the method of the region of interest marking technology described in step 4) is as follows:
1) multi-resolution image partitioning: each layer of the Gaussian Laplacian pyramid is initially divided equally into 4 sub-blocks 1(k),2(k),3(k),4(k)Searching a region of interest possibly containing ORB features, wherein k represents the depth number of the region of interest to be searched;
2) calculating the average gray level: for a given center point of any one sub-block, the average gray level in the (2a +1) × (2a +1) window of its neighboring region is calculated by the following formula:
Figure BDA0003054193430000031
wherein a represents a pixel value, m is one of four subblocks per every interval image in the kth depth search step, and (i, j) is a pixel coordinate in a (2a +1) × (2a +1) window;
3) calculating the sum of gray variance: in the k-th depth search step, the sum of the gray variances of each sub-block is obtained by the following formula
Figure BDA0003054193430000032
Figure BDA0003054193430000033
Then, the maximum and minimum sum of gray-scale variances is determined in the kth depth search step:
Figure BDA0003054193430000034
4) sub-block classification: in the k-th (k ═ 0) depth search step, sub-block classification is performed for each interval image, and the sub-block classification is divided into three classes in total, as follows:
Figure BDA0003054193430000035
N(0)in the form of an original interval image,
Figure BDA0003054193430000036
is the sum of the gray variances of each sub-block in the 0 th depth search step,
Figure BDA0003054193430000037
is the classification result of four sub-blocks in the 0 th depth search step;
5) sub-block evaluation: can be obtained from the above formula
Figure BDA0003054193430000038
Identifying the sub-block as a region of interest (RoI) at a 0 th depth search step; sub-blocks
Figure BDA0003054193430000039
Ignoring; the rest of the sub-blocks
Figure BDA00030541934300000310
An initial image input used as a kth (k ═ 1) depth search;
repeating the process from step (1) to step (4) to perform another loop process, wherein the classification result of the k (k is 1) th depth search step is as follows:
Figure BDA00030541934300000311
in the 1 st depth search step, N(1)Is an input of an original image, and,
Figure BDA00030541934300000312
is the sum of the gray variance of each sub-block,
Figure BDA00030541934300000313
is the classification result of four sub-blocks; similarly, a sub-block is selected
Figure BDA00030541934300000314
As a region of interest, while sub-blocks are ignored in this step
Figure BDA00030541934300000315
The rest of the sub-blocks
Figure BDA00030541934300000316
Again used as the initial image input for the kth (k ═ 2) depth search;
6) termination conditions were as follows: sum of gray variance at four sub-blocks
Figure BDA0003054193430000041
After sorting, in which case the value is smaller than the gray-scale variation threshold Δ Z, the depth search process is terminated and all the region-of-interest sub-blocks obtained in the k-step depth search can be accumulated into the entire region-of-interest
Figure BDA0003054193430000042
Which is used as a search scope to extract ORB features.
Preferably, the method for extracting ORB features by adjusting the template radius in step 5) is as follows:
1) in the obtained region of interest NRoIIn the method, the gray difference is calculated by using a 4-neighbor method between different adjacent pixels, so that
Figure BDA0003054193430000043
And
Figure BDA0003054193430000044
for the gray levels of the adjacent pixels, the gray level difference is calculated as follows
Figure BDA0003054193430000045
A 4-neighbor means that there are 4 horizontal and vertical neighbors to the center point p at coordinates (x, y) given by the formula, the group of pixels being called the 4-neighbor of the center point p
(x+1,y),(x-1,y),(x,y+1),(x,y-1) (10)
2) According to the gray difference delta and the distribution probability
Figure BDA0003054193430000046
Calculating the image contrast C, assuming P hereδIn the form of a normal distribution of the signals,
Figure BDA0003054193430000047
3) the template radius is adjusted according to the gray level difference and the image contrast as follows:
Ri,j=αC+βIi,j (12)
wherein α and β are adaptive coefficients in the range of (0, 1);
4) and (3) feature detection: let p be the center point at radius Ri,jFind all surrounding points N, e.g. 16 points around a circle with radius 3; the circle with the radius of 5 or 7 is surrounded by 28 or 40 points respectively, and the gray level I of N surrounding points on the circle is representedx,yGrey scale I from the centre point ppComparing, and checking whether the difference is larger than a threshold value T, wherein the formula is as follows;
Figure BDA0003054193430000048
if the number of surrounding points relative to the center point exceeds a threshold M, the center point p is identified as an ORB feature point.
The invention has the beneficial effects that:
according to the invention, a Gaussian Laplace transform method is adopted for a complex illumination natural image to realize multi-scale spatial expression of an original image, so that visual characteristics have scale invariance; any area where ORB features possibly exist is marked in each layer of the Gaussian Laplace pyramid by adopting an area-of-interest marking technology, so that excessive intensive collection of ORB features in a small area is avoided, and the range and detection time of subsequent feature detection are reduced; and (3) adjusting the radius of the template to extract the ORB features according to the illumination brightness and the image contrast by adopting a self-adaptive ORB feature extraction method, wherein the method has stronger complex illumination adaptability and improves the extraction quantity and the uniform distribution of the ORB feature points in the complex illumination environment.
Drawings
FIG. 1 is a system flow chart of the ORB feature extraction method based on the region of interest and the adaptive radius according to the present invention;
FIG. 2 is a schematic diagram of the multiresolution expression in the multiscale space after the Gaussian Laplace transform in the present invention;
FIG. 3 is a flow chart of a region of interest marking technique according to the present invention;
FIG. 4 is a block diagram of a multi-resolution image according to the present invention;
FIG. 5a is a schematic illustration of region of interest (RoI) labeling results for a first interval image of a first octave of a natural image with uniformly illuminated regions in accordance with the present invention;
FIG. 5b is a schematic representation of the region of interest (RoI) labeling results for a first interval image of a second octave of a natural image with a uniform illumination area according to the present invention;
FIG. 5c is a schematic illustration of region of interest (RoI) labeling results for a first interval image of a third octave of a natural image with a uniform illumination area according to the present invention;
FIG. 5d is a schematic illustration of the region of interest (RoI) labeling results for the first interval image of the fourth octave of the natural image with uniform illumination area in accordance with the present invention;
FIG. 6a is a schematic diagram of region of interest (RoI) labeling results for a first interval image of a first octave of a natural image with a complex illumination area according to the present invention;
FIG. 6b is a schematic diagram of the region of interest (RoI) labeling results for the first interval image of the second octave of the natural image with complex illumination areas in accordance with the present invention;
FIG. 6c is a schematic diagram of the region of interest (RoI) labeling results for the first interval image of the third octave of the natural image with complex illumination areas in accordance with the present invention;
FIG. 6d is a schematic diagram of the region of interest (RoI) labeling result of the first interval image of the fourth octave of the natural image with the complex illumination area according to the present invention;
FIG. 7a is a schematic diagram of feature detection with a template radius of 3 according to the present invention;
FIG. 7b is a schematic view of a template of the present invention having a radius of 3;
FIG. 8a is a schematic view of a template of the present invention having a radius of 5;
FIG. 8b is a schematic view of a template of the present invention having a radius of 7;
FIG. 8c is a schematic view of a template of the present invention having a radius of 9;
FIG. 9a is a diagram illustrating the ORB feature extraction result of a first complex-illumination natural image according to the present invention;
FIG. 9b is a diagram illustrating the ORB feature extraction result of a second complex-illumination natural image according to the present invention;
Detailed Description
For the understanding of those skilled in the art, the following detailed description will be made on an ORB feature extraction method based on a region of interest and an adaptive radius according to the present invention with reference to the accompanying drawings.
Referring to fig. 1, the ORB feature extraction method mainly includes three stages, namely, establishing a laplacian of gaussian pyramid to realize image multi-scale spatial expression, region-of-interest marking and adaptive template radius extraction of ORB features.
The ORB feature extraction method based on the region of interest and the adaptive radius specifically comprises the following steps:
1) the method comprises the steps of collecting RGB images as bottom layer images, namely a fundamental octave (octave 0), through a vehicle-mounted RGB-D camera, carrying out convolution on the RGB images through a Gaussian core, establishing a Gaussian pyramid, and generating a series of different octave images gradually reduced from bottom to top.
2) Aiming at each layer of image in the Gaussian pyramid, namely each octave, different convolution cores are utilized to carry out Laplacian transformation on the image, so that a group of gradually blurred images with the same size and different intervals from bottom to top are generated in each octave, and the Gaussian Laplacian pyramid is finally established.
3) Laplacian transformation and a Gaussian pyramid are combined to construct the Gaussian Laplacian pyramid, and multi-resolution expression in a multi-scale space is achieved.
4) And marking any region where ORB characteristics possibly exist according to the gray variance at each interval of each octave of the Gaussian Laplace pyramid by adopting a region-of-interest marking technology.
5) And aiming at the region where the ORB features possibly exist, namely the region of interest, adaptively adjusting the template radius according to the illumination brightness and the image contrast of the region, and extracting the potential ORB features.
Firstly, an RGB-D camera is mounted on a mobile robot, and a natural environment image I is carried out in a surrounding landmark-free laying environmentx,yCapture, image Ix,yHas a size of B1×B2Image Ix,yThe pixel coordinates of (1) are x and y, and the total frequency multiplication path number of the Gaussian pyramid is l:
l=[log2(min(B1,B2))](1) then, a gaussian function with variable scale is defined, as follows:
Figure BDA0003054193430000061
for the image octave (octave l) at the l-rd layer of the gaussian pyramid, its kernel scale σ is calculated as σ (l) ═ σ 02l/2,σ0Is the initial dimension.
Input image I by using a variable scale Gaussian function G (x, y, σ)x,yAnd performing convolution to obtain a multiresolution expression of the image octaves with the same size, wherein the convolution operation is as follows:
L(x,y,σ)=G(x,y,σ)*Ix,y (3)
a laplacian of gaussian pyramid is constructed by combining the laplacian transform with a gaussian pyramid, as shown in fig. 2. Multi-resolution representation in a multi-scale space is achieved. In multi-resolution representation in the multi-scale space, a set of images that become progressively more blurred at different intervals form an image octave (octave), while a set of progressively smaller image octaves at different levels form a gaussian pyramid. The image kernel scale sigma of the s-th interval in the pyramid l-th layer octave (octave l) is calculated as sigma(s)l=σ(l)(21/s)n-sAnd n is the total number of image intervals in the l-th layer octave.
As shown in fig. 3, the region of interest labeling technique based on gray variance narrows down the search for ORB features to some potential regions that are most likely to contain potential ORB features in each layer/interval of the laplacian of gaussian pyramid. The method comprises the following specific steps:
1) multi-resolution image partitioning: as shown in FIG. 4, each layer of the Gaussian Laplacian pyramid is initially divided equally into 4 sub-blocks 1(k),2(k),3(k),4(k)To search for a region of interest that may contain ORB features, and k represents the depth of the region of interest to search for.
2) Calculating the average gray level: for a given center point of any one sub-block, the average gray level in the (2a +1) × (2a +1) window of its neighboring region can be calculated by the following formula:
Figure BDA0003054193430000071
where a is a pixel value whose size is defined according to its own actual situation, m is one of four sub-blocks of each interval image in the kth depth search step, and (i, j) is a pixel coordinate in a (2a +1) × (2a +1) window.
3) Calculating the sum of gray variance: in the k-th depth search step, the sum of the gray variances of each sub-block is obtained by the following formula
Figure BDA0003054193430000072
Figure BDA0003054193430000073
Then, the maximum and minimum sum of gray-scale variances can be determined in the k-th depth search step.
Figure BDA0003054193430000074
4) Sub-block classification: in the k-th (k ═ 0) depth search step, sub-block classification is performed for each interval image, and the total is divided into three classes, as shown in the following formula.
Figure BDA0003054193430000075
N(0)In the form of an original interval image,
Figure BDA0003054193430000076
is the sum of the gray variances of each sub-block in the 0 th depth search step,
Figure BDA0003054193430000077
are the classification results of the four sub-blocks in the 0 th depth search step.
5) Sub-block evaluation: can be obtained from the above formula
Figure BDA0003054193430000078
Contains rich gray-scale variation information, which is the most likely region associated with ORB features. Therefore, the sub-block is identified as a region of interest (RoI) at the 0 th depth search step. On the contrary, since the sub-blocks
Figure BDA0003054193430000079
The gray scale change of (a) is not significant and thus is ignored. The rest of the sub-blocks
Figure BDA00030541934300000710
Used as an initial image input for the kth (k ═ 1) depth search. Repeating the process from step (1) to step (4) to perform another cyclic process. The classification result of the k (k is 1) th depth search step is
Figure BDA0003054193430000081
In the 1 st depth search step, N(1)Is an input of an original image, and,
Figure BDA0003054193430000082
is the sum of the gray variance of each sub-block,
Figure BDA0003054193430000083
is the result of the classification of four sub-blocks. Similarly, a sub-block is selected
Figure BDA0003054193430000084
As a region of interest, while sub-blocks are ignored in this step
Figure BDA0003054193430000085
The rest of the sub-blocks
Figure BDA0003054193430000086
Again used as the initial image input for the kth (k-2) depth search.
6) Termination conditions were as follows: sum of gray variance at four sub-blocks
Figure BDA0003054193430000087
After the ordering, where the value is less than the gray-scale change threshold Δ Z, this means that it is almost impossible to extract the potential ORB features in the input image for the next depth search. Therefore, it is reasonable to terminate the deep search process. All region of interest sub-blocks obtained in the k-step depth search may be accumulated as the entire region of interest
Figure BDA0003054193430000088
Which is used as a search scope to extract ORB features.
FIG. 5 is a schematic illustration of region of interest (RoI) labeling results for a first spaced image of four octaves of a natural image with uniformly illuminated regions; fig. 6 is a schematic diagram of region of interest (RoI) labeling results for the first spaced image of four octaves of a natural image with complex illumination areas.
Due to irregular gray scale variation in a complex lighting environment, it is difficult to ensure the number and uniform distribution of the feature points extracted in fig. 7a with a fixed template radius of 3 in fig. 7 b. Therefore, the template radius needs to be adaptively adjusted according to the illumination brightness and the image contrast to extract the ORB features, and the method comprises the following specific steps:
1) in the obtained region of interest NRoIIn the method, the gray difference is calculated by using a 4-neighbor method between different adjacent pixels, so that
Figure BDA0003054193430000089
And
Figure BDA00030541934300000810
the gray scale of the adjacent pixel. The gray difference is calculated as follows
Figure BDA00030541934300000811
A 4-neighbor means that there are 4 horizontal and vertical neighbors to the pixel p located at the coordinates (x, y), whose coordinates are given by the following equation, this group of pixels is called the 4-neighbor of p
(x+1,y),(x-1,y),(x,y+1),(x,y-1) (10)
2) According to the gray difference delta and the distribution probability
Figure BDA00030541934300000812
The image contrast C is calculated. Let P be assumed hereδIs normally distributed.
Figure BDA00030541934300000813
3) The template radius is adjusted according to the gray level difference and the image contrast as follows:
Ri,j=αC+βIi,j (12)
wherein α and β are adaptive coefficients in the range of (0, 1);
4) and (3) feature detection: let p be the center point at radius Ri,jFind all surrounding points N on the circle. As on a circle of radius 3, there are 16 points around, as shown in fig. 7 b; a circle with radius 5 or 7, with 28 or 40 points around it, respectively, as shown in fig. 8. The gray level I of N surrounding points on the circlex,yGrey scale I from the centre point ppComparing, and checking whether the difference is larger than a threshold value T, wherein the formula is as follows;
Figure BDA0003054193430000091
if the surrounding points have sufficient gray scale differences with respect to the center point (e.g., d or c is the result of equation (13)) and their number exceeds the threshold value M, the center point p is identified as an ORB feature point. T is generally taken as the gray level Ip20% of the total, M is typically 9, and it is clear that a smaller threshold T or M will detect more ORB feature points. Meanwhile, the larger template radius is more likely to check the center point p as a feature point because more peripheral points on the circle are selected to compare the gray scale variation. If the number of ORB features decreases due to feature degradation (lack of gray scale variation) under high or low light conditions, the template radius can be adaptively adjusted to maintain the number and quality of feature points.
Fig. 9 is a schematic diagram of ORB feature extraction results of a complex illumination natural image, and it can be seen that although loss of gray scale variation tends to bring great difficulty to feature extraction, the method of the present invention can still extract sufficient and uniformly distributed features under complex illumination conditions.
The specific application of the present invention is many, and the above is the preferred embodiment of the present invention, but the scope of the present invention is not limited thereto. Any changes or substitutions which are obvious to those skilled in the art and which are not made by the inventive idea in the technical scope of the present invention are included in the protection scope of the present invention, and therefore, the protection scope of the present invention is subject to the protection scope defined by the appended claims.

Claims (6)

1. An ORB feature extraction method based on an interested region and an adaptive radius is characterized by comprising the following steps:
1) collecting RGB images as the bottom layer image, namely a basic octave 0, by using an on-board RGB-D camera, performing convolution on the RGB images by using a Gaussian core, establishing a Gaussian pyramid, and generating a series of different octave images which gradually decrease from bottom to top;
2) aiming at each layer of image in the Gaussian pyramid, namely each octave, performing Laplace transform on the image by utilizing different convolution cores so as to generate a group of gradually blurred images with the same size and different intervals from bottom to top in each octave, and finally establishing the Gaussian Laplace pyramid;
3) combining Laplace transform and a Gaussian pyramid to construct a Gaussian Laplace pyramid, and realizing multi-resolution expression in a multi-scale space;
4) marking any area which is possible to have ORB characteristics according to the gray variance at each interval of each octave of the Gaussian Laplacian pyramid by adopting an interested area marking technology;
5) and aiming at the region where the ORB features possibly exist, namely the region of interest, adaptively adjusting the template radius according to the illumination brightness and the image contrast of the region, and extracting the potential ORB features.
2. The ORB feature extraction method based on the region of interest and the adaptive radius according to claim 1, wherein the RGB-D camera in step 1) is installed on a mobile robot, and the natural environment image I is performed in a surrounding landmark-free paving environmentx,yCapture, image Ix,yHas a size of B1×B2Image Ix,yThe pixel coordinates of (2) are x, y, and the total frequency multiplication path number of the Gaussian pyramid is:
l=[loq2(min(B1,B2))] (1)
3. the method of claim 1, wherein the laplace transform in step 2) is implemented as follows: first, a gaussian function with variable scale is defined, as follows:
Figure FDA0003054193420000011
for the image octave at the l-rd layer of the gaussian pyramid, its kernel scale σ is calculated as σ (l) ═ a02l/2,σ0Is an initial scale;
input image I by using a variable scale Gaussian function G (x, y, σ)x,yAnd performing convolution to obtain a multiresolution expression of the image octaves with the same size, wherein the convolution operation is as follows:
L(x,y,σ)=G(x,y,σ)*Ix,y (3)
4. the ORB feature extraction method based on a region of interest and an adaptive radius according to claim 1, wherein the laplacian of gaussian pyramid in step 3) is constructed by combining laplacian transformation with the laplacian pyramid, so as to realize multi-resolution expression in a multi-scale space, in the multi-resolution expression in the multi-scale space, a group of images gradually becoming more and more blurred at different intervals forms an image octave, a group of gradually shrinking image octaves forms the laplacian pyramid at different layers, and the kernel scale σ of the image at the s interval in the l-th octave of the laplacian pyramid is calculated as σ(s)l=σ(l)(21/s)n-sAnd n is the total number of image intervals in the l-th layer octave.
5. The method for extracting ORB features based on region of interest and adaptive radius according to claim 1, wherein the region of interest labeling technique in step 4) is as follows:
1) multi-resolution image partitioning: each layer of the Gaussian Laplacian pyramid is initially divided equally into 4 sub-blocks 1(k),2(k),3(k),4(k)Searching a region of interest possibly containing ORB features, wherein k represents the depth number of the region of interest to be searched;
2) calculating the average gray level: for a given center point of any one sub-block, the average gray level in the (2a +1) × (2a +1) window of its neighboring region is calculated by the following formula:
Figure FDA0003054193420000021
wherein a represents a pixel value, m is one of four subblocks per every interval image in the kth depth search step, and (i, j) is a pixel coordinate in a (2a +1) × (2a +1) window;
3) calculating the sum of gray variance: in the k-th depth search step, the sum of the gray variances of each sub-block is obtained by the following formula
Figure FDA0003054193420000022
Figure FDA0003054193420000023
Then, the maximum and minimum sum of gray-scale variances is determined in the kth depth search step:
Figure FDA0003054193420000024
4) sub-block classification: in the k (k ═ O) depth search step, sub-block classification is performed for each interval image, and the sub-block classification is divided into three classes in total, as follows:
Figure FDA0003054193420000025
N(0)in the form of an original interval image,
Figure FDA0003054193420000026
is the sum of the gray variances of each sub-block in the 0 th depth search step,
Figure FDA0003054193420000027
is the classification result of four sub-blocks in the 0 th depth search step;
5) sub-block evaluation: can be obtained from the above formula
Figure FDA0003054193420000028
Identifying the sub-block as a region of interest (RoI) at a 0 th depth search step; sub-blocks
Figure FDA0003054193420000029
Ignoring; the rest of the sub-blocks
Figure FDA00030541934200000210
An initial image input used as a kth (k ═ 1) depth search;
repeating the process from step (1) to step (4) to perform another loop process, wherein the classification result of the k (k is 1) th depth search step is as follows:
Figure FDA0003054193420000031
in the 1 st depth search step, N(1)Is an input of an original image, and,
Figure FDA0003054193420000032
is the sum of the gray variance of each sub-block,
Figure FDA0003054193420000033
is the classification result of four sub-blocks; similarly, a sub-block is selected
Figure FDA0003054193420000034
As a region of interest, while sub-blocks are ignored in this step
Figure FDA0003054193420000035
The rest of the sub-blocks
Figure FDA0003054193420000036
Again used as the initial image input for the kth (k ═ 2) depth search;
6) termination conditions were as follows: sum of gray variance at four sub-blocks
Figure FDA0003054193420000037
After a non-good order, where the value is less than the gray change threshold Δ Z, the depth search process is terminated and all the region-of-interest sub-blocks obtained in the k-step depth search can be accumulated into the entire region-of-interest
Figure FDA0003054193420000038
Which is used as a search scope to extract ORB features.
6. The method as claimed in claim 1, wherein the method for extracting ORB features based on region of interest and adaptive radius is as follows:
1) in the obtained region of interest NRoIIn the method, the gray difference is calculated by using a 4-neighbor method between different adjacent pixels, so that
Figure FDA0003054193420000039
And
Figure FDA00030541934200000310
for the gray levels of the adjacent pixels, the gray level difference is calculated as follows
Figure FDA00030541934200000311
A 4-neighbor means that there are 4 horizontal and vertical neighbors to the center point p at coordinates (x, y) given by the formula, the group of pixels being called the 4-neighbor of the center point p
(x+1,y),(x-1,y),(x,y+1),(x,y-1) (10)
2) According toGray difference Δ δ and distribution probability
Figure FDA00030541934200000312
Calculating the image contrast C, assuming P hereδIn the form of a normal distribution of the signals,
Figure FDA00030541934200000313
3) the template radius is adjusted according to the gray level difference and the image contrast as follows:
Ri,j=αC+βIi,j (12)
wherein α and β are adaptive coefficients in the range of (0, 1);
4) and (3) feature detection: let p be the center point at radius Ri,jFind all surrounding points N, e.g. 16 points around a circle with radius 3; the circle with the radius of 5 or 7 is surrounded by 28 or 40 points respectively, and the gray level I of N surrounding points on the circle is representedx,yGrey scale I from the centre point ppComparing, and checking whether the difference is larger than a threshold value T, wherein the formula is as follows;
Figure FDA00030541934200000314
if the number of surrounding points relative to the center point exceeds a threshold M, the center point p is identified as an ORB feature point.
CN202110495753.4A 2021-05-07 2021-05-07 ORB feature extraction method based on region of interest and adaptive radius Pending CN113191281A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110495753.4A CN113191281A (en) 2021-05-07 2021-05-07 ORB feature extraction method based on region of interest and adaptive radius

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110495753.4A CN113191281A (en) 2021-05-07 2021-05-07 ORB feature extraction method based on region of interest and adaptive radius

Publications (1)

Publication Number Publication Date
CN113191281A true CN113191281A (en) 2021-07-30

Family

ID=76984103

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110495753.4A Pending CN113191281A (en) 2021-05-07 2021-05-07 ORB feature extraction method based on region of interest and adaptive radius

Country Status (1)

Country Link
CN (1) CN113191281A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114763699A (en) * 2022-05-23 2022-07-19 中建四局安装工程有限公司 Embedded bolt positioning method, embedded bolt auxiliary fixing device and using method thereof
CN115147596A (en) * 2022-09-06 2022-10-04 南通鼎彩新材料科技有限公司 Irradiation dose control method and system for heat-shrinkable tube production process

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114763699A (en) * 2022-05-23 2022-07-19 中建四局安装工程有限公司 Embedded bolt positioning method, embedded bolt auxiliary fixing device and using method thereof
CN114763699B (en) * 2022-05-23 2022-11-08 中建四局安装工程有限公司 Embedded bolt positioning method, embedded bolt auxiliary fixing device and using method thereof
CN115147596A (en) * 2022-09-06 2022-10-04 南通鼎彩新材料科技有限公司 Irradiation dose control method and system for heat-shrinkable tube production process

Similar Documents

Publication Publication Date Title
Xie et al. Multilevel cloud detection in remote sensing images based on deep learning
CN113781402B (en) Method and device for detecting scratch defects on chip surface and computer equipment
You et al. Classification and segmentation of rotated and scaled textured images using texture “tuned” masks
CN111640157B (en) Checkerboard corner detection method based on neural network and application thereof
CN112950508A (en) Drainage pipeline video data restoration method based on computer vision
CN109325484B (en) Flower image classification method based on background prior significance
CN107967482A (en) Icon-based programming method and device
JP2014056572A (en) Template matching with histogram of gradient orientations
CN107564009B (en) Outdoor scene multi-target segmentation method based on deep convolutional neural network
Li et al. Road lane detection with gabor filters
CN110135438B (en) Improved SURF algorithm based on gradient amplitude precomputation
CN111369605B (en) Infrared and visible light image registration method and system based on edge features
JP2008217706A (en) Labeling device, labeling method and program
Babahajiani et al. Object recognition in 3D point cloud of urban street scene
CN112364865B (en) Method for detecting small moving target in complex scene
CN113191281A (en) ORB feature extraction method based on region of interest and adaptive radius
Xiang et al. Moving object detection and shadow removing under changing illumination condition
CN113888461A (en) Method, system and equipment for detecting defects of hardware parts based on deep learning
Wang et al. Fast vanishing point detection method based on road border region estimation
Wang et al. Road detection based on illuminant invariance and quadratic estimation
KR20170066014A (en) A feature matching method which is robust to the viewpoint change
CN112101283A (en) Intelligent identification method and system for traffic signs
Sari et al. Text extraction from historical document images by the combination of several thresholding techniques
Wu et al. Adaptive ORB feature detection with a variable extraction radius in RoI for complex illumination scenes
CN116824330A (en) Small sample cross-domain target detection method based on deep learning

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination