CN116385477A - Tower image registration method based on image segmentation - Google Patents

Tower image registration method based on image segmentation Download PDF

Info

Publication number
CN116385477A
CN116385477A CN202211618506.XA CN202211618506A CN116385477A CN 116385477 A CN116385477 A CN 116385477A CN 202211618506 A CN202211618506 A CN 202211618506A CN 116385477 A CN116385477 A CN 116385477A
Authority
CN
China
Prior art keywords
image
tower
algorithm
feature
points
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202211618506.XA
Other languages
Chinese (zh)
Inventor
沈阳
马培龙
冯广辉
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Jiayuan Technology Co Ltd
Original Assignee
Jiayuan Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Jiayuan Technology Co Ltd filed Critical Jiayuan Technology Co Ltd
Priority to CN202211618506.XA priority Critical patent/CN116385477A/en
Publication of CN116385477A publication Critical patent/CN116385477A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/194Segmentation; Edge detection involving foreground-background segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/30Determination of transform parameters for the alignment of images, i.e. image registration
    • G06T7/33Determination of transform parameters for the alignment of images, i.e. image registration using feature-based methods
    • G06T7/337Determination of transform parameters for the alignment of images, i.e. image registration using feature-based methods involving reference images or patches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30181Earth observation
    • G06T2207/30188Vegetation; Agriculture

Abstract

The invention discloses a pole and tower image registration method based on image segmentation, which comprises the following steps: reconstructing a backbone feature extraction network of the U-Net by using VGG16, and integrating the backbone feature extraction network into a attention mechanism CBAM module; training an improved U-Net network by using the marked pole tower data set; inputting the tower image into a trained improved U-Net network to obtain a high-precision tower segmentation map; manufacturing a standard tower image template, wherein a clear and complete tower to-be-detected target is arranged on the template; registering the tower segmentation map and the tower template map by using an SIFT-based image registration algorithm; the inspection robot performs intelligent shooting according to the registration result in the inspection process, reconstructs a trunk feature extraction network of the U-Net by using VGG-16, and fuses a attention mechanism CBAM module; the invention provides the method for carrying out image registration by using the tower segmentation image and the tower template, thereby avoiding the interference of complex background on registration precision.

Description

Tower image registration method based on image segmentation
Technical Field
The invention belongs to the technical field of image segmentation, and particularly relates to a pole and tower image registration method based on image segmentation.
Background
With the development of image recognition technology and the improvement of automation level, the automatic inspection by using an unmanned plane or a robot has gradually replaced the traditional manual inspection. In response to the plan of the national construction smart grid 2.0, researchers have begun turning their eyes to automated inspection techniques in recent years, and the national grid has begun using unmanned aerial vehicles or robots to conduct automated inspection operations. Through long-distance line inspection, no one can collect a large amount of tower image data, and if all the image video data are returned to technicians on the ground, the technicians still need to spend a large amount of time to judge faults.
Disclosure of Invention
In order to achieve the above purpose, the inspection robot can collect the tower images and identify and screen the tower images by combining the deep learning and image processing technology, and only the identified fault images and some unrecognizable images are reserved, so that most of inspection time can be saved. The robot inspection can ensure the safety and reliability of the line inspection process after the defect identification technology is added, so that the line inspection efficiency is greatly improved, and the robot inspection system has the advantages of low cost, small influence by environmental and geographic factors and the like.
In line inspection, robot inspection images often have complicated and changeable backgrounds like towns, mountains, forests and the like, and moreover, the acquisition visual angles are changed, the illumination intensity is changed, and the size scale of the detection target is changed, so that the traditional image segmentation is difficult to deal with. The image segmentation algorithm based on the deep learning can have a good segmentation effect in the face of a complex detection scene. In the transmission line inspection scene, the pole tower region is accurately extracted from a complex background, namely the pole tower edge is required to be segmented at the pixel level, namely the pole tower edge is segmented with high precision. The image segmentation algorithm based on deep learning is commonly used with U-Net and a segmentation algorithm Mask R-CNN which is improved on the basis of fast R-CNN. By extracting the electric power towers from complex backgrounds, interference of the complex backgrounds is avoided, accuracy of tower image registration can be improved to a great extent, and intelligent shooting of the towers is achieved.
The U-Net network structure mainly comprises three parts: encoder, decoder and jump connection structure. There are two more points to be noted in the U-Net network: the input and output sizes are inconsistent: U-Net is a symmetrical network, the left part is downsampled for feature extraction, the right part is upsampled to restore the condensed features to an image. When the U-Net input is a 572 x 572 image, the final output image size is 388 x 388, and the U-Net output is smaller than the input. Because the U-Net uses valid convolution strategy in the whole course, i.e. no padding is added, each convolution can reduce the feature map; overlay-tile strategy: when the input picture is too large, the original picture cannot be input into the network when the image is segmented, so that the original picture needs to be cut, and the large picture is changed into a Zhang Dexiao picture. In order to make the division of the picture splicing part more accurate, an overlay-tile strategy, that is, overlapping clipping, is adopted. From input to output, the U-Net network needs to provide more characteristic information by an overlap part, so that the influence of dividing a large graph into small graphs is minimized. Also, because of the overlay-tile strategy, the input picture size is 572 x 572.
The technical scheme of the invention is as follows: a tower image registration method based on image segmentation, the method comprising: reconstructing a backbone feature extraction network of the U-Net by using VGG16, and integrating the backbone feature extraction network into a attention mechanism CBAM module; training an improved U-Net network by using the marked pole tower data set; inputting the tower image into a trained improved U-Net network to obtain a high-precision tower segmentation map; marking all the targets to be detected on the towers with specific positions through a standard visual angle tower image with clear targets, which is acquired in advance, so that a standard tower image template is manufactured, and the clear and complete targets to be detected on the template are arranged; registering the tower segmentation map and the tower template map by using an SIFT-based image registration algorithm; and the inspection robot performs intelligent shooting according to the registration result in the inspection process.
Wherein CBAM (Convolution Block Attention Module) is a lightweight convolution attention module that incorporates both channel and spatial attention modules. The CBAM module may be used in a feed forward neural network, given an intermediate feature map, the CBAM module would infer the attention map sequentially along two independent dimensions (channel and space) and then multiply the attention map with the input feature map for adaptive feature optimization. Because CBAM is a lightweight generic module, the overhead of this module can be ignored and seamlessly integrated into any CNN architecture, and end-to-end training can be done with the underlying CNN.
As an improvement of the invention, the method comprises performing convolution of specific channels several times after inputting the image to obtain a preliminary effective feature layer, and performing maximum pooling to obtain a feature layer.
As an improvement of the invention, the method comprises SIFT feature detection, firstly, performing different scale changes on an original image and performing gaussian blur on the different scale image by using different gaussian kernel functions, wherein the formula is as follows:
L(x,y,σ)=G(x,y,σ)*I(x,y)
in the above formula, I (x, y) is an input image, x is the width of the image, y is the height of the image, σ is a scale space blur coefficient, L (x, y, σ) is a convolution operation, and L (x, y, σ) is an image subjected to a gaussian kernel G (x, y, σ) blur process. The specific formula of the Gaussian kernel function is as follows
Figure BDA0004001104060000031
After the Gaussian pyramid is obtained, adjacent images obtained in the same layer in the pyramid are subtracted, so that a difference image is obtained, and the following formula is calculated specifically:
D(x,y,σ)=(G(x,y,kσ)-G(x,y,σ))*I(x,y)
=L(x,y,kσ)-L(x,y,σ)
in the formula, k is the amplification factor of the space fuzzy coefficient, the operation is repeated in the Gaussian pyramid, the differential Gaussian image finally obtained by all layers forms the Gaussian differential pyramid, and the construction of the scale space in the SIFT algorithm is completed.
As an improvement of the invention, the method comprises the following step of positioning local extreme points in the scale space after the scale space is built by the SIFT algorithm.
As an improvement of the present invention, locating the local extremum points includes comparing each pixel point in the scale space with surrounding adjacent pixel points of the same size as the pixel point and with points of adjacent sizes up and down.
As an improvement of the present invention, the method includes that after the detection of the key points (the positioning step of the local extremum points) is completed, the SIFT algorithm is to determine the direction parameter of each key point, specifically as follows:
Figure BDA0004001104060000032
in the above formula, m (x, y) represents a modulus value of the gradient of the key point, and L (x, y) is position information of the key point in the scale space; l (x+/-1, y+/-1) is information about the key point, namely gradient information in the transverse direction and the longitudinal direction;
Figure BDA0004001104060000033
in the above formula, θ (x, y) represents the direction of the gradient.
The method further comprises the steps of extracting feature descriptors of the SIFT algorithm, dividing a neighborhood of the SIFT algorithm into sub-neighborhoods by taking each feature point as a center, calculating gradient features in a plurality of directions in each sub-neighborhood to obtain final feature vectors, and finally carrying out normalization processing on the feature vectors.
As an improvement of the invention, the method further comprises the steps of performing feature matching on the key points through a SIFT algorithm, and measuring similarity between the key points by adopting a euclidean distance, wherein the similarity is represented by the following formula:
Figure BDA0004001104060000041
wherein x and y are coordinate points, i is a determined coordinate point pair, a SIFT algorithm uses a nearest neighbor algorithm to key points in one graph, nearest neighbor key points and secondary neighbor key points are searched in the other graph according to Euclidean distances, and when the ratio of Euclidean distance of the nearest neighbor key point pair to Euclidean distance of the secondary neighbor key point pair is smaller than a set threshold, the nearest neighbor point pair is a pair of effective matching points.
As an improvement of the invention, next, the SIFT algorithm calculates a homography matrix H between two images, the mathematical expression of the matrix H being:
Figure BDA0004001104060000042
if the aligned coordinates of a pair of matching points are (x, y, 1) and (x ', y', 1), a matrix equation can be established through a homography matrix H:
Figure BDA0004001104060000043
after being unfolded, the method comprises the following steps:
Figure BDA0004001104060000044
when the homogeneous coordinate system is used, and the number of the matched key points is more than a certain value, an algorithm RANSAC algorithm is used for iteration to obtain an optimal model, and then an optimal homography matrix is obtained.
Compared with the prior art, the invention has the beneficial effects that: the invention uses VGG-16 to reconstruct the backbone feature extraction network of U-Net, and fuses the attention mechanism CBAM module; the method has the advantages that the image registration is carried out by using the tower segmentation image and the tower template, so that the interference of complex backgrounds on registration accuracy is avoided. By extracting the electric power towers from complex backgrounds, interference of the complex backgrounds is avoided, accuracy of tower image registration can be improved to a great extent, and intelligent shooting of the towers is achieved.
Drawings
Fig. 1 is a flowchart of the registration of a tower image in the present embodiment.
Fig. 2 is a schematic diagram of the attention mechanism CBAM module in this embodiment.
Fig. 3 is a schematic diagram of a U-Net network in this embodiment.
Detailed Description
The present invention is further illustrated in the following drawings and detailed description, which are to be understood as being merely illustrative of the invention and not limiting the scope of the invention.
Examples: as shown in fig. 1, a tower image registration method based on image segmentation, the method includes: the backbone feature extraction network of U-Net is reconstructed using VGG16 and incorporated into the attention mechanism CBAM module, as shown in FIG. 2, wherein Input feature map: input feature map, channel attention module: channel attention module, spatital attention module: spatial attention module, refined feature map: an optimized nominal graph; training an improved U-Net network by using the marked pole tower data set; inputting the tower image into a trained improved U-Net network to obtain a high-precision tower segmentation map; manufacturing a standard tower image template, wherein a clear and complete tower to-be-detected target is arranged on the template; registering the tower segmentation map and the tower template map by using an SIFT-based image registration algorithm; and the inspection robot performs intelligent shooting according to the registration result in the inspection process.
The above description briefly describes a U-Net network, as shown in FIG. 3, input image: input diagram, output segmentation map: output split picture, conv: convolution, relu: activation function, max pool: maximum pooling, up_conv: upsampling, copy and crop: and the characteristic diagrams are fused and spliced, and the applicability of the U-Net network in a transmission line inspection scene is analyzed. Firstly, analyzing the problems in the segmentation of the pole tower image: because the illumination intensity is changeable and the background is changeable, the boundary of the tower is difficult to distinguish from the background, and the gradient change is complex, the tower image segmentation needs high-resolution information to realize accurate segmentation; the tower skeleton structure is relatively fixed, can be described by simple semantic information, and the low-resolution information can be used for identifying a tower target; the transmission line inspection data sets are fewer, enough pole tower pictures are difficult to obtain for safety reasons, and the manufacturing process of the data sets is time-consuming and labor-consuming.
Because Skip-connect operation in U-Net combines low resolution information (providing object class identification basis) and high resolution information (providing accurate segmentation positioning basis), the U-Net segmentation is accurate and the details can be processed well. U-Net is applicable to data sets with few samples, and the original authors only use 30 pictures as training sets. U-Net is also a lightweight model, and the original U-Net model size is only about 28M. Based on the advantages of the U-Net network, the invention finds that the thought of U-Net can be applied to tower segmentation, but in consideration of the two problems of U-Net, the valid convolution increases the difficulty and universality of model design; the overlay-tile strategy is applicable to medical images and is not suitable for common scenes. Because the overall structure of the trunk feature extraction part of U-Net is similar to VGG-16. The present invention replaces the backbone feature extraction network with VGG-16 so that pre-training weights on ImageNet can be used.
When the input image size is 3×512×512, the specific implementation manner is as follows:
1) conv1: the convolution of 64 channels of [3,3] is carried out twice to obtain a preliminary effective characteristic layer of [512,512,64], and 2×2 max pooling is carried out to obtain a characteristic layer of [256,256,64 ].
2) conv2: the convolution of 128 channels of [3,3] is carried out twice to obtain a preliminary effective characteristic layer of [256,256,128], and 2×2 max pooling is carried out to obtain a characteristic layer of [128,128,128 ].
3) conv3: 256-channel convolution of [3,3] is carried out three times to obtain a preliminary effective characteristic layer of [128,128,256], and 2×2 maximum pooling is carried out to obtain a characteristic layer of [64,64,256 ].
4) conv4: the 512-channel convolution of [3,3] is performed three times to obtain a preliminary effective feature layer of [64,64,512], and then 2×2 max pooling is performed to obtain a feature layer of [32,32,256 ].
5) conv5: three convolutions of 512 channels [3,3] were performed to obtain a preliminary effective feature layer of [32,32,512 ].
The SIFT feature detection comprises three modules, namely scale space construction, key point positioning and key point direction determination. The method specifically comprises the following steps: detecting key points, generating local descriptors, matching according to similarity measurement, and purifying matching comparison values; in order to detect feature points in an image at different scales, SIFT firstly performs different-scale changes on an original image, and performs gaussian blur on the image at different scales by using different gaussian kernel functions.
L(x,y,σ)=G(x,y,σ)*I(x,y)
In the above formula, sigma represents a scale space fuzzy coefficient, and a specific formula of a Gaussian kernel function is as follows:
Figure BDA0004001104060000071
in the above expression, I (x, y) represents an input image, x represents a convolution operation, and L (x, y, σ) represents an image subjected to a gaussian kernel G (x, y, σ) blurring process, that is, a gaussian pyramid. After the Gaussian pyramid is obtained, subtraction operation is carried out on the adjacent images obtained in the same layer in the pyramid, so that a difference image is obtained, and the specific calculation is shown as the formula:
D(x,y,σ)=(G(x,y,kσ)-G(x,y,σ))*I(x,y)
=L(x,y,kσ)-L(x,y,σ)
and repeating the operation in the Gaussian pyramid, and constructing a scale space in the SIFT algorithm by forming the Gaussian difference pyramid by the differential Gaussian images finally obtained by all layers.
After the scale space is built by the SIFT algorithm, the key points are positioned in the scale space, and the local extreme points in the Gaussian differential pyramid are the key points to be positioned. To ensure that the SIFT algorithm does not miss extreme points, the SIFT locates the extreme points in the scale space, and a specific locating strategy is to compare each pixel point in the scale space with surrounding adjacent pixel points of the same scale as the pixel point and a total of 26 points of upper and lower adjacent scales.
After the detection of the key points is completed, the SIFT algorithm also determines the direction parameter of each key point.
Figure BDA0004001104060000072
Figure BDA0004001104060000073
Where m (x, y) represents the modulus of the gradient of the key point and θ (x, y) represents the direction of the gradient.
The SIFT algorithm then performs feature descriptor extraction. The SIFT algorithm takes each feature point as a center, divides the neighborhood of the SIFT algorithm into 4 multiplied by 4 sub-neighborhood, calculates gradient features in 8 directions in each sub-neighborhood, and finally obtains feature vectors of 128 dimensions in total for each key point. And finally, carrying out normalization processing on the feature vector, so that the influence caused by brightness change can be eliminated.
The last step of the SIFT algorithm is to perform feature matching on key points, and the Euclidean distance is used for measuring the similarity between the key points, wherein the similarity is shown in the formula:
Figure BDA0004001104060000081
the SIFT algorithm uses a nearest neighbor algorithm [75] on the keypoints in one graph, and finds nearest neighbor and next-neighbor keypoints according to euclidean distances in the other graph, and when the ratio of the euclidean distance of the nearest neighbor keypoint pair to the euclidean distance of the next-neighbor keypoint pair is smaller than a set threshold, the nearest neighbor pair is a valid matching point. Then, calculating a homography matrix H between two images by using a SIFT algorithm, wherein the mathematical expression of the matrix H is as follows:
Figure BDA0004001104060000082
if the aligned coordinates of a pair of matching points are (x, y, 1) and (x ', y', 1), a matrix equation can be established through a homography matrix H:
Figure BDA0004001104060000083
after being unfolded, the material is obtained:
Figure BDA0004001104060000084
when the homogeneous coordinate system is used, and the number of the matched key points is more than 4, an algorithm RANSAC algorithm is used for iteration to obtain an optimal model, and outlier matched point pairs are removed through the RANSAC algorithm to obtain an optimal homography matrix. And finally, carrying out registration model estimation and resampling the tower image to the same size as the template image.
It should be noted that the foregoing merely illustrates the technical idea of the present invention and is not intended to limit the scope of the present invention, and that a person skilled in the art may make several improvements and modifications without departing from the principles of the present invention, which fall within the scope of the claims of the present invention.

Claims (10)

1. A tower image registration method based on image segmentation, the method comprising:
reconstructing a backbone feature extraction network of the U-Net by using VGG16, and integrating the backbone feature extraction network into a attention mechanism CBAM module; training an improved U-Net network by using the marked pole tower data set; inputting the tower image into a trained improved U-Net network to obtain a high-precision tower segmentation map; manufacturing a standard tower image template, wherein a clear and complete tower to-be-detected target is arranged on the template; registering the tower segmentation map and the tower template map by using an SIFT-based image registration algorithm; and the inspection robot performs intelligent shooting according to the registration result in the inspection process.
2. The method for registering a tower image based on image segmentation according to claim 1, wherein the method comprises performing convolution of specific channels several times after inputting the image to obtain a preliminary effective feature layer, and performing maximum pooling to obtain a feature layer.
3. The method for registering a tower image based on image segmentation according to claim 2, wherein the method comprises SIFT feature detection, firstly, performing different scale changes on an original image and performing gaussian blur on the different scale image by using different gaussian kernel functions, and the formula is as follows:
L(x,y,σ)=G(x,y,σ)*I(x,y)
in the above formula, I (x, y) is an input image, x is the width of the image, y is the height of the image, σ is a scale space blur coefficient, L (x, y, σ) is a convolution operation, and the specific formula of the gaussian kernel function is as follows:
Figure FDA0004001104050000011
subtracting the adjacent images obtained in the same layer in the pyramid to obtain a difference image, wherein the specific formula is as follows:
D(x,y,σ)=(G(x,y,kσ)-G(x,y,σ))*I(x,y)
=L(x,y,kσ)-L(x,y,σ)
in the above formula, k is the amplification factor of the space fuzzy coefficient, the operation is repeated in the Gaussian pyramid, the differential Gaussian image finally obtained by all layers forms the Gaussian differential pyramid, and the scale space construction in the SIFT algorithm is completed.
4. A tower image registration method based on image segmentation according to claim 3, wherein the method comprises the following step of locating local extreme points in a scale space after the scale space is constructed by a SIFT algorithm.
5. The method of claim 4, wherein locating local extremal points comprises comparing each pixel in the scale space with surrounding pixels of the same size as the pixel and with points of adjacent sizes.
6. The method for registering a tower image based on image segmentation according to claim 5, wherein the method comprises determining a direction parameter of each key point by a SIFT algorithm after the key point detection is completed, and the method is characterized by comprising the following specific formula:
Figure FDA0004001104050000021
in the above formula, m (x, y) represents the modulus of the key point gradient;
Figure FDA0004001104050000022
in the above formula, θ (x, y) represents the direction of the gradient.
7. The method for registering a tower image based on image segmentation according to claim 6, further comprising extracting feature descriptors from a SIFT algorithm, wherein the SIFT algorithm uses each feature point as a center, divides a neighborhood of the SIFT algorithm into sub-neighborhoods, calculates gradient features in a plurality of directions in each sub-neighborhood to obtain final feature vectors, and finally normalizes the feature vectors.
8. The method for registering a tower image based on image segmentation according to claim 7, further comprising feature matching key points by a SIFT algorithm, and measuring similarity between the key points by using euclidean distance, wherein the similarity is represented by the following formula:
Figure FDA0004001104050000023
the SIFT algorithm uses a nearest neighbor algorithm for key points in one graph, and searches nearest neighbor key points and secondary neighbor key points in the other graph according to Euclidean distances, and when the ratio of the Euclidean distance of the nearest neighbor key point pair to the Euclidean distance of the secondary neighbor key point pair is smaller than a set threshold, the nearest neighbor point pair is a pair of effective matching points.
9. The method for registering a tower image based on image segmentation according to claim 8, wherein a SIFT algorithm is used to calculate a homography matrix H between two images, and the mathematical expression of the matrix H is:
Figure FDA0004001104050000031
if the aligned coordinates of a pair of matching points are (x, y, 1) and (x ', y', 1), a matrix equation can be established through a homography matrix H:
Figure FDA0004001104050000032
10. the method for registering the pole and tower images based on image segmentation according to claim 9, wherein when the homogeneous coordinate system is used, when the number of the matched key points is more than a certain value, an algorithm RANSAC algorithm is used for iteration to obtain an optimal homography matrix.
CN202211618506.XA 2022-12-15 2022-12-15 Tower image registration method based on image segmentation Pending CN116385477A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202211618506.XA CN116385477A (en) 2022-12-15 2022-12-15 Tower image registration method based on image segmentation

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202211618506.XA CN116385477A (en) 2022-12-15 2022-12-15 Tower image registration method based on image segmentation

Publications (1)

Publication Number Publication Date
CN116385477A true CN116385477A (en) 2023-07-04

Family

ID=86960281

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202211618506.XA Pending CN116385477A (en) 2022-12-15 2022-12-15 Tower image registration method based on image segmentation

Country Status (1)

Country Link
CN (1) CN116385477A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116993644A (en) * 2023-09-27 2023-11-03 广东工业大学 Multi-focus image fusion method and device based on image segmentation

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116993644A (en) * 2023-09-27 2023-11-03 广东工业大学 Multi-focus image fusion method and device based on image segmentation
CN116993644B (en) * 2023-09-27 2024-01-19 广东工业大学 Multi-focus image fusion method and device based on image segmentation

Similar Documents

Publication Publication Date Title
CN108961235B (en) Defective insulator identification method based on YOLOv3 network and particle filter algorithm
CN110084850B (en) Dynamic scene visual positioning method based on image semantic segmentation
CN106683119B (en) Moving vehicle detection method based on aerial video image
CN105335973A (en) Visual processing method for strip steel processing production line
CN112766136B (en) Space parking space detection method based on deep learning
CN112819748B (en) Training method and device for strip steel surface defect recognition model
CN110245566B (en) Infrared target remote tracking method based on background features
CN114913498A (en) Parallel multi-scale feature aggregation lane line detection method based on key point estimation
CN111914596B (en) Lane line detection method, device, system and storage medium
CN116385477A (en) Tower image registration method based on image segmentation
CN114494373A (en) High-precision rail alignment method and system based on target detection and image registration
CN113989604A (en) Tire DOT information identification method based on end-to-end deep learning
CN106682668A (en) Power transmission line geological disaster monitoring method using unmanned aerial vehicle to mark images
CN109544608B (en) Unmanned aerial vehicle image acquisition characteristic registration method
KR102416714B1 (en) System and method for city-scale tree mapping using 3-channel images and multiple deep learning
CN116310902A (en) Unmanned aerial vehicle target detection method and system based on lightweight neural network
CN116188943A (en) Solar radio spectrum burst information detection method and device
CN113723447B (en) End-to-end template matching method for multi-mode image
CN114332814A (en) Parking frame identification method and device, electronic equipment and storage medium
CN113642430A (en) High-precision visual positioning method and system for underground parking lot based on VGG + NetVLAD
CN115019306A (en) Embedding box label batch identification method and system based on deep learning and machine vision
CN115147644A (en) Method, system, device and storage medium for training and describing image description model
CN115019201A (en) Weak and small target detection method based on feature refined depth network
CN111046861B (en) Method for identifying infrared image, method for constructing identification model and application
CN113379738A (en) Method and system for detecting and positioning epidemic trees based on images

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination