CN111985503A - Target detection method and device based on improved characteristic pyramid network structure - Google Patents

Target detection method and device based on improved characteristic pyramid network structure Download PDF

Info

Publication number
CN111985503A
CN111985503A CN202010825554.0A CN202010825554A CN111985503A CN 111985503 A CN111985503 A CN 111985503A CN 202010825554 A CN202010825554 A CN 202010825554A CN 111985503 A CN111985503 A CN 111985503A
Authority
CN
China
Prior art keywords
feature
layer
characteristic
layers
resolution
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202010825554.0A
Other languages
Chinese (zh)
Other versions
CN111985503B (en
Inventor
李百成
张翊
黎嘉朗
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Whale Cloud Technology Co Ltd
Original Assignee
Whale Cloud Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Whale Cloud Technology Co Ltd filed Critical Whale Cloud Technology Co Ltd
Priority to CN202010825554.0A priority Critical patent/CN111985503B/en
Publication of CN111985503A publication Critical patent/CN111985503A/en
Application granted granted Critical
Publication of CN111985503B publication Critical patent/CN111985503B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2415Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on parametric or probabilistic models, e.g. based on likelihood ratio or false acceptance rate versus a false rejection rate
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/253Fusion techniques of extracted features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/29Graphical models, e.g. Bayesian networks

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Probability & Statistics with Applications (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)

Abstract

The scheme of the invention obtains a final characteristic layer by carrying out fusion, selection and residual operation on multi-scale characteristics extracted from a backbone network, and carries out target classification and position regression on the final characteristic layer to obtain a final result. The method can correctly detect the target even under the condition that the target is partially shielded in the target detection, and has high robustness and high performance.

Description

Target detection method and device based on improved characteristic pyramid network structure
Technical Field
The invention belongs to the technical field of target detection, and particularly relates to a target detection method and device based on an improved characteristic pyramid network structure.
Background
In the target detection of pictures, the feature pyramid network is a structure capable of greatly improving the network performance at a low cost, and has been used in various mainstream target detection network structures due to its excellent performance.
In the feature pyramid network, features from the backbone network are upsampled through a top-down path to a higher-level pyramid to generate a coarser feature map with stronger semantic information. These features are then connected horizontally with a bottom-up feature map of the same spatial size, enhancing the semantic information at a lower level. However, such sequentially integrated features may focus more on features at adjacent resolutions and dilute features at non-adjacent resolutions, limiting the performance of the network. Therefore, how to better integrate the features of the backbone network becomes a hot issue in academic and industrial fields.
Disclosure of Invention
In view of the above-mentioned deficiencies of the prior art, it is an object of the present invention to provide a method for controlling a motor.
The embodiment of the invention discloses a target detection method based on an improved feature pyramid network structure, which comprises the steps of obtaining a plurality of first feature layers with different sizes of a detected picture through a feature extraction network, and zooming the first feature layers to a preset resolution; fusing the plurality of zoomed first feature layers to obtain a second feature layer, and performing global average pooling and dimension reduction operations on the second feature layer to obtain a third feature layer; obtaining a selected fourth characteristic layer according to the weights of the third characteristic layer and different characteristic layers; processing the fourth characteristic layer in a residual error mode to obtain a pyramid characteristic diagram; and performing target classification and position regression on the pyramid feature map to output a detection frame.
In one possible embodiment, a bilinear interpolation method is adopted to perform amplification operation on a first feature layer with a resolution smaller than a preset resolution to reach the preset resolution; performing reduction operation on the first characteristic layer with the resolution greater than the preset resolution by adopting a maximum pooling method to achieve the preset resolution; wherein the preset resolution is centered in resolution among the plurality of first feature layers.
In one possible embodiment, the statistical information of the whole channel domain of the second feature layer is calculated by using global average pooling, and the dependency relationship of the channel domain is calculated by using the full connection layer by using the statistical information, so as to obtain the third feature layer.
In one possible embodiment, the third feature layer is expanded to n × d dimensions by 1 × 1 convolution, where n is the number of feature pyramid layers used and d is the dimension of each feature layer; and the weight of each channel is obtained through softmax operation, the weight is multiplied by the third feature layer to obtain the selected feature layer, and the feature layers are added pixel by pixel to obtain a finally selected fourth feature map.
In one possible embodiment, the pyramid feature map is obtained by scaling the fourth feature layer to a resolution corresponding to the plurality of first feature layers by an inverse operation of rescaling and adding the fourth feature layer to the corresponding first feature layer.
An object detection apparatus based on an improved feature pyramid network structure, comprising: the zooming module is used for acquiring a plurality of first feature layers with different sizes of the detected picture through a feature extraction network and zooming the first feature layers to a preset resolution; the fusion module is used for fusing the plurality of zoomed first feature layers to obtain a second feature layer, and performing global average pooling and dimension reduction operation on the second feature layer to obtain a third feature layer; the selection module is used for obtaining a selected fourth characteristic layer according to the weights of the third characteristic layer and different characteristic layers; the residual error module is used for processing the fourth characteristic layer in a residual error mode to obtain a pyramid characteristic diagram; and the position regression module is used for carrying out target classification and position regression on the pyramid feature map so as to output a detection frame.
In one possible embodiment, the scaling module is further configured to: performing amplification operation on the first characteristic layer with the resolution smaller than the preset resolution by adopting a bilinear interpolation method to achieve the preset resolution; performing reduction operation on the first characteristic layer with the resolution greater than the preset resolution by adopting a maximum pooling method to achieve the preset resolution; wherein the preset resolution is centered in resolution among the plurality of first feature layers.
In one possible embodiment, the fusion module is further configured to: and calculating the statistical information of the whole channel domain of the second characteristic layer by adopting global average pooling, and calculating the dependency relationship of the channel domain by adopting the full connection layer by utilizing the statistical information to obtain a third characteristic layer.
In one possible embodiment, the selection module is further configured to: amplifying the third characteristic layer to n x d dimension by 1 x 1 convolution, wherein n is the number of the adopted characteristic pyramid layers, and d is the dimension of each characteristic layer; and the weight of each channel is obtained through softmax operation, the weight is multiplied by the third feature layer to obtain the selected feature layer, and the feature layers are added pixel by pixel to obtain a finally selected fourth feature map.
In a possible embodiment, the residual module is further configured to scale the fourth feature layer to a resolution corresponding to the plurality of first feature layers by an inverse operation of the rescaling and add the fourth feature layer to the corresponding first feature layers to obtain the pyramid feature map.
A computer storage medium storing a computer program which, when executed, implements the method as hereinbefore described.
Compared with the prior art, the invention has the following beneficial effects:
the scheme of the invention has high robustness, and the network has stronger extraction capability on the target characteristics by dynamically selecting the characteristics of the backbone network. Under extreme conditions, such as the condition that the target characteristics are damaged by weak light, shielding and the like, the network can realize the enhancement of the target characteristics by aggregating the characteristics of characteristic layers with different resolutions, the detection accuracy is improved, and the target can still be correctly detected under the condition that the target is partially shielded. The method has the advantages of stronger network performance, smaller introduced parameter quantity, no obvious reduction of reasoning speed and improvement of accuracy.
Drawings
FIG. 1 is a flow chart of a method according to an embodiment of the present invention;
FIG. 2 is a network structure diagram of a convergence phase according to an embodiment of the present invention;
FIG. 3 is a network architecture diagram of a selection phase according to an embodiment of the present invention;
FIG. 4 is a diagram of a network structure in a residual error phase according to an embodiment of the present invention;
fig. 5(a) and 5(b) are comparison graphs of the testing results on the ODF port testing data set of the optical distribution frame according to the embodiment of the present invention.
Detailed Description
In order to facilitate understanding of those skilled in the art, the present invention will be further described with reference to the following examples and drawings, which are not intended to limit the present invention.
The method can dynamically select the feature layer, and specifically comprises the steps of performing fusion, selection and residual operation on multi-scale features extracted from a backbone network to obtain a final feature layer, and performing target classification and position regression on the final feature layer to obtain a final result. The method has high robustness and high performance in target detection, and can correctly detect the target even under the condition that the target is partially shielded.
Specifically, with reference to fig. 1, an embodiment of the present invention discloses a target detection method based on an improved feature pyramid network structure, including:
s101, a plurality of first feature layers with different sizes of the detected picture are obtained through a feature extraction network, and the first feature layers are scaled to a preset resolution.
And inputting the picture to be detected into a backbone network for feature extraction, wherein the backbone network, namely the feature extraction network, can comprise ResNet, VggNet and the like.
Next, referring to FIG. 2, for the multi-scale features { C ] extracted from the backbone network3,C4,C5,C6,C7Scaling a plurality of first feature layers with different sizes to the size of the middle layer, namely, for the feature layer with the resolution less than the C5 layer, adopting a bilinear interpolation method to enlarge the size, and for the feature layer with the resolution greater than the C5 layer, reducing the size by using maximum pooling to obtain { R }3,R4,R5,R6,R7}。
And S102, fusing the scaled first feature layers to obtain a second feature layer, and performing global average pooling and dimension reduction operations on the second feature layer to obtain a third feature layer.
Referring to FIG. 2, the scaled features are then simply information fused using a pixel-by-pixel addition operation to generate a fused feature layer RsI.e. the second feature layer, and then use the global average pooling to obtain the statistical information of the whole channel domain, as shown in formula (1):
Figure BDA0002636075780000051
wherein H and W represent the height and width of the feature map respectively, (i, j) represents the coordinates of the pixel, and zcRepresenting the statistical information of the c-th channel,
Figure BDA0002636075780000052
and (4) fusing the characteristic layers. In order to fully utilize the information in z, a fully connected layer is used to calculate the dependency relationship of the channel domain, and meanwhile, the dimensionality is reduced to improve the efficiency of the network, so that a third characteristic layer p is obtained, as shown in formula (2)
p=Fc(zc) (2)
And S103, obtaining the selected fourth characteristic layer according to the weights of the third characteristic layer and different characteristic layers.
In order to dynamically select the appropriate feature layer information, as shown in fig. 3, the network needs to adaptively assign weights of different feature layers, and to achieve this goal, the third feature layer p is first expanded to n × d dimension by 1 × 1 convolution, where n is the number of feature pyramid layers used, and is set to 5 here, and d is the dimension of each feature layer. Then, the weight of each channel is obtained by softmax operation, and is represented by a, that is:
A=softmax(conv(p)) (3)
after the weights of all channels are obtained, multiplying the weights by the original rescaled features to obtain selected feature layers, and then adding the selected feature layers pixel by pixel to obtain a finally selected fourth feature layer q:
q=sum(AiRi) (4)
wherein R isi∈{R3,R4,R5,R6,R7}。
And S104, processing the fourth characteristic layer in a residual error mode to obtain a pyramid characteristic diagram.
Referring to fig. 4, the selected features enhance the original features by means of residual errors, which makes the training speed of the network faster and the learned features more robust. The specific operation is to scale q by the inverse of rescalingTo the corresponding first feature layer CiAnd adding the resolution ratios to obtain the final pyramid feature map. As shown in formula (5):
Pi=Rescale(q)+Ci (5)
and S105, performing target classification and position regression on the pyramid feature map to output a detection frame.
Taking the single target detection network FCOS as an example, the characteristic pyramid network structure is replaced by the method, and the port detection data set is tested, and the results are shown in fig. 5(a) and fig. 5 (b). Because the ODF ports are densely arranged, the ports are often shielded from each other, which makes the detection of the shielded ports difficult, but the FCOS using the method shown in fig. 5(b) can still correctly detect the target under the condition that the target is partially shielded, as compared with the original FCOS shown in fig. 5(a), and the detection effect is better.
The embodiment of the invention also discloses a target detection device based on the improved characteristic pyramid network structure, which comprises a zooming module, a feature extraction module and a target detection module, wherein the zooming module is used for acquiring a plurality of first characteristic layers with different sizes of a detected picture through a characteristic extraction network and zooming the plurality of first characteristic layers to a preset resolution; the fusion module is used for fusing the plurality of zoomed first feature layers to obtain a second feature layer, and performing global average pooling and dimension reduction operation on the second feature layer to obtain a third feature layer; the selection module is used for obtaining a selected fourth characteristic layer according to the weights of the third characteristic layer and different characteristic layers; the residual error module is used for processing the fourth characteristic layer in a residual error mode to obtain a pyramid characteristic diagram; and the position regression module is used for carrying out target classification and position regression on the pyramid feature map so as to output a detection frame. The embodiments of the present invention correspond to the embodiments of the method described above, and specific contents thereof may refer to the embodiments of the method.
In the several embodiments provided in the present application, it should be understood that the disclosed apparatus and method may be implemented in other ways. For example, the above-described embodiments are merely illustrative, and for example, a division of a unit is merely a division of a logic function, and an actual implementation may have another division, for example, a plurality of units or components may be combined or integrated into another system, or some features may be omitted, or not executed.
Units described as separate parts may or may not be physically separate, and parts displayed as units may or may not be physical units, may be located in one place, or may be distributed on a plurality of network units. Some or all of the units can be selected according to actual needs to achieve the purpose of the solution of the embodiment of the present invention.
In addition, functional units in the embodiments of the present invention may be integrated into one processing unit, or each unit may exist alone physically, or two or more units are integrated into one unit. The integrated unit can be realized in a form of hardware, and can also be realized in a form of a software functional unit.
While the invention has been described in terms of its preferred embodiments, it will be understood by those skilled in the art that various changes in form and details may be made therein without departing from the spirit and scope of the invention.

Claims (10)

1. A target detection method based on an improved characteristic pyramid network structure is characterized in that,
acquiring a plurality of first feature layers with different sizes of a detected picture through a feature extraction network, and scaling the plurality of first feature layers to a preset resolution;
fusing the plurality of zoomed first feature layers to obtain a second feature layer, and performing global average pooling and dimension reduction operations on the second feature layer to obtain a third feature layer;
obtaining a selected fourth characteristic layer according to the weights of the third characteristic layer and different characteristic layers;
processing the fourth characteristic layer in a residual error mode to obtain a pyramid characteristic diagram;
and performing target classification and position regression on the pyramid feature map to output a detection frame.
2. The method of claim 1, wherein scaling the plurality of first feature layers to a preset resolution size comprises: performing amplification operation on the first characteristic layer with the resolution smaller than the preset resolution by adopting a bilinear interpolation method to achieve the preset resolution; performing reduction operation on the first characteristic layer with the resolution greater than the preset resolution by adopting a maximum pooling method to achieve the preset resolution; wherein the preset resolution is centered in resolution among the plurality of first feature layers.
3. The method of claim 1, wherein performing global average pooling and dimensionality reduction operations on the second feature layer to obtain a third feature layer comprises: and calculating the statistical information of the whole channel domain of the second characteristic layer by adopting global average pooling, and calculating the dependency relationship of the channel domain by adopting the full connection layer by utilizing the statistical information to obtain a third characteristic layer.
4. The method of claim 1, wherein obtaining the selected fourth feature layer based on the weights of the third feature layer and the different feature layers comprises: amplifying the third characteristic layer to n x d dimension by 1 x 1 convolution, wherein n is the number of the adopted characteristic pyramid layers, and d is the dimension of each characteristic layer; and the weight of each channel is obtained through softmax operation, the weight is multiplied by the third feature layer to obtain the selected feature layer, and the feature layers are added pixel by pixel to obtain a finally selected fourth feature map.
5. The method of claim 1, wherein the pyramid feature map is obtained by scaling the fourth feature layer to a resolution corresponding to the plurality of first feature layers by an inverse operation of rescaling and adding to the corresponding first feature layers.
6. An object detection device based on an improved characteristic pyramid network structure is characterized in that,
the zooming module is used for acquiring a plurality of first feature layers with different sizes of the detected picture through a feature extraction network and zooming the first feature layers to a preset resolution;
the fusion module is used for fusing the plurality of zoomed first feature layers to obtain a second feature layer, and performing global average pooling and dimension reduction operation on the second feature layer to obtain a third feature layer;
the selection module is used for obtaining a selected fourth characteristic layer according to the weights of the third characteristic layer and different characteristic layers;
the residual error module is used for processing the fourth characteristic layer in a residual error mode to obtain a pyramid characteristic diagram;
and the position regression module is used for carrying out target classification and position regression on the pyramid feature map so as to output a detection frame.
7. The apparatus of claim 6, wherein the scaling module is further to: performing amplification operation on the first characteristic layer with the resolution smaller than the preset resolution by adopting a bilinear interpolation method to achieve the preset resolution; performing reduction operation on the first characteristic layer with the resolution greater than the preset resolution by adopting a maximum pooling method to achieve the preset resolution; wherein the preset resolution is centered in resolution among the plurality of first feature layers.
8. The apparatus of claim 6, wherein the fusion module is further to: and calculating the statistical information of the whole channel domain of the second characteristic layer by adopting global average pooling, and calculating the dependency relationship of the channel domain by adopting the full connection layer by utilizing the statistical information to obtain a third characteristic layer.
9. The apparatus of claim 6, wherein the selection module is further to: amplifying the third characteristic layer to n x d dimension by 1 x 1 convolution, wherein n is the number of the adopted characteristic pyramid layers, and d is the dimension of each characteristic layer; and the weight of each channel is obtained through softmax operation, the weight is multiplied by the third feature layer to obtain the selected feature layer, and the feature layers are added pixel by pixel to obtain a finally selected fourth feature map.
10. The apparatus of claim 6, wherein the residual module is further configured to scale a fourth feature layer to a resolution corresponding to the plurality of first feature layers by an inverse of the rescaling operation and to add the fourth feature layer to the corresponding first feature layer to obtain a pyramid feature map.
CN202010825554.0A 2020-08-17 2020-08-17 Target detection method and device based on improved feature pyramid network structure Active CN111985503B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010825554.0A CN111985503B (en) 2020-08-17 2020-08-17 Target detection method and device based on improved feature pyramid network structure

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010825554.0A CN111985503B (en) 2020-08-17 2020-08-17 Target detection method and device based on improved feature pyramid network structure

Publications (2)

Publication Number Publication Date
CN111985503A true CN111985503A (en) 2020-11-24
CN111985503B CN111985503B (en) 2024-04-26

Family

ID=73434028

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010825554.0A Active CN111985503B (en) 2020-08-17 2020-08-17 Target detection method and device based on improved feature pyramid network structure

Country Status (1)

Country Link
CN (1) CN111985503B (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20220318541A1 (en) * 2021-04-05 2022-10-06 Microsoft Technology Licensing, Llc Dynamic head for object detection
CN116257038A (en) * 2023-05-15 2023-06-13 深圳市瓴鹰智能科技有限公司 Steering engine control and diagnosis method and device based on lightweight convolutional neural network

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106845510A (en) * 2016-11-07 2017-06-13 中国传媒大学 Chinese tradition visual culture Symbol Recognition based on depth level Fusion Features
CN110633661A (en) * 2019-08-31 2019-12-31 南京理工大学 Semantic segmentation fused remote sensing image target detection method
CN111210443A (en) * 2020-01-03 2020-05-29 吉林大学 Deformable convolution mixing task cascading semantic segmentation method based on embedding balance
CN111242071A (en) * 2020-01-17 2020-06-05 陕西师范大学 Attention remote sensing image target detection method based on anchor frame
CN111507359A (en) * 2020-03-09 2020-08-07 杭州电子科技大学 Self-adaptive weighting fusion method of image feature pyramid

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106845510A (en) * 2016-11-07 2017-06-13 中国传媒大学 Chinese tradition visual culture Symbol Recognition based on depth level Fusion Features
CN110633661A (en) * 2019-08-31 2019-12-31 南京理工大学 Semantic segmentation fused remote sensing image target detection method
CN111210443A (en) * 2020-01-03 2020-05-29 吉林大学 Deformable convolution mixing task cascading semantic segmentation method based on embedding balance
CN111242071A (en) * 2020-01-17 2020-06-05 陕西师范大学 Attention remote sensing image target detection method based on anchor frame
CN111507359A (en) * 2020-03-09 2020-08-07 杭州电子科技大学 Self-adaptive weighting fusion method of image feature pyramid

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
JIANGMIAO PANG ET AL.: ""Libra R-CNN: Towards Balanced Learning for Object Detection"", 《IEEE》, pages 821 - 830 *

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20220318541A1 (en) * 2021-04-05 2022-10-06 Microsoft Technology Licensing, Llc Dynamic head for object detection
US11989956B2 (en) * 2021-04-05 2024-05-21 Microsoft Technology Licensing, Llc Dynamic head for object detection
CN116257038A (en) * 2023-05-15 2023-06-13 深圳市瓴鹰智能科技有限公司 Steering engine control and diagnosis method and device based on lightweight convolutional neural network

Also Published As

Publication number Publication date
CN111985503B (en) 2024-04-26

Similar Documents

Publication Publication Date Title
Amanatiadis et al. A survey on evaluation methods for image interpolation
JP6075295B2 (en) Dictionary creation device, image processing device, image processing system, dictionary creation method, image processing method, and program
CN110136056A (en) The method and apparatus of image super-resolution rebuilding
US20200184697A1 (en) Image Modification Using Detected Symmetry
CN109934773A (en) A kind of image processing method, device, electronic equipment and computer-readable medium
Munoz-Mejias et al. A low-complexity pre-processing system for restoring low-quality QR code images
CN103390275B (en) The method of dynamical image joining
CN111985503A (en) Target detection method and device based on improved characteristic pyramid network structure
CN108876716B (en) Super-resolution reconstruction method and device
Patel et al. Accelerated seam carving for image retargeting
US9076232B2 (en) Apparatus and method for interpolating image, and apparatus for processing image using the same
JP2020098455A (en) Object identification system, object identification method, and image identification program
Pan et al. LPSRGAN: Generative adversarial networks for super-resolution of license plate image
Zhou et al. Cross-scale collaborative network for single image super resolution
US20220020113A1 (en) Image resizing using seam carving
CN113536971B (en) Target detection method based on incremental learning
CN113807354B (en) Image semantic segmentation method, device, equipment and storage medium
Wang et al. A multi-scale attentive recurrent network for image dehazing
CN113516697A (en) Image registration method and device, electronic equipment and computer-readable storage medium
WO2022247394A1 (en) Image splicing method and apparatus, and storage medium and electronic device
CN114372944B (en) Multi-mode and multi-scale fused candidate region generation method and related device
CN112070853A (en) Image generation method and device
CN115619678A (en) Image deformation correction method and device, computer equipment and storage medium
CN101989350A (en) Image enhancement method, image enhancement device and image processing circuit
CN114897919B (en) Deplabv3+ semantic segmentation method based on edge gradient guide interpolation

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant