CN111985503A - Target detection method and device based on improved characteristic pyramid network structure - Google Patents
Target detection method and device based on improved characteristic pyramid network structure Download PDFInfo
- Publication number
- CN111985503A CN111985503A CN202010825554.0A CN202010825554A CN111985503A CN 111985503 A CN111985503 A CN 111985503A CN 202010825554 A CN202010825554 A CN 202010825554A CN 111985503 A CN111985503 A CN 111985503A
- Authority
- CN
- China
- Prior art keywords
- feature
- layer
- characteristic
- layers
- resolution
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000001514 detection method Methods 0.000 title claims abstract description 26
- 238000000034 method Methods 0.000 claims abstract description 25
- 230000004927 fusion Effects 0.000 claims abstract description 7
- 238000011176 pooling Methods 0.000 claims description 17
- 238000000605 extraction Methods 0.000 claims description 10
- 238000010586 diagram Methods 0.000 claims description 9
- 230000003321 amplification Effects 0.000 claims description 4
- 238000003199 nucleic acid amplification method Methods 0.000 claims description 4
- 230000004931 aggregating effect Effects 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000004590 computer program Methods 0.000 description 1
- 230000007812 deficiency Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000002708 enhancing effect Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
- G06F18/2415—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on parametric or probabilistic models, e.g. based on likelihood ratio or false acceptance rate versus a false rejection rate
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/25—Fusion techniques
- G06F18/253—Fusion techniques of extracted features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/29—Graphical models, e.g. Bayesian networks
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Evolutionary Biology (AREA)
- Evolutionary Computation (AREA)
- Bioinformatics & Computational Biology (AREA)
- General Engineering & Computer Science (AREA)
- Artificial Intelligence (AREA)
- Life Sciences & Earth Sciences (AREA)
- Probability & Statistics with Applications (AREA)
- Multimedia (AREA)
- Image Analysis (AREA)
Abstract
The scheme of the invention obtains a final characteristic layer by carrying out fusion, selection and residual operation on multi-scale characteristics extracted from a backbone network, and carries out target classification and position regression on the final characteristic layer to obtain a final result. The method can correctly detect the target even under the condition that the target is partially shielded in the target detection, and has high robustness and high performance.
Description
Technical Field
The invention belongs to the technical field of target detection, and particularly relates to a target detection method and device based on an improved characteristic pyramid network structure.
Background
In the target detection of pictures, the feature pyramid network is a structure capable of greatly improving the network performance at a low cost, and has been used in various mainstream target detection network structures due to its excellent performance.
In the feature pyramid network, features from the backbone network are upsampled through a top-down path to a higher-level pyramid to generate a coarser feature map with stronger semantic information. These features are then connected horizontally with a bottom-up feature map of the same spatial size, enhancing the semantic information at a lower level. However, such sequentially integrated features may focus more on features at adjacent resolutions and dilute features at non-adjacent resolutions, limiting the performance of the network. Therefore, how to better integrate the features of the backbone network becomes a hot issue in academic and industrial fields.
Disclosure of Invention
In view of the above-mentioned deficiencies of the prior art, it is an object of the present invention to provide a method for controlling a motor.
The embodiment of the invention discloses a target detection method based on an improved feature pyramid network structure, which comprises the steps of obtaining a plurality of first feature layers with different sizes of a detected picture through a feature extraction network, and zooming the first feature layers to a preset resolution; fusing the plurality of zoomed first feature layers to obtain a second feature layer, and performing global average pooling and dimension reduction operations on the second feature layer to obtain a third feature layer; obtaining a selected fourth characteristic layer according to the weights of the third characteristic layer and different characteristic layers; processing the fourth characteristic layer in a residual error mode to obtain a pyramid characteristic diagram; and performing target classification and position regression on the pyramid feature map to output a detection frame.
In one possible embodiment, a bilinear interpolation method is adopted to perform amplification operation on a first feature layer with a resolution smaller than a preset resolution to reach the preset resolution; performing reduction operation on the first characteristic layer with the resolution greater than the preset resolution by adopting a maximum pooling method to achieve the preset resolution; wherein the preset resolution is centered in resolution among the plurality of first feature layers.
In one possible embodiment, the statistical information of the whole channel domain of the second feature layer is calculated by using global average pooling, and the dependency relationship of the channel domain is calculated by using the full connection layer by using the statistical information, so as to obtain the third feature layer.
In one possible embodiment, the third feature layer is expanded to n × d dimensions by 1 × 1 convolution, where n is the number of feature pyramid layers used and d is the dimension of each feature layer; and the weight of each channel is obtained through softmax operation, the weight is multiplied by the third feature layer to obtain the selected feature layer, and the feature layers are added pixel by pixel to obtain a finally selected fourth feature map.
In one possible embodiment, the pyramid feature map is obtained by scaling the fourth feature layer to a resolution corresponding to the plurality of first feature layers by an inverse operation of rescaling and adding the fourth feature layer to the corresponding first feature layer.
An object detection apparatus based on an improved feature pyramid network structure, comprising: the zooming module is used for acquiring a plurality of first feature layers with different sizes of the detected picture through a feature extraction network and zooming the first feature layers to a preset resolution; the fusion module is used for fusing the plurality of zoomed first feature layers to obtain a second feature layer, and performing global average pooling and dimension reduction operation on the second feature layer to obtain a third feature layer; the selection module is used for obtaining a selected fourth characteristic layer according to the weights of the third characteristic layer and different characteristic layers; the residual error module is used for processing the fourth characteristic layer in a residual error mode to obtain a pyramid characteristic diagram; and the position regression module is used for carrying out target classification and position regression on the pyramid feature map so as to output a detection frame.
In one possible embodiment, the scaling module is further configured to: performing amplification operation on the first characteristic layer with the resolution smaller than the preset resolution by adopting a bilinear interpolation method to achieve the preset resolution; performing reduction operation on the first characteristic layer with the resolution greater than the preset resolution by adopting a maximum pooling method to achieve the preset resolution; wherein the preset resolution is centered in resolution among the plurality of first feature layers.
In one possible embodiment, the fusion module is further configured to: and calculating the statistical information of the whole channel domain of the second characteristic layer by adopting global average pooling, and calculating the dependency relationship of the channel domain by adopting the full connection layer by utilizing the statistical information to obtain a third characteristic layer.
In one possible embodiment, the selection module is further configured to: amplifying the third characteristic layer to n x d dimension by 1 x 1 convolution, wherein n is the number of the adopted characteristic pyramid layers, and d is the dimension of each characteristic layer; and the weight of each channel is obtained through softmax operation, the weight is multiplied by the third feature layer to obtain the selected feature layer, and the feature layers are added pixel by pixel to obtain a finally selected fourth feature map.
In a possible embodiment, the residual module is further configured to scale the fourth feature layer to a resolution corresponding to the plurality of first feature layers by an inverse operation of the rescaling and add the fourth feature layer to the corresponding first feature layers to obtain the pyramid feature map.
A computer storage medium storing a computer program which, when executed, implements the method as hereinbefore described.
Compared with the prior art, the invention has the following beneficial effects:
the scheme of the invention has high robustness, and the network has stronger extraction capability on the target characteristics by dynamically selecting the characteristics of the backbone network. Under extreme conditions, such as the condition that the target characteristics are damaged by weak light, shielding and the like, the network can realize the enhancement of the target characteristics by aggregating the characteristics of characteristic layers with different resolutions, the detection accuracy is improved, and the target can still be correctly detected under the condition that the target is partially shielded. The method has the advantages of stronger network performance, smaller introduced parameter quantity, no obvious reduction of reasoning speed and improvement of accuracy.
Drawings
FIG. 1 is a flow chart of a method according to an embodiment of the present invention;
FIG. 2 is a network structure diagram of a convergence phase according to an embodiment of the present invention;
FIG. 3 is a network architecture diagram of a selection phase according to an embodiment of the present invention;
FIG. 4 is a diagram of a network structure in a residual error phase according to an embodiment of the present invention;
fig. 5(a) and 5(b) are comparison graphs of the testing results on the ODF port testing data set of the optical distribution frame according to the embodiment of the present invention.
Detailed Description
In order to facilitate understanding of those skilled in the art, the present invention will be further described with reference to the following examples and drawings, which are not intended to limit the present invention.
The method can dynamically select the feature layer, and specifically comprises the steps of performing fusion, selection and residual operation on multi-scale features extracted from a backbone network to obtain a final feature layer, and performing target classification and position regression on the final feature layer to obtain a final result. The method has high robustness and high performance in target detection, and can correctly detect the target even under the condition that the target is partially shielded.
Specifically, with reference to fig. 1, an embodiment of the present invention discloses a target detection method based on an improved feature pyramid network structure, including:
s101, a plurality of first feature layers with different sizes of the detected picture are obtained through a feature extraction network, and the first feature layers are scaled to a preset resolution.
And inputting the picture to be detected into a backbone network for feature extraction, wherein the backbone network, namely the feature extraction network, can comprise ResNet, VggNet and the like.
Next, referring to FIG. 2, for the multi-scale features { C ] extracted from the backbone network3,C4,C5,C6,C7Scaling a plurality of first feature layers with different sizes to the size of the middle layer, namely, for the feature layer with the resolution less than the C5 layer, adopting a bilinear interpolation method to enlarge the size, and for the feature layer with the resolution greater than the C5 layer, reducing the size by using maximum pooling to obtain { R }3,R4,R5,R6,R7}。
And S102, fusing the scaled first feature layers to obtain a second feature layer, and performing global average pooling and dimension reduction operations on the second feature layer to obtain a third feature layer.
Referring to FIG. 2, the scaled features are then simply information fused using a pixel-by-pixel addition operation to generate a fused feature layer RsI.e. the second feature layer, and then use the global average pooling to obtain the statistical information of the whole channel domain, as shown in formula (1):
wherein H and W represent the height and width of the feature map respectively, (i, j) represents the coordinates of the pixel, and zcRepresenting the statistical information of the c-th channel,and (4) fusing the characteristic layers. In order to fully utilize the information in z, a fully connected layer is used to calculate the dependency relationship of the channel domain, and meanwhile, the dimensionality is reduced to improve the efficiency of the network, so that a third characteristic layer p is obtained, as shown in formula (2)
p=Fc(zc) (2)
And S103, obtaining the selected fourth characteristic layer according to the weights of the third characteristic layer and different characteristic layers.
In order to dynamically select the appropriate feature layer information, as shown in fig. 3, the network needs to adaptively assign weights of different feature layers, and to achieve this goal, the third feature layer p is first expanded to n × d dimension by 1 × 1 convolution, where n is the number of feature pyramid layers used, and is set to 5 here, and d is the dimension of each feature layer. Then, the weight of each channel is obtained by softmax operation, and is represented by a, that is:
A=softmax(conv(p)) (3)
after the weights of all channels are obtained, multiplying the weights by the original rescaled features to obtain selected feature layers, and then adding the selected feature layers pixel by pixel to obtain a finally selected fourth feature layer q:
q=sum(AiRi) (4)
wherein R isi∈{R3,R4,R5,R6,R7}。
And S104, processing the fourth characteristic layer in a residual error mode to obtain a pyramid characteristic diagram.
Referring to fig. 4, the selected features enhance the original features by means of residual errors, which makes the training speed of the network faster and the learned features more robust. The specific operation is to scale q by the inverse of rescalingTo the corresponding first feature layer CiAnd adding the resolution ratios to obtain the final pyramid feature map. As shown in formula (5):
Pi=Rescale(q)+Ci (5)
and S105, performing target classification and position regression on the pyramid feature map to output a detection frame.
Taking the single target detection network FCOS as an example, the characteristic pyramid network structure is replaced by the method, and the port detection data set is tested, and the results are shown in fig. 5(a) and fig. 5 (b). Because the ODF ports are densely arranged, the ports are often shielded from each other, which makes the detection of the shielded ports difficult, but the FCOS using the method shown in fig. 5(b) can still correctly detect the target under the condition that the target is partially shielded, as compared with the original FCOS shown in fig. 5(a), and the detection effect is better.
The embodiment of the invention also discloses a target detection device based on the improved characteristic pyramid network structure, which comprises a zooming module, a feature extraction module and a target detection module, wherein the zooming module is used for acquiring a plurality of first characteristic layers with different sizes of a detected picture through a characteristic extraction network and zooming the plurality of first characteristic layers to a preset resolution; the fusion module is used for fusing the plurality of zoomed first feature layers to obtain a second feature layer, and performing global average pooling and dimension reduction operation on the second feature layer to obtain a third feature layer; the selection module is used for obtaining a selected fourth characteristic layer according to the weights of the third characteristic layer and different characteristic layers; the residual error module is used for processing the fourth characteristic layer in a residual error mode to obtain a pyramid characteristic diagram; and the position regression module is used for carrying out target classification and position regression on the pyramid feature map so as to output a detection frame. The embodiments of the present invention correspond to the embodiments of the method described above, and specific contents thereof may refer to the embodiments of the method.
In the several embodiments provided in the present application, it should be understood that the disclosed apparatus and method may be implemented in other ways. For example, the above-described embodiments are merely illustrative, and for example, a division of a unit is merely a division of a logic function, and an actual implementation may have another division, for example, a plurality of units or components may be combined or integrated into another system, or some features may be omitted, or not executed.
Units described as separate parts may or may not be physically separate, and parts displayed as units may or may not be physical units, may be located in one place, or may be distributed on a plurality of network units. Some or all of the units can be selected according to actual needs to achieve the purpose of the solution of the embodiment of the present invention.
In addition, functional units in the embodiments of the present invention may be integrated into one processing unit, or each unit may exist alone physically, or two or more units are integrated into one unit. The integrated unit can be realized in a form of hardware, and can also be realized in a form of a software functional unit.
While the invention has been described in terms of its preferred embodiments, it will be understood by those skilled in the art that various changes in form and details may be made therein without departing from the spirit and scope of the invention.
Claims (10)
1. A target detection method based on an improved characteristic pyramid network structure is characterized in that,
acquiring a plurality of first feature layers with different sizes of a detected picture through a feature extraction network, and scaling the plurality of first feature layers to a preset resolution;
fusing the plurality of zoomed first feature layers to obtain a second feature layer, and performing global average pooling and dimension reduction operations on the second feature layer to obtain a third feature layer;
obtaining a selected fourth characteristic layer according to the weights of the third characteristic layer and different characteristic layers;
processing the fourth characteristic layer in a residual error mode to obtain a pyramid characteristic diagram;
and performing target classification and position regression on the pyramid feature map to output a detection frame.
2. The method of claim 1, wherein scaling the plurality of first feature layers to a preset resolution size comprises: performing amplification operation on the first characteristic layer with the resolution smaller than the preset resolution by adopting a bilinear interpolation method to achieve the preset resolution; performing reduction operation on the first characteristic layer with the resolution greater than the preset resolution by adopting a maximum pooling method to achieve the preset resolution; wherein the preset resolution is centered in resolution among the plurality of first feature layers.
3. The method of claim 1, wherein performing global average pooling and dimensionality reduction operations on the second feature layer to obtain a third feature layer comprises: and calculating the statistical information of the whole channel domain of the second characteristic layer by adopting global average pooling, and calculating the dependency relationship of the channel domain by adopting the full connection layer by utilizing the statistical information to obtain a third characteristic layer.
4. The method of claim 1, wherein obtaining the selected fourth feature layer based on the weights of the third feature layer and the different feature layers comprises: amplifying the third characteristic layer to n x d dimension by 1 x 1 convolution, wherein n is the number of the adopted characteristic pyramid layers, and d is the dimension of each characteristic layer; and the weight of each channel is obtained through softmax operation, the weight is multiplied by the third feature layer to obtain the selected feature layer, and the feature layers are added pixel by pixel to obtain a finally selected fourth feature map.
5. The method of claim 1, wherein the pyramid feature map is obtained by scaling the fourth feature layer to a resolution corresponding to the plurality of first feature layers by an inverse operation of rescaling and adding to the corresponding first feature layers.
6. An object detection device based on an improved characteristic pyramid network structure is characterized in that,
the zooming module is used for acquiring a plurality of first feature layers with different sizes of the detected picture through a feature extraction network and zooming the first feature layers to a preset resolution;
the fusion module is used for fusing the plurality of zoomed first feature layers to obtain a second feature layer, and performing global average pooling and dimension reduction operation on the second feature layer to obtain a third feature layer;
the selection module is used for obtaining a selected fourth characteristic layer according to the weights of the third characteristic layer and different characteristic layers;
the residual error module is used for processing the fourth characteristic layer in a residual error mode to obtain a pyramid characteristic diagram;
and the position regression module is used for carrying out target classification and position regression on the pyramid feature map so as to output a detection frame.
7. The apparatus of claim 6, wherein the scaling module is further to: performing amplification operation on the first characteristic layer with the resolution smaller than the preset resolution by adopting a bilinear interpolation method to achieve the preset resolution; performing reduction operation on the first characteristic layer with the resolution greater than the preset resolution by adopting a maximum pooling method to achieve the preset resolution; wherein the preset resolution is centered in resolution among the plurality of first feature layers.
8. The apparatus of claim 6, wherein the fusion module is further to: and calculating the statistical information of the whole channel domain of the second characteristic layer by adopting global average pooling, and calculating the dependency relationship of the channel domain by adopting the full connection layer by utilizing the statistical information to obtain a third characteristic layer.
9. The apparatus of claim 6, wherein the selection module is further to: amplifying the third characteristic layer to n x d dimension by 1 x 1 convolution, wherein n is the number of the adopted characteristic pyramid layers, and d is the dimension of each characteristic layer; and the weight of each channel is obtained through softmax operation, the weight is multiplied by the third feature layer to obtain the selected feature layer, and the feature layers are added pixel by pixel to obtain a finally selected fourth feature map.
10. The apparatus of claim 6, wherein the residual module is further configured to scale a fourth feature layer to a resolution corresponding to the plurality of first feature layers by an inverse of the rescaling operation and to add the fourth feature layer to the corresponding first feature layer to obtain a pyramid feature map.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010825554.0A CN111985503B (en) | 2020-08-17 | 2020-08-17 | Target detection method and device based on improved feature pyramid network structure |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010825554.0A CN111985503B (en) | 2020-08-17 | 2020-08-17 | Target detection method and device based on improved feature pyramid network structure |
Publications (2)
Publication Number | Publication Date |
---|---|
CN111985503A true CN111985503A (en) | 2020-11-24 |
CN111985503B CN111985503B (en) | 2024-04-26 |
Family
ID=73434028
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010825554.0A Active CN111985503B (en) | 2020-08-17 | 2020-08-17 | Target detection method and device based on improved feature pyramid network structure |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111985503B (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20220318541A1 (en) * | 2021-04-05 | 2022-10-06 | Microsoft Technology Licensing, Llc | Dynamic head for object detection |
CN116257038A (en) * | 2023-05-15 | 2023-06-13 | 深圳市瓴鹰智能科技有限公司 | Steering engine control and diagnosis method and device based on lightweight convolutional neural network |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106845510A (en) * | 2016-11-07 | 2017-06-13 | 中国传媒大学 | Chinese tradition visual culture Symbol Recognition based on depth level Fusion Features |
CN110633661A (en) * | 2019-08-31 | 2019-12-31 | 南京理工大学 | Semantic segmentation fused remote sensing image target detection method |
CN111210443A (en) * | 2020-01-03 | 2020-05-29 | 吉林大学 | Deformable convolution mixing task cascading semantic segmentation method based on embedding balance |
CN111242071A (en) * | 2020-01-17 | 2020-06-05 | 陕西师范大学 | Attention remote sensing image target detection method based on anchor frame |
CN111507359A (en) * | 2020-03-09 | 2020-08-07 | 杭州电子科技大学 | Self-adaptive weighting fusion method of image feature pyramid |
-
2020
- 2020-08-17 CN CN202010825554.0A patent/CN111985503B/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106845510A (en) * | 2016-11-07 | 2017-06-13 | 中国传媒大学 | Chinese tradition visual culture Symbol Recognition based on depth level Fusion Features |
CN110633661A (en) * | 2019-08-31 | 2019-12-31 | 南京理工大学 | Semantic segmentation fused remote sensing image target detection method |
CN111210443A (en) * | 2020-01-03 | 2020-05-29 | 吉林大学 | Deformable convolution mixing task cascading semantic segmentation method based on embedding balance |
CN111242071A (en) * | 2020-01-17 | 2020-06-05 | 陕西师范大学 | Attention remote sensing image target detection method based on anchor frame |
CN111507359A (en) * | 2020-03-09 | 2020-08-07 | 杭州电子科技大学 | Self-adaptive weighting fusion method of image feature pyramid |
Non-Patent Citations (1)
Title |
---|
JIANGMIAO PANG ET AL.: ""Libra R-CNN: Towards Balanced Learning for Object Detection"", 《IEEE》, pages 821 - 830 * |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20220318541A1 (en) * | 2021-04-05 | 2022-10-06 | Microsoft Technology Licensing, Llc | Dynamic head for object detection |
US11989956B2 (en) * | 2021-04-05 | 2024-05-21 | Microsoft Technology Licensing, Llc | Dynamic head for object detection |
CN116257038A (en) * | 2023-05-15 | 2023-06-13 | 深圳市瓴鹰智能科技有限公司 | Steering engine control and diagnosis method and device based on lightweight convolutional neural network |
Also Published As
Publication number | Publication date |
---|---|
CN111985503B (en) | 2024-04-26 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Amanatiadis et al. | A survey on evaluation methods for image interpolation | |
JP6075295B2 (en) | Dictionary creation device, image processing device, image processing system, dictionary creation method, image processing method, and program | |
CN110136056A (en) | The method and apparatus of image super-resolution rebuilding | |
US20200184697A1 (en) | Image Modification Using Detected Symmetry | |
CN109934773A (en) | A kind of image processing method, device, electronic equipment and computer-readable medium | |
Munoz-Mejias et al. | A low-complexity pre-processing system for restoring low-quality QR code images | |
CN103390275B (en) | The method of dynamical image joining | |
CN111985503A (en) | Target detection method and device based on improved characteristic pyramid network structure | |
CN108876716B (en) | Super-resolution reconstruction method and device | |
Patel et al. | Accelerated seam carving for image retargeting | |
US9076232B2 (en) | Apparatus and method for interpolating image, and apparatus for processing image using the same | |
JP2020098455A (en) | Object identification system, object identification method, and image identification program | |
Pan et al. | LPSRGAN: Generative adversarial networks for super-resolution of license plate image | |
Zhou et al. | Cross-scale collaborative network for single image super resolution | |
US20220020113A1 (en) | Image resizing using seam carving | |
CN113536971B (en) | Target detection method based on incremental learning | |
CN113807354B (en) | Image semantic segmentation method, device, equipment and storage medium | |
Wang et al. | A multi-scale attentive recurrent network for image dehazing | |
CN113516697A (en) | Image registration method and device, electronic equipment and computer-readable storage medium | |
WO2022247394A1 (en) | Image splicing method and apparatus, and storage medium and electronic device | |
CN114372944B (en) | Multi-mode and multi-scale fused candidate region generation method and related device | |
CN112070853A (en) | Image generation method and device | |
CN115619678A (en) | Image deformation correction method and device, computer equipment and storage medium | |
CN101989350A (en) | Image enhancement method, image enhancement device and image processing circuit | |
CN114897919B (en) | Deplabv3+ semantic segmentation method based on edge gradient guide interpolation |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |