CN111914904B - Image classification method fusing DarkNet and Capsule eNet models - Google Patents

Image classification method fusing DarkNet and Capsule eNet models Download PDF

Info

Publication number
CN111914904B
CN111914904B CN202010652781.8A CN202010652781A CN111914904B CN 111914904 B CN111914904 B CN 111914904B CN 202010652781 A CN202010652781 A CN 202010652781A CN 111914904 B CN111914904 B CN 111914904B
Authority
CN
China
Prior art keywords
capsule
darknet
network
model
layer
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202010652781.8A
Other languages
Chinese (zh)
Other versions
CN111914904A (en
Inventor
李钢
张玲
王飞龙
李晶
冯军鹏
郝中良
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Taiyuan University of Technology
Original Assignee
Taiyuan University of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Taiyuan University of Technology filed Critical Taiyuan University of Technology
Priority to CN202010652781.8A priority Critical patent/CN111914904B/en
Publication of CN111914904A publication Critical patent/CN111914904A/en
Application granted granted Critical
Publication of CN111914904B publication Critical patent/CN111914904B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Abstract

The invention discloses an image classification method fusing DarkNet and Capsule model, belongs to the technical field of image processing, and solves the problem of poor classification effect caused by unbalanced data; the technical scheme comprises the following steps: constructing a DarkNet-Capsule network fusion classification model, realizing definition of a fusion classification model loss function, inputting an image to be classified in the fusion classification model, performing forward training by using the DarkNet, and extracting a feature map of a target image; further processing the characteristic map of the target image, and completing error back transmission through loss to update parameters of the whole network; after multi-round iterative learning, classifying the images by utilizing a fusion classification model; the invention can further improve the classification accuracy when the data are unbalanced in the field of image classification, and simultaneously lays a firmer foundation for the research of machine vision.

Description

Image classification method fusing DarkNet and Capsule eNet models
Technical Field
The invention belongs to the technical field of image processing, and particularly relates to an image classification method fusing DarkNet and Capsule eNet models.
Background
Image classification is an important technology in the field of image processing, and in recent years, with the development of deep learning, image classification technology has been developed greatly.
The DarkNet model belongs to an image feature extraction model improved based on a residual error concept in a YOLO detection framework, and not only has the property of avoiding network degradation of a residual error network, but also reduces the parameter quantity of the model. However, as the amount of training data is reduced, the generalization performance of the model is deteriorated, and the classification accuracy is sharply reduced.
The CapsuleNet model proposes the problem that the convolutional neural network extracts image features by using convolutional kernels, so that the neural network cannot learn representations of the same features under all different view angles. In order to solve the problem, the model expresses image features by scalar quantity of vectors representing neurons, so that the generalization performance of the model is improved, a relatively small data volume can be used for obtaining a good classification effect, and the classification accuracy of the model is poor when the model faces a relatively complex data set.
In reality, the magnitude of the data set owned by us is often extremely inclined, and the effect of the model owned by us at present is often not ideal when the image features of a certain small category are extracted, so that the final classification effect is poor. Based on this, it is necessary to provide a model that can improve the classification result caused by data imbalance.
Disclosure of Invention
The invention overcomes the defects of the prior art, provides the image classification method fusing the DarkNet and the Capsule eNet model, and solves the problem of poor classification effect caused by unbalanced data.
In order to achieve the above object, the present invention is achieved by the following technical solutions.
An image classification method fusing DarkNet and Capsule eNet models comprises the following specific steps:
step S1: constructing a DarkNet-Capsule network fusion classification model, including the definition of the DarkNet model and the definition of the Capsule eNet model; specifically, a plurality of DarkNet blocks form a DarkNet model, and a CapsuleNet network is built after the last DarkNet block.
Step S2: realizing the definition of a fusion classification model loss function; the loss of the fusion model is composed of margin loss of the capsule network and L2 weight regularization loss of DarkNet, and the loss function formula is as follows: l ═ Tk max(0,m+-||vk||)2+λ(1-Tk)max(0,||vk||-m-)2+||wD||2
Wherein, TkIndicates whether a class exists, m+=0.9,m-0.1, λ is a hyperparameter set to 0.5, | | vkI represents the probability that a capsule unit belongs to this category, WD||2The loss function is the sum of the L2 regularization losses of each layer weight of the DarkNet network, and the loss function is the sum of the MarginLoss of CapsuleNet plus the L2 regularization losses of each layer weight of the DarkNet network.
Step S3: inputting images to be classified in the fusion classification model, performing forward training by using DarkNet, and extracting a feature map of a target image;
step S4: further processing the characteristic map of the target image by using a capsule network, and completing error back transmission through loss to update parameters of the whole network;
step S5: and after multiple rounds of iterative learning, the images are classified by utilizing a fusion classification model.
Preferably, the step S1 specifically includes:
step S11: building a DarkNet model, wherein the DarkNet model consists of a plurality of DarkNet blocks, all the DarkNet blocks consist of two convolutions and a residual error link, and the convolution kernels of the two convolutions are respectively 1 and 3; and an independent convolution layer is arranged between every two DarkNet blocks, the downsampling convolution is adopted, the original input is supplemented with 0, and the convolution is realized through VALID convolution with the step length of 2 and the convolution kernel of 3.
Step S12: building a capsuleNet network after the last DarkNet block, wherein the capsuleNet network consists of a main capsule network and a digital capsule network, the main capsule network changes the characteristics into a plurality of capsule units, and each capsule unit is a vector; the digital capsule layer separates each capsule unit, traverses each capsule unit, performs dynamic routing and outputs the final result of the network.
Preferably, the images to be classified in step S3 include training set, validation set, and test set images and labels for all classes, and all images are adjusted to a consistent size according to the fused classification model during input.
Preferably, the process of extracting the feature map by using the DarkNet in step S3 is as follows:
(1) inputting the image into a DarkNet network, and returning the results of the last 2 DarkNet blocks to provide richer visual field characteristics;
(2) and adjusting the return result of the small scale to the size same as that of the return result of the large scale by using a nearest neighbor interpolation method, combining the features of the return result of the small scale and the return result of the large scale, and performing convolution to extract the features.
Preferably, the step S4 specifically includes:
step S41: inputting the characteristic map processed by the DarkNet model into a main capsule layer of the capsule network, and converting the characteristic map into capsule units with each unit being an 8-dimensional vector.
Step S42: when the digital capsule layer divides the input according to capsule units, b is initialized firstly for all the capsule units i of the layer I and the capsule units j of the layer (l +1)ijWith 0, r iterations are performed with the dynamic routing algorithm:
c, calculating capsule units i of the layer Iij=soft max(bij);
② to capsule sheet with (l +1) layerThe element j calculates:
Figure BDA0002575610740000031
③ for the (l +1) layer of capsule units j:
Figure BDA0002575610740000032
fourthly, updating parameters of the layer I of capsule units and the layer (l +1) of capsule units j:
Figure BDA0002575610740000033
wherein, bijIs a temporary variable whose value is updated in the course of iteration, and when the whole algorithm is finished, its value is stored in cij(ii) a Calculating the vector cijThe value of (d), i.e. all weights of the l layers of capsules i;
Figure BDA0002575610740000034
is the output of the layer of capsule units, sjUsed for measuring the importance degree of the capsules with l layers to the (l +1) capsules; the squarh function is a nonlinear function to obtain an output capsule vjThe function ensures that the direction of the vector is preserved, while the length is limited to 1 or less, the modulo length of the vector represents the probability of belonging to this category, the greater the modulo length, the greater the probability, and the vector itself represents the information of the individual features in the image, including position, pose information.
Step S43: and solving the loss corresponding to the network output, and updating the parameters of the fusion classification model according to the solved loss.
Preferably, the step S5 specifically includes:
step S51: preprocessing data, inputting the data into a fusion classification model, and extracting image features;
step S52: after 5 rounds of iterative learning, updating the learning rate and the learning attenuation rate, and continuing iterative optimization;
step S53: and after the model accuracy rate and the loss are converged, performing an experiment by using a verification set and a test set, and storing check points.
Compared with the prior art, the invention has the beneficial effects that.
The image classification method fusing the DarkNet and the Capsule eNet model provided by the invention can combine the excellent feature extraction capability of the DarkNet aiming at complex images and the advantages of the Capsule eNet on the same deformation and good generalization of the feature extraction, make up for the shortages, and solve the problem of poor classification effect caused by unbalanced data from the network model level. The model can improve the classification accuracy of the model when data are unbalanced, and the learning capability of the model is improved, so that the model can be quickly converged. The method can further improve the classification accuracy when the data are unbalanced in the field of image classification, and simultaneously lays a firmer foundation for the research of machine vision.
Drawings
FIG. 1 is a flow chart of image classification fusing DarkNet and Capsule eNet models, implemented in accordance with the present invention.
FIG. 2 is a block diagram of DarkNet in a fused DarkNet and Capsule eNet model, implemented in accordance with the present invention.
Detailed Description
In order to make the technical problems, technical solutions and advantageous effects to be solved by the present invention more clearly apparent, the present invention is further described in detail with reference to the embodiments and the accompanying drawings. It should be understood that the specific embodiments described herein are merely illustrative of the invention and are not intended to limit the invention. The technical solution of the present invention is described in detail below with reference to the embodiments and the drawings, but the scope of protection is not limited thereto.
As shown in fig. 1-2, an image classification method fusing a DarkNet and a CapsuleNet model specifically includes the following steps:
step S1: processing image data to be loaded, wherein the image to be loaded comprises the following steps: training sets, validation sets, and test set images and labels for all classes. And all the images are adjusted to be consistent in size according to the fusion classification model, so that the model can be conveniently processed.
Step S2: and the definition for realizing the DarkNet-Capsule network fusion classification model comprises the definition of the DarkNet model and the definition of the Capsule eNet model.
Step S21: building a DarkNet model, wherein the DarkNet model consists of a plurality of DarkNet blocks, and all the DarkNet blocks consist of two convolutions (convolution kernels are 1 and 3 respectively) and a residual error link; and an independent convolution layer is arranged between every two DarkNet blocks, the downsampling convolution is adopted, the original input is supplemented with 0, and the convolution is realized through VALID convolution with the step length of 2 and the convolution kernel of 3.
Step S22: and building a capsuleNet network after the last DarkNet block, wherein the capsuleNet network consists of a main capsule network and a digital capsule network. The main capsule network changes the characteristics into a plurality of capsule units, and each capsule unit is a vector; the digital capsule layer separates each capsule unit, traverses each capsule unit, performs dynamic routing and outputs the final result of the network.
Step S3: realizing the definition of a fusion model loss function; the loss of the fusion model is composed of margin loss of the capsule network plus L2 weight regularization loss of DarkNet;
L=Tk max(0,m+-||vk||)2+λ(1-Tk)max(0,||vk||-m-)2+||wD||2
wherein, TkIndicates whether a class exists, m+=0.9,m-λ is a hyperparameter, typically set to 0.5, | | v |, 0.1kI represents the probability that the capsule unit belongs to this category, WD||2The sum of the L2 regularization losses for each layer weight of the DarkNet network. The loss function is the sum of the MarginLoss of CapsuleNet plus the L2 regularization loss of each layer weight of the DarkNet network.
Step S4: inputting images to be classified in the fusion classification model, performing forward training by using DarkNet, and extracting a feature map of a target image; the method specifically comprises the following steps:
step S41: inputting the image into a DarkNet network, and returning the results of the last 2 DarkNet blocks to provide richer visual field characteristics;
step S42: adjusting the return result of the small scale to the size same as the return result of the large scale by using a nearest neighbor interpolation method, combining the characteristics of the return result of the small scale and the return result of the large scale, and then performing convolution to extract the characteristics;
step S5: utilizing a capsule network to further process the characteristic map of the target image, and completing error back transmission through loss to update the parameters of the whole network;
step S51: inputting the characteristic map processed by the DarkNet model into a main capsule layer of the capsule network, and changing the characteristic map into capsule units with each unit being an 8-dimensional vector;
step S52: when the digital capsule layer divides the input according to capsule units, b is initialized firstly for all the capsule units i of the layer I and the capsule units j of the layer (l +1)ijWhen the value is equal to 0, carrying out r iterations by using a dynamic routing algorithm;
c is calculated for the capsule unit i of the layer lij=soft max(bij);
For capsule unit j of (l +1) layer, calculate:
Figure BDA0002575610740000051
for capsule unit j of (l +1) layer, calculate:
Figure BDA0002575610740000052
updating parameters for the layer i of capsule units and the layer (l +1) of capsule units j:
Figure BDA0002575610740000053
wherein, bijIs a temporary variable whose value is updated in the course of iteration, and when the whole algorithm is finished, its value is stored in cij(ii) a Calculating vector cijThe value of (i), i.e. all weights of the l layers of capsules i;
Figure BDA0002575610740000054
is the output of the layer of capsule units, sjUsed for measuring the importance degree of the capsules with l layers to the (l +1) capsules; the squarh function is a nonlinear function to obtain an output capsule vjThe function ensures that the direction of the vector is preserved, while the lengthConstrained to 1 or less, the modulo length of the vector represents the probability of belonging to this category, the larger the modulo length, the greater the probability, and the vector itself may represent information of various features in the image (e.g., position, pose, etc.).
And calculating the loss corresponding to the network output, and updating the parameters of the previous network according to the calculated loss.
Step S6: after multi-round iterative learning, classifying the images by using a fusion model;
step S61: preprocessing data, inputting the data into a fusion network model, and extracting image characteristics;
after 5 rounds of iterative learning, updating the learning rate and the learning attenuation rate, and continuing iterative optimization;
and after the model accuracy rate and the loss are converged, performing an experiment by using a verification set and a test set, and storing check points.
The DarkNet network in the fusion model has the characteristics of avoiding network degradation, relatively small network parameters and strong characteristic extraction capability for complex data sets; the Capsule network is improved aiming at the convolutional network, so that the network can obtain a better effect on fewer data sets and has stronger generalization; the advantages of the two can be complemented, and the DarkNet can make up the problem that the Capsule eNet has poor performance in the complex data set aiming at the strong feature extraction capability of the complex data set; the Capsule eNet uses the vector to represent the scalar quantity of the neuron, makes up the defect of convolution, converts the characteristic map extracted by DarkNet into the vector and improves the generalization performance. Therefore, the DarkNet and the Capsule eNet are fused into the same deep learning model, the defects of the DarkNet and the Capsule eNet are overcome, the new model still shows stronger classification capability when the image categories are unbalanced, and the new model inherits the advantages of the Capsule eNet, so that the robustness of the model is stronger.
While the invention has been described in further detail with reference to specific preferred embodiments thereof, it will be understood by those skilled in the art that various changes in form and details may be made therein without departing from the spirit and scope of the invention as defined by the appended claims.

Claims (6)

1. An image classification method fusing DarkNet and Capsule eNet models is characterized by comprising the following specific steps:
step S1: constructing a DarkNet-Capsule network fusion classification model, including the definition of the DarkNet model and the definition of the Capsule model; specifically, a plurality of DarkNet blocks form a DarkNet model, and a CapsuleNet network is built after the last DarkNet block;
step S2: realizing the definition of a fusion classification model loss function; the loss of the fusion model is composed of the marginloss of the capsule network plus the L2 weight regularization loss of DarkNet, and the loss function formula is as follows:
L=Tkmax(0,m+-||vk||)2+λ(1-Tk)max(0,||vk||-m-)2+||wD||2
wherein, TkIndicates whether a class exists, m+=0.9,m-0.1, λ is a hyperparameter set to 0.5, | | v |, andki represents the probability that a capsule unit belongs to this category, WD||2The sum of L2 regularization loss of each layer weight of the DarkNet network, and the loss function is the sum of MarginLoss of CapsuleNet and L2 regularization loss of each layer weight of the DarkNet network;
step S3: inputting images to be classified in the fusion classification model, performing forward training by using DarkNet, and extracting a feature map of a target image;
step S4: further processing the characteristic map of the target image by using a capsule network, and completing error back transmission through loss to update parameters of the whole network;
step S5: and after multiple rounds of iterative learning, the images are classified by utilizing a fusion classification model.
2. The image classification method fusing the DarkNet and the Capsule Net models according to claim 1, wherein the step S1 specifically comprises:
step S11: building a DarkNet model, wherein the DarkNet model consists of a plurality of DarkNet blocks, all the DarkNet blocks consist of two convolutions and a residual error link, and the convolution kernels of the two convolutions are respectively 1 and 3; and an independent convolution layer is arranged between every two DarkNet blocks, downsampling convolution is adopted, the original input is supplemented with 0, and then the convolution is realized through VALID convolution with the step length of 2 and the convolution kernel of 3;
step S12: building a capsuleNet network after the last DarkNet block, wherein the capsuleNet network consists of a main capsule network and a digital capsule network, the main capsule network changes the characteristics into a plurality of capsule units, and each capsule unit is a vector; the digital capsule layer separates each capsule unit, traverses each capsule unit, performs dynamic routing and outputs the final result of the network.
3. The method of claim 1, wherein the images to be classified in step S3 include training set, verification set and test set images and labels of all classes, and all images are adjusted to a consistent size according to the fused classification model during the input process.
4. The image classification method based on the DarkNet and Capsule Net models as claimed in claim 1, wherein the process of extracting the feature map by using DarkNet in step S3 is as follows:
(1) inputting the image into a DarkNet network, and returning the results of the last 2 DarkNet blocks to provide richer visual field characteristics;
(2) and adjusting the return result of the small scale to the size same as that of the return result of the large scale by using a nearest neighbor interpolation method, combining the features of the return result of the small scale and the return result of the large scale, and performing convolution to extract the features.
5. The image classification method fusing the DarkNet and the Capsule Net models according to claim 1, wherein the step S4 specifically comprises:
step S41: inputting the characteristic spectrum processed by the DarkNet model into a main capsule layer of a capsule network, and converting the characteristic spectrum into capsule units with each unit being an 8-dimensional vector;
step S42: when the digital capsule layer divides the input according to capsule units, b is initialized firstly for all the capsule units i of the layer I and the capsule units j of the layer (l +1)ijWith 0, r iterations are performed with the dynamic routing algorithm:
c, calculating capsule units i of the layer Iij=soft max(bij);
Calculating the capsule unit j of the (l +1) layer:
Figure FDA0002575610730000021
③ for the (l +1) layer of capsule units j:
Figure FDA0002575610730000022
fourthly, updating parameters of the layer I of capsule units and the layer (l +1) of capsule units j:
Figure FDA0002575610730000023
wherein, bijIs a temporary variable whose value is updated in the course of iteration, and when the whole algorithm is finished, its value is stored in cij(ii) a Calculating vector cijThe value of (i), i.e. all weights of the l layers of capsules i;
Figure FDA0002575610730000024
is the output of the layer of capsule units, sjUsed for measuring the importance degree of the capsules with l layers to the (l +1) capsules; the squarh function is a nonlinear function to obtain an output capsule vjThe function ensures that the direction of the vector is preserved and the length is limited to 1 or less, the modulo length of the vector represents the probability of belonging to this category, the greater the modulo length, the greater the probability, and the vector itself represents the information of the individual features in the image, including position, pose information;
step S43: and solving the loss corresponding to the network output, and updating the parameters of the fusion classification model according to the solved loss.
6. The method for classifying images by fusing DarkNet and Capsule Net models according to claim 1, wherein the step S5 is executed
Step S51: preprocessing data, inputting the data into a fusion classification model, and extracting image features;
step S52: after 5 rounds of iterative learning, updating the learning rate and the learning attenuation rate, and continuing iterative optimization;
step S53: and after the model accuracy rate and the loss are converged, performing an experiment by using a verification set and a test set, and storing check points.
CN202010652781.8A 2020-07-08 2020-07-08 Image classification method fusing DarkNet and Capsule eNet models Active CN111914904B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010652781.8A CN111914904B (en) 2020-07-08 2020-07-08 Image classification method fusing DarkNet and Capsule eNet models

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010652781.8A CN111914904B (en) 2020-07-08 2020-07-08 Image classification method fusing DarkNet and Capsule eNet models

Publications (2)

Publication Number Publication Date
CN111914904A CN111914904A (en) 2020-11-10
CN111914904B true CN111914904B (en) 2022-07-01

Family

ID=73227625

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010652781.8A Active CN111914904B (en) 2020-07-08 2020-07-08 Image classification method fusing DarkNet and Capsule eNet models

Country Status (1)

Country Link
CN (1) CN111914904B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116188792B (en) * 2023-02-23 2023-10-20 四川大学 Quantitative analysis method and system for whole blood cell scatter diagram

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108805223A (en) * 2018-05-18 2018-11-13 厦门大学 A kind of recognition methods of seal character text and system based on Incep-CapsNet networks
CN108830826A (en) * 2018-04-28 2018-11-16 四川大学 A kind of system and method detecting Lung neoplasm
CN109816024A (en) * 2019-01-29 2019-05-28 电子科技大学 A kind of real-time automobile logo detection method based on multi-scale feature fusion and DCNN
CN110232406A (en) * 2019-05-28 2019-09-13 厦门大学 A kind of liquid crystal display panel CF image identification method based on statistical learning
AU2019101142A4 (en) * 2019-09-30 2019-10-31 Dong, Qirui MR A pedestrian detection method with lightweight backbone based on yolov3 network
CN110533004A (en) * 2019-09-07 2019-12-03 哈尔滨理工大学 A kind of complex scene face identification system based on deep learning
CN111191660A (en) * 2019-12-30 2020-05-22 浙江工业大学 Rectal cancer pathology image classification method based on multi-channel collaborative capsule network
CN111292305A (en) * 2020-01-22 2020-06-16 重庆大学 Improved YOLO-V3 metal processing surface defect detection method

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11010902B2 (en) * 2018-06-04 2021-05-18 University Of Central Florida Research Foundation, Inc. Capsules for image analysis
US11886989B2 (en) * 2018-09-10 2024-01-30 International Business Machines Corporation System for measuring information leakage of deep learning models
US11100352B2 (en) * 2018-10-16 2021-08-24 Samsung Electronics Co., Ltd. Convolutional neural network for object detection

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108830826A (en) * 2018-04-28 2018-11-16 四川大学 A kind of system and method detecting Lung neoplasm
CN108805223A (en) * 2018-05-18 2018-11-13 厦门大学 A kind of recognition methods of seal character text and system based on Incep-CapsNet networks
CN109816024A (en) * 2019-01-29 2019-05-28 电子科技大学 A kind of real-time automobile logo detection method based on multi-scale feature fusion and DCNN
CN110232406A (en) * 2019-05-28 2019-09-13 厦门大学 A kind of liquid crystal display panel CF image identification method based on statistical learning
CN110533004A (en) * 2019-09-07 2019-12-03 哈尔滨理工大学 A kind of complex scene face identification system based on deep learning
AU2019101142A4 (en) * 2019-09-30 2019-10-31 Dong, Qirui MR A pedestrian detection method with lightweight backbone based on yolov3 network
CN111191660A (en) * 2019-12-30 2020-05-22 浙江工业大学 Rectal cancer pathology image classification method based on multi-channel collaborative capsule network
CN111292305A (en) * 2020-01-22 2020-06-16 重庆大学 Improved YOLO-V3 metal processing surface defect detection method

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
Dynamic Routing Between Capsules;Sara Sabour等;《arXiv》;20171107;1-11 *
YOLOv3: An Incremental Improvement;Joseph Redmon等;《arXiv》;20180408;1-6 *
基于胶囊网络的甲状腺结节超声图像分类研究;刘凯;《中国优秀博硕士学位论文全文数据库(硕士) 医药卫生科技辑》;20200615(第(2020)06期);E060-70 *
改进胶囊网络的小样本图像分类算法;王飞龙 等;《计算机科学与探索》;20210426;1-9 *

Also Published As

Publication number Publication date
CN111914904A (en) 2020-11-10

Similar Documents

Publication Publication Date Title
CN110188685B (en) Target counting method and system based on double-attention multi-scale cascade network
CN111666836B (en) High-resolution remote sensing image target detection method of M-F-Y type light convolutional neural network
CN111489358B (en) Three-dimensional point cloud semantic segmentation method based on deep learning
CN107316066B (en) Image classification method and system based on multi-channel convolutional neural network
CN109685152B (en) Image target detection method based on DC-SPP-YOLO
CN107977932B (en) Face image super-resolution reconstruction method based on discriminable attribute constraint generation countermeasure network
CN111652321B (en) Marine ship detection method based on improved YOLOV3 algorithm
CN108230278B (en) Image raindrop removing method based on generation countermeasure network
CN108681752B (en) Image scene labeling method based on deep learning
CN107103285B (en) Face depth prediction method based on convolutional neural network
CN110287777B (en) Golden monkey body segmentation algorithm in natural scene
CN110309835B (en) Image local feature extraction method and device
CN110223234A (en) Depth residual error network image super resolution ratio reconstruction method based on cascade shrinkage expansion
Zhang et al. Automatic velocity picking based on deep learning
CN112801169A (en) Camouflage target detection method based on improved YOLO algorithm
CN109711401A (en) A kind of Method for text detection in natural scene image based on Faster Rcnn
CN115222946B (en) Single-stage instance image segmentation method and device and computer equipment
CN109948575A (en) Eyeball dividing method in ultrasound image
CN114004333A (en) Oversampling method for generating countermeasure network based on multiple false classes
CN114283320A (en) Target detection method based on full convolution and without branch structure
CN113205103A (en) Lightweight tattoo detection method
CN111914904B (en) Image classification method fusing DarkNet and Capsule eNet models
CN111382684B (en) Angle robust personalized facial expression recognition method based on antagonistic learning
CN108470209B (en) Convolutional neural network visualization method based on gram matrix regularization
CN113869503B (en) Data processing method and storage medium based on depth matrix decomposition completion

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant