CN108416270B - Traffic sign identification method based on multi-attribute combined characteristics - Google Patents

Traffic sign identification method based on multi-attribute combined characteristics Download PDF

Info

Publication number
CN108416270B
CN108416270B CN201810117900.2A CN201810117900A CN108416270B CN 108416270 B CN108416270 B CN 108416270B CN 201810117900 A CN201810117900 A CN 201810117900A CN 108416270 B CN108416270 B CN 108416270B
Authority
CN
China
Prior art keywords
layer
size
convolutional
feature
layers
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201810117900.2A
Other languages
Chinese (zh)
Other versions
CN108416270A (en
Inventor
孙伟
杜宏吉
张小瑞
赵玉舟
施顺顺
杨翠芳
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
NANJING LOONG SHIELD INTELLIGENT TECHNOLOGY CO.,LTD.
Original Assignee
Nanjing University of Information Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanjing University of Information Science and Technology filed Critical Nanjing University of Information Science and Technology
Priority to CN201810117900.2A priority Critical patent/CN108416270B/en
Publication of CN108416270A publication Critical patent/CN108416270A/en
Application granted granted Critical
Publication of CN108416270B publication Critical patent/CN108416270B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/56Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
    • G06V20/58Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads
    • G06V20/582Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads of traffic signs
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • G06F18/2155Generating training patterns; Bootstrap methods, e.g. bagging or boosting characterised by the incorporation of unlabelled data, e.g. multiple instance learning [MIL], semi-supervised techniques using expectation-maximisation [EM] or naïve labelling
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks

Abstract

The invention relates to a traffic sign recognition method based on multi-attribute combined characteristics, which comprises the steps of preprocessing an image, designing a convolutional neural network structure, automatically extracting traffic sign characteristics through a CNN network, extracting final 3-layer characteristic graphs in order to fully utilize useful information of CNN multi-layer characteristics, forming a characteristic matrix of 3 scales by performing multi-scale pooling operation on each extracted layer of characteristic graph, and expanding and cascading the characteristic matrices of 3 scales into column vectors according to columns; then, cascading the obtained 3 column vectors into a combined feature vector with multiple scales and multiple attributes; and finally, classifying the combined feature vectors through an ELM classifier, thereby efficiently finishing the identification and classification of the traffic signs.

Description

Traffic sign identification method based on multi-attribute combined characteristics
Technical Field
The invention relates to a traffic sign identification method based on multi-attribute combined characteristics, and belongs to the field of traffic sign identification in intelligent traffic systems.
Background
In recent years, traffic sign recognition is widely applied to aspects such as driver assistance systems, unmanned intelligent automobiles, highway maintenance and the like, and the traditional traffic sign recognition method is difficult to meet the requirements of high accuracy and real-time performance.
In recent two years, traffic sign recognition methods based on deep learning become popular to research, for example, Convolutional Neural Networks (CNNs) have been successfully applied to traffic sign recognition systems, but it is common to use the last layer of features of CNNs for classifier training, and the features may not contain enough useful information to realize the classification of traffic signs. Therefore, if the features extracted by the multilayer network can be fully utilized for classifier training, the recognition accuracy can be improved, and the training time and the computation amount can be reduced to a certain extent.
Most of the current CNN researches focus on improving the classification precision, the learning speed is neglected, and the learning speed is just required to be improved for the traffic sign identification to ensure the real-time requirement. Meanwhile, research shows that the generalization capability of the CNN has certain limitation, and an Extreme Learning Machine (ELM) has good generalization performance. ELM belongs to a single hidden layer feedforward neural network, and the learning speed is faster than other traditional learning algorithms under the condition of ensuring the learning precision.
Disclosure of Invention
The technical problem to be solved by the invention patent is as follows: aiming at the problems and the defects in the background technology, the traffic sign identification method based on the multi-attribute combined characteristics is provided, the traffic sign can be accurately identified, effective information can be extracted, and technical support is provided for the aspects of unmanned intelligent automobiles, driver assistance systems, road maintenance and the like.
The technical scheme provided by the invention is as follows: a traffic sign identification method based on multi-attribute combined features comprises the following steps:
step 1, preprocessing a traffic sign image and normalizing image data;
step 2, designing a convolutional neural network structure;
step 3, training the convolutional neural network in the step 2, and extracting the preprocessed traffic sign image features by using the network;
step 4, extracting the feature maps of the last three layers of the convolutional neural network, forming feature matrixes of three scales by multi-scale pooling operation on the extracted feature maps of the last three layers, expanding and cascading the feature matrixes of the three scales according to columns to form column vectors, and then cascading the obtained three column vectors to form a combined feature vector with multi-scale and multi-attribute;
step 5, designing an ELM classifier model;
and 6, classifying the combined feature vectors through an ELM classifier.
The technical scheme is further designed as follows: the convolutional neural network structure in the step 2 is composed of 8 layers, and comprises 1 input layer, 1 full-connection layer, 3 convolutional layers and 3 pooling layers, wherein the convolutional layers and the pooling layers are arranged in a cross mode.
The number of iterations in training the convolutional neural network in step 3 is set to 30.
The ELM classifier model is as follows:
Figure BDA0001571197470000021
the labeled feature vector input by the ELM classifier is represented as (f)i,ti) N, N represents the number of training samples; wherein f isi=[fi1,fi2,...,fin]T∈RnRepresenting the joint feature vector of the ith sample, wherein n is the number of input neurons of the ELM model; t is ti=[ti1,ti2,...,tim]T∈RmRepresenting a label vector of the ith sample, wherein m is the number of output neurons of the ELM model; j 1, 2.. M, M denotes the number of hidden layer neurons, βjRepresenting a weight value connecting the jth hidden node and the output node; w is aijRepresenting a weight vector connecting the ith sample and the jth hidden node; bjRepresents the bias of the jth hidden node; oiAn output vector representing the ith sample; t is tiA label vector representing the ith sample; g (-) represents the activation function.
The invention has the beneficial effects that:
(1) the invention adopts the convolutional neural network to extract the features and classify the extreme learning machine, fully combines the advantages of multilayer feature extraction of the convolutional neural network and the advantages of high generalization performance and learning speed of the extreme learning machine, and the like, not only can improve the identification accuracy of the traffic sign, but also can ensure the real-time requirement of the traffic sign identification.
(2) The invention adopts the convolutional neural network to extract the multilayer traffic sign characteristic diagram, the multilayer characteristic contains more detailed information, the structure is more stable, the useful information of the multilayer characteristic is fully utilized, and the identification accuracy of the traffic sign can be effectively improved.
(3) The method adopts multi-scale pooling to sample the extracted characteristic graph, so that the convolutional neural network can process input with any size, output with the same dimensionality can be obtained aiming at input with different sizes, and the problem that an input image in the traditional convolutional neural network needs to be fixed in size is solved; meanwhile, the invariance of the characteristic extracted by the convolutional neural network can be enhanced by performing multi-scale pooling operation on the characteristic graph after the convolutional operation, so that the accuracy and the robustness of target identification are improved.
(4) Compared with the existing traffic sign identification technology, the method only needs simple normalization processing in the preprocessing process, and greatly simplifies the preprocessing flow which is complicated in the previous period.
Drawings
FIG. 1 is a flow chart of the identification method of the present invention.
Fig. 2 is a schematic structural diagram of CNN extracting features of the last three layers.
FIG. 3 is a schematic diagram of a multi-scale pooling operation.
FIG. 4 is a diagram of an ELM classifier model.
Detailed Description
The traffic sign identification method based on the multi-attribute combined characteristic is further explained with reference to the attached drawings.
As shown in fig. 1, the traffic sign recognition method of the present invention includes the following steps:
step 1: preprocessing a database image;
the traffic sign image adopted in this embodiment mainly comes from the GTSRB data set and the traffic sign image in the natural scene shot by the intelligent camera, and total 5000 training samples and 1000 test samples contain 43 types of traffic signs.
First, the picture size is uniformly adjusted to 48 × 48 pixels, and normalization processing is performed. The training samples need to carry labels, and the labeled training samples are denoted as (x)i,ti) N, N represents the number of training samples;xifeature vector representing the ith sample, tiThe label vector representing the ith sample.
Step 2: designing a CNN network structure;
the CNN network structure designed by the patent comprises 8 layers, including 1 input layer, 1 full-connection layer, 3 convolutional layers (C1, C3, C5) and 3 pooling layers (P2, P4, P6), wherein the convolutional layers and the pooling layers are arranged in a cross mode. The input layer is a traffic sign image sample with the pixel value size of 48 multiplied by 48, and input is divided into three color dimensions of RGB. Convolutional layer C1 has 100 feature maps of size 46 × 46, convolutional kernel size 3 × 3, and convolutional step size 1; pooling layer P2 has 100 feature maps of 23 × 23, pooled kernel size of 2 × 2, and stride of 2; convolutional layer C3 has 150 feature maps of size 20 × 20, convolutional kernel size 4 × 4, and stride 1; pooling layer P4 has 150 feature maps of size 10 × 10, pooled kernel size 2 × 2, and stride 2; convolutional layer C5 has 250 feature maps of size 8 × 8, convolutional kernel size 3 × 3, and stride 1; pooling layer P6 has 250 feature maps of size 4X 4, pooled kernel size 2X 2, and stride 2. Since the CNN network is only used for feature extraction and not for classification, the final fully-connected layer is equivalent to a common forward neural network (SLFN) classifier, having a total of 43 neurons, representing 43 different traffic sign classes.
The activation function of each neuron in this patent uses a hyperbolic tangent function:
f(x)=tanh(x)=(ex-e-x)/(ex+e-x)。
and step 3: training a CNN network, and automatically extracting the traffic sign image characteristics of each level of a sample;
the embodiment trains the CNN network designed in step 2 by using an online gradient descent method. The training iteration number is set to 30 in the embodiment, overfitting can occur when the iteration number is too high, and underfitting can occur when the iteration number is too low. The network initial weight is initialized within the range of [ -0.01,0.01 ].
And 4, step 4: extracting feature maps of the last three layers of the CNN network, and combining the feature maps to form a multi-scale multi-attribute combined feature vector by using multi-scale pooling operation;
the traditional CNN usually uses the last layer of characteristics of the network for identification and classification, can not completely show the detailed characteristics of the traffic sign, and can more comprehensively and more abundantly show the multi-attribute characteristics of the traffic sign by combining the multi-layer characteristics. Firstly, extracting a characteristic diagram of the last three layers (P4, C5 and P6) during feedforward training; then, using multi-scale pooling operation to form 3 feature matrices, finally expanding and cascading the 3 feature matrices by columns into a column vector with multi-scale and multi-features, and the structural diagram of CNN extracting the features of the last three layers is shown in fig. 2.
The multi-scale pooling adopts a plurality of sampling sizes and sampling step sizes, and no matter how large the size of the extracted feature map is, the multi-scale pooling outputs feature matrixes of 3 different scales, namely 1 × 1 × a, 2 × 2 × a and 3 × 3 × a, wherein a represents the number of the extracted feature maps, and the multi-scale pooling operation is schematically shown in fig. 3.
In this embodiment, the P4 layers are subjected to multi-scale pooling to obtain 3 feature matrices with sizes of 1 × 150, 4 × 150, and 9 × 150, and are sequentially cascaded to obtain a fixed-size 14 × 150-2100 × 1 feature column vector; obtaining 3 characteristic matrixes with the sizes of 1 × 250, 4 × 250 and 9 × 250 by using the C5 layer after multi-scale pooling, and sequentially cascading to obtain a 14 × 250 ═ 3500 × 1 characteristic column vector with a fixed size; the P6 layers are subjected to multi-scale pooling to obtain 3 feature matrices with the sizes of 1 × 250, 4 × 250 and 9 × 250, and the feature matrices are sequentially cascaded to obtain a 14 × 250 ═ 3500 × 1 feature column vector with a fixed size.
And 5: designing an ELM classifier;
the extracted feature maps of the P4, C5 and P6 layers are cascaded by 3 column vectors formed after multi-scale pooling operation to form a combined feature vector f with multi-scale and multi-attributesiAnd serves as an input of the ELM model. The labeled feature vector input by the ELM classifier is represented as (f)i,ti) N, N represents the number of training samples. Wherein f isi=[fi1,fi2,...,fin]T∈RnRepresenting the joint feature vector of the ith sample, wherein n is the number of input neurons of the ELM model; t is ti=[ti1,ti2,...,tim]T∈RmAnd m is the number of output neurons of the ELM model. The ELM classifier model is shown in FIG. 4.
The ELM model is expressed as:
Figure BDA0001571197470000041
wherein M represents the number of hidden layer neurons, j ═ 1, 2.., M; beta is ajRepresenting a weight value connecting the jth hidden node and the output node; w is aijRepresenting a weight vector connecting the ith sample and the jth hidden node; bjRepresents the bias of the jth hidden node; oiAn output vector representing the ith sample; t is tiA label vector representing the ith sample; g (-) represents the activation function.
Joint feature vector { f) of all training samples is inputiH, with yiRepresenting the actual output vector, equation (1) is further simplified to:
Hw,b,fβ=Y (2)
wherein beta is the output weight between the hidden layer and the output neuron,
Figure BDA0001571197470000042
h is the output matrix of hidden layer neurons,
Figure BDA0001571197470000043
the training goal is to minimize the training error | | | T-H β | | | Y2And the norm of the output weight | | | β | | |, so the constraint optimization formula in the training process is:
Figure BDA0001571197470000051
wherein, the constant C is a cost parameter and represents a regularization factor, and C is 2000; ξ represents the error tolerance parameter introduced to ensure that the ELM model fits all the training samples. Solving equation (3) by using a Lagrange multiplier method to obtain:
Figure BDA0001571197470000052
step 6: training an ELM classifier;
the method comprises the following specific steps:
6.1 input Joint features of training samples (f)i,ti) (ii) a Using sigmoid function as activation function in the form of
Figure BDA0001571197470000053
The hidden layer node M is set to 10000;
6.2, randomly generating parameters (w, b) of the hidden layer;
6.3, calculating a hidden layer output matrix H;
6.4, calculating an output weight beta according to a formula (4);
6.5 calculating the output vector o according to the formula (1)i,oiIs a binary target vector; the output neuron m represents the type of the traffic sign, and m is 43 in the patent. If the ith training sample xiBelonging to the k-th traffic sign, then oiThe k-th element in (1) and the other elements are 0.
And 7: traffic sign detection
According to the step 1, preprocessing a traffic sign image sample to be detected, extracting a characteristic diagram of P4, C5 and P6 layers through a trained CNN network after preprocessing, cascading characteristic column vectors to form a combined characteristic vector with multiple scales and multiple attributes after using multi-scale pooling operation, then identifying the traffic sign by taking the combined characteristic vector as the input of an ELM classifier, and if the output vector o is outputiIf the kth element of (1) is 1, it indicates that the sample to be detected belongs to the kth traffic sign.
The technical solutions of the present invention are not limited to the above embodiments, and all technical solutions obtained by using equivalent substitution modes fall within the scope of the present invention.

Claims (3)

1. A traffic sign identification method based on multi-attribute combined features is characterized by comprising the following steps:
step 1, preprocessing a traffic sign image;
step 2, designing a convolutional neural network structure; the convolutional neural network comprises 1 input layer, 1 full-connection layer, 3 convolutional layers C1, C3, C5 and 3 pooling layers P2, P4 and P6;
convolutional layer C1 has 100 feature maps of size 46 × 46, convolutional kernel size 3 × 3, and convolutional step size 1; pooling layer P2 has 100 feature maps of 23 × 23, pooled kernel size of 2 × 2, and stride of 2; convolutional layer C3 has 150 feature maps of size 20 × 20, convolutional kernel size 4 × 4, and stride 1; pooling layer P4 has 150 feature maps of size 10 × 10, pooled kernel size 2 × 2, and stride 2; convolutional layer C5 has 250 feature maps of size 8 × 8, convolutional kernel size 3 × 3, and stride 1; pooling layer P6 has 250 feature maps of size 4 × 4, pooled kernel size 2 × 2, and stride 2;
step 3, training the convolutional neural network in the step 2, and extracting the preprocessed traffic sign image features by using the network;
step 4, extracting the feature maps of the last three layers of the convolutional neural network, forming feature matrixes of three scales by multi-scale pooling operation on the extracted feature maps of the last three layers, expanding and cascading the feature matrixes of the three scales according to columns to form column vectors, and then cascading the obtained three column vectors to form a combined feature vector with multi-scale and multi-attribute;
the P4 layers are subjected to multi-scale pooling to obtain 3 characteristic matrixes with the sizes of 1 × 150, 4 × 150 and 9 × 150, and are sequentially cascaded to obtain a characteristic column vector of which the size is 14 × 150 which is 2100 × 1 and the fixed size; obtaining 3 characteristic matrixes with the sizes of 1 × 250, 4 × 250 and 9 × 250 by using the C5 layer after multi-scale pooling, and sequentially cascading to obtain a 14 × 250 ═ 3500 × 1 characteristic column vector with a fixed size; obtaining 3 feature matrixes with the sizes of 1 × 250, 4 × 250 and 9 × 250 by using the P6 layer after multi-scale pooling, and sequentially cascading to obtain a 14 × 250 ═ 3500 × 1 feature column vector with a fixed size;
step 5, designing an ELM classifier model;
the constraint optimization formula of the ELM classifier model in the training process is as follows:
Figure FDA0003047118470000011
st.Hβ=T-ξ
wherein, the constant C is a cost parameter and represents a regularization factor; ξ represents an error tolerant parameter to ensure that the ELM model fits all training samples;
solving the above equation by using a Lagrange multiplier method to obtain:
Figure FDA0003047118470000012
and 6, classifying the combined feature vectors through an ELM classifier.
2. The method for recognizing the traffic sign based on the multi-attribute combined characteristic as claimed in claim 1, wherein: the convolutional neural network structure in the step 2 is composed of 8 layers, and comprises 1 input layer, 1 full-connection layer, 3 convolutional layers and 3 pooling layers, wherein the convolutional layers and the pooling layers are arranged in a cross mode.
3. The method for recognizing traffic signs based on multi-attribute combined features as claimed in claim 2, wherein: the ELM classifier model is as follows:
Figure FDA0003047118470000021
the labeled feature vector input by the ELM classifier is represented as (f)i,ti) N, N represents the number of training samples; wherein f isi=[fi1,fi2,...,fin]T∈RnRepresenting the joint feature vector of the ith sample, n is an ELM modelInputting the number of neurons; t is ti=[ti1,ti2,...,tim]T∈RmRepresenting a label vector of the ith sample, wherein m is the number of output neurons of the ELM model; j 1, 2.. M, M denotes the number of hidden layer neurons, βjRepresenting a weight value connecting the jth hidden node and the output node; w is aijRepresenting a weight vector connecting the ith sample and the jth hidden node; bjRepresents the bias of the jth hidden node; oiAn output vector representing the ith sample; t is tiA label vector representing the ith sample; g (-) represents the activation function.
CN201810117900.2A 2018-02-06 2018-02-06 Traffic sign identification method based on multi-attribute combined characteristics Active CN108416270B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810117900.2A CN108416270B (en) 2018-02-06 2018-02-06 Traffic sign identification method based on multi-attribute combined characteristics

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810117900.2A CN108416270B (en) 2018-02-06 2018-02-06 Traffic sign identification method based on multi-attribute combined characteristics

Publications (2)

Publication Number Publication Date
CN108416270A CN108416270A (en) 2018-08-17
CN108416270B true CN108416270B (en) 2021-07-06

Family

ID=63127773

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810117900.2A Active CN108416270B (en) 2018-02-06 2018-02-06 Traffic sign identification method based on multi-attribute combined characteristics

Country Status (1)

Country Link
CN (1) CN108416270B (en)

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109002562B (en) * 2018-08-30 2021-04-13 北京信立方科技发展股份有限公司 Instrument recognition model training method and device and instrument recognition method and device
CN109492661B (en) * 2018-09-27 2021-08-13 桂林电子科技大学 Traffic sign identification method based on WWCNN model and TVPDE algorithm
CN109993058A (en) * 2019-02-27 2019-07-09 北京大学 The recognition methods of road signs based on multi-tag classification
CN111178153A (en) * 2019-12-09 2020-05-19 武汉光庭信息技术股份有限公司 Traffic sign detection method and system
CN111767860A (en) * 2020-06-30 2020-10-13 阳光学院 Method and terminal for realizing image recognition through convolutional neural network
CN116912518B (en) * 2023-09-12 2024-01-05 深圳须弥云图空间科技有限公司 Image multi-scale feature processing method and device

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104992165A (en) * 2015-07-24 2015-10-21 天津大学 Extreme learning machine based traffic sign recognition method
CN105956524A (en) * 2016-04-22 2016-09-21 北京智芯原动科技有限公司 Method and device for identifying traffic signs
CN107122776A (en) * 2017-04-14 2017-09-01 重庆邮电大学 A kind of road traffic sign detection and recognition methods based on convolutional neural networks
CN107301383A (en) * 2017-06-07 2017-10-27 华南理工大学 A kind of pavement marking recognition methods based on Fast R CNN

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9811775B2 (en) * 2012-12-24 2017-11-07 Google Inc. Parallelizing neural networks during training
CN104680144B (en) * 2015-03-02 2018-06-05 华为技术有限公司 Based on the lip reading recognition methods and device for projecting very fast learning machine

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104992165A (en) * 2015-07-24 2015-10-21 天津大学 Extreme learning machine based traffic sign recognition method
CN105956524A (en) * 2016-04-22 2016-09-21 北京智芯原动科技有限公司 Method and device for identifying traffic signs
CN107122776A (en) * 2017-04-14 2017-09-01 重庆邮电大学 A kind of road traffic sign detection and recognition methods based on convolutional neural networks
CN107301383A (en) * 2017-06-07 2017-10-27 华南理工大学 A kind of pavement marking recognition methods based on Fast R CNN

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
Joint Multiple Fully Connected Convolutional Neural Network with Extreme Learning Machine for Hepatocellular Carcinoma Nuclei Grading;Siqi Li etc;《Computers in Biology and Medicine》;20170321;正文第11-14页 *

Also Published As

Publication number Publication date
CN108416270A (en) 2018-08-17

Similar Documents

Publication Publication Date Title
CN108416270B (en) Traffic sign identification method based on multi-attribute combined characteristics
Lin et al. Transfer learning based traffic sign recognition using inception-v3 model
CN110298266B (en) Deep neural network target detection method based on multiscale receptive field feature fusion
CN110334705B (en) Language identification method of scene text image combining global and local information
CN109241817B (en) Crop image recognition method shot by unmanned aerial vehicle
CN112446388A (en) Multi-category vegetable seedling identification method and system based on lightweight two-stage detection model
CN110321967B (en) Image classification improvement method based on convolutional neural network
CN111191583B (en) Space target recognition system and method based on convolutional neural network
CN108985252B (en) Improved image classification method of pulse depth neural network
CN107451565B (en) Semi-supervised small sample deep learning image mode classification and identification method
CN111460980B (en) Multi-scale detection method for small-target pedestrian based on multi-semantic feature fusion
Girisha et al. Uvid-net: Enhanced semantic segmentation of uav aerial videos by embedding temporal information
CN109241995B (en) Image identification method based on improved ArcFace loss function
CN111709311A (en) Pedestrian re-identification method based on multi-scale convolution feature fusion
Hasan An application of pre-trained CNN for image classification
Roecker et al. Automatic vehicle type classification with convolutional neural networks
CN108230330B (en) Method for quickly segmenting highway pavement and positioning camera
CN112215296B (en) Infrared image recognition method based on transfer learning and storage medium
Yang et al. Local label descriptor for example based semantic image labeling
CN114241053A (en) FairMOT multi-class tracking method based on improved attention mechanism
CN114821014A (en) Multi-mode and counterstudy-based multi-task target detection and identification method and device
CN116258990A (en) Cross-modal affinity-based small sample reference video target segmentation method
CN115063832A (en) Global and local feature-based cross-modal pedestrian re-identification method for counterstudy
CN111310820A (en) Foundation meteorological cloud chart classification method based on cross validation depth CNN feature integration
Bui et al. Deep learning architectures for hard character classification

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
TR01 Transfer of patent right

Effective date of registration: 20211118

Address after: 210000 Building 1, 116 Shiyang Road, Qinhuai District, Nanjing City, Jiangsu Province

Patentee after: NANJING LOONG SHIELD INTELLIGENT TECHNOLOGY CO.,LTD.

Address before: 210044, No. 219, Ning six road, Pukou District, Jiangsu, Nanjing

Patentee before: NANJING University OF INFORMATION SCIENCE & TECHNOLOGY

TR01 Transfer of patent right