CN107301383B - Road traffic sign identification method based on Fast R-CNN - Google Patents

Road traffic sign identification method based on Fast R-CNN Download PDF

Info

Publication number
CN107301383B
CN107301383B CN201710421849.XA CN201710421849A CN107301383B CN 107301383 B CN107301383 B CN 107301383B CN 201710421849 A CN201710421849 A CN 201710421849A CN 107301383 B CN107301383 B CN 107301383B
Authority
CN
China
Prior art keywords
layer
layers
fast
training
pooling
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201710421849.XA
Other languages
Chinese (zh)
Other versions
CN107301383A (en
Inventor
刘兰馨
李巍华
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
South China University of Technology SCUT
Original Assignee
South China University of Technology SCUT
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by South China University of Technology SCUT filed Critical South China University of Technology SCUT
Priority to CN201710421849.XA priority Critical patent/CN107301383B/en
Publication of CN107301383A publication Critical patent/CN107301383A/en
Application granted granted Critical
Publication of CN107301383B publication Critical patent/CN107301383B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/56Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
    • G06V20/58Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads
    • G06V20/582Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads of traffic signs
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/07Target detection

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Multimedia (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Engineering & Computer Science (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a road traffic sign identification method based on Fast R-CNN, which comprises the following steps: carrying out image acquisition and pretreatment to manufacture a sample set; inputting a training set, and performing multitask training on a Fast R-CNN network; the picture to be identified passes through a plurality of convolution layers and pooling layers to obtain a characteristic diagram; obtaining a corresponding feature frame from the candidate frame, and obtaining two output vectors of classification score and window regression respectively through the ROI pooling layer and the full-connection layer; and (4) performing non-maximum suppression processing on all the results to generate final target detection and identification results, and identifying the traffic sign. The invention adopts a Fast R-CNN deep learning method, avoids redundant feature extraction operation in the regional convolutional neural network R-CNN, realizes multi-task training, does not need additional feature storage space, and improves the detection speed and precision. Compared with a shallow learning classifier, the method has higher learning efficiency and recognition accuracy.

Description

Road traffic sign identification method based on Fast R-CNN
Technical Field
The invention belongs to the field of image processing and automobile safety assistant driving, and particularly relates to a road traffic sign detection and identification method based on Fast R-CNN, which is used for solving the problem of low identification precision in the road traffic sign identification problem.
Background
Traffic Sign Recognition (TSR) is one of the problems that have not been solved at present, and is an important branch in a vehicle-mounted auxiliary system. Because the traffic sign contains a lot of important traffic information, such as speed prompt of current driving, change of road condition in front and restriction of driver behavior, in the auxiliary system, how to quickly, accurately and effectively identify the traffic sign in the road and feed the traffic sign back to the driver or a control system has very important research significance for ensuring driving safety and avoiding traffic accidents.
The common methods for recognizing the road traffic signs comprise a shape-based recognition method, a method combining feature extraction and a classifier and a deep learning recognition method. The shape-based recognition method has poor robustness and poor effect in a complex environment. The method combining the feature extraction and the classifier has good identification effect, but has high calculation cost and poor environment adaptability. The deep learning can directly identify the original image, extract the recessive characteristics reflecting the essence of the data and has enough learning depth. The convolutional neural network has the characteristic of local weight sharing, and has certain real-time performance and robustness for the conditions of complex environment, multi-angle change and the like. Therefore, it is necessary to design an identification method capable of accurately acquiring the road traffic sign in the road scene. A Fast R-CNN algorithm is proposed by Ross B.Girshick in 2015, redundant feature extraction operation in R-CNN (Region-based connected Neural Network) is avoided, multitask training is realized, an additional feature storage space is not needed, and detection speed and precision are improved.
Disclosure of Invention
In order to solve the problems in the prior art, the invention designs a road traffic sign identification method based on Fast R-CNN, which can accurately acquire road traffic signs in a road scene so as to help a driver to better sense the environment outside a vehicle under complex conditions and prevent traffic accidents.
In order to achieve the purpose, the technical scheme of the invention is as follows:
a road traffic sign identification method based on Fast R-CNN comprises the following steps:
carrying out image acquisition and pretreatment to manufacture a sample set;
inputting a training set, and performing multitask training on a Fast R-CNN network;
inputting the picture to be identified into a Fast R-CNN network, and obtaining a characteristic diagram through a plurality of convolution layers and pooling layers;
extracting a plurality of candidate frames by adopting a Selective Search algorithm, finding a feature frame corresponding to each candidate frame in the feature map according to the mapping relation between the candidate frames in the original image and the feature map, and pooling each feature frame to a fixed size in an ROI pooling layer;
the feature frame is processed by a full connection layer to obtain feature vectors with fixed size, and the feature vectors are processed by respective full connection layers to respectively obtain two output vectors of classification scores and window regression;
and performing non-maximum suppression processing on all the results to generate a final target detection and identification result, so that the road traffic sign can be identified.
Further, the image acquisition step specifically includes:
starting a vehicle-mounted automobile data recorder, and shooting road traffic video information in real time;
performing framing processing on video information shot by a camera to obtain an image set sequence;
and screening the image set, and selecting the images containing the road traffic signs.
Further, the image preprocessing and sample set preparation steps specifically include:
in the selected image, taking out the target area, zooming to a fixed size of 224 multiplied by 224 to enhance the contrast, then carrying out contrast enhancement processing on the target area to obtain an original training set, and processing the test set in the same way;
rotating the original training set to the degree of 12 degrees and 12 degrees, zooming to the degree of 0.4 and 1.6, and adding the original training set into the original data set to form a new training set;
and randomly taking samples with the number equivalent to that of the test set in the new data set to form a verification set, and forming the rest samples into a final training set.
Further, the Fast R-CNN network structure comprises: 13 convolutional layers, 4 pooling layers, 1 ROI pooling layer, 2 fully-connected layers and two planar layers.
Further, each feature frame is pooled to a fixed size of 7 x 7 in the ROI pooling layer.
Further, the fully-connected output of the multitask training Fast R-CNN network comprises two branches: a cls _ score layer for classification and a bbox _ pred layer for adjusting candidate box positions.
Further, when the feature vectors pass through the respective full connection layers, the feature vectors are accelerated by Singular Value Decomposition (SVD) to obtain two output vectors, namely, classification score of Softmax and Bounding-box window regression.
Further, for two branches of full-connection output, a classification layer and a regression layer of an output layer are trained by using a random gradient descent method until loss functions of classification and regression are converged.
Further, the step of performing non-maximum suppression processing on all the results specifically includes: according to the two output branches, non-maximum value inhibition is respectively carried out on each class of objects by using window scores to remove overlapping candidate frames, and finally, a window with the highest score after regression correction in each class is obtained.
Further, the road traffic sign comprises a straight arrow, a turning arrow, a left arrow, a right arrow, a straight left arrow, a straight right arrow and a diamond-shaped marked line.
Compared with the prior art, the invention provides a traffic sign detection and identification method based on Fast R-CNN in order to solve at least some problems in the prior art. The method automatically makes a road traffic sign data set, learns the characteristics from the sample through deep learning, can extract the recessive characteristics reflecting the essence of the data, has higher learning efficiency and recognition precision, improves the robustness of the detection algorithm, and effectively improves the accuracy of the detection of the road traffic sign.
Drawings
The accompanying drawings, which are included to provide a further understanding of the disclosure, are incorporated in and constitute a part of this application, illustrate only some non-limiting examples of the invention embodying the inventive concept and are not intended to be limiting in any way.
FIG. 1 is a flow diagram of a Fast R-CNN based method of pavement marking identification in accordance with some exemplary embodiments of the present invention.
FIG. 2 is a diagram of a Fast R-CNN network architecture in accordance with some exemplary embodiments of the present invention.
FIG. 3 is a diagram of a multitasking training cost function according to some demonstrative embodiments of the invention.
Fig. 4 is a schematic illustration of a portion of a sample set of traffic signs according to some exemplary embodiments of the invention.
FIG. 5 is a graphical representation of the detection results of Fast R-CNN based pavement marker identification methods according to some exemplary embodiments of the present invention.
Detailed Description
The invention is described in detail below with reference to the drawings and technical solutions.
As shown in the attached figure 1, the invention is a flow chart of a road traffic sign identification method based on Fast R-CNN, and the specific implementation mode of the invention is as follows:
a road traffic sign identification method based on Fast R-CNN comprises the following steps:
carrying out image acquisition and pretreatment to manufacture a sample set;
inputting a training set, and performing multitask training on a Fast R-CNN network;
inputting the picture to be identified into a Fast R-CNN network, and obtaining a characteristic diagram through a plurality of convolution layers and pooling layers;
extracting about 2000 candidate frames by adopting a Selective Search algorithm, finding a feature frame corresponding to each candidate frame in the feature map according to the mapping relation between the candidate frame and the feature map in the original image, and pooling each feature frame to a fixed size in an ROI (Region of Interest) pooling layer;
the feature frames are processed by full connection layers to obtain feature vectors with fixed sizes, and two output vectors of classification scores and window regression are obtained through respective full connection layers;
and (4) performing non-maximum suppression processing on all the results to generate final target detection and identification results, and identifying the traffic sign.
In some embodiments, the step of image acquisition specifically comprises:
starting a vehicle-mounted automobile data recorder, shooting road traffic video information in real time, and selecting a video image with a video resolution of 1280 x 720 shot by the automobile data recorder;
performing frame processing on the shot video image to obtain an image set sequence;
and screening the image set, and selecting 7 road traffic signs with high occurrence frequency from the image set.
Specifically, the image preprocessing and sample set preparation steps specifically include:
in the selected image, the target area is taken out and is zoomed to a fixed size of 224 multiplied by 224 to enhance the contrast, then the target area is subjected to contrast enhancement processing to form an original training set, and the test set is processed in the same way;
rotating the original training set to the degree of 12 degrees and 12 degrees, zooming to the degree of 0.4 and 1.6, and adding the original training set into the original data set to form a new training set;
and randomly taking samples with the number equivalent to that of the test set in the new data set to form a verification set, and forming the rest samples into a final training set.
In the example shown in fig. 4, the selected traffic signs can be classified into 7 types, which are respectively straight arrow, u-turn arrow, left arrow, right arrow, straight left arrow, straight right arrow, and diamond-shaped marked line, which are respectively numbered 01, 02, 03, 04, 05, 06, 07, and the recognition results are output in this way.
A VGG16 network proposed in the document "k.simony and a.zisserman.very deep conditional networks for large-scale image registration, 2015" by simony et al, comprising 13 convolutional layers, 5 pooling layers and 3 fully connected layers. On the basis of VGG16, the last layer of the VGG-16 network is replaced by an ROI pooling layer, and the last layer of the VGG-16 network and a softmax layer are replaced by two parallel layers.
As shown in FIG. 2, the resulting Fast R-CNN network structure is as follows: including 13 convolutional layers, 4 pooling layers, 1 ROI pooling layer, 2 fully-connected layers and two planar layers. Image samples of size 224 x 224 are input into the network via the input layer; for all convolutional layers, the convolutional kernel size is 3 × 3, step size is 1; for all pooling layers, a 2 × 2 sample size was used, and the pooling approach employed maximum pooling. The activation function adopts a modified Linear unit activation (ReLU) function, has the capability of guiding moderate sparsity, can accelerate the training speed of the network, improves the precision and avoids the problem of gradient disappearance.
The original layer parameters are initialized by a pre-training mode. Initializing a full-connected layer for classification by Gaussian distribution with the mean value of 0 and the standard deviation of 0.01; the fully connected layers for regression were initialized with a gaussian distribution with a mean of 0 and a standard deviation of 0.001, with the bias all initialized to 0.
During tuning training, N complete pictures are added firstly, and then R candidate frames selected from the N pictures are added. The R/N candidate frames of the same image are convolved to share calculation and memory, and the operation cost is reduced. The constitution of the R candidate frames is as follows: candidate boxes that overlap with some true value at [0.5,1] we define as foreground, accounting for 25% of the total; candidate boxes with a maximum value of [0.1,0.5) overlapping with the true value are defined as background and account for 75% of the total.
The ROI pooling layer averagely divides each feature frame according to a fixed size, performs maximum pooling on each block, can convert feature frames with different sizes on the feature map into data with uniform size, and sends the data into the next layer.
The Fast R-CNN network was multi-tasked with classification and regression losses as shown in FIG. 3. The cls _ score layer is used for classifying and outputting a K + 1-dimensional array p, wherein the K + 1-dimensional array p represents the probability of belonging to K types of objects and backgrounds, and K is set to be 7 according to the number of detection types; the bbox _ pred layer is used for adjusting the position of the candidate region, and outputs a 4 x K-dimensional array which represents parameters which should be translated and scaled when the candidate region belongs to the K classes respectively, and a separate regressor is trained for each class.
loss _ cls layer evaluation classification cost LclsProbability p corresponding to the true class uuDetermining:
Lcls=-log pu (1)
loss _ bbox estimation regression loss cost LlocComparing the predicted translation scaling parameters corresponding to the true class u
Figure BDA0001315162160000071
And the true pan zoom parameter v ═ (v ═ vx,vy,vw,vh) The difference between (1) and (2):
Figure BDA0001315162160000072
Figure BDA0001315162160000073
combining the classification loss and the regression loss, the total loss function in the network fine tuning stage is:
Figure BDA0001315162160000074
appointing u to be 0 for background classification, wherein a background candidate region, namely a negative sample, does not participate in regression loss, and does not need to perform regression operation on the candidate region; λ controls the balance of losses and regression losses, λ 1.
And training the network by using a random gradient descent method according to the loss function until L converges.
SVD decomposition is adopted in a Fast R-CNN network to accelerate calculation of a full connection layer; object classification and window regression are all realized through the full articulamentum, establish the full articulamentum input data as x, and the output data is y, and full articulamentum parameter is W, and once forward propagation is promptly:
y=Wx (5)
and (3) carrying out SVD on W, so that the original forward propagation is decomposed into two steps:
y=Wx=U·(∑·VT)·x=U·z (6)
u and V are intermediate variables, and the decomposition can greatly reduce the calculated amount, thereby realizing the acceleration calculation of the full-connection layer.
Fig. 5 is a schematic diagram of the detection results of the road traffic sign, and it can be seen that the detection and recognition effects are good under general road conditions.
In summary, the invention provides a traffic sign detection and identification method based on Fast R-CNN. The method learns the features from the samples through deep learning, can extract the recessive features reflecting the essence of the data, has higher learning efficiency and recognition precision, improves the robustness of the detection algorithm, and effectively improves the accuracy of the detection of the road traffic signs. The problem of detection difficulty caused by factors such as serious shielding, serious abrasion, serious deformation, serious illumination change and the like of a road traffic sign can be solved to a great extent. Some of the method steps and flows herein may need to be performed by a computer to be implemented in hardware, software, firmware, or any combination thereof.
The above description is only a preferred embodiment of the present invention and should not be taken as limiting the scope of the invention, which is intended to include all equivalent changes, modifications, substitutions and the like in the appended claims. Those skilled in the art will recognize that changes and modifications may be made in the broader aspects without departing from the scope and spirit of the invention.

Claims (1)

1. A road traffic sign identification method based on Fast R-CNN comprises the following steps:
carrying out image acquisition and pretreatment to manufacture a sample set;
inputting a training set, and performing multitask training on a Fast R-CNN network;
inputting the picture to be identified into a Fast R-CNN network, and obtaining a characteristic diagram through a plurality of convolution layers and pooling layers;
extracting a plurality of candidate frames by adopting a Selective Search algorithm, finding a feature frame corresponding to each candidate frame in the feature map according to the mapping relation between the candidate frames in the original image and the feature map, and pooling each feature frame to a fixed size in an ROI (Region of Interest) pooling layer;
the feature frames are processed by full connection layers to obtain feature vectors with fixed sizes, and two output vectors of classification scores and window regression are obtained through respective full connection layers;
all the results are subjected to non-maximum suppression processing to generate final target detection and identification results, and the traffic signs are identified;
the image acquisition step specifically comprises:
starting a vehicle-mounted automobile data recorder, shooting road traffic video information in real time, and selecting a video image with a set video resolution ratio shot by the automobile data recorder;
performing frame processing on the shot video image to obtain an image set sequence;
screening the image set, and selecting 7 road traffic signs with more occurrence times from the image set;
the image preprocessing and sample set making steps specifically include:
in the selected image, the target area is taken out and zoomed to a fixed size to enhance the contrast, then the target area is subjected to contrast enhancement processing to form an original training set, and the test set is processed in the same way;
rotating the original training set to the degree of 12 degrees and 12 degrees, zooming to the degree of 0.4 and 1.6, and adding the original training set into the original data set to form a new training set;
randomly taking out samples with the number equivalent to that of the test set from the new data set to form a verification set, and forming a final training set by the rest samples;
the selected traffic signs can be divided into 7 types, namely straight arrow, turning arrow, left arrow, right arrow, straight left arrow, straight right arrow and diamond mark line, wherein the numbers of the traffic signs are respectively 01, 02, 03, 04, 05, 06 and 07, and the identification results are output in the mode;
on the basis of VGG16, replacing the last pooling layer of the VGG-16 network with an ROI pooling layer, and replacing the last full-link layer and the softmax layer of the VGG-16 network with two parallel layers;
the resulting Fast R-CNN network structure is as follows: the multilayer structure comprises 13 convolutional layers, 4 pooling layers, 1 ROI pooling layer, 2 full-connection layers and two flat layers; inputting image samples with fixed size into a network through an input layer; for all convolutional layers, the convolutional kernel size is 3 × 3, step size is 1; for all pooling layers, 2 × 2 sampling sizes are used, and the pooling mode adopts maximum pooling; the activation function adopts a modified Linear unit activation (ReLU) function; initializing original layer parameters in a pre-training mode; initializing a full-connected layer for classification by Gaussian distribution with the mean value of 0 and the standard deviation of 0.01; the full-link layers for regression were initialized with a gaussian distribution with a mean of 0 and a standard deviation of 0.001, with offsets all initialized to 0;
during tuning training, adding N complete pictures, and then adding R candidate frames selected from the N pictures; the R/N candidate frames of the same image are subjected to convolution sharing calculation and memory, so that the calculation overhead is reduced; the constitution of the R candidate frames is as follows: candidate boxes that overlap with some true value at [0.5,1] we define as foreground, accounting for 25% of the total; candidate boxes with a maximum value of [0.1,0.5 ] overlapping with the true value are defined as the background and account for 75% of the total;
the ROI pooling layer averagely divides each feature frame according to a fixed size, performs maximum pooling on each block, can convert feature frames with different sizes on a feature map into data with uniform size, and sends the data into the next layer;
performing multitask training on a Fast R-CNN network, wherein a cls _ score layer is used for classifying, outputting a K + 1-dimensional array p which represents the probability of belonging to K-class objects and backgrounds, and setting K to be 7 according to the number of detection classes; the bbox _ pred layer is used for adjusting the position of the candidate region, outputting 4 x K-dimensional arrays which represent parameters which should be translated and scaled when the candidate region belongs to the K classes respectively, and training a single regressor for each class;
loss _ cls layer evaluation classification cost LclsProbability p corresponding to the true class uuDetermining:
Lcls=-logpu (1)
loss _ bbox estimation regression loss cost LlocComparing the predicted translation scaling parameters corresponding to the true class u
Figure FDA0002634320490000031
And the true pan zoom parameter v ═ (v ═ vx,vy,vw,vh) The difference between (1) and (2):
Figure FDA0002634320490000032
Figure FDA0002634320490000033
combining the classification loss and the regression loss, the total loss function in the network fine tuning stage is:
Figure FDA0002634320490000034
appointing u to be 0 for background classification, wherein a background candidate region, namely a negative sample, does not participate in regression loss, and does not need to perform regression operation on the candidate region; λ controls the balance of losses and regression losses, λ is 1;
training the network by using a random gradient descent method according to the loss function until L converges;
SVD decomposition is adopted in a Fast R-CNN network to accelerate calculation of a full connection layer; object classification and window regression are all realized through the full articulamentum, establish the full articulamentum input data as x, and the output data is y, and full articulamentum parameter is W, and once forward propagation is promptly:
y=Wx (5)
and (3) carrying out SVD on W, so that the original forward propagation is decomposed into two steps:
y=Wx=U·(∑·VT)·x=U·z (6)
u and V are intermediate variables, and the decomposition can greatly reduce the calculated amount, thereby realizing the acceleration calculation of the full-connection layer.
CN201710421849.XA 2017-06-07 2017-06-07 Road traffic sign identification method based on Fast R-CNN Active CN107301383B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710421849.XA CN107301383B (en) 2017-06-07 2017-06-07 Road traffic sign identification method based on Fast R-CNN

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710421849.XA CN107301383B (en) 2017-06-07 2017-06-07 Road traffic sign identification method based on Fast R-CNN

Publications (2)

Publication Number Publication Date
CN107301383A CN107301383A (en) 2017-10-27
CN107301383B true CN107301383B (en) 2020-11-24

Family

ID=60134793

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710421849.XA Active CN107301383B (en) 2017-06-07 2017-06-07 Road traffic sign identification method based on Fast R-CNN

Country Status (1)

Country Link
CN (1) CN107301383B (en)

Families Citing this family (37)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107895191B (en) 2017-10-30 2022-02-22 上海寒武纪信息科技有限公司 Information processing method and related product
CN107909044B (en) * 2017-11-22 2020-04-28 天津大学 People counting method combining convolutional neural network and track prediction
CN108009581A (en) * 2017-11-30 2018-05-08 中国地质大学(武汉) A kind of method for crack based on CNN, equipment and storage device
CN108095746A (en) * 2017-12-21 2018-06-01 安徽省星灵信息科技有限公司 A kind of automatic light concentrator and automatic beam photosystem
CN108275524B (en) * 2018-01-12 2019-08-09 东北大学 A kind of elevator maintenance operation monitoring and guiding device based on the assessment of the first multi-view video series of operations
CN108346144B (en) * 2018-01-30 2021-03-16 哈尔滨工业大学 Automatic bridge crack monitoring and identifying method based on computer vision
CN108416270B (en) * 2018-02-06 2021-07-06 南京信息工程大学 Traffic sign identification method based on multi-attribute combined characteristics
CN108388641B (en) * 2018-02-27 2022-02-01 广东方纬科技有限公司 Traffic facility map generation method and system based on deep learning
CN108416283A (en) * 2018-02-28 2018-08-17 华南理工大学 A kind of pavement marking recognition methods based on SSD
CN108694829B (en) * 2018-03-27 2021-08-13 西安科技大学 Traffic flow identification monitoring network and method based on unmanned aerial vehicle group mobile platform
CN108491889A (en) * 2018-04-02 2018-09-04 深圳市易成自动驾驶技术有限公司 Image, semantic dividing method, device and computer readable storage medium
CN110390344B (en) * 2018-04-19 2021-10-26 华为技术有限公司 Alternative frame updating method and device
CN108520286A (en) * 2018-04-24 2018-09-11 青岛科技大学 Infrared dark dim light small target deteection system based on convolutional Neural and candidate region
CN108776834B (en) * 2018-05-07 2021-08-06 上海商汤智能科技有限公司 System reinforcement learning method and device, electronic equipment and computer storage medium
CN110555877B (en) * 2018-05-31 2022-05-31 杭州海康威视数字技术股份有限公司 Image processing method, device and equipment and readable medium
CN108960308A (en) * 2018-06-25 2018-12-07 中国科学院自动化研究所 Traffic sign recognition method, device, car-mounted terminal and vehicle
CN109214441A (en) * 2018-08-23 2019-01-15 桂林电子科技大学 A kind of fine granularity model recognition system and method
US20210133854A1 (en) 2018-09-13 2021-05-06 Shanghai Cambricon Information Technology Co., Ltd. Information processing method and terminal device
CN109492661B (en) * 2018-09-27 2021-08-13 桂林电子科技大学 Traffic sign identification method based on WWCNN model and TVPDE algorithm
CN111222387B (en) * 2018-11-27 2023-03-03 北京嘀嘀无限科技发展有限公司 System and method for object detection
CN109559536B (en) * 2018-12-10 2021-06-08 百度在线网络技术(北京)有限公司 Traffic light, traffic light identification method, traffic light identification device, traffic light identification equipment and storage medium
CN109636881A (en) * 2018-12-19 2019-04-16 沈阳天择智能交通工程有限公司 Based on AI identification technology traffic accident situ sketch drafting method
CN109858372B (en) * 2018-12-29 2021-04-27 浙江零跑科技有限公司 Lane-level precision automatic driving structured data analysis method
CN109886279B (en) * 2019-01-24 2023-09-29 平安科技(深圳)有限公司 Image processing method, device, computer equipment and storage medium
CN110188705B (en) * 2019-06-02 2022-05-06 东北石油大学 Remote traffic sign detection and identification method suitable for vehicle-mounted system
CN110163187B (en) * 2019-06-02 2022-09-02 东北石油大学 F-RCNN-based remote traffic sign detection and identification method
CN110312103A (en) * 2019-07-06 2019-10-08 辽宁大学 A kind of high-speed equipment anti-thefting monitoring method for processing video frequency based on cloud computing platform
CN112329497A (en) * 2019-07-18 2021-02-05 杭州海康威视数字技术股份有限公司 Target identification method, device and equipment
CN112784084B (en) * 2019-11-08 2024-01-26 阿里巴巴集团控股有限公司 Image processing method and device and electronic equipment
CN111160336A (en) * 2019-12-09 2020-05-15 平安科技(深圳)有限公司 Target detection method, device and computer readable storage medium
CN111126571B (en) * 2019-12-25 2023-07-14 福建天晴数码有限公司 R-CNN network optimization method based on DHT network and storage medium
CN111401466A (en) * 2020-03-26 2020-07-10 广州紫为云科技有限公司 Traffic sign detection and identification marking method and device and computer equipment
CN111444976A (en) * 2020-04-02 2020-07-24 Oppo广东移动通信有限公司 Target detection method and device, electronic equipment and readable storage medium
CN111814873A (en) * 2020-07-07 2020-10-23 广州市运通水务有限公司 Method for distinguishing drainage pipeline defect types and automatically identifying defect grades
CN112929665A (en) * 2021-01-28 2021-06-08 北京博雅慧视智能技术研究院有限公司 Target tracking method, device, equipment and medium combining super-resolution and video coding
CN113344880A (en) * 2021-06-09 2021-09-03 浙江国研智能电气有限公司 Fast-RCNN-based low-voltage electrical appliance transfer printing pattern defect detection method
CN116901975B (en) * 2023-09-12 2023-11-21 深圳市九洲卓能电气有限公司 Vehicle-mounted AI security monitoring system and method thereof

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20120016461A (en) * 2010-08-16 2012-02-24 주식회사 이미지넥스트 Pavement marking recogniton system and method
CN104766042A (en) * 2014-01-06 2015-07-08 现代摩比斯株式会社 Method and apparatus for and recognizing traffic sign board
CN105930830A (en) * 2016-05-18 2016-09-07 大连理工大学 Road surface traffic sign recognition method based on convolution neural network
CN106295707A (en) * 2016-08-17 2017-01-04 北京小米移动软件有限公司 Image-recognizing method and device
CN106372571A (en) * 2016-08-18 2017-02-01 宁波傲视智绘光电科技有限公司 Road traffic sign detection and identification method

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106372577A (en) * 2016-08-23 2017-02-01 北京航空航天大学 Deep learning-based traffic sign automatic identifying and marking method
CN106682569A (en) * 2016-09-28 2017-05-17 天津工业大学 Fast traffic signboard recognition method based on convolution neural network

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20120016461A (en) * 2010-08-16 2012-02-24 주식회사 이미지넥스트 Pavement marking recogniton system and method
CN104766042A (en) * 2014-01-06 2015-07-08 现代摩比斯株式会社 Method and apparatus for and recognizing traffic sign board
CN105930830A (en) * 2016-05-18 2016-09-07 大连理工大学 Road surface traffic sign recognition method based on convolution neural network
CN106295707A (en) * 2016-08-17 2017-01-04 北京小米移动软件有限公司 Image-recognizing method and device
CN106372571A (en) * 2016-08-18 2017-02-01 宁波傲视智绘光电科技有限公司 Road traffic sign detection and identification method

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
Fast R-CNN论文详解;xuanyuyt;《博客园》;20161222;第1-6页 *
Road Surface Traffic Sign Detection with Hybrid Region Proposal and Fast R-CNN;Rongqiang Q. 等;《2016 12th International Conference on Natural Computation, Fuzzy Systems and Knowledge Discovery (ICNC-FSKD)》;20161024;第555-559页 *

Also Published As

Publication number Publication date
CN107301383A (en) 2017-10-27

Similar Documents

Publication Publication Date Title
CN107301383B (en) Road traffic sign identification method based on Fast R-CNN
US11734786B2 (en) Low- and high-fidelity classifiers applied to road-scene images
CN108875608B (en) Motor vehicle traffic signal identification method based on deep learning
CN106683119B (en) Moving vehicle detection method based on aerial video image
Zhang et al. DAGN: A real-time UAV remote sensing image vehicle detection framework
US20170206434A1 (en) Low- and high-fidelity classifiers applied to road-scene images
CN111160249A (en) Multi-class target detection method of optical remote sensing image based on cross-scale feature fusion
CN108280460B (en) SAR vehicle target identification method based on improved convolutional neural network
Hoang et al. Enhanced detection and recognition of road markings based on adaptive region of interest and deep learning
CN111738055B (en) Multi-category text detection system and bill form detection method based on same
CN106815323B (en) Cross-domain visual retrieval method based on significance detection
CN109658442B (en) Multi-target tracking method, device, equipment and computer readable storage medium
Wang et al. Spatial attention for multi-scale feature refinement for object detection
CN111340855A (en) Road moving target detection method based on track prediction
CN113205026B (en) Improved vehicle type recognition method based on fast RCNN deep learning network
Xing et al. Traffic sign recognition using guided image filtering
Rafique et al. Smart traffic monitoring through pyramid pooling vehicle detection and filter-based tracking on aerial images
Liao et al. Unsupervised cluster guided object detection in aerial images
US20240161304A1 (en) Systems and methods for processing images
Zang et al. Traffic lane detection using fully convolutional neural network
CN114037640A (en) Image generation method and device
Dorbe et al. FCN and LSTM based computer vision system for recognition of vehicle type, license plate number, and registration country
Asgarian Dehkordi et al. Vehicle type recognition based on dimension estimation and bag of word classification
CN111274964A (en) Detection method for analyzing water surface pollutants based on visual saliency of unmanned aerial vehicle
CN112949453A (en) Training method of smoke and fire detection model, smoke and fire detection method and smoke and fire detection equipment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant