CN111079753A - License plate recognition method and device based on deep learning and big data combination - Google Patents
License plate recognition method and device based on deep learning and big data combination Download PDFInfo
- Publication number
- CN111079753A CN111079753A CN201911325648.5A CN201911325648A CN111079753A CN 111079753 A CN111079753 A CN 111079753A CN 201911325648 A CN201911325648 A CN 201911325648A CN 111079753 A CN111079753 A CN 111079753A
- Authority
- CN
- China
- Prior art keywords
- training
- license plate
- deep learning
- layer
- model
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 58
- 238000013135 deep learning Methods 0.000 title claims abstract description 33
- 238000012549 training Methods 0.000 claims abstract description 150
- 238000013136 deep learning model Methods 0.000 claims abstract description 47
- 238000004422 calculation algorithm Methods 0.000 claims abstract description 33
- 230000003044 adaptive effect Effects 0.000 claims abstract description 9
- 238000004590 computer program Methods 0.000 claims description 20
- 230000006870 function Effects 0.000 claims description 20
- 238000011176 pooling Methods 0.000 claims description 16
- 230000008569 process Effects 0.000 claims description 15
- 230000009466 transformation Effects 0.000 claims description 15
- 238000013528 artificial neural network Methods 0.000 claims description 13
- 238000000605 extraction Methods 0.000 claims description 12
- 238000001514 detection method Methods 0.000 claims description 11
- 239000013598 vector Substances 0.000 claims description 8
- 230000006872 improvement Effects 0.000 claims description 7
- 230000000306 recurrent effect Effects 0.000 claims description 6
- 238000012545 processing Methods 0.000 claims description 5
- 230000001133 acceleration Effects 0.000 claims description 4
- 230000002457 bidirectional effect Effects 0.000 claims description 2
- 238000006243 chemical reaction Methods 0.000 claims description 2
- 230000008034 disappearance Effects 0.000 claims description 2
- 125000004122 cyclic group Chemical group 0.000 claims 1
- 238000010586 diagram Methods 0.000 description 8
- 238000004364 calculation method Methods 0.000 description 4
- 238000010606 normalization Methods 0.000 description 4
- 230000011218 segmentation Effects 0.000 description 4
- 230000008878 coupling Effects 0.000 description 3
- 238000010168 coupling process Methods 0.000 description 3
- 238000005859 coupling reaction Methods 0.000 description 3
- 238000012360 testing method Methods 0.000 description 3
- 230000006978 adaptation Effects 0.000 description 2
- 238000013459 approach Methods 0.000 description 2
- 210000004027 cell Anatomy 0.000 description 2
- 210000005266 circulating tumour cell Anatomy 0.000 description 2
- 238000004891 communication Methods 0.000 description 2
- 230000003247 decreasing effect Effects 0.000 description 2
- 230000004044 response Effects 0.000 description 2
- 238000013527 convolutional neural network Methods 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 239000011159 matrix material Substances 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/60—Type of objects
- G06V20/62—Text, e.g. of license plates, overlay texts or captions on TV images
- G06V20/63—Scene text, e.g. street names
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/46—Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
- G06V10/462—Salient features, e.g. scale invariant feature transforms [SIFT]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/60—Type of objects
- G06V20/62—Text, e.g. of license plates, overlay texts or captions on TV images
- G06V20/625—License plates
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02T—CLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
- Y02T10/00—Road transport of goods or passengers
- Y02T10/10—Internal combustion engine [ICE] based vehicles
- Y02T10/40—Engine management systems
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- General Engineering & Computer Science (AREA)
- Evolutionary Computation (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Computational Linguistics (AREA)
- Multimedia (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Mathematical Physics (AREA)
- General Health & Medical Sciences (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Evolutionary Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Image Analysis (AREA)
Abstract
The invention provides a license plate recognition method and device based on deep learning and big data combination, wherein the method comprises the following steps: carrying out iterative training on the deep learning model of the big data vehicle image through an improved random gradient descent iterative algorithm; batch training the loaded more vehicle image data through an improved linear scaling and preheating strategy to improve the accuracy of training the deep learning model of the big data vehicle image; training a large batch of training layers in the batch training through an improved adaptive rate scaling algorithm to obtain a rapid training model; inputting the vehicle image trained by the deep learning model and the rapid training model into the RNN, and detecting whether the ROI of the vehicle image is a license plate or not; if yes, license plate recognition is carried out on the ROI through the integrated depth network model, and therefore efficiency and accuracy of large-data license plate recognition can be improved.
Description
Technical Field
The invention belongs to the technical field of computer vision and intelligent traffic, and particularly relates to a license plate recognition method and device based on deep learning and big data combination, terminal equipment and a computer readable medium.
Background
At present, distributed training of a model based on big data and deep learning is an important research foundation of a deep learning network in the field of computer vision. In general, for deep learning applications, a larger data set and a larger model can result in a significant increase in accuracy, but at the cost of taking longer training times. With the rise of deep learning in recent years, many researchers have tried to construct a deep learning network training model based on the rise of deep learning, and both accuracy and effectiveness can be achieved. The method aims to train the real vehicle images, pedestrian images and the like, so that the distributed training method has wide application value in real scenes. The existing training method of the big data vehicle image deep learning model has the defects of low training speed, high training cost and the like, for example, the training of a residual error network-50 (ResNet-50) of millions of vehicle images by using a GPU (image processor) of M40 of great invida takes nearly 14 days. This training requires a total of 10 to the power of 18 single-precision operations. This is clearly disadvantageous both in terms of time and cost. Moreover, in the existing license plate recognition technology, a large amount of calculation of license plate character segmentation is required, so that the problems of slow and inaccurate license plate recognition, poor real-time performance and the like are caused.
Disclosure of Invention
In view of this, embodiments of the present invention provide a license plate recognition method and apparatus based on deep learning and big data combination, a terminal device, and a computer readable medium, which can improve efficiency and accuracy of big data license plate recognition.
The first aspect of the embodiment of the invention provides a license plate recognition method based on deep learning and big data combination, which comprises the following steps:
carrying out iterative training on the deep learning model of the big data vehicle image through an improved random gradient descent iterative algorithm; wherein, each time the deep learning model of the big-data vehicle image is iteratively trained, more vehicle image data is loaded using more processors than the previous iterative training;
batch training the loaded more vehicle image data to improve the accuracy of training the deep learning model of the large data vehicle images by an improved linear scaling and warm-up strategy, wherein the improved linear scaling comprises increasing the batch from B to kB while increasing the learning rate from η to k η, the improved warm-up strategy comprises increasing the relatively small learning rate η to the relatively large learning rate k η for the first several time periods from the relatively small learning rate η value if the relatively large learning rate k η is used;
training a large batch of training layers in the batch training through an improved adaptive rate scaling algorithm to obtain a rapid training model;
inputting the vehicle images trained by the deep learning model and the rapid training model into an RNN (navigation network) to detect whether the RoI (RoI) of the vehicle images is a license plate or not;
if the license plate is the license plate, performing license plate identification on the RoI through an integrated deep network model; the integrated depth network model comprises a convolution layer, a BRNN layer, a linear transformation layer and a CTC layer.
A second aspect of the embodiments of the present invention provides a license plate recognition device based on deep learning and big data combination, including:
the iterative training module is used for performing iterative training on the deep learning model of the big data vehicle image through an improved random gradient descent iterative algorithm; wherein, each time the deep learning model of the big data vehicle image is iteratively trained, more processors are used to load more vehicle image data than the previous iterative training;
a accuracy training module for batch training the loaded more vehicle image data to improve the accuracy of training the deep learning model for the large data vehicle images by an improved linear scaling and warm-up strategy, the improved linear scaling including increasing the batch from B to kB while increasing the learning rate from η to k η, the improved warm-up strategy including increasing the relatively small learning rate η to the relatively large learning rate k η for the first several time periods starting from a relatively small learning rate η value if a relatively large learning rate k η is used;
the scaling improvement module is used for training a large batch of training layers in the batch training through an improved adaptive rate scaling algorithm to obtain a fast training model;
the license plate detection module is used for inputting the vehicle image trained by the deep learning model and the rapid training model into the RNN so as to detect whether the RoI of the vehicle image is a license plate or not;
the license plate recognition module is used for recognizing the license plate of the RoI through an integrated deep network model when the license plate is detected; the integrated deep network model comprises a convolution layer, a BRNN layer, a linear transformation layer and a CTC layer.
A third aspect of the embodiments of the present invention provides a terminal device, which includes a memory, a processor, and a computer program stored in the memory and executable on the processor, where the processor implements the steps of the license plate recognition method based on deep learning and big data combination when executing the computer program.
A fourth aspect of the embodiments of the present invention provides a computer-readable medium, where a computer program is stored, and when the computer program is processed and executed, the steps of the license plate recognition method based on deep learning and big data combination are implemented.
In the license plate recognition method based on deep learning and big data combination provided by the embodiment of the invention, a big data license plate image can be trained by combining a relevant model through an improved random gradient descent iterative algorithm, a linear scaling and preheating strategy adaptation rate scaling algorithm, the trained vehicle data is input into an RNN to detect whether the RoI of the vehicle image is a license plate or not, and when the RoI is detected to be the license plate, the license plate recognition is carried out on the RoI through an integrated deep network model, so that the efficiency and the accuracy of the big data license plate recognition can be improved.
Drawings
In order to more clearly illustrate the technical solutions in the embodiments of the present invention, the following briefly introduces the embodiments or drawings used in the description of the prior art, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and it is obvious for those skilled in the art that other drawings can be obtained according to these drawings without inventive labor.
Fig. 1 is a flowchart of a license plate recognition method based on deep learning and big data combination according to an embodiment of the present invention;
fig. 2 is a schematic diagram of a process of identifying the RoI of a license plate image through an integrated deep network model according to an embodiment of the present invention;
FIG. 3 is a schematic structural diagram of a license plate recognition device based on deep learning and big data combination according to an embodiment of the present invention;
FIG. 4 is a schematic diagram of a detailed structure of the license plate detection module in FIG. 3;
FIG. 5 is a schematic diagram of a detailed structure of the license plate recognition module in FIG. 3;
fig. 6 is a schematic diagram of a terminal device according to an embodiment of the present invention.
Detailed Description
In the following description, for purposes of explanation and not limitation, specific details are set forth, such as particular system structures, techniques, etc. in order to provide a thorough understanding of the embodiments of the invention. It will be apparent, however, to one skilled in the art that the present invention may be practiced in other embodiments that depart from these specific details. In other instances, detailed descriptions of well-known systems, devices, circuits, and methods are omitted so as not to obscure the description of the present invention with unnecessary detail.
In order to explain the technical means of the present invention, the following description will be given by way of specific examples.
Referring to fig. 1, fig. 1 is a diagram illustrating a license plate recognition method based on deep learning and big data combination according to an embodiment of the present invention. As shown in fig. 1, the license plate recognition method based on deep learning and big data combination of the embodiment includes the following steps:
s101: and carrying out iterative training on the deep learning model of the big data vehicle image through an improved random gradient descent iterative algorithm.
In the embodiments of the present invention, generally speaking, an asynchronous method using a parameter server cannot guarantee stability on a large system. For very large Deep Neural Network (DNN) training, the data parallel synchronization method is more stable. The idea is also simple-by using large batch sizes for random gradient descent (SGD), the work per iteration can be easily distributed to multiple processors. In the ideal vehicle image training case, ResNet-50 requires 772 billions of single precision operations to process one 225x225 vehicle image. If an epoch (time period) is run 90 times for an image mesh (ImageNet) dataset, the operand is 90 x 128 x 77.2 million (18 th power of 10). Currently, the most powerful supercomputers can complete 200 × 1015 single-precision operations per second. If an algorithm is available that makes full use of the supercomputer, the training of ResNet-50 can theoretically be completed in 5 seconds. For this reason, it is necessary to have the algorithm use more processors and load more vehicle image data at each iteration. Thereby reducing the total training time. Generally, to the extent that larger batches will result in sheetsThe speed of the GPU is higher (as shown in fig. 2). The reason is that the low-level matrix computation library will be more efficient. For training the Res-Net 50 model using ImageNet, the optimal batch size for each GPU is 512. If it is desired to use many GPUs and have each GPU active, a larger batch size is required. For example, if there are 16 GPUs, then the batch size should be set to 16 × 512 — 8192. Ideally, if the total number of accesses is fixed, and the batch size is linearly increased as the number of processors increases, the number of improved SGD iterations is linearly decreased, the time cost per iteration remains the same, and thus the total time is linearly decreased with the number of processors. The specific improved SGD iterative algorithm is as follows: let w represent the weight of (deep neural network) DNN, X represent the training data, n be the number of samples in X, and Y represent the label of the training data X. Let us order xiSample for X, l (X)i,yiW) is xiAnd its label yi(i ∈ {1, 2.., n)) calculated losses. Embodiments of the present invention use a loss function like the cross-entropy function, and the goal of DNN training is to minimize the loss function in equation (1), with the following formula:
wherein w represents the weight of DNN, X is training data, n is the number of samples in X, Y represents the label of the training data X, and X represents the weight of DNNiAre samples in the training data X.
In the t-th iteration, embodiments of the invention use forward and backward propagation to find the gradient of the loss function versus the weight. This gradient is then used to update the weights, with equation (2) for updating the weights according to the gradient as follows:
wherein, wtIs the weight after the t-1 th iteration, wt+1Weight after the t-th iteration, η learning rate, batch size of the t-th iteration BtAnd B istIs largeAnd b is small. In the embodiment of the invention, the batch size of the t iteration is BtAnd B istB, the weights may then be updated based on the following equation (3):
wherein, wtIs the weight after the t-1 th iteration, wt+1Weight after the t-th iteration, η learning rate, batch size of the t-th iteration BtAnd B istThe size of (a) is b.
To simplify the expression, we can say that the update rule in equation (4) represents that we use the gradient of the weightUpdate the weight wtIs wt+1。
By adopting the method, iteration is carried out, and more image data are loaded by using the processor as much as possible, so that the training time can be reduced linearly and greatly. In addition, before the iterative training of the deep learning model is performed through the improved stochastic gradient descent iterative algorithm, the deep learning model is established, and the method for establishing the deep learning model is the same as the prior art, so that the detailed description is omitted.
S102: and training the loaded more image data in batch through an improved linear scaling and preheating strategy so as to improve the accuracy of training the deep learning model.
In the embodiment of the present invention, when training a large batch, we need to ensure that the test accuracy is as good as that of a small batch under the condition of running the same number of time periods (epochs). Here we fix the number of time periods (epochs) because statistically one time period (epoch) means that the algorithm will touch the entire data set once; and, in calculation, it is fixedThe present invention trains large batches of data using an improved linear scaling and warm-up strategy 1. Linear scaling, which increases batches from B to kB, while also increasing the learning rate from η to k η. the warm-up strategy, which, if a relatively large learning rate k η is used, starts with a relatively small learning rate η value and increases the relatively small learning rate η to the relatively large learning rate k η for the first few time periodsThe weights are updated. Using the data parallel approach, multiple machine versions can be handled in the same way. Each layer of the deep learning model has own weight w and gradientThe standard SGD algorithm uses the same LR for all layers (η), however, from routine experimentation, it can be observed that different layers may require different LR's due to the weight norm | | w | | and the weight gradient normThe ratio between them is very different in different layers. Embodiments of the present invention solve this problem using a modified LARS algorithm (a new updated learning rate rule), the basic LR rule being defined in equation (1). L in equation (1) is a scaling factor, and in embodiments of the present invention, l can be set to 0.001 during AlexNet and ResNet training. Gamma is the user's adjustment parameter. Usually a good gamma, all values being [1, 50]]BetweenIn this equation, different layers may have different LR. adding momentum (denoted as μ) and weight attenuation (denoted as β) to the SGD and using the method steps of obtaining a local learning rate η for each learnable parameter in a large batch of training layers in the batch, obtaining a true learning rate η ' for each layer in a large batch of training layers in the batch, the true learning rate being η ' γ × α × η, where γ is a user's tuning parameter and γ ranges from [1, 50] y]α is an acceleration term, by formulaUpdating the weight gradient; wherein,is weight gradient, w is weight, β is weight decay, and the weight gradient is determined by the formulaThe Local Response Normalization (LRN) needs to be changed to Batch Normalization (BN) in order to extend to larger batch sizes (e.g., 32 k). The present method adds BN. the improved LARS provided by embodiments of the present invention after each convolution layer of the deep neural network can help ResNet-50 maintain high test accuracy.
S103: and training the large-batch training layers in the batch training through an improved adaptive rate scaling algorithm to obtain a fast training model.
Specifically, in order to improve the accuracy of mass training, the embodiment of the method of the inventionA new update Learning Rate (LR) rule is used. The stand-alone case must be considered here, the use of whichThe weights are updated. Using the data parallel approach, multiple machine versions can be handled in the same manner. Each layer of the deep learning model has own weight w and gradientThe standard SGD algorithm uses the same LR for all layers (η), however, from routine experimentation, it can be observed that different layers may require different LRs due to the weight norm | | w | | and the weight gradient normThe ratio between them varies greatly in different layers. Embodiments of the present invention solve this problem using a modified LARS algorithm (a new updated learning rate rule), the basic LR rule being defined in equation (1). L in equation (1) is a scaling factor, and in the embodiment of the present invention, l can be set to 0.001 in AlexNet and ResNet training. Gamma is the user's adjustment parameter. Usually a good gamma, all values being [1, 50]]In this equation, different layers may have different LR. adding momentum (denoted as μ) and weight attenuation (denoted as β) to the SGD and using the method steps of obtaining a local learning rate η for each learnable parameter in a large batch of training layers in the batch training, obtaining a true learning rate η ' for each layer in a large batch of training layers in the batch training, the true learning rate being η ' γ × α × η, where γ is a user's adjustment parameter and γ has a range of values [1, 50 × η ]]α is an acceleration term, by formulaUpdating the weight gradient; wherein,is weight gradient, w is weight, β is weight decay, byFormula (II)The Local Response Normalization (LRN) needs to be changed to Batch Normalization (BN) in order to scale to a larger batch size (e.g., 32 k). The method of the present invention adds BN. the improved LARS provided by embodiments of the present invention after each convolution layer of the deep neural network to help ResNet-50 maintain high test accuracy.
S104: inputting the vehicle image trained by the deep learning model and the fast training model into a Recurrent Neural Network (RNN) to detect whether a region of interest (RoI) of the vehicle image is a license plate.
In the embodiment of the invention, the vehicle images trained by the deep learning model and the fast training model can be input into the RNN, and the trained vehicle images are subjected to RoI pooling by using the RNN; then, an extraction layer is added in two Full Connection (FC) layers in the RNN to convert the pooled features (or called region features) into feature vectors; and then, scoring and frame regression can be carried out on the RoI through the characteristic vector, and whether the RoI is the license plate or not is judged according to the scoring and frame regression. If the RoI license plate is detected, the process goes to S105. It can be understood that, since one extraction layer is added to two FC layers and the license plate is detected by way of scoring and frame regression, a new RNN different from the prior art is constructed in the embodiments of the present invention.
S105: and carrying out license plate identification on the RoI through an integrated deep network model.
In the embodiment of the invention, because the character segmentation of the license plate image requires a large amount of calculation, in order to improve the recognition efficiency, the invention provides an integrated deep network model for license plate recognition, and the integrated deep network model comprises a convolutional layer, a Bidirectional Recurrent Neural Network (BRNN) layer, a linear transformation layer and a connected significance time classification (CTC) layer. In particular, a specific method for identifying the region of interest of the vehicle image by the integrated depth network model can be understood with reference to fig. 2, as follows:
a first step of subjecting a region of interest (e.g., bay a.02u10) of the vehicle image to feature extraction after RoI pooling, processing the extracted features (e.g., as region features C × X × Y) through two convolutional layers and a rectangular pooling layer between the two convolutional layers to transform the extracted features into a feature sequence D × L; wherein D-512 and L-19, and said signature sequence is represented by V-V (V1, V2, VL).
And secondly, applying the characteristic sequence V at a BRNN layer to form two mutually separated RNNs, wherein one RNN forwards processes the characteristic sequence V, the other RNN backwards processes the characteristic sequence V, two implicit states are concatenated together and input into a linear transformation layer with 37 outputs, the linear transformation layer is switched to a Softmax layer, the 37 outputs are converted into probabilities, the probabilities correspond to the probabilities of 26 letters, 10 numbers and a non-character class, the probabilities are coded by the BRNN layer, the characteristic sequence V is converted into a probability estimation q which has the same length as that of L, the probability is (q1, q 2.,. q L.), and meanwhile, a long-short-term memory network (LSTM) is used for defining memory cells comprising three multiplication gates so as to selectively store related information and solve the problem of gradient disappearance in RNN training.
Thirdly, performing sequence decoding on the probability estimation q through a CTC layer, and searching an approximate optimal path with the maximum probability through the decoded probability estimation q:
where pi is the near-optimal path with the highest probability (e.g., a02U10), B operator is used for one repeated token and non-character token, P is a probability operation, an example is: b (a-a-B-) (B (-a-bb) ═ aab), and the specific details of CTCs are the structure of existing CTCs and are not described herein.
And fourthly, determining a loss function of the integrated depth network model through the approximate optimal path, and performing license plate recognition on the RoI through the loss function (for example, recognizing that the license plate is Gui A.02U10). The method for identifying the RoI through the overall loss function of the model is the same as that in the prior art, and therefore, the details are not repeated herein. It should be noted that the integrated deep network model may include a Softmax layer and a rectangular pooling layer between two convolutional layers in addition to the main convolutional layers (two), the BRNN layer, the linear transformation layer, and the CTC layer, and the convolutional layers may also be regarded as convolutional neural networks.
In the license plate recognition method based on deep learning and big data combination provided in fig. 1, a deep learning model of a big data vehicle image can be iteratively trained through an improved stochastic gradient descent iterative algorithm, more processors are used to load more image data than the previous iterative training each time the deep learning model of the big data vehicle image is iteratively trained, the loaded more image data is batch trained through an improved linear scaling and preheating strategy to adjust the training accuracy, a large batch of training layers in the batch training are trained through an improved adaptation rate scaling algorithm to obtain a fast training model, and the vehicle images trained through the deep learning model and the fast training model are license plate detected and recognized through an improved RNN and an integrated depth network model, therefore, the training cost of license plate recognition can be reduced, character segmentation with large calculation amount is avoided, and the efficiency and effectiveness of large-data license plate recognition are improved.
Referring to fig. 3, fig. 3 is a block diagram illustrating a license plate recognition device based on deep learning and big data combination according to an embodiment of the present invention. As shown in fig. 3, the license plate recognition device 30 based on deep learning and big data combination of the present embodiment includes an iterative training module 301, a precision training module 302, a scaling improvement module 303, a license plate detection module 304, and a license plate recognition module 305. The iterative training module 301, the accuracy training module 302, the scaling improvement module 303, the license plate detection module 304, and the license plate recognition module 305 are respectively configured to perform the specific methods in S101 to S105 in fig. 1, and details can be referred to the related introduction of fig. 1, which is only briefly described here:
the iterative training module 301 is configured to perform iterative training on the depth learning model of the big-data vehicle image through an improved stochastic gradient descent iterative algorithm; wherein, each time the deep learning model of the big data vehicle image is iteratively trained, more processors are used to load more vehicle image data than the previous iterative training.
A accuracy training module 302 for batch training the loaded more vehicle image data to improve the accuracy of training the deep learning model for the large data vehicle images by an improved linear scaling and warm-up strategy, the improved linear scaling including increasing the batch from B to kB while increasing the learning rate from η to k η, the improved warm-up strategy including increasing the relatively small learning rate η to the relatively large learning rate k η for the previous several time periods, starting from a relatively small learning rate η value if a relatively large learning rate k η is used.
And a scaling improvement module 303, configured to train a large batch of training layers in the batch training through an improved adaptive scaling algorithm to obtain a fast training model.
And the license plate detection module 304 is configured to input the vehicle image trained by the deep learning model and the fast training model into the RNN, so as to detect whether the RoI of the vehicle image is a license plate.
The license plate recognition module 305 is used for recognizing the license plate of the RoI through an integrated deep network model when the license plate is detected; the integrated deep network model comprises a convolution layer, a BRNN layer, a linear transformation layer and a CTC layer.
Further, referring to fig. 4, the license plate detection module 304 may specifically include a pooling unit 3041, a conversion unit 3042, and a determination unit 3043:
a pooling unit 3041, configured to input the vehicle images trained by the deep learning model and the fast training model into an RNN, and perform RoI pooling on the trained vehicle images by using the RNN.
A converting unit 3042, configured to convert the pooled features into feature vectors by adding an extraction layer to two fully-connected layers in the RNN.
The determining unit 3043 is configured to score the RoI by using the feature vector and perform border regression, and determine whether the RoI is a license plate according to the score and the border regression.
Further, referring to fig. 5, the license plate recognition module 305 may specifically include a feature extraction unit 3051, a probability estimation unit 3052, an optimal path unit 3053, and a recognition unit 3054:
a feature extraction unit 3051, configured to perform feature extraction after the RoI is subjected to the RoI pooling, and process the extracted features through two convolutional layers and a rectangular pooling layer between the two convolutional layers, so as to transform the extracted features into a feature sequence dx L; wherein D-512 and L-19, and said signature sequence is represented by V-V (V1, V2, VL).
A probability estimation unit 3052, configured to apply the feature sequence V at the BRNN layer to form two mutually separated recurrent neural networks RNN, where one RNN processes the feature sequence V forward and the other RNN processes the feature sequence V backward, concatenates two implicit states together, inputs the two implicit states into a linear transformation layer having 37 outputs, and switches the linear transformation layer to a Softmax layer, converts the 37 outputs into probabilities corresponding to the probabilities of 26 letters, 10 numbers and a non-character class, the probabilities are encoded by the BRNN layer, so that the feature sequence V is converted into a probability estimate q ═ having the same length as L (q1, q 2.., qL), and simultaneously defines a memory cell including three multiplication gates using LSTM to selectively store related information, thereby solving a gradient vanishing problem in RNN training.
An optimal path unit 3053, configured to perform sequence decoding on the probability estimate q through a CTC layer, and find an approximate optimal path with a maximum probability through the decoded probability estimate q:
wherein pi is an approximate optimal path with the maximum probability, the B operator is used for repeated marks and non-character marks at one position, and P is probability operation.
And the identification unit 3054 is configured to determine a loss function of the integrated deep network model according to the approximate optimal path, and perform license plate identification on the RoI through the loss function.
The license plate recognition device based on deep learning and big data combination provided by fig. 3 can perform iterative training on a deep learning model of a big data vehicle image through an improved random gradient descent iterative algorithm, when the iterative training is performed on the deep learning model of the big data vehicle image each time, more processors are used for loading more image data than the previous iterative training, the loaded more image data are subjected to batch training through an improved linear scaling and preheating strategy to adjust the training accuracy, and license plate detection and recognition are performed on the vehicle image through an improved RNN and an integrated deep network model, so that the training cost of license plate recognition can be reduced, large-calculation-amount character segmentation is avoided, and the efficiency and effectiveness of the license plate recognition of the big data vehicle are improved.
Fig. 6 is a schematic diagram of a terminal device according to an embodiment of the present invention. As shown in fig. 6, the terminal device 6 of this embodiment includes: a processor 60, a memory 61 and a computer program 62 stored in said memory 61 and executable on said processor 60, such as a program for license plate recognition based on deep learning combined with big data. The processor 60, when executing the computer program 62, implements the steps in the above-described method embodiments, e.g., S101 to S105 shown in fig. 1. Alternatively, the processor 60, when executing the computer program 62, implements the functions of the modules/units in the above-mentioned device embodiments, such as the functions of the modules 301 to 305 shown in fig. 2.
Illustratively, the computer program 62 may be partitioned into one or more modules/units that are stored in the memory 61 and executed by the processor 60 to implement the present invention. The one or more modules/units may be a series of computer program instruction segments capable of performing specific functions, which are used to describe the execution of the computer program 62 in the terminal device 6. For example, the computer program 62 may be partitioned into an iterative training module 301, a accuracy training module 302, a scaling improvement module 303, a license plate detection module 304, and a license plate recognition module 305. (modules in the virtual device), the specific functions of each module are as follows:
the iterative training module 301 is configured to perform iterative training on the depth learning model of the big-data vehicle image through an improved stochastic gradient descent iterative algorithm; wherein, each time the deep learning model of the big data vehicle image is iteratively trained, more processors are used to load more vehicle image data than the previous iterative training.
A accuracy training module 302 for batch training the loaded more vehicle image data to improve the accuracy of training the deep learning model for the large data vehicle images by an improved linear scaling and warm-up strategy, the improved linear scaling including increasing the batch from B to kB while increasing the learning rate from η to k η, the improved warm-up strategy including increasing the relatively small learning rate η to the relatively large learning rate k η for the previous several time periods, starting from a relatively small learning rate η value if a relatively large learning rate k η is used.
And a scaling improvement module 303, configured to train a large batch of training layers in the batch training through an improved adaptive scaling algorithm to obtain a fast training model.
And the license plate detection module 304 is configured to input the vehicle image trained by the deep learning model and the fast training model into the RNN, so as to detect whether the RoI of the vehicle image is a license plate.
The license plate recognition module 305 is used for recognizing the license plate of the RoI through an integrated deep network model when the license plate is detected; the integrated deep network model comprises a convolution layer, a BRNN layer, a linear transformation layer and a CTC layer.
The terminal device 6 may be a desktop computer, a notebook, a palm computer, a cloud server, or other computing devices. Terminal device 6 may include, but is not limited to, a processor 60, a memory 61. Those skilled in the art will appreciate that fig. 6 is merely an example of a terminal device 6 and does not constitute a limitation of terminal device 6 and may include more or less components than those shown, or some components in combination, or different components, for example, the terminal device may also include input output devices, network access devices, buses, etc.
The Processor 60 may be a Central Processing Unit (CPU), other general purpose Processor, a Digital Signal Processor (DSP), an Application Specific Integrated Circuit (ASIC), an off-the-shelf Programmable Gate Array (FPGA) or other Programmable logic device, discrete Gate or transistor logic device, discrete hardware component, etc. A general purpose processor may be a microprocessor or the processor may be any conventional processor or the like.
The memory 61 may be an internal storage unit of the terminal device 6, such as a hard disk or a memory of the terminal device 6. The memory 61 may also be an external storage device of the terminal device 6, such as a plug-in hard disk, a Smart Media Card (SMC), a Secure Digital (SD) Card, a Flash memory Card (Flash Card), and the like provided on the terminal device 6. Further, the memory 61 may also include both an internal storage unit of the terminal device 6 and an external storage device. The memory 61 is used for storing the computer programs and other programs and data required by the terminal device 6. The memory 61 may also be used to temporarily store data that has been output or is to be output.
It will be clear to those skilled in the art that, for convenience and simplicity of description, the foregoing functional units and modules are merely illustrated in terms of their division, and in practical applications, the foregoing functional allocation may be performed by different functional units and modules as needed, that is, the internal structure of the device is divided into different functional units or modules to perform all or part of the above described functions. Each functional unit and module in the embodiments may be integrated in one processing unit, or each unit may exist alone physically, or two or more units are integrated in one unit, and the integrated unit may be implemented in a form of hardware, or in a form of software functional unit. In addition, specific names of the functional units and modules are only for convenience of distinguishing from each other, and are not used for limiting the protection scope of the present application. The specific working process of the units and modules in the system can refer to the corresponding process in the foregoing method embodiments, and is not described herein again.
In the above embodiments, the descriptions of the respective embodiments have respective emphasis, and reference may be made to the related descriptions of other embodiments for parts that are not described or illustrated in a certain embodiment.
Those of ordinary skill in the art will appreciate that the various illustrative elements and algorithm steps described in connection with the embodiments disclosed herein may be implemented as electronic hardware or combinations of computer software and electronic hardware. Whether such functionality is implemented as hardware or software depends upon the particular application and design constraints imposed on the implementation. Skilled artisans may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the present invention.
In the embodiments provided in the present invention, it should be understood that the disclosed apparatus/terminal device and method may be implemented in other ways. For example, the above-described embodiments of the apparatus/terminal device are merely illustrative, and for example, the division of the modules or units is only one logical division, and there may be other divisions when actually implemented, for example, a plurality of units or components may be combined or integrated into another system, or some features may be omitted, or not executed. In addition, the shown or discussed mutual coupling or direct coupling or communication connection may be an indirect coupling or communication connection of devices or units through some interfaces, and may be in an electrical, mechanical or other form.
The units described as separate parts may or may not be physically separate, and parts displayed as units may or may not be physical units, may be located in one place, or may be distributed on a plurality of network units. Some or all of the units can be selected according to actual needs to achieve the purpose of the solution of the embodiment.
In addition, functional units in the embodiments of the present invention may be integrated into one processing unit, or each unit may exist alone physically, or two or more units are integrated into one unit. The integrated unit can be realized in a form of hardware, and can also be realized in a form of a software functional unit.
The integrated modules/units, if implemented in the form of software functional units and sold or used as separate products, may be stored in a computer readable storage medium. Based on such understanding, all or part of the flow of the method according to the embodiments of the present invention may also be implemented by a computer program, which may be stored in a computer-readable storage medium, and when the computer program is executed by a processor, the steps of the method embodiments may be implemented. Wherein the computer program comprises computer program code, which may be in the form of source code, object code, an executable file or some intermediate form, etc. The computer-readable medium may include: any entity or device capable of carrying the computer program code, recording medium, usb disk, removable hard disk, magnetic disk, optical disk, computer Memory, Read-Only Memory (ROM), Random Access Memory (RAM), electrical carrier wave signals, telecommunications signals, software distribution medium, and the like. It should be noted that the computer readable medium may contain content that is subject to appropriate increase or decrease as required by legislation and patent practice in jurisdictions, for example, in some jurisdictions, computer readable media does not include electrical carrier signals and telecommunications signals in accordance with legislation and patent practice.
The above-mentioned embodiments are only used for illustrating the technical solutions of the present invention, and not for limiting the same; although the present invention has been described in detail with reference to the foregoing embodiments, it will be understood by those of ordinary skill in the art that: the technical solutions described in the foregoing embodiments may still be modified, or some technical features may be equivalently replaced; such modifications and substitutions do not substantially depart from the spirit and scope of the embodiments of the present invention, and are intended to be included within the scope of the present invention.
Claims (10)
1. A license plate recognition method based on deep learning and big data combination is characterized by comprising the following steps:
carrying out iterative training on the deep learning model of the big data vehicle image through an improved random gradient descent iterative algorithm; wherein, each time the deep learning model of the big data vehicle image is iteratively trained, more processors are used to load more vehicle image data than the previous iterative training;
batch training the loaded more vehicle image data to improve the accuracy of training the deep learning model of the large data vehicle images by an improved linear scaling and warm-up strategy, wherein the improved linear scaling comprises increasing the batch from B to kB while increasing the learning rate from η to k η, the improved warm-up strategy comprises increasing the relatively small learning rate η to the relatively large learning rate k η for the first several time periods from the relatively small learning rate η value if the relatively large learning rate k η is used;
training a large batch of training layers in the batch training through an improved adaptive rate scaling algorithm to obtain a rapid training model;
inputting the vehicle images trained by the deep learning model and the rapid training model into a Recurrent Neural Network (RNN) to detect whether the region of interest (RoI) of the vehicle images is a license plate;
if the license plate is the license plate, performing license plate identification on the RoI through an integrated deep network model; the integrated deep network model comprises a convolutional layer, a bidirectional cyclic neural network BRNN layer, a linear transformation layer and a connection meaning time classification CTC layer.
2. The license plate recognition method based on deep learning and big data combination of claim 1, wherein the iterative training of the deep learning model of the big data vehicle image through the improved stochastic gradient descent iterative algorithm comprises:
constructing a loss function L (W) of the deep learning model of the big data vehicle image:
wherein w represents the weight of the deep neural network DNN, X is training data, n is the number of samples in X, Y represents the label of the training data X, and X represents the weight of the deep neural network DNNiFor samples in the training data X, l (X)i,yiW) is for xiAnd its label yi(i ∈ {1, 2.., n));
updating the weight of the DNN according to the gradient of the loss function to the weight when the deep learning model of the big data vehicle image is iteratively trained each time:
wherein, wtIs the weight after the t-1 th iteration, wt+1Weight after the t-th iteration, η learning rate, batch size of the t-th iteration BtAnd B istThe size of (a) is b; more processors are used to load more image data per iteration of training than in the previous iteration.
3. The license plate recognition method based on deep learning and big data combination of claim 1, wherein the training of the large batch of training layers in the batch training through the improved adaptive scaling algorithm to obtain the fast training model comprises:
obtaining a local learning rate η for each learnable parameter in a bulk training layer in the bulk training;
acquiring a real learning rate η 'of each layer in a large batch of training layers in the batch training, wherein the real learning rate is η' ═ γ × α × η, γ is an adjustment parameter of a user, the numeric range of γ is [1, 50], and α is an acceleration item;
by the formulaUpdating the weight gradient; wherein,is weight gradient, w is weight, β is weight decay;
the weights are updated with the formula w- α to arrive at a fast training model.
4. The license plate recognition method based on deep learning and big data combination of claim 1, wherein the step of inputting the vehicle image trained by the deep learning model and the fast training model into an integrated deep network model to detect whether the RoI of the vehicle image is a license plate comprises the steps of:
inputting the vehicle images trained by the deep learning model and the fast training model into an RNN, and performing RoI pooling on the trained vehicle images by using the RNN,
converting the pooled features into feature vectors by adding an extraction layer to two fully-connected layers in the RNN;
and scoring and frame regression are carried out on the RoI through the characteristic vector, and whether the RoI is the license plate or not is judged according to the scoring and frame regression.
5. The license plate recognition method based on deep learning and big data combination of claim 1, wherein the license plate recognition of the ROI through an integrated deep network model comprises:
performing feature extraction after the region of interest is subjected to RoI pooling, and processing the extracted features through two convolution layers and a rectangular pooling layer between the two convolution layers so as to transform the extracted features into a feature sequence DxL; wherein D-512 and L-19, said signature sequence is represented by V-1, V2, VL;
applying the characteristic sequence V at a BRNN layer to form two mutually separated recurrent neural networks RNN, wherein one RNN processes the characteristic sequence V forward, the other RNN processes the characteristic sequence V backward, two implicit states are concatenated together, the input is input into a linear transformation layer with 37 outputs, the output is transferred to a Softmax layer, the 37 outputs are converted into probabilities, the probabilities correspond to the probabilities of 26 letters, 10 numbers and a non-character class, the probabilities are coded by the BRNN layer, the characteristic sequence V is converted into a probability estimation q (q1, q 2.., q L.) with the same length as L, and a long-short term memory network LSTM is used to define a memory cell containing three multiplication gates so as to selectively store related information and solve the problem of gradient disappearance in RNN training,
performing sequence decoding on the probability estimation q through a CTC layer, and searching an approximate optimal path with the maximum probability through the decoded probability estimation q:
wherein pi is an approximate optimal path with the maximum probability, the B operator is used for repeated marks and non-character marks at one position, and P is probability operation;
and determining a loss function of the integrated depth network model through the approximate optimal path, and identifying the license plate of the RoI through the loss function.
6. A license plate recognition device based on deep learning and big data combination, characterized by includes:
the iterative training module is used for performing iterative training on the deep learning model of the big data vehicle image through an improved random gradient descent iterative algorithm; wherein, each time the deep learning model of the big data vehicle image is iteratively trained, more processors are used to load more vehicle image data than the previous iterative training;
a accuracy training module for batch training the loaded more vehicle image data to improve the accuracy of training the deep learning model for the large data vehicle images by an improved linear scaling and warm-up strategy, the improved linear scaling including increasing the batch from B to kB while increasing the learning rate from η to k η, the improved warm-up strategy including increasing the relatively small learning rate η to the relatively large learning rate k η for the first several time periods starting from a relatively small learning rate η value if a relatively large learning rate k η is used;
the scaling improvement module is used for training a large batch of training layers in the batch training through an improved adaptive rate scaling algorithm to obtain a fast training model;
the license plate detection module is used for inputting the vehicle image trained by the deep learning model and the rapid training model into the RNN so as to detect whether the RoI of the vehicle image is a license plate or not;
the license plate recognition module is used for recognizing the license plate of the ROI through an integrated depth network model when the license plate is detected; the integrated deep network model comprises a convolution layer, a BRNN layer, a linear transformation layer and a CTC layer.
7. The deep learning and big data combination-based license plate recognition device of claim 6, wherein the license plate detection module comprises:
a pooling unit for inputting the vehicle images trained by the deep learning model and the fast training model into an RNN, and performing RoI pooling on the trained vehicle images by using the RNN,
a conversion unit, configured to add an extraction layer to two fully-connected layers in the RNN to convert the pooled features into feature vectors;
and the judging unit is used for scoring and frame regression on the Roi through the feature vector and judging whether the RoI is the license plate or not according to the scoring and frame regression.
8. The deep learning and big data based license plate recognition device of claim 6, wherein the recognition module comprises:
a feature extraction unit, configured to perform feature extraction after RoI pooling on the region of interest, and process the extracted features through two convolutional layers and a rectangular pooling layer between the two convolutional layers to transform the extracted features into a feature sequence D × L; wherein D-512 and L-19, said signature sequence is represented by V-1, V2, VL;
a probability estimation unit, configured to apply the feature sequence V at a BRNN layer to form two mutually separated recurrent neural networks RNN, where one RNN processes the feature sequence V forward, and the other RNN processes the feature sequence V backward, concatenates two implicit states, inputs the concatenated states into a linear transformation layer having 37 outputs, and switches the linear transformation layer to a Softmax layer, and converts the 37 outputs into probabilities, where the probabilities correspond to probabilities of 26 letters, 10 numbers, and a non-character class, and the probabilities are encoded by the BRNN layer, so that the feature sequence V is converted into a probability estimate q (q1, q 2.., qL) having the same length as L, and at the same time, LSTM is used to define a memory cell including three multiplication gates, so as to selectively store related information and solve a gradient vanishing problem in RNN training;
the optimal path unit is used for performing sequence decoding on the probability estimation q through a CTC layer, and searching an approximate optimal path with the maximum probability through the decoded probability estimation q:
wherein pi is an approximate optimal path with the maximum probability, the B operator is used for repeated marks and non-character marks at one position, and P is probability operation;
and the identification unit is used for determining a loss function of the integrated depth network model through the approximate optimal path and identifying the license plate of the RoI through the loss function.
9. A terminal device comprising a memory, a processor and a computer program stored in the memory and executable on the processor, characterized in that the processor implements the steps of the method according to any of claims 1-5 when executing the computer program.
10. A computer-readable medium, in which a computer program is stored which, when being processed and executed, carries out the steps of the method according to any one of claims 1 to 5.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201911325648.5A CN111079753B (en) | 2019-12-20 | 2019-12-20 | License plate recognition method and device based on combination of deep learning and big data |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201911325648.5A CN111079753B (en) | 2019-12-20 | 2019-12-20 | License plate recognition method and device based on combination of deep learning and big data |
Publications (2)
Publication Number | Publication Date |
---|---|
CN111079753A true CN111079753A (en) | 2020-04-28 |
CN111079753B CN111079753B (en) | 2023-08-22 |
Family
ID=70316200
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201911325648.5A Active CN111079753B (en) | 2019-12-20 | 2019-12-20 | License plate recognition method and device based on combination of deep learning and big data |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111079753B (en) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111832568A (en) * | 2020-06-12 | 2020-10-27 | 北京百度网讯科技有限公司 | License plate recognition method, and training method and device of license plate recognition model |
CN111931801A (en) * | 2020-05-28 | 2020-11-13 | 浙江大学 | Dynamic routing network learning method based on path diversity and consistency |
CN112069717A (en) * | 2020-08-19 | 2020-12-11 | 五邑大学 | Magnetic storm prediction method and device based on multi-mode representation learning and storage medium |
WO2022267000A1 (en) * | 2021-06-25 | 2022-12-29 | Intel Corporation | Methods and apparatus for scale recovery from monocular video |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109461495A (en) * | 2018-11-01 | 2019-03-12 | 腾讯科技(深圳)有限公司 | A kind of recognition methods of medical image, model training method and server |
US20190095730A1 (en) * | 2017-09-25 | 2019-03-28 | Beijing University Of Posts And Telecommunications | End-To-End Lightweight Method And Apparatus For License Plate Recognition |
CN109784476A (en) * | 2019-01-12 | 2019-05-21 | 福州大学 | A method of improving DSOD network |
CN110096982A (en) * | 2019-04-22 | 2019-08-06 | 长沙千视通智能科技有限公司 | A kind of video frequency vehicle big data searching method based on deep learning |
-
2019
- 2019-12-20 CN CN201911325648.5A patent/CN111079753B/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20190095730A1 (en) * | 2017-09-25 | 2019-03-28 | Beijing University Of Posts And Telecommunications | End-To-End Lightweight Method And Apparatus For License Plate Recognition |
CN109461495A (en) * | 2018-11-01 | 2019-03-12 | 腾讯科技(深圳)有限公司 | A kind of recognition methods of medical image, model training method and server |
CN109784476A (en) * | 2019-01-12 | 2019-05-21 | 福州大学 | A method of improving DSOD network |
CN110096982A (en) * | 2019-04-22 | 2019-08-06 | 长沙千视通智能科技有限公司 | A kind of video frequency vehicle big data searching method based on deep learning |
Non-Patent Citations (1)
Title |
---|
赵珊;黄强强;曲宏山;刘相利;: "改进的多标签深度学习车辆属性识别研究" * |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111931801A (en) * | 2020-05-28 | 2020-11-13 | 浙江大学 | Dynamic routing network learning method based on path diversity and consistency |
CN111931801B (en) * | 2020-05-28 | 2024-03-12 | 浙江大学 | Dynamic route network learning method based on path diversity and consistency |
CN111832568A (en) * | 2020-06-12 | 2020-10-27 | 北京百度网讯科技有限公司 | License plate recognition method, and training method and device of license plate recognition model |
CN111832568B (en) * | 2020-06-12 | 2024-01-12 | 北京百度网讯科技有限公司 | License plate recognition method, training method and device of license plate recognition model |
CN112069717A (en) * | 2020-08-19 | 2020-12-11 | 五邑大学 | Magnetic storm prediction method and device based on multi-mode representation learning and storage medium |
WO2022267000A1 (en) * | 2021-06-25 | 2022-12-29 | Intel Corporation | Methods and apparatus for scale recovery from monocular video |
Also Published As
Publication number | Publication date |
---|---|
CN111079753B (en) | 2023-08-22 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN111079753A (en) | License plate recognition method and device based on deep learning and big data combination | |
CN112184508B (en) | Student model training method and device for image processing | |
CN111382868B (en) | Neural network structure searching method and device | |
US11934949B2 (en) | Composite binary decomposition network | |
US20170061279A1 (en) | Updating an artificial neural network using flexible fixed point representation | |
CN111275107A (en) | Multi-label scene image classification method and device based on transfer learning | |
CN111832440B (en) | Face feature extraction model construction method, computer storage medium and equipment | |
CN110175641B (en) | Image recognition method, device, equipment and storage medium | |
CN111026544B (en) | Node classification method and device for graph network model and terminal equipment | |
CN110689045A (en) | Distributed training method and device for deep learning model | |
CN113298152B (en) | Model training method, device, terminal equipment and computer readable storage medium | |
CN111553215A (en) | Personnel association method and device, and graph convolution network training method and device | |
WO2023231954A1 (en) | Data denoising method and related device | |
EP4361843A1 (en) | Neural network searching method and related device | |
WO2021253941A1 (en) | Neural network model training method, image classification method, text translation method and apparatus, and device | |
CN113011532B (en) | Classification model training method, device, computing equipment and storage medium | |
EP4060526A1 (en) | Text processing method and device | |
CN110874627A (en) | Data processing method, data processing apparatus, and computer readable medium | |
Chen et al. | HRCP: High-ratio channel pruning for real-time object detection on resource-limited platform | |
CN112085175A (en) | Data processing method and device based on neural network calculation | |
CN114298289A (en) | Data processing method, data processing equipment and storage medium | |
CN116432608A (en) | Text generation method and device based on artificial intelligence, computer equipment and medium | |
WO2021081854A1 (en) | Convolution operation circuit and convolution operation method | |
US20230073835A1 (en) | Structured Pruning of Vision Transformer | |
CN114155388B (en) | Image recognition method and device, computer equipment and storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |