WO2021196873A1 - 车牌字符识别方法、装置、电子设备和存储介质 - Google Patents
车牌字符识别方法、装置、电子设备和存储介质 Download PDFInfo
- Publication number
- WO2021196873A1 WO2021196873A1 PCT/CN2021/074915 CN2021074915W WO2021196873A1 WO 2021196873 A1 WO2021196873 A1 WO 2021196873A1 CN 2021074915 W CN2021074915 W CN 2021074915W WO 2021196873 A1 WO2021196873 A1 WO 2021196873A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- license plate
- network
- layer
- image
- network structure
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 61
- 238000013527 convolutional neural network Methods 0.000 claims abstract description 61
- 230000002457 bidirectional effect Effects 0.000 claims abstract description 29
- 238000000605 extraction Methods 0.000 claims abstract description 23
- 238000003062 neural network model Methods 0.000 claims abstract description 21
- 125000004122 cyclic group Chemical group 0.000 claims description 28
- 230000015654 memory Effects 0.000 claims description 25
- 230000009466 transformation Effects 0.000 claims description 23
- 238000011176 pooling Methods 0.000 claims description 22
- 230000007787 long-term memory Effects 0.000 claims description 9
- 238000013528 artificial neural network Methods 0.000 claims description 8
- 230000006403 short-term memory Effects 0.000 claims description 8
- 230000011218 segmentation Effects 0.000 abstract description 19
- 230000008569 process Effects 0.000 abstract description 15
- 230000000306 recurrent effect Effects 0.000 abstract 1
- PXFBZOLANLWPMH-UHFFFAOYSA-N 16-Epiaffinine Natural products C1C(C2=CC=CC=C2N2)=C2C(=O)CC2C(=CC)CN(C)C1C2CO PXFBZOLANLWPMH-UHFFFAOYSA-N 0.000 description 13
- 238000010586 diagram Methods 0.000 description 11
- 230000008034 disappearance Effects 0.000 description 10
- 230000006870 function Effects 0.000 description 10
- 238000004891 communication Methods 0.000 description 7
- 238000012545 processing Methods 0.000 description 7
- 238000012937 correction Methods 0.000 description 6
- 239000011159 matrix material Substances 0.000 description 5
- 238000001514 detection method Methods 0.000 description 4
- 238000005070 sampling Methods 0.000 description 4
- 230000000694 effects Effects 0.000 description 3
- 238000005516 engineering process Methods 0.000 description 3
- 238000004458 analytical method Methods 0.000 description 2
- 239000000284 extract Substances 0.000 description 2
- 238000007726 management method Methods 0.000 description 2
- 238000013507 mapping Methods 0.000 description 2
- 239000000463 material Substances 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000010606 normalization Methods 0.000 description 2
- 230000004913 activation Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 239000003086 colorant Substances 0.000 description 1
- 230000008094 contradictory effect Effects 0.000 description 1
- 238000013135 deep learning Methods 0.000 description 1
- 239000006185 dispersion Substances 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 239000000835 fiber Substances 0.000 description 1
- 230000004927 fusion Effects 0.000 description 1
- 238000003064 k means clustering Methods 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000005457 optimization Methods 0.000 description 1
- 230000001151 other effect Effects 0.000 description 1
- 230000002093 peripheral effect Effects 0.000 description 1
- 238000011897 real-time detection Methods 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 230000002441 reversible effect Effects 0.000 description 1
- 238000000926 separation method Methods 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 230000002123 temporal effect Effects 0.000 description 1
- 238000012549 training Methods 0.000 description 1
- 238000012795 verification Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/60—Type of objects
- G06V20/62—Text, e.g. of license plates, overlay texts or captions on TV images
- G06V20/63—Scene text, e.g. street names
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/60—Type of objects
- G06V20/62—Text, e.g. of license plates, overlay texts or captions on TV images
- G06V20/625—License plates
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/044—Recurrent networks, e.g. Hopfield networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/0464—Convolutional networks [CNN, ConvNet]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/084—Backpropagation, e.g. using gradient descent
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/24—Aligning, centring, orientation detection or correction of the image
- G06V10/242—Aligning, centring, orientation detection or correction of the image by image rotation, e.g. by 90 degrees
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/44—Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
- G06V10/443—Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components by matching or filtering
- G06V10/449—Biologically inspired filters, e.g. difference of Gaussians [DoG] or Gabor filters
- G06V10/451—Biologically inspired filters, e.g. difference of Gaussians [DoG] or Gabor filters with interaction between the filter responses, e.g. cortical complex cells
- G06V10/454—Integrating the filters into a hierarchical structure, e.g. convolutional neural networks [CNN]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/764—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/7715—Feature extraction, e.g. by transforming the feature space, e.g. multi-dimensional scaling [MDS]; Mappings, e.g. subspace methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V30/00—Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
- G06V30/10—Character recognition
Definitions
- This application relates to the field of image technology, and in particular to a method, device, electronic device, and storage medium for recognizing license plate characters.
- the method of recognizing the license plate in the related intelligent license plate recognition system is generally as follows: firstly, it is necessary to perform character segmentation on the located license plate, then extract the character characteristics of a single character, and finally perform the license plate character recognition.
- the above-mentioned license plate is based on the character segmentation. In the way of recognition, the accuracy of character segmentation is difficult to guarantee, resulting in low accuracy of license plate recognition.
- This application proposes a license plate character recognition method, device, electronic equipment and storage medium, which does not require character segmentation of the license plate, and the license plate characters on the license plate can be obtained by directly recognizing the entire license plate, avoiding segmentation and separation of the characters of the license plate Recognition improves the technical effect of recognition speed and recognition accuracy.
- An embodiment of the present application proposes a license plate character recognition method, including: acquiring a vehicle image collected by an image acquisition device; positioning the license plate area of the vehicle image to obtain a license plate image; and analyzing the license plate through a convolutional neural network.
- Image feature extraction is performed to obtain the feature information of the license plate image, wherein the convolutional neural network includes a residual network structure; the feature information is parsed through a bidirectional cyclic neural network model to obtain the license plate corresponding to the license plate image character.
- the license plate character recognition method of the embodiment of the present application locates the vehicle area in the license plate image after acquiring the vehicle image collected by the image acquisition device to obtain the license plate image, and passes the convolutional neural network including the residual network structure
- the feature extraction of the license plate image can effectively avoid the gradient disappearance and reduce the feature loss in the convolution process of the convolutional neural network, so that the bidirectional cyclic neural network model can accurately identify the license plate characters of the license plate image based on the feature information of the license plate image. Therefore, there is no need to perform character segmentation on the license plate, and the license plate characters on the license plate can be obtained by directly recognizing the entire license plate, avoiding segmentation and separate recognition of the characters of the license plate, and improving the recognition speed and recognition accuracy.
- a license plate character recognition device which includes: an image acquisition module for acquiring a vehicle image collected by an image acquisition device; a license plate positioning module for locating a license plate area on the vehicle image to Obtain a license plate image; a feature extraction module for feature extraction of the license plate image through a convolutional neural network to obtain feature information of the license plate image, wherein the convolutional neural network includes a residual network structure; a recognition module , Used to analyze the feature information through the bidirectional cyclic neural network model to obtain the license plate characters corresponding to the license plate image.
- the license plate character recognition device of the embodiment of the application after acquiring the vehicle image collected by the image acquisition device, locates the vehicle area in the license plate image to obtain the license plate image, and passes the convolutional neural network including the residual network structure
- the feature extraction of the license plate image can effectively avoid the gradient disappearance and reduce the feature loss in the convolution process of the convolutional neural network, so that the bidirectional cyclic neural network model can accurately identify the license plate characters of the license plate image based on the feature information of the license plate image. Therefore, there is no need to perform character segmentation on the license plate, and the license plate characters on the license plate can be obtained by directly recognizing the entire license plate, avoiding segmentation and separate recognition of the characters of the license plate, and improving the recognition speed and recognition accuracy.
- Another embodiment of the present application proposes an electronic device, including: an electronic device including: a memory and a processor; the memory stores computer instructions, and when the computer instructions are executed by the processor, Realize the license plate character recognition method of the embodiment of the present application.
- Another embodiment of the present application proposes a non-transitory computer-readable storage medium storing computer instructions for causing the computer to execute the license plate character recognition method disclosed in the embodiments of the present application.
- Fig. 1 is a schematic flowchart of a method for recognizing characters on a license plate according to an embodiment of the present application
- FIG. 2 is a schematic diagram of a network structure of a convolutional neural network including a first residual network structure and a second residual network structure;
- FIG. 3 is a schematic diagram of the network structure of the first residual network structure
- Figure 4 is a schematic diagram of the network structure of the bidirectional cyclic neural network model
- FIG. 5 is a schematic flowchart of a method for recognizing characters on a license plate according to another embodiment of the present application.
- Fig. 6 is a schematic structural diagram of a license plate character recognition device according to an embodiment of the present application.
- Fig. 7 is a schematic structural diagram of a license plate character recognition device according to another embodiment of the present application.
- Fig. 8 is a block diagram of an electronic device used to implement the method for recognizing license plate characters in an embodiment of the present application.
- Fig. 1 is a schematic flowchart of a method for recognizing license plate characters according to an embodiment of the present application.
- the execution body of the license plate character recognition method provided in this embodiment is a license plate character recognition device.
- the license plate character recognition device can be configured in a smart license plate recognition system, and the smart license plate recognition system can be installed in an electronic device.
- the electronic device may be a hardware device such as a terminal device and a server.
- the license plate character recognition method may include:
- Step 101 Acquire a vehicle image collected by an image collection device.
- the image acquisition device may be a camera.
- step 102 the license plate area is located on the vehicle image to obtain the license plate image.
- the YOLOV3 network in order to improve the ability to quickly locate the license plate image in the license plate image, after obtaining the license plate image, the YOLOV3 network can be used to locate the license plate area of the vehicle image to obtain the license plate image.
- YOLOV3 network is a kind of deep learning image detection algorithm, which is mainly based on the regression idea.
- the two steps of target recognition and target positioning are carried out at the same time, which can improve the speed of target detection and meet the requirements of real-time detection.
- the YOLOV3 network uses a 53-layer convolutional network and uses a softmax layer.
- the YOLOV3 network can also use three-scale feature fusion for object detection and positioning, and use K-means clustering to generate nine a priori boxes as bounding boxes (that is, a compact rectangle containing license plates).
- the initial size of the frame can thus achieve the effect of accelerating the convergence speed and improving the detection accuracy.
- the Yolov3 network can accurately locate the license plates of vehicles of various colors in various complex environments, and the recognition time is about 20 ms.
- Step 103 Perform feature extraction on the license plate image through a convolutional neural network to obtain feature information of the license plate image, where the convolutional neural network includes a residual network structure.
- the convolutional layer in the convolutional neural network usually performs feature extraction on the license plate image to obtain the feature image corresponding to the license plate image.
- the convolutional layer performs feature extraction on the license plate image, it is easy to cause the characteristics of the license plate image Loss, and when the depth of the convolutional neural network increases, the problem of gradient disappearance occurs.
- a residual network structure can be set in the convolutional neural network, so that the residual network structure can effectively prevent the disappearance of the gradient caused by the increase in depth in the convolutional neural network. And reduce the problem of feature loss in the convolution process.
- the aforementioned convolutional neural network may also include a convolutional layer and a pooling layer.
- the convolutional neural network in this embodiment may be connected in sequence according to the convolutional layer, the residual network structure, and the pooling layer.
- the convolutional neural network of this embodiment includes a convolutional layer, a residual network structure, and a pooling layer that are sequentially connected.
- the convolutional layer in the convolutional neural network will perform feature extraction on the license plate image to obtain the first feature image of the license plate image.
- the first feature can be The image is input to the residual network structure, and the characteristic image output by the residual structure is input to the pooling layer for corresponding processing.
- three convolutional layers may be set in the convolutional neural network to perform feature extraction on the license plate image.
- the first convolutional layer, the residual network structure, the second convolutional layer, the third convolutional layer, and the pooling layer are sequentially connected to form the convolutional neural network of this embodiment, thereby minimizing the feature loss of the convolution process.
- a batch normalization layer is added after the first convolutional layer in the convolutional neural network, thereby effectively improving network performance and speeding up convergence.
- the convolutional neural network in this embodiment may use Leaky ReLU as the activation function.
- the residual network structure of this embodiment includes the first residual network structure and The second residual network structure.
- the convolutional neural network includes a first convolutional layer, a first residual network structure, a second residual network structure, a second convolutional layer, a third convolutional layer, and a pooling layer that are sequentially connected.
- the network structure diagram of the convolutional neural network including the first residual network structure and the second residual network structure is shown in FIG. 2.
- the network structure of the convolutional neural network in FIG. 2 is specifically: a first convolutional layer, a first residual network structure connected to the first convolutional layer, and a second residual network structure connected to the first residual network structure , A second convolutional layer connected to the second residual network structure, a third convolutional layer connected to the second convolutional layer, and a maximum pooling layer connected to the third convolutional layer.
- the convolutional neural network parameters are shown in Table 1, where Filters is the number of convolution kernels, K is the size of the convolution kernel, S is the step size, and P is the padding parameter.
- Residual-1 in Table 1 represents the first residual network structure
- Residual-2 represents the second residual network structure
- ConvolutionN represents the Nth convolutional layer, where N is any integer from 1 to 3.
- a Batch Normalization layer may be set between the second residual network structure in the convolutional neural network and Leaky ReLU.
- the network structures of the first residual network structure and the second residual network structure are the same, and the network structures both include the first network substructure, the second network substructure and the adder, and the first network substructure
- the structure includes the largest pooling layer, the first convolution sublayer, the second convolution sublayer, the third convolution sublayer, and the fourth convolution sublayer that are sequentially connected.
- the second network substructure includes the fifth convolution sublayer. Network, the input ends of the maximum pooling layer and the fifth convolution sub-network are all connected to the output end of the output layer of the first convolutional layer, and the output ends of the fourth convolution sub-layer and the fifth convolution sub-network are both connected to the addition The input terminal of the device is connected.
- FIG. 3 the network structure diagram of the first residual network structure is shown in FIG. 3.
- the characteristic image can be input into the first residual network structure, and the first residual network structure The output is input into the second residual network structure again, and then the characteristic information output by the second residual network structure is input to the corresponding processing layer connected to the second residual network structure for subsequent processing through the corresponding processing layer.
- the network structures of the first residual network structure and the second residual network structure in this embodiment are the same, the network parameters used by the two are different.
- the residual network structure of this embodiment can also make the forward and backward propagation of information smoother.
- the number of convolution kernels corresponding to the convolution layer in the first residual network structure and the second residual network structure are different.
- Convolution in the table represents the convolutional layer
- ConvolutionN represents the Nth convolutional sublayer, where N is any integer from 1 to 5.
- Step 104 Analyze the feature information through the bidirectional cyclic neural network model to obtain license plate characters corresponding to the license plate image.
- the above-mentioned bidirectional cyclic neural network model may include a bidirectional cyclic neural network and a Connectionist Temporal Classification (CTC) network.
- CTC Connectionist Temporal Classification
- the bidirectional cyclic neural network of this embodiment may include a first layer of bidirectional long-term memory (BLSTM) sequentially connected. And the second layer two-way long and short-term memory network.
- BLSTM bidirectional long-term memory
- each LSTM network in the bidirectional long and short-term memory network in this embodiment contains 512 hidden units. Therefore, the output depth of each layer of BLSTM is 1024, and the output of the last layer of BLSTM is connected to two fully connected layers. The depth of the full connection of the layer is 1024, and the depth of the full connection of the second layer is the number of license plate character types (for example, 66).
- a dropout layer is provided between the two fully connected layers in this embodiment. That is, the second layer of the bidirectional long-short-term memory network of this embodiment includes: a first fully connected layer, a dropout layer connected to the first fully connected layer, and a second fully connected layer connected to the dropout layer.
- FIG. 4 the schematic diagram of the network structure of the bidirectional cyclic neural network model in this embodiment is shown in FIG. 4.
- the feature extraction of the entire license plate image is performed directly through the convolutional neural network, and the extracted feature information is analyzed through the bidirectional cyclic neural network model to Get the license plate characters in the license plate image. From this, it can be seen that this embodiment does not need to perform character segmentation on the license plate image, so the time used for character segmentation in the license plate recognition process can be reduced, and the license plate recognition speed can be improved.
- the license plate character recognition method of the embodiment of the present application locates the vehicle area in the license plate image after acquiring the vehicle image collected by the image acquisition device to obtain the license plate image, and passes the convolutional neural network including the residual network structure
- the feature extraction of the license plate image can effectively avoid the gradient disappearance and reduce the feature loss in the convolution process of the convolutional neural network, so that the bidirectional cyclic neural network model can accurately identify the license plate characters of the license plate image based on the feature information of the license plate image. Therefore, there is no need to perform character segmentation on the license plate, and the license plate characters on the license plate can be obtained by directly recognizing the entire license plate, avoiding segmentation and separate recognition of the characters of the license plate, and improving the recognition speed and recognition accuracy.
- Fig. 5 is a schematic flowchart of a method for recognizing license plate characters according to another embodiment of the present application. Among them, it should be noted that this embodiment is a further refinement or optimization of the above-mentioned embodiment.
- the license plate character recognition method may include:
- Step 501 Acquire a vehicle image collected by an image collection device.
- Step 502 Use YOLOV3 network to locate the license plate area of the vehicle image to obtain the license plate image.
- Step 503 Use the spatial transformation network to perform tilt correction on the license plate image.
- the spatial transform network may include a local network, a grid generator, and a sampler.
- the license plate image can be input into the local network to obtain the affine transformation parameters used to correct the license plate image; the affine transformation parameters are input into the grid generator to obtain the corresponding affine Transformation matrix: A sampler is used to perform affine transformation on each pixel coordinate in the license plate image based on the affine transformation matrix to correct the license plate image.
- the above-mentioned local network is a network used to regress and transform the parameter ⁇ .
- the specific parameters of the network are shown in Table 4, where Filters is the number of convolution kernels, K is the size of the convolution kernel, S is the step size, and P is The padding parameter; its input is a grayscale image, and then it passes through a series of hidden network layers, and finally outputs the spatial transformation parameters.
- the form of ⁇ can be various, if 2D affine transformation is required, ⁇ is the output of a 6-dimensional (2x3) vector.
- the size of ⁇ depends on the type of transformation.
- the input is a 2D image, and 2D affine transformation needs to be implemented.
- the network outputs a 2*3 matrix as shown in formula 2-1, expressed as A ⁇ .
- the aforementioned grid generator constructs a sampling grid based on predicted transformation parameters, which is an output obtained after sampling and transformation of points in a group of input images. What the grid generator actually gets is a mapping relationship T ⁇ . Assume that the coordinates of each pixel of the input image are The coordinate of each pixel output by STN is The space transformation function T ⁇ is a two-dimensional affine transformation function, then with The corresponding relationship can be expressed as 2-2:
- the sampler uses the sampling grid and the input image as input at the same time to obtain the transformed result of the input image.
- the above-mentioned sampler samples the input image features on the basis of the mapping relationship obtained by the network generator, and obtains the image features that have undergone spatial transformation to achieve a correction effect.
- the sampling method used in this embodiment is bilinear interpolation.
- the noise interference around the license plate image can also be filtered through the local network to make the license plate easier to recognize.
- Step 504 Perform feature extraction on the license plate image through a convolutional neural network to obtain feature information of the license plate image, where the convolutional neural network includes a residual network structure.
- Step 505 Analyze the feature information through the bidirectional cyclic neural network model to obtain license plate characters corresponding to the license plate image.
- the license plate characters can include Chinese characters, letters, and numbers.
- the vehicle area in the license plate image is located to obtain the license plate image, and the license plate image is tilted and corrected.
- the convolutional neural network of the difference network structure extracts the characteristics of the corrected license plate image, effectively avoiding the disappearance of the gradient and reducing the feature loss in the convolution process of the convolutional neural network, so that the bidirectional cyclic neural network model can be based on the characteristics of the license plate image
- the information accurately recognizes the license plate characters of the license plate image.
- this embodiment does not need to perform character segmentation on the license plate after directly locating the license plate in the picture vehicle, and directly input the license plate image obtained after the positioning into the convolutional neural network including the residual network structure Carry out feature extraction, and use the bidirectional cyclic neural network model to recognize the license plate characters based on the feature information of the license plate image, which can directly recognize the entire license plate, avoid segmentation and separate recognition of the license plate characters, and improve the recognition speed.
- This embodiment conducts experiments on the license plate recognition method using STN and the license plate recognition method without STN. According to the experimental results, it is determined that the accuracy rate of the license plate recognition with the STN network removed is 94.6%, which is lower than the accuracy of the license plate recognition model with the STN network. The rate is 96.1%. The average loss of license plate recognition that removes the STN network is 0.53, which is higher than the average loss of 0.40 for license plate recognition that joins the STN network. From this, it can be seen that performing tilt correction before inputting the license plate image to the convolutional neural network can further improve the recognition accuracy of the license plate characters.
- Table 5 shows that when the network of this application has no STN network, the number of hidden units of one-layer BLSTM, two-layer BLSTM and three-layer BLSTM is At 128, 256, 512, 1024, the accuracy of license plate recognition and the average loss of license plate recognition. Since the necessity of the STN network has been verified above, in order to improve the training convergence speed, the STN network will no longer be added during network verification.
- the obtained experimental parameters reflect the rationality of the BLSTM parameters selected in this application.
- the number before the slash in the above table is the value of the license plate recognition accuracy rate ACC
- the number after the slash represents the value of the average loss loss of the license plate recognition.
- 94.6/0.53 indicates that the license plate recognition is accurate.
- the ACC rate is 94.6%
- the average loss of license plate recognition is 0.53.
- this embodiment also analyzes the fully connected layer FC_layer.
- Table 6 shows the license plate recognition accuracy rate and the average number of license plates recognized by the model when there is no STN network in the first-layer fully-connected, two-layer fully-connected, and three-layer fully-connected Loss, the experimental parameters obtained through this can reflect the rationality of the fully connected parameters selected by the network of this application.
- ACC represents the accuracy of license plate recognition under the corresponding model
- loss represents the average loss of license plate recognition under the corresponding model
- F1 represents the output depth of the first layer of fully connected
- F2 represents the output depth of the second layer of fully connected
- F3 represents the third The output depth of the layer fully connected.
- an embodiment of the present application also provides a license plate character recognition device, because the license plate character recognition device provided by the embodiment of the application is different from the license plate character recognition device provided by the foregoing embodiments.
- the license plate character recognition method corresponds, so the implementation of the license plate character recognition method is also applicable to the license plate character recognition device provided in this embodiment, and will not be described in detail in this embodiment.
- Fig. 6 is a schematic structural diagram of a license plate character recognition device according to an embodiment of the present application.
- the license plate character recognition device 600 includes an image acquisition module 110, a license plate location module 120, a feature extraction module 130, and a recognition module 140, wherein:
- the image acquisition module 110 is used to acquire the vehicle image collected by the image acquisition device.
- the license plate location module 120 is used to locate the license plate area of the vehicle image to obtain the license plate image.
- the feature extraction module 130 is configured to perform feature extraction on the license plate image through a convolutional neural network to obtain feature information of the license plate image, where the convolutional neural network includes a residual network structure.
- the recognition module 140 is used to analyze the feature information through the bidirectional cyclic neural network model to obtain the license plate characters corresponding to the license plate image.
- the convolutional neural network includes a convolutional layer, a residual network structure, and a pooling layer that are sequentially connected.
- the convolutional neural network includes a first convolutional layer, a residual network structure, a second convolutional layer, a third convolutional layer, and a pooling layer that are sequentially connected.
- the convolutional layer includes 3 layers
- the residual network structure includes a first residual network structure and a second residual network structure
- the convolutional neural network includes sequentially connected The first convolutional layer, the first residual network structure, the second residual network structure, the second convolutional layer, the third convolutional layer, and the pooling layer.
- the first residual network structure and the second residual network structure have the same network structure, and the network structure includes a first network substructure, a second network substructure, and an adder.
- the first network substructure Including the largest pooling layer, the first convolution sublayer, the second convolution sublayer, the third convolution sublayer, and the fourth convolution sublayer that are sequentially connected.
- the second network substructure includes the fifth convolution subnetwork. , The input ends of the maximum pooling layer and the fifth convolution subnetwork are both connected to the output ends of the output layer of the first convolution layer, and the output ends of the fourth convolution sublayer and the fifth convolution subnetwork are both connected to the adder The input terminal is connected.
- the device further includes:
- the tilt correction module 150 is configured to use the spatial transformation network to perform tilt correction on the license plate image.
- the spatial transformation network includes a local network, a grid generator, and a sampler.
- the above-mentioned tilt correction module 150 is specifically used to: input the license plate image into the local network to be used to correct the license plate image Affine transformation parameters; input the affine transformation parameters to the grid generator to obtain the corresponding affine transformation matrix; use the sampler to perform affine transformation on each pixel coordinate in the license plate image based on the affine transformation matrix to Correct the license plate image.
- the bidirectional cyclic neural network model includes a bidirectional cyclic neural network and a CTC network for connection timing classification.
- the bidirectional cyclic neural network includes a first layer of bidirectional long and short-term memory network and a second layer of bidirectional long and short-term memory network that are sequentially connected.
- the second-layer bidirectional long-short-term memory network includes: a first fully connected layer; a dropout layer connected to the first fully connected layer; and a second fully connected layer connected to the dropout layer.
- the license plate character recognition device of the embodiment of the application after acquiring the vehicle image collected by the image acquisition device, locates the vehicle area in the license plate image to obtain the license plate image, and passes the convolutional neural network including the residual network structure
- the feature extraction of the license plate image can effectively avoid the gradient disappearance and reduce the feature loss in the convolution process of the convolutional neural network, so that the bidirectional cyclic neural network model can accurately identify the license plate characters of the license plate image based on the feature information of the license plate image. Therefore, there is no need to perform character segmentation on the license plate, and the license plate characters on the license plate can be obtained by directly recognizing the entire license plate, avoiding segmentation and separate recognition of the characters of the license plate, and improving the recognition speed and recognition accuracy.
- the present application also provides an electronic device and a readable storage medium.
- FIG. 8 it is a block diagram of an electronic device of a method for recognizing license plate characters according to an embodiment of the present application.
- the electronic device includes:
- the processor 1002 implements the database management method provided in the foregoing embodiment when executing instructions.
- the electronic equipment also includes:
- the communication interface 1003 is used for communication between the memory 1001 and the processor 1002.
- the memory 1001 is used to store computer instructions that can run on the processor 1002.
- the memory 1001 may include a high-speed RAM memory, and may also include a non-volatile memory (non-volatile memory), for example, at least one disk memory.
- the processor 1002 is used to implement the license plate character recognition method of the foregoing embodiment when executing a program.
- the bus may be an Industry Standard Architecture (ISA) bus, Peripheral Component (PCI) bus, or Extended Industry Standard Architecture (EISA) bus, etc.
- ISA Industry Standard Architecture
- PCI Peripheral Component
- EISA Extended Industry Standard Architecture
- the bus can be divided into address bus, data bus, control bus and so on. For ease of representation, only one thick line is used in FIG. 8, but it does not mean that there is only one bus or one type of bus.
- the memory 1001, the processor 1002, and the communication interface 1003 are integrated on a single chip, the memory 1001, the processor 1002, and the communication interface 1003 can communicate with each other through internal interfaces.
- the processor 1002 may be a central processing unit (Central Processing Unit, referred to as CPU), or a specific integrated circuit (Application Specific Integrated Circuit, referred to as ASIC), or configured to implement one or more of the embodiments of the present application integrated circuit.
- CPU Central Processing Unit
- ASIC Application Specific Integrated Circuit
- first and second are only used for descriptive purposes, and cannot be understood as indicating or implying relative importance or implicitly indicating the number of indicated technical features. Therefore, the features defined with “first” and “second” may explicitly or implicitly include at least one of the features. In the description of the present application, "a plurality of” means at least two, such as two, three, etc., unless specifically defined otherwise.
- a "computer-readable medium” can be any device that can contain, store, communicate, propagate, or transmit a program for use by an instruction execution system, device, or device or in combination with these instruction execution systems, devices, or devices.
- computer readable media include the following: electrical connections (electronic devices) with one or more wiring, portable computer disk cases (magnetic devices), random access memory (RAM), Read only memory (ROM), erasable and editable read only memory (EPROM or flash memory), fiber optic devices, and portable compact disk read only memory (CDROM).
- the computer-readable medium may even be paper or other suitable medium on which the program can be printed, because it can be used, for example, by optically scanning the paper or other medium, followed by editing, interpretation, or other suitable media if necessary. The program is processed in a manner to obtain the program electronically, and then stored in the computer memory.
- each part of this application can be implemented by hardware, software, firmware, or a combination thereof.
- multiple steps or methods can be implemented by software or firmware stored in a memory and executed by a suitable instruction execution system.
- Discrete logic gate circuits with logic functions for data signals Logic circuits, application specific integrated circuits with suitable combinational logic gates, programmable gate array (PGA), field programmable gate array (FPGA), etc.
- a person of ordinary skill in the art can understand that all or part of the steps carried in the method of the foregoing embodiments can be implemented by a program instructing relevant hardware to complete.
- the program can be stored in a computer-readable storage medium. When executed, it includes one of the steps of the method embodiment or a combination thereof.
- the functional units in the various embodiments of the present application may be integrated into one processing module, or each unit may exist alone physically, or two or more units may be integrated into one module.
- the above-mentioned integrated modules can be implemented in the form of hardware or software functional modules. If the integrated module is implemented in the form of a software function module and sold or used as an independent product, it can also be stored in a computer readable storage medium.
- the aforementioned storage medium may be a read-only memory, a magnetic disk or an optical disk, etc.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Evolutionary Computation (AREA)
- Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- General Health & Medical Sciences (AREA)
- Multimedia (AREA)
- Computing Systems (AREA)
- Software Systems (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Molecular Biology (AREA)
- Life Sciences & Earth Sciences (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Mathematical Physics (AREA)
- Data Mining & Analysis (AREA)
- General Engineering & Computer Science (AREA)
- Computational Linguistics (AREA)
- Databases & Information Systems (AREA)
- Medical Informatics (AREA)
- Biodiversity & Conservation Biology (AREA)
- Image Analysis (AREA)
- Character Discrimination (AREA)
- Traffic Control Systems (AREA)
Abstract
Description
类型 | 参数 |
Input | GrayScaleImage,W:220,H:32 |
Convolution1 | Filters:8,K=3,S=1,P=1x1 |
Leaky ReLU | - |
Max pooling1 | K=2,S=2,P=0*0 |
Convolution1 | Filters:16,K=3,S=1,P=1x1 |
Leaky ReLU | - |
Convolution1 | Filters:32,K=3,S=1,P=1x1 |
Leaky ReLU | - |
FC Layer | Filters:64 |
Leaky ReLU | - |
Dropout | - |
FC Layer | Filters:6 |
Claims (15)
- 一种车牌字符识别方法,其特征在于,所述方法包括:获取图像采集设备所采集到的车辆图像;对车辆图像进行车牌区域定位,以得到车牌图像;通过卷积神经网络对所述车牌图像进行特征提取,以得到所述车牌图像的特征信息,其中,所述卷积神经网络包括残差网络结构;通过双向循环神经网络模型解析所述特征信息,以得到所述车牌图像对应的车牌字符。
- 根据权利要求1所述的方法,其特征在于,所述卷积神经网络包括顺次连接的卷积层、所述残差网络结构和池化层。
- 根据权利要求2所述的方法,其特征在于,所述卷积层包括3层,所述残差网络结构包括第一残差网络结构和第二残差网络结构,所述卷积神经网络包括顺次相连的第一卷积层、第一残差网络结构、第二残差网络结构、第二卷积层、第三卷积层和池化层。
- 根据权利要求3所述的方法,其特征在于,所述第一残差网络结构和所述第二残差网络结构的网络结构相同,所述网络结构包括第一网络子结构、第二网络子结构和加法器,所述第一网络子结构包括顺次相连的最大池化层、第一卷积子层、第二卷积子层、第三卷积子层和第四卷积子层,所述第二网络子结构包括第五卷积子网络,所述最大池化层和所述第五卷积子网络的输入端均与所述第一卷积层的输出层的输出端连接,所述第四卷积子层和所述第五卷积子网络的输出端均与加法器的输入端相连。
- 根据权利要求1-4中任一项所述的方法,其特征在于,在所述通过卷积神经网络对所述车牌图像进行特征提取,以得到所述车牌图像的特征信息之前,所述方法还包括:利用空间变换网络,对所述车牌图像进行倾斜校正。
- 根据权利要求1-5中任一项所述的方法,其特征在于,所述双向循环神经网络模型包括双向循环神经网络和连接时序分类CTC网络。
- 根据权利要求6所述的方法,其特征在于,所述双向循环神经网络包括顺次连接的第一层双向长短时记忆网络和第二层双向长短时记忆网络。
- 一种车牌字符识别装置,其特征在于,所述装置包括:图像采集模块,用于获取图像采集设备所采集到的车辆图像;车牌定位模块,用于对车辆图像进行车牌区域定位,以得到车牌图像;特征提取模块,用于通过卷积神经网络对所述车牌图像进行特征提取,以得到所述车牌图像的特征信息,其中,所述卷积神经网络包括残差网络结构;识别模块,用于通过双向循环神经网络模型解析所述特征信息,以得到所述车牌图像 对应的车牌字符。
- 根据权利要求8所述的装置,其特征在于,所述卷积神经网络包括顺次连接的卷积层、所述残差网络结构和池化层。
- 根据权利要求9所述的装置,其特征在于,所述卷积层包括3层,所述残差网络结构包括第一残差网络结构和第二残差网络结构,所述卷积神经网络包括顺次相连的第一卷积层、第一残差网络结构、第二残差网络结构、第二卷积层、第三卷积层和池化层。
- 根据权利要求10所述的装置,其特征在于,所述第一残差网络结构和所述第二残差网络结构的网络结构相同,所述网络结构包括第一网络子结构、第二网络子结构和加法器,所述第一网络子结构包括顺次相连的最大池化层、第一卷积子层、第二卷积子层、第三卷积子层和第四卷积子层,所述第二网络子结构包括第五卷积子网络,所述最大池化层和所述第五卷积子网络的输入端均与所述第一卷积层的输出层的输出端连接,所述第四卷积子层和所述第五卷积子网络的输出端均与加法器的输入端相连。
- 根据权利要求8-11中任一项所述的装置,其特征在于,所述双向循环神经网络模型包括双向循环神经网络和连接时序分类CTC网络。
- 根据权利要求12所述的装置,其特征在于,所述双向循环神经网络包括顺次连接的第一层双向长短时记忆网络和第二层双向长短时记忆网络。
- 一种电子设备,包括:存储器,处理器;所述存储器中存储有计算机指令,当所述计算机指令被所述处理器执行时,实现如权利要求1-7中任一项所述的车牌字符识别方法。
- 一种存储有计算机指令的非瞬时计算机可读存储介质,其特征在于,所述计算机指令用于使所述计算机执行权利要求1-7中任一项所述的车牌字符识别方法。
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US17/914,993 US20230154210A1 (en) | 2020-03-30 | 2021-02-02 | Method for recognizing license plate characters, electronic device and storage medium |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010234926.2A CN111444917A (zh) | 2020-03-30 | 2020-03-30 | 车牌字符识别方法、装置、电子设备和存储介质 |
CN202010234926.2 | 2020-03-30 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2021196873A1 true WO2021196873A1 (zh) | 2021-10-07 |
Family
ID=71654027
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2021/074915 WO2021196873A1 (zh) | 2020-03-30 | 2021-02-02 | 车牌字符识别方法、装置、电子设备和存储介质 |
Country Status (3)
Country | Link |
---|---|
US (1) | US20230154210A1 (zh) |
CN (1) | CN111444917A (zh) |
WO (1) | WO2021196873A1 (zh) |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114092926A (zh) * | 2021-10-20 | 2022-02-25 | 杭州电子科技大学 | 一种复杂环境下的车牌定位和识别方法 |
CN114299486A (zh) * | 2021-12-30 | 2022-04-08 | 郑州信大先进技术研究院 | 嵌入式端实时高效的车牌识别方法、设备及介质 |
CN114495106A (zh) * | 2022-04-18 | 2022-05-13 | 电子科技大学 | 一种应用于dfb激光器芯片的深度学习mocr方法 |
CN116958954A (zh) * | 2023-07-27 | 2023-10-27 | 匀熵智能科技(无锡)有限公司 | 基于关键点与旁路矫正的车牌识别方法、装置及存储介质 |
CN117523543A (zh) * | 2024-01-08 | 2024-02-06 | 成都大学 | 一种基于深度学习的金属压印字符识别方法 |
Families Citing this family (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111444917A (zh) * | 2020-03-30 | 2020-07-24 | 合肥京东方显示技术有限公司 | 车牌字符识别方法、装置、电子设备和存储介质 |
CN112149661A (zh) * | 2020-08-07 | 2020-12-29 | 珠海欧比特宇航科技股份有限公司 | 车牌识别方法、装置及介质 |
CN112016432A (zh) * | 2020-08-24 | 2020-12-01 | 高新兴科技集团股份有限公司 | 基于深度学习的车牌字符识别方法、存储介质及电子设备 |
CN112215245A (zh) * | 2020-11-05 | 2021-01-12 | 中国联合网络通信集团有限公司 | 一种图像识别方法及装置 |
CN112966631A (zh) * | 2021-03-19 | 2021-06-15 | 浪潮云信息技术股份公司 | 一种无限制安防场景下的车牌检测识别系统及其方法 |
CN113705561A (zh) * | 2021-09-02 | 2021-11-26 | 北京云蝶智学科技有限公司 | 特殊符号的识别方法及装置 |
CN117409298B (zh) * | 2023-12-15 | 2024-04-02 | 西安航空学院 | 针对路面车辆识别的多尺寸目标精确识别方法及设备 |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109815956A (zh) * | 2018-12-28 | 2019-05-28 | 安徽清新互联信息科技有限公司 | 一种基于自适应位置分割的车牌字符识别方法 |
CN110321755A (zh) * | 2018-03-28 | 2019-10-11 | 中移(苏州)软件技术有限公司 | 一种识别方法及装置 |
CN110414451A (zh) * | 2019-07-31 | 2019-11-05 | 深圳市捷顺科技实业股份有限公司 | 一种基于端对端的车牌识别方法、装置、设备及存储介质 |
CN111444917A (zh) * | 2020-03-30 | 2020-07-24 | 合肥京东方显示技术有限公司 | 车牌字符识别方法、装置、电子设备和存储介质 |
Family Cites Families (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110363049B (zh) * | 2018-04-10 | 2024-01-12 | 阿里巴巴集团控股有限公司 | 图形元素检测识别和类别确定的方法及装置 |
CN109934864B (zh) * | 2019-03-14 | 2023-01-20 | 东北大学 | 面向机械臂抓取位姿估计的残差网络深度学习方法 |
-
2020
- 2020-03-30 CN CN202010234926.2A patent/CN111444917A/zh active Pending
-
2021
- 2021-02-02 US US17/914,993 patent/US20230154210A1/en active Pending
- 2021-02-02 WO PCT/CN2021/074915 patent/WO2021196873A1/zh active Application Filing
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110321755A (zh) * | 2018-03-28 | 2019-10-11 | 中移(苏州)软件技术有限公司 | 一种识别方法及装置 |
CN109815956A (zh) * | 2018-12-28 | 2019-05-28 | 安徽清新互联信息科技有限公司 | 一种基于自适应位置分割的车牌字符识别方法 |
CN110414451A (zh) * | 2019-07-31 | 2019-11-05 | 深圳市捷顺科技实业股份有限公司 | 一种基于端对端的车牌识别方法、装置、设备及存储介质 |
CN111444917A (zh) * | 2020-03-30 | 2020-07-24 | 合肥京东方显示技术有限公司 | 车牌字符识别方法、装置、电子设备和存储介质 |
Cited By (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114092926A (zh) * | 2021-10-20 | 2022-02-25 | 杭州电子科技大学 | 一种复杂环境下的车牌定位和识别方法 |
CN114299486A (zh) * | 2021-12-30 | 2022-04-08 | 郑州信大先进技术研究院 | 嵌入式端实时高效的车牌识别方法、设备及介质 |
CN114299486B (zh) * | 2021-12-30 | 2024-04-16 | 郑州信大先进技术研究院 | 嵌入式端实时高效的车牌识别方法、设备及介质 |
CN114495106A (zh) * | 2022-04-18 | 2022-05-13 | 电子科技大学 | 一种应用于dfb激光器芯片的深度学习mocr方法 |
CN116958954A (zh) * | 2023-07-27 | 2023-10-27 | 匀熵智能科技(无锡)有限公司 | 基于关键点与旁路矫正的车牌识别方法、装置及存储介质 |
CN116958954B (zh) * | 2023-07-27 | 2024-03-22 | 匀熵智能科技(无锡)有限公司 | 基于关键点与旁路矫正的车牌识别方法、装置及存储介质 |
CN117523543A (zh) * | 2024-01-08 | 2024-02-06 | 成都大学 | 一种基于深度学习的金属压印字符识别方法 |
CN117523543B (zh) * | 2024-01-08 | 2024-03-19 | 成都大学 | 一种基于深度学习的金属压印字符识别方法 |
Also Published As
Publication number | Publication date |
---|---|
CN111444917A (zh) | 2020-07-24 |
US20230154210A1 (en) | 2023-05-18 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2021196873A1 (zh) | 车牌字符识别方法、装置、电子设备和存储介质 | |
US9965719B2 (en) | Subcategory-aware convolutional neural networks for object detection | |
US20220092351A1 (en) | Image classification method, neural network training method, and apparatus | |
WO2021238281A1 (zh) | 一种神经网络的训练方法、图像分类系统及相关设备 | |
CN108898086B (zh) | 视频图像处理方法及装置、计算机可读介质和电子设备 | |
WO2020164270A1 (zh) | 基于深度学习的行人检测方法、系统、装置及存储介质 | |
CN109241982B (zh) | 基于深浅层卷积神经网络的目标检测方法 | |
WO2021244621A1 (zh) | 基于全局引导选择性上下文网络的场景语义解析方法 | |
WO2019201035A1 (zh) | 对图像中的对象节点的识别方法、装置、终端及计算机可读存储介质 | |
JP2022515895A (ja) | 物体認識方法及び装置 | |
CN109766805B (zh) | 一种基于深度学习的双层车牌字符识别方法 | |
US20220148291A1 (en) | Image classification method and apparatus, and image classification model training method and apparatus | |
CN111428664B (zh) | 一种基于深度学习技术的计算机视觉的实时多人姿态估计方法 | |
WO2021047587A1 (zh) | 手势识别方法、电子设备、计算机可读存储介质和芯片 | |
CN113537211B (zh) | 一种基于非对称iou的深度学习车牌框定位方法 | |
CN112307853A (zh) | 航拍图像的检测方法、存储介质和电子装置 | |
WO2024001123A1 (zh) | 基于神经网络模型的图像识别方法、装置及终端设备 | |
CN109710788A (zh) | 图像样本标注和管理方法及设备 | |
WO2024077781A1 (zh) | 基于卷积神经网络模型的图像识别方法、装置及终端设备 | |
CN115862045B (zh) | 基于图文识别技术的病例自动识别方法、系统、设备及存储介质 | |
CN111209873A (zh) | 一种基于深度学习的高精度人脸关键点定位方法及系统 | |
US20220198224A1 (en) | Face recognition method, terminal device using the same, and computer readable storage medium | |
CN116229406B (zh) | 车道线检测方法、系统、电子设备及存储介质 | |
JP7246104B2 (ja) | テキスト行識別に基づくナンバープレート識別方法 | |
CN114067186B (zh) | 一种行人检测方法、装置、电子设备及存储介质 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 21779976 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 21779976 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 15.05.2023) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 21779976 Country of ref document: EP Kind code of ref document: A1 |