WO2016155371A1 - 识别交通标志的方法和装置 - Google Patents
识别交通标志的方法和装置 Download PDFInfo
- Publication number
- WO2016155371A1 WO2016155371A1 PCT/CN2015/098903 CN2015098903W WO2016155371A1 WO 2016155371 A1 WO2016155371 A1 WO 2016155371A1 CN 2015098903 W CN2015098903 W CN 2015098903W WO 2016155371 A1 WO2016155371 A1 WO 2016155371A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- traffic sign
- window image
- neural network
- trained
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 54
- 238000004422 calculation algorithm Methods 0.000 claims abstract description 79
- 238000001514 detection method Methods 0.000 claims abstract description 52
- 238000013527 convolutional neural network Methods 0.000 claims description 75
- 230000010354 integration Effects 0.000 claims description 40
- 238000012549 training Methods 0.000 claims description 32
- 230000006870 function Effects 0.000 claims description 22
- 238000000605 extraction Methods 0.000 claims description 13
- 238000010606 normalization Methods 0.000 claims description 7
- 238000003062 neural network model Methods 0.000 claims 2
- 238000013528 artificial neural network Methods 0.000 abstract 2
- 238000010586 diagram Methods 0.000 description 16
- 238000012545 processing Methods 0.000 description 10
- 238000005516 engineering process Methods 0.000 description 7
- 238000004590 computer program Methods 0.000 description 5
- 238000004891 communication Methods 0.000 description 4
- 230000008569 process Effects 0.000 description 4
- 238000003708 edge detection Methods 0.000 description 3
- 230000009466 transformation Effects 0.000 description 3
- 238000004364 calculation method Methods 0.000 description 2
- 238000012360 testing method Methods 0.000 description 2
- 238000011161 development Methods 0.000 description 1
- 230000018109 developmental process Effects 0.000 description 1
- 238000003709 image segmentation Methods 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 239000003550 marker Substances 0.000 description 1
- 210000002569 neuron Anatomy 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/56—Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
- G06V20/58—Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads
- G06V20/582—Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads of traffic signs
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/30—Determination of transform parameters for the alignment of images, i.e. image registration
- G06T7/33—Determination of transform parameters for the alignment of images, i.e. image registration using feature-based methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20021—Dividing image into blocks, subimages or windows
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
Definitions
- the present application relates to the field of computer technologies, and in particular, to the field of computer object recognition technologies, and in particular, to a method and apparatus for identifying a traffic sign.
- the traditional method of identifying traffic signs relies mainly on manual processing.
- the specific program software and basemap data are deployed to the collection vehicle in advance, and the trained field personnel work with the vehicle: after observing the traffic signage, the operator will The type and visual distance are manually entered into the software; after the field collection is completed, the internal staff compares the base map before and after the operation to update the valid information to the road network database.
- This process requires the support of a variety of software tools, but the accuracy of the data results mainly depends on the individual ability and concentration of the operator, while the complex acquisition process reduces the update efficiency of the road network data.
- the purpose of the present application is to propose a method and apparatus for identifying a traffic sign to solve the technical problems mentioned in the background section above.
- the present application provides a method for identifying a traffic sign, the method comprising: acquiring a feature value obtained by a predetermined feature channel by a predetermined feature algorithm by a scan window image obtained by dividing a panoramic spherical image; according to the scan window image The eigenvalue and the pre-trained detection classifier model, detecting the scanned window image, and obtaining the traffic sign window to be confirmed An image, wherein the pre-trained detection classifier model is trained according to samples of the scan window image and its eigenvalues; according to the pre-trained convolutional neural network model, the traffic sign window image to be confirmed is identified, and the traffic sign category is obtained, wherein The pre-trained convolutional neural network model is trained based on the samples of the confirmed traffic sign window image and its traffic sign categories.
- the present application provides an apparatus for identifying a traffic sign, the apparatus comprising: a feature value acquiring module, configured to acquire a feature of a scan window image obtained by dividing a panoramic spherical image by a predetermined feature algorithm in a predetermined integration channel; a traffic sign detecting module, configured to detect a scan window image according to a feature value of the scan window image and a pre-trained test classifier model, to obtain a to-be-confirmed traffic sign window image, wherein the pre-trained detection classifier The model is trained according to the sample of the scan window image and its characteristic value; the traffic sign recognition module is configured to identify the image of the traffic sign window to be confirmed according to the pre-trained convolutional neural network model, and obtain a traffic sign category, wherein the pre-training The convolutional neural network model is trained based on the samples of the confirmed traffic sign window image and its traffic sign categories.
- the method and apparatus for identifying a traffic sign obtained by the present application obtains a feature value obtained by a predetermined feature algorithm by a predetermined feature algorithm by acquiring a scan window image obtained by dividing a panoramic spherical image, and then according to the feature value of the scan window image and pre-training
- the detection classifier model detects the scanning window image, obtains the image of the traffic sign window to be confirmed, and then identifies the image of the traffic sign window to be confirmed according to the pre-trained convolutional neural network model, obtains the traffic sign category, and realizes the detection through the pre-training.
- the classifier model detects the image of the traffic sign window to be confirmed, identifies the traffic sign category through the pre-trained convolutional neural network model, improves the accuracy of detecting and identifying the traffic sign in the panoramic image, and improves the update efficiency of the road network data. .
- FIG. 1 illustrates an exemplary flow chart of a method of identifying a traffic sign in accordance with an embodiment of the present application
- FIG. 2 illustrates an exemplary flow chart of a method of training a pre-trained detection model in accordance with an embodiment of the present application
- FIG. 3 illustrates an exemplary flowchart of a method of acquiring an image of a traffic sign window to be confirmed according to an embodiment of the present application
- FIG. 4 illustrates an exemplary flow chart of a method of training a pre-trained convolutional neural network model in accordance with an embodiment of the present application
- FIG. 5 illustrates an exemplary structural diagram of a preset convolutional neural network model in accordance with an embodiment of the present application
- FIG. 6 is a diagram showing an example of the structure of an apparatus for identifying a traffic sign according to an embodiment of the present application
- FIG. 7 shows an exemplary structural diagram of a detection classifier model training device according to an embodiment of the present application.
- FIG. 8 shows an exemplary structural diagram of an apparatus for acquiring an image of a traffic sign window to be confirmed
- FIG. 9 illustrates an exemplary structural diagram of an apparatus for training a pre-trained convolutional neural network model in accordance with an embodiment of the present application
- FIG. 10 is a schematic structural diagram of a computer system according to an embodiment of the present application.
- FIG. 1 illustrates an exemplary flow chart of a method of identifying a traffic sign in accordance with an embodiment of the present application.
- a method 100 of identifying a traffic sign can include:
- Step 110 Acquire a feature value obtained by the predetermined feature channel by a predetermined feature algorithm of the scan window image obtained by dividing the panoramic spherical image.
- the panoramic spherical image that needs to be identified by the traffic sign can be first obtained, and then the panoramic spherical image is scanned by the window of the preset size to obtain the scanned window image, and then the scanned window image is acquired through the predetermined feature algorithm in the predetermined integration channel. The resulting feature value.
- the panoramic spherical image refers to an image that can be displayed by combining a plurality of fisheye images.
- a window of a preset size is usually a window that is preset according to the size of the target object.
- the predetermined integration channel may be any one or more integration channels in the prior art in the field of image processing. Any one or more of the integration channels that may be developed in the future technology; the predetermined feature algorithm may be any one or more of the feature algorithms in the prior art in the field of image processing, or may be any possible development in the future technology. One or more feature algorithms.
- the application does not limit the specific manner of obtaining the feature value, and the user can select according to the actual use requirement.
- a haar eigenvalue can be obtained by a Hal algorithm in a gray image channel
- a histogram eigenvalue can be obtained by a histogram eigenvalue algorithm at a gradient image channel of different angle parameters, while being in red, green, and blue monochrome.
- the image channel obtains a random pair of feature values and the like by randomly eigenvalue algorithm.
- the feature value of the scan window image may be acquired by the integral map: optionally, in the above embodiment, step 110 may include: step 111, acquiring the image by the panoramic spherical image at a predetermined time The integral map of the integration channel divides the integrated map of the scanned window image; and in step 112, the feature value of the scanned window image is obtained according to the integral map of the scanned window image.
- the feature value of the scan window image can be accelerated, and the calculation efficiency of the feature value of the scan window image can be improved.
- the sample of the traffic sign window image may be determined according to the traffic sign category.
- the feature value and a predetermined integration channel and a predetermined feature algorithm for acquiring the feature values of the scanned window image For example, for the ban class Marks such as speed limit can enhance the description on the red channel.
- the predetermined integration channel may include one or more of the following: a grayscale image channel, a red, green, and blue monochrome image channel, a gradient image channel of different angle parameters, and an edge detection image channel; and a predetermined feature algorithm may It includes one or more of the following: Haar feature algorithm, local binary pattern feature algorithm, histogram feature algorithm and random pair feature algorithm.
- a predetermined feature algorithm may It includes one or more of the following: Haar feature algorithm, local binary pattern feature algorithm, histogram feature algorithm and random pair feature algorithm.
- the predetermined integral channel and the predetermined feature algorithm for acquiring the feature value of the sample of the traffic sign window image and the feature value of the scan window image determined according to the traffic sign category can describe the characteristics of the target from a plurality of different angles, thereby overcoming the target because Changes from different angles and lighting.
- Step 120 Detecting a scan window image according to the feature value of the scan window image and the pre-trained detection classifier model, and obtaining a traffic sign window image to be confirmed, wherein the pre-trained detection classifier model is based on the sample of the scan window image and its feature Value training is obtained.
- the pre-trained detection classifier model first receives the manual annotation, and determines a scan window image containing the traffic sign and a scan window image not including the traffic sign as the sample of the traffic sign window image in the scan window image, and then acquires the sample.
- the eigenvalue is then used to train the parameters of the classifier model set according to the actual demand using the sample and its eigenvalues, thereby obtaining a pre-trained detection classifier model.
- the scan window image may be detected according to the feature value of the scan window image and the pre-trained detection classifier model to obtain a window in which the traffic sign is suspected.
- the image serves as a window image of the traffic sign to be confirmed.
- Step S130 identifying a traffic sign window image to be confirmed according to the pre-trained convolutional neural network model, and obtaining a traffic sign category, wherein the pre-trained convolutional neural network model is based on the confirmed traffic sign window image sample and its traffic sign Category training is available.
- the pre-trained convolutional neural network model first receives the manual annotation, and divides the sample of the to-be-confirmed traffic sign window image detected by the detection classifier model. Class, marked as a specific traffic sign category or non-traffic category, and then use the sample of the traffic sign window image to be confirmed and its traffic sign category to train the convolutional neural network model set according to actual needs, thus obtaining pre-training Convolutional neural network model.
- the traffic sign window image to be confirmed may be identified according to the to-be-confirmed traffic sign window image and the pre-trained convolutional neural network model to obtain the traffic to be confirmed.
- the traffic sign category of the maximum probability corresponding to the flag window image is used as the identified traffic sign category.
- the method for identifying a traffic sign in the above embodiment of the present application improves the accuracy of detecting and identifying a traffic sign in a panoramic image, and improves the update efficiency of the road network data.
- FIG. 2 illustrates an exemplary flow chart of a method of training a pre-trained detection model in accordance with an embodiment of the present application.
- the method 200 of training a pre-trained detection classifier model can include:
- a positive sample and a negative sample in a sample of the scan window image are acquired, wherein the positive sample includes a window image of the traffic sign or a window image including the traffic sign and the surrounding extended preset pixels, and the negative sample includes the positive sample removed. Scan the window image.
- the positive sample and the negative sample in the sample of the scan window image are obtained by first acquiring the sample in the scan window image, and then including the window image of the traffic sign in the sample of the scan window image or including the traffic according to the received manual annotation.
- the window image of the marker and its surrounding preset pixels is used as a positive sample, and the scan window image in the sample from which the positive sample is removed is taken as a negative sample.
- the above positive sample can improve the efficiency of calculating the feature value of the positive sample when only the window image of the traffic sign is included. Considering that the pixel around the target object can also effectively describe the target object itself, the positive sample can improve the accuracy of describing the feature value of the positive sample when the window image of the preset pixel is expanded including the traffic sign and the surrounding.
- step 202 the feature values obtained by the predetermined feature and the negative sample in the predetermined integration channel by the predetermined feature algorithm are obtained.
- the predetermined integration channel when acquiring the feature values obtained by the predetermined feature algorithm in the predetermined integration channel by the positive sample and the negative sample, the predetermined integration channel may be in the prior art in the image processing field. Any one or more of the integration channels may also be any one or more integration channels that may be developed in the future technology; the predetermined feature algorithm may be any one or more of the prior art algorithms in the image processing field. It can also be any one or more feature algorithms that may be developed in future technologies.
- the present application does not limit the specific pre-integration channel and the predetermined feature algorithm for acquiring feature values, which can be selected by the user according to actual use requirements.
- the haar eigenvalues obtained by the Haar feature algorithm in the gray image channel the histogram eigenvalues obtained by the gradient image channel histogram feature algorithm at different angle parameters, in the red, green, and blue monochrome images
- the channel passes the random pair of feature values obtained by the feature algorithm randomly.
- the traffic sign window may be determined according to the traffic sign category.
- a predetermined integration channel and a predetermined feature algorithm for the samples of the image For example, for ban-like signs such as speed limit, the description on the red channel can be enhanced.
- the predetermined integration channel may include one or more of the following: a grayscale image channel, a red, green, and blue monochrome image channel, a gradient image channel of different angle parameters, and an edge detection image channel; and the predetermined feature algorithm may include One or more of the following: a Hal feature algorithm, a local binary pattern feature algorithm, a histogram feature algorithm, and a random pair feature algorithm. It should be understood by those skilled in the art that the predetermined integration channel and the predetermined feature algorithm in the above embodiments may be combined as needed to obtain the feature values obtained by the predetermined feature and the negative sample in the predetermined integration channel by the predetermined feature algorithm.
- step 203 the classifier model is trained by the boosting algorithm according to the positive sample and the negative sample and the feature values obtained by the predetermined feature algorithm in the predetermined integration channel, and the pre-trained detection classifier model is obtained.
- the classifier model can be trained by the boosting algorithm to obtain a high-accuracy result classifier as the pre-training. Detection classifier model.
- FIG. 3 illustrates an exemplary flow chart of a method of acquiring a to-be-confirmed traffic sign window image in accordance with an embodiment of the present application.
- the method 300 for obtaining an image of a traffic sign window to be confirmed includes:
- step 301 the panoramic spherical image is continuously downsampled to obtain an image pyramid.
- the panoramic spherical image may be subjected to pyramid transformation, such as Laplacian pyramid transformation or the like, to transform the panoramic spherical image from the original size to a different scale space, thereby obtaining an image pyramid including the multi-level image.
- pyramid transformation such as Laplacian pyramid transformation or the like
- Step 302 Acquire an integral graph of the scan window image of each level image in the image pyramid obtained by dividing the image pyramid in the integral map of the predetermined integration channel.
- each level image of the image pyramid may be first integrated in a predetermined integration channel to obtain an integral map of the image pyramid, and then the integral map of the image pyramid is divided by a window of a preset size to obtain An integral map of the scanned window image for each level of image.
- Step 303 Acquire feature values of the scan window image of each level image according to the integral map of the scan window image of each level image.
- the integral map of the scan window image of each level image may be used to perform the operation according to the integral map of the scan window image of each level image, and the image of each level is quickly obtained. Scans the feature values of the window image.
- Step 304 Detect a scan window image of each level image according to a feature value of the scan window image of each level image and a pre-trained detection classifier model, and obtain a traffic sign window image to be confirmed.
- the scan window image of each level image may be detected by the pre-trained detection classifier model according to the feature value of the scan window image of each level image, and the obtained The suspected traffic sign window image is used as the to-be-confirmed traffic sign window image, thereby obtaining all the to-be-confirmed traffic sign window images in the image pyramid.
- the method for obtaining a to-be-confirmed traffic sign window image in the above embodiment of the present application acquires a scan window image in a pyramid image acquired by a pyramidal image by a pyramid transformation, and then detects a scan window image of each level image to obtain a to-be-confirmed traffic sign window image.
- the possibility of missing the traffic sign window to be confirmed in the panoramic spherical image is reduced, thereby improving the accuracy of obtaining the traffic sign window to be confirmed.
- the integral value is used to acquire the feature value of the scan window image of each level image, the acquisition speed of the feature value is improved.
- FIG. 4 illustrates an exemplary flow chart of a method of training a pre-trained convolutional neural network model in accordance with an embodiment of the present application.
- a method 400 of training a pre-trained convolutional neural network model includes:
- Step 401 Initialize weights of the convolutional layer and the fully connected layer of the preset convolutional neural network model according to the Gaussian distribution, where the preset convolutional neural network model includes convolution layers, extraction layers, and Fully connected layer and normalized layer.
- the successively connected convolution layer, extraction layer, fully connected layer, and normalization layer may include: one convolution layer, one extraction layer, one full connection layer, and one normalization layer; or may include multiple volumes The layer and the extraction layer corresponding to the convolution layer, one or more fully connected layers and one normalized layer.
- the convolution layer set therein can enhance the original signal feature and reduce the noise by convolution operation; the set extraction layer can use the principle of image local correlation to subsample the image while reducing the amount of data processing. Keep useful information.
- FIG. 5 shows an exemplary structural diagram of a preset convolutional neural network model in accordance with an embodiment of the present application.
- the preset convolutional neural network model 500 includes: a convolutional layer conv1, an extraction layer pool1, a convolutional layer conv2, an extraction layer pool2, a convolution layer conv3, an extraction layer pool3, and a full connection layer.
- Fc1 fully connected layer fc2.
- conv1 has 16 convolution kernels with size 5*5*3
- conv2 has 32 convolution kernels with size 5*5*16
- conv3 has 64 convolution kernels with size 5*5*32.
- the two fully connected layers each have 512 and 120 neurons.
- the preset convolutional neural network model may further include a loss function layer.
- the loss function can be calculated using a multi-class logarithmic loss logarithmic loss function, ie
- L is the loss function
- N is the number of training samples
- M is the number of categories
- p i,j is the probability that the i-th sample of the network output is j-th class
- y i,j is the true value of the sample, if the ith sample It belongs to the jth class, it is 1, otherwise it is 0.
- step 402 according to the confirmed sample of the traffic sign window image and its traffic sign category, the weight of both the convolutional layer and the fully connected layer is iterated by the error backpropagation BP algorithm.
- the sample of the confirmed traffic sign window image is a sample of the window image including the traffic sign confirmed according to the received manual mark in the sample extracted from the image of the traffic sign window to be confirmed.
- the sample of the confirmed traffic sign window image may include: a sample of the original confirmed traffic sign window image, and normalizing the sample of the original confirmed traffic sign window image to the preset window after one or more of the following processing Dimensional images: rotate, pan, and zoom.
- iterating through the error back propagation BP algorithm for the weights of both the convolution layer and the fully connected layer may include: through the loss function and the BP algorithm, the convolution layer and the full connection layer
- the weights of the individuals are iterated separately. For example, after initializing the weight of the Gaussian distribution, the network model is iterated multiple times by using a stochastic gradient descent algorithm. Each time, the loss function L of the network is calculated by the loss function formula (a), and then reversed. The layer calculates the gradient of L relative to each layer weight W i , and then updates the weight W i according to the gradient of each layer weight W i , ie
- ⁇ is the learning rate of the preset gradient descent, It is the gradient of the loss function L with respect to the weight W i .
- the convolutional neural network model of the optimal weight may be determined by step S403 or step S404.
- Step S403 If the difference between the weight after the current iteration and the weight after the last iteration is less than a preset value, the weight after the current iteration is determined as the optimal weight.
- Step S404 If the weight after the iteration with the lowest error rate occurs, the weight after the iteration with the lowest error rate is determined as the optimal weight.
- Step S405 The convolutional neural network model including the optimal weight is set as a pre-trained convolutional neural network model.
- step 103 in FIG. 1 may include: inputting the image of the traffic sign window to be confirmed into the convolutional neural network model to obtain The traffic sign category with the largest weight output from the normalized layer; the traffic sign category with the largest weight is set as the identified traffic sign category.
- FIG. 6 shows an example of the structure of an apparatus for identifying a traffic sign according to an embodiment of the present application.
- the apparatus 600 for identifying a traffic sign may include: a feature value acquisition module 610, a traffic sign detection module 620, and a traffic sign recognition module 630.
- the feature value obtaining module 610 is configured to acquire the feature value obtained by the predetermined feature channel by the predetermined feature channel by the scan window image obtained by the panoramic spherical image segmentation.
- the feature value obtaining module 610 may include: an integral map acquiring first sub-module 611, configured to acquire an integral map of the scan window image obtained by dividing the panoramic spherical image in the integral map of the predetermined integration channel; and acquiring the feature value A sub-module 612 is configured to acquire a feature value of the scan window image according to the integral map of the scan window image.
- the traffic sign detecting module 620 is configured to detect a scan window image according to the feature value of the scan window image and the pre-trained test classifier model, and obtain a to-be-confirmed traffic sign window image, wherein the pre-trained detection classifier model is based on the scan window image
- the samples and their eigenvalues are trained.
- the traffic sign recognition module 630 is configured to identify a traffic sign window image to be confirmed according to the pre-trained convolutional neural network model, and obtain a traffic sign category, wherein the pre-trained convolutional neural network model is based on the confirmed traffic sign window image.
- the sample and its traffic sign categories were trained.
- the device for identifying a traffic sign in the above embodiment of the present application improves the accuracy of detecting and identifying a traffic sign in a panoramic image, and improves the update efficiency of the road network data.
- the pre-trained detection classifier model described above may be obtained by acquiring a positive sample and a negative sample in a sample of the scan window image, wherein the positive sample includes a window image of the traffic sign or includes a traffic sign and its surrounding expansion a window image of a preset pixel, the negative sample includes a scan window image from which the positive sample is removed; the feature value obtained by the predetermined feature algorithm in the predetermined integration channel by the positive sample and the negative sample; and the positive sample and the negative sample and the The feature value obtained by the predetermined feature algorithm is trained by the boosting algorithm to detect the classifier model, and a pre-trained detection classifier model is obtained.
- the above method for obtaining a pre-trained detection classifier model can be illustrated by FIG.
- the detection classifier model training device is implemented.
- FIG. 7 shows an exemplary structural diagram of a detection classifier model training device according to an embodiment of the present application.
- the detection classifier model training apparatus 700 may include a sample acquisition module 701, a sample feature value acquisition module 702, and a detection model training module 703.
- the sample obtaining module 701 is configured to obtain a positive sample and a negative sample in the sample of the scan window image, where the positive sample includes a window image of the traffic sign or a window image including the traffic sign and the surrounding extended preset pixels, and the negative sample includes the positive sample The scanned window image of the sample.
- the sample feature value obtaining module 702 is configured to obtain the feature values obtained by the predetermined feature channel by the predetermined feature algorithm in the predetermined integration channel.
- the predetermined integration channel includes one or more of the following: a grayscale image channel, a red, green, and blue monochrome image channel, a gradient image channel with different angle parameters, and an edge detection image channel; and the predetermined feature algorithm includes the following one Or multiple: Hal feature algorithm, local binary pattern feature algorithm, histogram feature algorithm and random pair feature algorithm.
- the detection model training module 703 is configured to train the detection classifier model by using a boosting algorithm according to the positive sample and the negative sample and the feature values obtained by the predetermined feature algorithm in the predetermined integration channel, to obtain a pre-trained detection classifier model.
- FIG. 8 shows an exemplary structural diagram of an apparatus for acquiring an image of a traffic sign window to be confirmed.
- the apparatus 800 for acquiring a to-be-confirmed traffic sign window image may include: a downsampling sub-module 801, an integral map acquisition second sub-module 802, a feature value second acquisition sub-module 803, and a multi-level detection sub-module. 804.
- the downsampling sub-module 801 is configured to continuously downsample the panoramic spherical image to obtain an image pyramid.
- the integral map acquires a second sub-module 802 for acquiring an integral map of the scan window image of each level image in the image pyramid obtained by dividing the image pyramid by the integral map of the predetermined integration channel.
- Feature value second acquisition sub-module 803 for scanning window image according to each level of image The integral map of the image of the scanned window image of each level of image is obtained.
- the multi-level detection sub-module 804 is configured to detect a scan window image of each level image according to the feature value of the scan window image of each level image and the pre-trained detection classifier model, and obtain a to-be-confirmed traffic sign window image.
- the device for obtaining the image of the traffic sign window to be confirmed by the foregoing embodiment of the present application can further improve the accuracy of obtaining the image of the traffic sign window to be confirmed and the speed of detecting the image of the traffic sign window to be confirmed.
- the pre-trained convolutional neural network model in FIG. 6 above may be obtained by: initializing the weights of both the convolutional layer and the fully connected layer of the preset convolutional neural network model according to a Gaussian distribution, wherein The preset convolutional neural network model includes a convolution layer, an extraction layer, a full connection layer, and a normalization layer connected in sequence; according to the confirmed sample of the traffic sign window image and its traffic sign category, the error is reversed
- the propagation BP algorithm iterates the weights of the convolution layer and the fully connected layer.
- the current iteration is The weight is determined as the optimal weight, and the convolutional neural network model including the optimal weight is set as the pre-trained convolutional neural network model; or if the weight after the iteration with the lowest error rate occurs, the error rate is the lowest.
- the iterative weight is determined as the optimal weight, and the convolutional neural network model including the optimal weight is set as the pre-trained convolutional neural network model.
- the sample of the confirmed traffic sign window image is a sample of the window image including the traffic sign confirmed according to the received manual mark in the sample extracted from the image of the traffic sign window to be confirmed.
- sample of the confirmed traffic sign window image may include: a sample of the original confirmed traffic sign window image, and normalizing the sample of the original confirmed traffic sign window image to the preset window after one or more of the following processing Dimensional images: rotate, pan, and zoom.
- the convolution layer, the decimation layer, the full connection layer and the normalization layer sequentially connected include: a plurality of convolution layers and an extraction layer corresponding to the convolution layer, one or more fully connected layers and one normalized layer .
- the preset convolutional neural network model may further include a loss function layer.
- the convolutional neural network model corresponding to the preset also includes a loss function layer, which is reversed by error. Iterating to the weighting of the convolutional layer and the fully connected layer by the propagation BP algorithm may include iterating the weights of both the convolutional layer and the fully connected layer by a loss function and a BP algorithm, respectively.
- the above method of obtaining a pre-trained convolutional neural network model can be implemented by the apparatus for training a pre-trained convolutional neural network model shown in FIG.
- FIG. 9 illustrates an exemplary structural diagram of an apparatus for training a pre-trained convolutional neural network model in accordance with an embodiment of the present application.
- an apparatus 900 for training a pre-trained convolutional neural network model includes:
- the initialization module 901 is configured to initialize weights of the convolutional layer and the fully connected layer of the preset convolutional neural network model according to the Gaussian distribution, where the preset convolutional neural network model includes convolution layers connected in sequence, Extract layer, fully connected layer, and normalized layer.
- the convolution layer, the decimation layer, the full connection layer and the normalization layer sequentially connected include: a plurality of convolution layers and an extraction layer corresponding to the convolution layer, one or more fully connected layers and one normalized layer .
- the preset convolutional neural network model further includes a loss function layer.
- the weight iteration module 902 is configured to iterate the weights of the convolutional layer and the fully connected layer by using an error backpropagation BP algorithm according to the sample of the confirmed traffic sign window image and its traffic sign category;
- the first optimal weight determining module 903 is configured to determine the weight after the current iteration as the optimal weight if the difference between the weight after the current iteration and the weight after the last iteration is less than a preset value.
- the second optimal weight determination module 904 is configured to determine, after the iterative weight with the lowest error rate, the weight of the iteration with the lowest error rate as the optimal weight.
- the model setting module 905 is configured to set the convolutional neural network model including the optimal weight as a pre-trained convolutional neural network model.
- the convolutional neural network model corresponding to the preset further includes a loss function layer, and the weight determination module 902 is further configured to iterate the weights of both the convolutional layer and the fully connected layer by the loss function and the BP algorithm, respectively.
- the device 900 for training a pre-trained convolutional neural network model is trained in advance After the trained convolutional neural network model, the traffic sign recognition module 630 in FIG. 6 may include: a maximum weight identification module (not shown) for inputting the image of the traffic sign window to be confirmed into the convolutional neural network model. A traffic sign category having the largest weight of the output of the layer; and a flag category setting module (not shown) for setting the traffic sign category having the largest weight as the identified traffic sign class.
- the units recited in apparatus 600 correspond to the various steps in the method described with reference to FIG.
- the units described in apparatus 700 correspond to the various steps in the method described with reference to FIG. 2.
- the units described in apparatus 800 correspond to the various steps in the method described with reference to FIG.
- the units described in apparatus 900 correspond to the various steps in the method described with reference to FIG.
- the operations and features described above for the method of identifying traffic signs are equally applicable to the apparatus 600 and the units contained therein, and the operations and features described above for the method of training the pre-trained detection model are equally applicable to the apparatus 700 and therein.
- the included elements, the operations and features described above for the method of obtaining the image of the traffic sign window to be confirmed are equally applicable to the apparatus 800 and the units contained therein, the operations described above for the method of training the pre-trained convolutional neural network model and The features are equally applicable to the device 900 and the units contained therein, and are not described herein again.
- Corresponding units in devices 600, 700, 800, and 900 can cooperate with units in the terminal device and/or server to implement the solution of the embodiments of the present application.
- the modules involved in the embodiments of the present application may be implemented by software or by hardware.
- the described modules may also be provided in the processor, for example, as described in one of the following: a processor includes a feature value acquisition module, a detection module, and an identification module.
- the name of these modules does not constitute a limitation on the module itself in some cases.
- the feature value acquisition module may also be described as "for acquiring a scan window image obtained by dividing the panoramic spherical image in a predetermined integration channel.
- a module of feature values obtained by a predetermined feature algorithm may be implemented by software or by hardware.
- the described modules may also be provided in the processor, for example, as described in one of the following: a processor includes a feature value acquisition module, a detection module, and an identification module.
- the name of these modules does not constitute a limitation on the module itself in some cases.
- the feature value acquisition module may also be described as "for acquiring a scan window image obtained by dividing the panoramic spherical image in a pre
- FIG. 10 is a schematic structural diagram of a computer system according to an embodiment of the present application.
- FIG. 10 there is shown a block diagram of a computer system 700 suitable for use in implementing the apparatus of the embodiments of the present application.
- computer system 1000 includes a central processing unit (CPU) 1001 that can be based on a program stored in read only memory (ROM) 1002 or from a storage portion. 708 loads the program into random access memory (RAM) 1003 to perform various appropriate actions and processes.
- CPU central processing unit
- RAM random access memory
- the CPU 1001 executes the method described in the present application.
- various programs and data required for the operation of the system 1000 are also stored.
- the CPU 1001, the ROM 1002, and the RAM 1003 are connected to each other through a bus 1004.
- An input/output (I/O) interface 1005 is also coupled to bus 1004.
- the following components are connected to the I/O interface 1005: an input portion 1006 including a keyboard, a mouse, etc.; an output portion 1007 including a cathode ray tube (CRT), a liquid crystal display (LCD), and the like, and a speaker; a storage portion 1008 including a hard disk or the like And a communication portion 1009 including a network interface card such as a LAN card, a modem, or the like.
- the communication section 1009 performs communication processing via a network such as the Internet.
- Driver 1010 is also coupled to I/O interface 1005 as needed.
- a removable medium 1011 such as a magnetic disk, an optical disk, a magneto-optical disk, a semiconductor memory or the like is mounted on the drive 1010 as needed so that a computer program read therefrom is installed into the storage portion 1008 as needed.
- embodiments of the present disclosure include a computer program product comprising a computer program tangibly embodied on a machine readable medium, the computer program comprising program code for executing the method illustrated in the flowchart.
- the computer program can be downloaded and installed from the network via the communication portion 1009, and/or installed from the removable medium 1011.
- each block in the flowchart or block diagram can represent a module, a program segment, or a portion of code, and a module, a program segment, or a portion of code includes one or more Executable instructions.
- the functions noted in the blocks may also occur in a different order than that illustrated in the drawings. For example, two successively represented blocks may in fact be executed substantially in parallel, and they may sometimes be executed in the reverse order, depending upon the functionality involved.
- each block of the block diagrams and/or flowcharts, and combinations of blocks in the block diagrams and/or flowcharts can be implemented in a dedicated hardware-based system that performs the specified function or operation. Or you can use dedicated hardware Implemented in combination with computer instructions.
- the present application further provides a computer readable storage medium, which may be a computer readable storage medium included in the apparatus described in the foregoing embodiment, or may exist separately, not A computer readable storage medium that is assembled into a terminal.
- the computer readable storage medium stores one or more programs that are used by one or more processors to perform the method of identifying a traffic sign as described herein.
Abstract
Description
Claims (16)
- 一种识别交通标志的方法,其特征在于,包括:获取由全景球面图像划分得到的扫描窗口图像在预定积分通道通过预定特征算法得到的特征值;根据所述扫描窗口图像的特征值和预先训练的检测分类器模型,检测扫描窗口图像,得到待确认交通标志窗口图像,其中,所述预先训练的检测分类器模型根据扫描窗口图像的样本及其特征值训练得到;根据预先训练的卷积神经网络模型,识别待确认交通标志窗口图像,得到交通标志类别,其中,所述预先训练的卷积神经网络模型根据已确认的交通标志窗口图像的样本及其交通标志类别训练得到。
- 根据权利要求1所述的方法,其特征在于,通过如下方法获得所述预先训练的检测分类器模型:获取扫描窗口图像的样本中的正样本和负样本,其中所述正样本包括交通标志的窗口图像或包括交通标志及其周围扩充预设像素的窗口图像,所述负样本包括去除正样本的扫描窗口图像;获取所述正样本和负样本在预定积分通道通过预定特征算法得到的特征值;根据所述正样本和负样本及其在预定积分通道通过预定特征算法得到的特征值,通过boosting算法训练检测分类器模型,得到所述预先训练的检测分类器模型。
- 根据权利要求1所述的方法,其特征在于,所述获取由全景球面图像划分得到的扫描窗口图像在预定积分通道通过预定特征算法得到的特征值包括:获取由全景球面图像在预定积分通道的积分图划分得到的扫描窗口图像的积分图;根据所述扫描窗口图像的积分图,获取所述扫描窗口图像的特征 值。
- 根据权利要求3所述的方法,其特征在于,所述获取由全景球面图像在预定积分通道的积分图划分得到的扫描窗口图像的积分图包括:对所述全景球面图像连续降采样,获得图像金字塔;获取由图像金字塔在预定积分通道的积分图划分得到的图像金字塔中每级图像的扫描窗口图像的积分图;所述根据所述扫描窗口图像的积分图,获取所述扫描窗口图像的特征值包括:根据所述每级图像的扫描窗口图像的积分图,获取每级图像的扫描窗口图像的特征值;以及所述根据所述扫描窗口图像的特征值和预先训练的检测分类器模型,检测扫描窗口图像,得到待确认交通标志窗口图像包括:根据所述每级图像的扫描窗口图像的特征值和预先训练的检测分类器模型,检测所述每级图像的扫描窗口图像,得到待确认交通标志窗口图像。
- 根据权利要求1-4之一所述的方法,其特征在于,通过如下方法获得所述预先训练的卷积神经网络模型:按照高斯分布,初始化预设的卷积神经网络模型的卷积层和全连接层两者的权重,其中,所述预设的卷积神经网络模型包括依次连接的卷积层、抽取层、全连接层和归一化层;根据已确认的交通标志窗口图像的样本及其交通标志类别,通过误差反向传播BP算法对所述卷积层和所述全连接层两者的权重进行迭代,若当前迭代后的权重与上一次迭代后的权重的差值小于预设值,则将当前迭代后的权重确定为最优权重,将包括最优权重的卷积神经网络模型设定为所述预先训练的卷积神经网络模型;或若出现错误率最低的迭代后的权重,则将错误率最低的迭代后的权重确定为最优权重,将包括最优权重的卷积神经网络模型设定为所述预先训练的卷积神经网络模型。
- 根据权利要求5所述的方法,其特征在于,其特征在于,所述 预设的卷积神经网络模型还包括损失函数层;以及所述通过误差反向传播BP算法对所述卷积层和所述全连接层两者的权重进行迭代包括:通过损失函数和BP算法对所述卷积层和所述全连接层两者的权重分别进行迭代。
- 根据权利要求5所述的方法,其特征在于,所述根据预先训练的卷积神经网络模型,识别待确认交通标志窗口图像,得到交通标志类别包括:将所述待确认交通标志窗口图像输入所述卷积神经网络模型,得到归一化层输出的权重最大的交通标志类别;将所述权重最大的交通标志类别设定为识别得到的交通标志类别。
- 一种识别交通标志的装置,其特征在于,包括:特征值获取模块,用于获取由全景球面图像划分得到的扫描窗口图像在预定积分通道通过预定特征算法得到的特征值;交通标志检测模块,用于根据所述扫描窗口图像的特征值和预先训练的检测分类器模型,检测扫描窗口图像,得到待确认交通标志窗口图像,其中,所述预先训练的检测分类器模型根据扫描窗口图像的样本及其特征值训练得到;交通标志识别模块,用于根据预先训练的卷积神经网络模型,识别待确认交通标志窗口图像,得到交通标志类别,其中,所述预先训练的卷积神经网络模型根据已确认的交通标志窗口图像的样本及其交通标志类别训练得到。
- 根据权利要求8所述的装置,其特征在于,所述预先训练的检测分类器模型通过如下方法获得:获取扫描窗口图像的样本中的正样本和负样本,其中所述正样本包括交通标志的窗口图像或包括交通标志及其周围扩充预设像素的窗口图像,所述负样本包括去除正样本的扫描窗口图像;获取所述正样本和负样本在预定积分通道通过预定特征算法得到的特征值;根据所述正样本和负样本及其在预定积分通道通过预定特征算法得到的特征值,通过boosting算法训练检测分类器模型,得到所述预先训练的检测分类器模型。
- 根据权利要求8所述的装置,其特征在于,所述特征值获取模块包括:积分图获取第一子模块,用于获取由全景球面图像在预定积分通道的积分图划分得到的扫描窗口图像的积分图;特征值获取第一子模块,用于根据所述扫描窗口图像的积分图,获取所述扫描窗口图像的特征值。
- 根据权利要求10所述的装置,其特征在于,所述积分图获取第一子模块包括:降采样子模块,用于对所述全景球面图像连续降采样,获得图像金字塔;积分图获取第二子模块,用于获取由图像金字塔在预定积分通道的积分图划分得到的图像金字塔中每级图像的扫描窗口图像的积分图;所述特征值获取第一子模块包括:特征值第二获取子模块,用于根据所述每级图像的扫描窗口图像的积分图,获取每级图像的扫描窗口图像的特征值;以及所述检测模块包括:多级检测子模块,用于根据所述每级图像的扫描窗口图像的特征值和预先训练的检测分类器模型,检测所述每级图像的扫描窗口图像,得到待确认交通标志窗口图像。
- 根据权利要求8-11之一所述的装置,其特征在于,所述预先训练的卷积神经网络模型通过如下方法获得:按照高斯分布,初始化预设的卷积神经网络模型的卷积层和全连接层两者的权重,其中,所述预设的卷积神经网络模型包括依次连接的卷积层、抽取层、全连接层和归一化层;根据已确认的交通标志窗口图像的样本及其交通标志类别,通过误差反向传播BP算法对所述卷积层和所述全连接层两者的权重进行迭代,若当前迭代后的权重与上一次迭代后的权重的差值小于预设值,则将当前迭代后的权重确定为最优权重,将包括最优权重的卷积神经网络模型设定为所述预先训练的卷积神经网络模型;或若出现错误率最低的迭代后的权重,则将错误率最低的迭代后的权重确定为最优权重,将包括最优权重的卷积神经网络模型设定为所述预先训练的卷积神经网络模型。
- 根据权利要求12所述的装置,其特征在于,其特征在于,所述预设的卷积神经网络模型还包括损失函数层;以及所述权重确定模块进一步用于通过损失函数和BP算法对所述卷积层和所述全连接层两者的权重分别进行迭代。
- 根据权利要求12所述的装置,其特征在于,所述交通标志识别模块包括:最大权重识别模块,用于将所述待确认交通标志窗口图像输入所述卷积神经网络模型,得到归一化层输出的权重最大的交通标志类别;标志类别设定模块,用于将所述权重最大的交通标志类别设定为识别得到的交通标志类。
- 一种设备,包括:处理器;和存储器,所述存储器中存储有能够被所述处理器执行的计算机可读指令,在所述计算机可读指令被执行时,所述处理器执行权利要求1至7中任一项所述的方法。
- 一种非易失性计算机存储介质,所述计算机存储介质存储有能够被处理器执行的计算机可读指令,当所述计算机可读指令被处理 器执行时,所述处理器执行权利要求1至7中任一项所述的方法。
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2016559945A JP6400117B2 (ja) | 2015-03-31 | 2015-12-25 | 交通標識を認識する方法及び装置 |
KR1020167027213A KR101856584B1 (ko) | 2015-03-31 | 2015-12-25 | 교통 표지의 인식 방법 및 장치 |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201510150525.8A CN104700099B (zh) | 2015-03-31 | 2015-03-31 | 识别交通标志的方法和装置 |
CN201510150525.8 | 2015-03-31 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2016155371A1 true WO2016155371A1 (zh) | 2016-10-06 |
Family
ID=53347198
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2015/098903 WO2016155371A1 (zh) | 2015-03-31 | 2015-12-25 | 识别交通标志的方法和装置 |
Country Status (4)
Country | Link |
---|---|
JP (1) | JP6400117B2 (zh) |
KR (1) | KR101856584B1 (zh) |
CN (1) | CN104700099B (zh) |
WO (1) | WO2016155371A1 (zh) |
Cited By (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2018173946A (ja) * | 2017-03-06 | 2018-11-08 | バイドゥ ユーエスエイ エルエルシーBaidu USA LLC | 自動運転車両に用いる畳み込みニューラルネットワークモデルのコンボリューション/デコンボリューション層とバッチ正規化層のオフライン組み合わせ |
CN109255279A (zh) * | 2017-07-13 | 2019-01-22 | 深圳市凯立德科技股份有限公司 | 一种交通标志检测识别的方法及系统 |
CN109919017A (zh) * | 2019-01-28 | 2019-06-21 | 平安科技(深圳)有限公司 | 人脸识别优化方法、装置、计算机设备和存储介质 |
CN110175500A (zh) * | 2019-04-03 | 2019-08-27 | 平安科技(深圳)有限公司 | 指静脉比对方法、装置、计算机设备及存储介质 |
US20190286917A1 (en) * | 2018-03-19 | 2019-09-19 | Kabushiki Kaisha Toshiba | Recognition device, vehicle system and storage medium |
CN110348396A (zh) * | 2019-07-15 | 2019-10-18 | 南京信息工程大学 | 一种基于深度学习的道路上方文字交通标志识别方法及装置 |
CN110889378A (zh) * | 2019-11-28 | 2020-03-17 | 湖南率为控制科技有限公司 | 一种多视图融合交通标志检测识别方法及其系统 |
CN111062885A (zh) * | 2019-12-09 | 2020-04-24 | 中国科学院自动化研究所 | 基于多阶段迁移学习的标志检测模型训练及标志检测方法 |
CN111199217A (zh) * | 2020-01-09 | 2020-05-26 | 上海应用技术大学 | 一种基于卷积神经网络的交通标志识别方法和系统 |
CN111310844A (zh) * | 2020-02-26 | 2020-06-19 | 广州华工邦元信息技术有限公司 | 车辆识别模型构建方法及装置、识别方法及装置 |
CN111597875A (zh) * | 2020-04-01 | 2020-08-28 | 吉利汽车研究院(宁波)有限公司 | 一种交通标志识别方法、装置、设备及存储介质 |
CN111881944A (zh) * | 2020-07-08 | 2020-11-03 | 贵州无忧天空科技有限公司 | 图像鉴别的方法、电子设备和计算机可读介质 |
CN113409393A (zh) * | 2019-05-17 | 2021-09-17 | 百度在线网络技术(北京)有限公司 | 用于识别交通标志牌的方法及装置 |
Families Citing this family (39)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104700099B (zh) * | 2015-03-31 | 2017-08-11 | 百度在线网络技术(北京)有限公司 | 识别交通标志的方法和装置 |
CN105069472B (zh) * | 2015-08-03 | 2018-07-27 | 电子科技大学 | 一种基于卷积神经网络自适应的车辆检测方法 |
CN105335754A (zh) * | 2015-10-29 | 2016-02-17 | 小米科技有限责任公司 | 文字识别方法及装置 |
CN105678322A (zh) * | 2015-12-31 | 2016-06-15 | 百度在线网络技术(北京)有限公司 | 样本标注方法和装置 |
CN105869016A (zh) * | 2016-03-28 | 2016-08-17 | 天津中科智能识别产业技术研究院有限公司 | 一种基于卷积神经网络的点击通过率预估方法 |
CN105956524A (zh) * | 2016-04-22 | 2016-09-21 | 北京智芯原动科技有限公司 | 一种交通标识识别方法及装置 |
CN106022380A (zh) * | 2016-05-25 | 2016-10-12 | 中国科学院自动化研究所 | 基于深度学习的个体身份识别方法 |
CN106778835B (zh) * | 2016-11-29 | 2020-03-24 | 武汉大学 | 融合场景信息和深度特征的遥感图像机场目标识别方法 |
US10282999B2 (en) * | 2017-03-17 | 2019-05-07 | GM Global Technology Operations LLC | Road construction detection systems and methods |
CN106886801B (zh) | 2017-04-14 | 2021-12-17 | 北京图森智途科技有限公司 | 一种图像语义分割方法及装置 |
CN107067043B (zh) * | 2017-05-25 | 2020-07-24 | 哈尔滨工业大学 | 一种农作物病虫害检测方法 |
KR101834778B1 (ko) | 2017-05-30 | 2018-03-07 | 만도헬라일렉트로닉스(주) | 교통 표지판 인식장치 및 방법 |
WO2019009490A1 (ko) | 2017-07-06 | 2019-01-10 | 삼성전자 주식회사 | 영상을 부호화/복호화 하는 방법 및 그 장치 |
US10986356B2 (en) | 2017-07-06 | 2021-04-20 | Samsung Electronics Co., Ltd. | Method for encoding/decoding image and device therefor |
WO2019009447A1 (ko) * | 2017-07-06 | 2019-01-10 | 삼성전자 주식회사 | 영상을 부호화/복호화 하는 방법 및 그 장치 |
US9947103B1 (en) * | 2017-10-03 | 2018-04-17 | StradVision, Inc. | Learning method and learning device for improving image segmentation and testing method and testing device using the same |
CN107679577A (zh) * | 2017-10-12 | 2018-02-09 | 理光图像技术(上海)有限公司 | 基于深度学习的图像检测方法 |
CN109726728B (zh) * | 2017-10-31 | 2020-12-15 | 阿里巴巴(中国)有限公司 | 一种训练数据生成方法及装置 |
CN107944437B (zh) * | 2017-12-31 | 2018-12-14 | 广州二元科技有限公司 | 一种基于神经网络和积分图像的人脸定位方法 |
CN110096937A (zh) * | 2018-01-31 | 2019-08-06 | 北京四维图新科技股份有限公司 | 一种用于辅助车辆自动驾驶的图像识别的方法及装置 |
CN108549927A (zh) * | 2018-03-16 | 2018-09-18 | 深圳市商汤科技有限公司 | 物体检测网络的训练方法和特征监督图生成方法 |
CN108520212A (zh) * | 2018-03-27 | 2018-09-11 | 东华大学 | 基于改进的卷积神经网络的交通标志检测方法 |
KR102102161B1 (ko) * | 2018-05-18 | 2020-04-20 | 오드컨셉 주식회사 | 이미지 내 객체의 대표 특성을 추출하는 방법, 장치 및 컴퓨터 프로그램 |
CN108960308A (zh) * | 2018-06-25 | 2018-12-07 | 中国科学院自动化研究所 | 交通标志识别方法、装置、车载终端及车辆 |
KR102106029B1 (ko) * | 2018-07-19 | 2020-05-04 | 네이버랩스 주식회사 | 간판 검출 성능 향상을 위한 방법 및 시스템 |
KR102096926B1 (ko) * | 2018-07-19 | 2020-04-06 | 네이버랩스 주식회사 | 관심 영역 변화를 검출하는 방법 및 시스템 |
CN109102017B (zh) * | 2018-08-09 | 2021-08-03 | 百度在线网络技术(北京)有限公司 | 神经网络模型处理方法、装置、设备及可读存储介质 |
CN110879950A (zh) * | 2018-09-06 | 2020-03-13 | 北京市商汤科技开发有限公司 | 多级目标分类及交通标志检测方法和装置、设备、介质 |
CN109492628B (zh) * | 2018-10-08 | 2021-11-16 | 杭州电子科技大学 | 一种应用于课堂违纪抓取的球形定位系统的实现方法 |
WO2020080765A1 (en) | 2018-10-19 | 2020-04-23 | Samsung Electronics Co., Ltd. | Apparatuses and methods for performing artificial intelligence encoding and artificial intelligence decoding on image |
WO2020080665A1 (en) | 2018-10-19 | 2020-04-23 | Samsung Electronics Co., Ltd. | Methods and apparatuses for performing artificial intelligence encoding and artificial intelligence decoding on image |
US10402977B1 (en) * | 2019-01-25 | 2019-09-03 | StradVision, Inc. | Learning method and learning device for improving segmentation performance in road obstacle detection required to satisfy level 4 and level 5 of autonomous vehicles using laplacian pyramid network and testing method and testing device using the same |
KR102436512B1 (ko) | 2019-10-29 | 2022-08-25 | 삼성전자주식회사 | 부호화 방법 및 그 장치, 복호화 방법 및 그 장치 |
CN110796109A (zh) * | 2019-11-05 | 2020-02-14 | 哈尔滨理工大学 | 一种基于模型融合的驾驶员分心行为识别方法 |
CN111178153A (zh) * | 2019-12-09 | 2020-05-19 | 武汉光庭信息技术股份有限公司 | 一种交通标志检测方法及系统 |
CN111178371B (zh) * | 2019-12-17 | 2023-12-01 | 深圳市优必选科技股份有限公司 | 目标检测方法、装置和计算机存储介质 |
KR102287942B1 (ko) | 2020-02-24 | 2021-08-09 | 삼성전자주식회사 | 전처리를 이용한 영상의 ai 부호화 및 ai 복호화 방법, 및 장치 |
KR102189926B1 (ko) * | 2020-03-30 | 2020-12-11 | 네이버랩스 주식회사 | 관심 영역 변화를 검출하는 방법 및 시스템 |
CN112085056B (zh) * | 2020-08-05 | 2023-12-29 | 深圳市优必选科技股份有限公司 | 目标检测模型生成方法、装置、设备及存储介质 |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130272575A1 (en) * | 2011-11-01 | 2013-10-17 | Intel Corporation | Object detection using extended surf features |
CN103366177A (zh) * | 2012-03-28 | 2013-10-23 | 佳能株式会社 | 对象检测分类器生成方法和设备、图像对象检测方法和设备 |
CN104036323A (zh) * | 2014-06-26 | 2014-09-10 | 叶茂 | 一种基于卷积神经网络的车辆检测方法 |
CN104700099A (zh) * | 2015-03-31 | 2015-06-10 | 百度在线网络技术(北京)有限公司 | 识别交通标志的方法和装置 |
Family Cites Families (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH08221378A (ja) * | 1995-02-10 | 1996-08-30 | Ricoh Co Ltd | 学習機械 |
JP2004191339A (ja) * | 2002-12-13 | 2004-07-08 | Sharp Corp | 位置情報検索方法、位置情報検索装置、位置情報検索端末、及び、位置情報検索システム |
JP2005311691A (ja) * | 2004-04-21 | 2005-11-04 | Toyota Central Res & Dev Lab Inc | 物体検出装置及び方法 |
JP5364578B2 (ja) * | 2006-07-12 | 2013-12-11 | コファックス, インコーポレイテッド | トランスダクティブデータ分類のための方法およびシステム、ならびに機械学習手法を用いたデータ分類方法 |
JP2013196454A (ja) * | 2012-03-21 | 2013-09-30 | Honda Elesys Co Ltd | 画像処理装置、画像処理方法および画像処理プログラム |
JP2014010633A (ja) * | 2012-06-29 | 2014-01-20 | Honda Elesys Co Ltd | 画像認識装置、画像認識方法、及び画像認識プログラム |
JP2014071832A (ja) * | 2012-10-01 | 2014-04-21 | Toshiba Corp | 物体検出装置及びその検出方法 |
-
2015
- 2015-03-31 CN CN201510150525.8A patent/CN104700099B/zh active Active
- 2015-12-25 WO PCT/CN2015/098903 patent/WO2016155371A1/zh active Application Filing
- 2015-12-25 KR KR1020167027213A patent/KR101856584B1/ko active IP Right Grant
- 2015-12-25 JP JP2016559945A patent/JP6400117B2/ja active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130272575A1 (en) * | 2011-11-01 | 2013-10-17 | Intel Corporation | Object detection using extended surf features |
CN103366177A (zh) * | 2012-03-28 | 2013-10-23 | 佳能株式会社 | 对象检测分类器生成方法和设备、图像对象检测方法和设备 |
CN104036323A (zh) * | 2014-06-26 | 2014-09-10 | 叶茂 | 一种基于卷积神经网络的车辆检测方法 |
CN104700099A (zh) * | 2015-03-31 | 2015-06-10 | 百度在线网络技术(北京)有限公司 | 识别交通标志的方法和装置 |
Cited By (19)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2018173946A (ja) * | 2017-03-06 | 2018-11-08 | バイドゥ ユーエスエイ エルエルシーBaidu USA LLC | 自動運転車両に用いる畳み込みニューラルネットワークモデルのコンボリューション/デコンボリューション層とバッチ正規化層のオフライン組み合わせ |
CN109255279A (zh) * | 2017-07-13 | 2019-01-22 | 深圳市凯立德科技股份有限公司 | 一种交通标志检测识别的方法及系统 |
US10984256B2 (en) * | 2018-03-19 | 2021-04-20 | Kabushiki Kaisha Toshiba | Recognition device, vehicle system and storage medium |
US20190286917A1 (en) * | 2018-03-19 | 2019-09-19 | Kabushiki Kaisha Toshiba | Recognition device, vehicle system and storage medium |
CN109919017A (zh) * | 2019-01-28 | 2019-06-21 | 平安科技(深圳)有限公司 | 人脸识别优化方法、装置、计算机设备和存储介质 |
CN109919017B (zh) * | 2019-01-28 | 2023-09-01 | 平安科技(深圳)有限公司 | 人脸识别优化方法、装置、计算机设备和存储介质 |
CN110175500A (zh) * | 2019-04-03 | 2019-08-27 | 平安科技(深圳)有限公司 | 指静脉比对方法、装置、计算机设备及存储介质 |
CN110175500B (zh) * | 2019-04-03 | 2024-01-19 | 平安科技(深圳)有限公司 | 指静脉比对方法、装置、计算机设备及存储介质 |
CN113409393B (zh) * | 2019-05-17 | 2023-10-03 | 百度在线网络技术(北京)有限公司 | 用于识别交通标志牌的方法及装置 |
CN113409393A (zh) * | 2019-05-17 | 2021-09-17 | 百度在线网络技术(北京)有限公司 | 用于识别交通标志牌的方法及装置 |
CN110348396A (zh) * | 2019-07-15 | 2019-10-18 | 南京信息工程大学 | 一种基于深度学习的道路上方文字交通标志识别方法及装置 |
CN110889378A (zh) * | 2019-11-28 | 2020-03-17 | 湖南率为控制科技有限公司 | 一种多视图融合交通标志检测识别方法及其系统 |
CN111062885A (zh) * | 2019-12-09 | 2020-04-24 | 中国科学院自动化研究所 | 基于多阶段迁移学习的标志检测模型训练及标志检测方法 |
CN111062885B (zh) * | 2019-12-09 | 2023-09-12 | 中国科学院自动化研究所 | 基于多阶段迁移学习的标志检测模型训练及标志检测方法 |
CN111199217B (zh) * | 2020-01-09 | 2023-03-28 | 上海应用技术大学 | 一种基于卷积神经网络的交通标志识别方法和系统 |
CN111199217A (zh) * | 2020-01-09 | 2020-05-26 | 上海应用技术大学 | 一种基于卷积神经网络的交通标志识别方法和系统 |
CN111310844A (zh) * | 2020-02-26 | 2020-06-19 | 广州华工邦元信息技术有限公司 | 车辆识别模型构建方法及装置、识别方法及装置 |
CN111597875A (zh) * | 2020-04-01 | 2020-08-28 | 吉利汽车研究院(宁波)有限公司 | 一种交通标志识别方法、装置、设备及存储介质 |
CN111881944A (zh) * | 2020-07-08 | 2020-11-03 | 贵州无忧天空科技有限公司 | 图像鉴别的方法、电子设备和计算机可读介质 |
Also Published As
Publication number | Publication date |
---|---|
KR101856584B1 (ko) | 2018-05-10 |
JP2017516197A (ja) | 2017-06-15 |
CN104700099A (zh) | 2015-06-10 |
KR20160132331A (ko) | 2016-11-17 |
CN104700099B (zh) | 2017-08-11 |
JP6400117B2 (ja) | 2018-10-03 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2016155371A1 (zh) | 识别交通标志的方法和装置 | |
CN108197326B (zh) | 一种车辆检索方法及装置、电子设备、存储介质 | |
CN111898523A (zh) | 一种基于迁移学习的遥感图像特种车辆目标检测方法 | |
WO2016150240A1 (zh) | 身份认证方法和装置 | |
CN111461213B (zh) | 一种目标检测模型的训练方法、目标快速检测方法 | |
CN110033040B (zh) | 一种火焰识别方法、系统、介质和设备 | |
CN113177456B (zh) | 基于单阶段全卷积网络和多特征融合的遥感目标检测方法 | |
CN110705565A (zh) | 淋巴结肿瘤区域识别方法与装置 | |
JP2017033197A (ja) | 変化領域検出装置、方法、及びプログラム | |
CN113052170B (zh) | 一种无约束场景下的小目标车牌识别方法 | |
CN111860106B (zh) | 一种无监督的桥梁裂缝识别方法 | |
CN109858438B (zh) | 一种基于模型拟合的车道线检测方法 | |
CN111881944A (zh) | 图像鉴别的方法、电子设备和计算机可读介质 | |
CN109584206B (zh) | 零件表面瑕疵检测中神经网络的训练样本的合成方法 | |
CN111539456B (zh) | 一种目标识别方法及设备 | |
CN114495010A (zh) | 一种基于多特征学习的跨模态行人重识别方法及系统 | |
CN110458019B (zh) | 稀缺认知样本条件下的排除倒影干扰的水面目标检测方法 | |
CN111444816A (zh) | 一种基于Faster RCNN的多尺度密集行人检测方法 | |
CN111832463A (zh) | 一种基于深度学习的交通标志检测方法 | |
CN111046861B (zh) | 识别红外影像的方法、构建识别模型的方法及应用 | |
CN114445788A (zh) | 车辆停放检测方法、装置、终端设备和可读存储介质 | |
CN107679528A (zh) | 一种基于AdaBoost‑SVM集成学习算法的行人检测方法 | |
CN114627424A (zh) | 一种基于视角转化的步态识别方法和系统 | |
JP2021009493A (ja) | 画像処理装置、画像処理装置の制御方法及びプログラム | |
JP2007139421A (ja) | 形態分類装置及び方法、並びに変化領域抽出装置及び方法 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
ENP | Entry into the national phase |
Ref document number: 2016559945 Country of ref document: JP Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 20167027213 Country of ref document: KR Kind code of ref document: A |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 15887321 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 15887321 Country of ref document: EP Kind code of ref document: A1 |