CN106250812B - A kind of model recognizing method based on quick R-CNN deep neural network - Google Patents

A kind of model recognizing method based on quick R-CNN deep neural network Download PDF

Info

Publication number
CN106250812B
CN106250812B CN201610563184.1A CN201610563184A CN106250812B CN 106250812 B CN106250812 B CN 106250812B CN 201610563184 A CN201610563184 A CN 201610563184A CN 106250812 B CN106250812 B CN 106250812B
Authority
CN
China
Prior art keywords
network
layer
vehicle
training
region
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201610563184.1A
Other languages
Chinese (zh)
Other versions
CN106250812A (en
Inventor
汤一平
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hangzhou Yixun Technology Service Co ltd
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Priority to CN201610563184.1A priority Critical patent/CN106250812B/en
Publication of CN106250812A publication Critical patent/CN106250812A/en
Application granted granted Critical
Publication of CN106250812B publication Critical patent/CN106250812B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/56Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
    • G06V20/58Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads
    • G06V20/584Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads of vehicle lights or traffic lights
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Computational Linguistics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Evolutionary Biology (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)

Abstract

The present invention discloses a kind of model recognizing method based on quick R-CNN deep neural network, mainly includes unsupervised deep learning, the CNN convolutional neural networks of multilayer, region suggestion network, network share, softmax classifier;Realize the frame of one quick R-CNN network implementations of use a vehicle detection and identification end to end truly, and there is the Morphological Diversity for being suitable for vehicle target, illumination variation diversity, under the environment such as background diversity quickly, the vehicle subclass identification of high-precision and robustness.

Description

A kind of model recognizing method based on quick R-CNN deep neural network
Technical field
The present invention relates to computer technology, pattern-recognition, artificial intelligence, applied mathematics and biological vision technologies in intelligence The application of field of traffic more particularly to a kind of model recognizing method based on quick R-CNN deep neural network.
Background technique
Core function in intelligent transportation system is to the accurate detection of vehicular traffic and correct vehicle cab recognition.It is current right Mainly there are two important technologies for the research of vehicle detection sorting technique: i.e. automatic vehicle identification and automobile automatic recognition.
Automatic vehicle identification is that progress is mutually known using mobile unit and ground base station equipment, which is mainly used for charge system Wider in some technology developed countries use scope in system, AE-PASS system, the ETC system of Japan such as the U.S., the whole world is defended Star GPS positioning etc..
Automobile automatic recognition is the parameter by detection vehicle inherently, with appropriate under certain vehicle classification standard Classification and identification algorithm, parting initiatively is carried out to vehicle, than wide, oneself is this kind of technical applications through have many maturations System is applied in real life, such technology can be known automatically by modes such as frequency microwave, feux rouges, laser, surface acoustic waves The mode of video image processing can be used also to identify the information of vehicles such as license plate, vehicle in other information of vehicles.
Automobile automatic recognition comparative maturity technology has Data mining, swashs for infrared detection, ultrasonic wave/microwave inspection Survey, geomagnetism detecting etc., but these types of method respectively has superiority and inferiority, and advantage is that identification is accurate relatively high, but disadvantage is it is also obvious that main lack Point has construction and installation process sufficiently complex, influences normal traffic order, and difficult in maintenance, capital equipment is easy to damage, spends larger Deng.
In recent years, video detection technology oneself become the most important information collection means of intelligent transportation field, synthetical comparison and assessment, will Video detection technology, which is applied to highway and urban road, has great practical application value, is based on video vehicle cab recognition system System, the level of the information collection of General Promotion urban road and safety management can play increasingly in intelligent transportation system Important role.
The visual identity of vehicle, lot of domestic and international scholar have carried out correlative study.Paper " Robert T.Collins, Alan J.Lipton,Hironobu Fujiyoshi,and Takeo Kanade.Algorith rns For Cooperative multisensor surveillanee.In Proceedings of the IEEE " discloses a road The detection, tracking, identifying system for moving up moving-target identify that moving target is people, crowd, vehicle with the neural network trained Or it interferes, the input characteristics amount of network has dispersibility measurement, target sizes target surface size and the video camera of target to monitor The relative value of area size.Vehicle further divides into different type and color.Paper " TieniuN.Tan and Keith D.Baker Efficient image gradient based vehicle localization.IEEE Transaction On Iimage Proeessing, 2000,9 (8): 1343-1356. " describes a kind of vehicle location and knows method for distinguishing, one In a wicket, this method is carried out according to image gradient.Using surface constraints and most of contour of the vehicle by two straight lines about The posture of vehicle can be obtained in the fact that beam.Paper " George Fung, NelsonYung, and Granthajm Pang.Vehicle shape approximation from motion for visual traffic Surveillance.In Proc.IEEE Conf.Intelligent Transport System, 2001,608-613. " is used High-precision video camera observes the movement of vehicle to estimate vehicle shape, by estimating that characteristic point obtains vehicle's contour.Basic thought It is that the movement speed of high characteristic point is greater than the movement speed of low characteristic point, because high characteristic point is close from video camera, vehicle's contour can With with vehicle identification.Paper " Ferryman, A.Worral, G.Sullivan, K.Baker, A generic deformable model for Vehicle recognition,Proeeeding of British Machine Vision Conference, 1995,127-136. " proposes the deformable three-dimensional template of a parametrization, which, which passes through, develops, it is said that It can be applied to various vehicle identifications.Paper " Larry Davis, V.Philomin and R.Duralswami.Tracking humans from a moving platform.In Proc.Intematl on Conference on Pattem Recognition, 2000. " study vehicle identification with deforming template, firstly, establishing the side view of target vehicle vehicle head part And the deforming template of front view.By histogram intersection, the RGB histogram of vehicle must also compare, suitable vehicle template The point set on side is compared also by the Hausdorff distance between point set with other car modals.Above-mentioned technology substantially also needs Feature extraction is manually completed, bigger problem is: 1) be influenced too big, all kinds of detection algorithm requirements by specific application environment Condition it is too harsh;2) vehicle class is various but difference is little, without apparent distinguishing characteristics;3) it is influenced by visible change Greatly, the automobile characteristic difference taken the photograph from different perspectives is big;4) too big by natural environment influence, especially illumination effect is serious Illumination reflection is so that vehicle wheel profile fuzzy, color deviation, variation are too big, it is difficult to recognize;5) shape of automobile updates too fast, Changing features are too fast, so that algorithm adaptability is poor.The country remains in research state in most of vehicle cab recognition technical aspect, Such as Chinese Academy of Sciences, Xi'an Highway house, Shanghai Communications University, Xi'an Communications University, some research achievements of Sichuan University.It is closed Key problem is that mankind itself limit the knowledge of vehicle cab recognition process.
It needs to have scale, rotation and certain angular transformation, illumination variation for carrying out the feature of vehicle classification Good robustness, the computer vision technique in preceding deep learning epoch are typically all that labor costs' plenty of time and energy go to set Count suitable feature.In order to allow computer to automatically select suitable characteristics, artificial neural network just comes into being, early in 1999 Year just has external researcher to classify using neural network to object, includes the side such as fuzzy neural network and BP network Method;But since there are problems for its performance, the contradiction as being difficult to solve complexity present in pattern-recognition and generalization, mind Although there is powerful modeling ability through network, but classify to such as vehicle large-scale image, huge parameter space makes Excellent optimized initial value more difficulty etc. is found, therefore has been treated coldly for a long time by people, until the proposition ability of deep learning Become research hotspot again.
In image classification field, to the image classifications of substantial amounts, there are mainly two types of methods: one kind is extracts every photo Local feature, by the feature of extraction carry out cluster and coding obtain a high dimension vector, then it is classified with classifier.Its The method of middle coding has vision bag of words coding, sparse coding and Fei Sheer vector coding etc., from the point of view of current result of study The performance of Fei Sheer vector coding will be got well compared with other several coding modes.The very wide image classification method of another kind application is depth Neural network, deep learning are a new hot spots in neural network research, and its object is to pass through non-supervisory pre-training Excellent initial parameter value is provided for neural network, by way of greedy, training in layer classifies to large-scale image Obtain extraordinary effect.
The concept of deep learning start to attract much attention be before and after about 2006, paper " Hinton, G.E.and R.R.Salakhutdinov,Reducing the dimensionality of data with neural Networks.Science, 2006.313 (5786): the feedforward neural network that 504-507 " proposes a kind of multilayer can be layer-by-layer Efficient training early period is done, each layer is trained using unsupervised restricted Boltzmann machine, finally has prison in utilization The back-propagating superintended and directed is finely tuned, and provides a kind of new side to solve complexity present in pattern-recognition and the contradiction of generalization Method and thinking have thus pulled open the computer vision technique prelude in deep learning epoch.
Convolutional neural networks, i.e. CNN are one kind of deep learning algorithm, are that the mode in special disposal image domains is known It not, while being also the algorithm that achievement is most surprising in current image steganalysis.Convolutional neural networks algorithm is advantageous in that training It is not needed when model using any manual features, algorithm can explore the feature that image implies automatically.
Application No. is 201610019285.2 Chinese patent applications to disclose a kind of model recognizing method and system, including Machine training generates classifier process and treats the differentiation process of mapping piece, generates classifier during, based on license plate It determines required image range in training set picture, the image range having determined is divided into region, is selected in each region Characteristic information investment machine training all in selected respective region are generated point for corresponding each region by characteristic information respectively Class device treats mapping piece by the classifier generated and carries out single area judging, according to single area judging result using more Region confidence fusion judgement obtains vehicle cab recognition result.The invention carries out vehicle cab recognition using random deep woods classifier, maximum The problem of be the absence of unsupervised learning process.
Application No. is 201510639752.7 Chinese patent applications to disclose a kind of model recognizing method, the method packet It includes: obtaining picture to be detected;The picture to be detected is detected using the first default classifier;If the picture to be detected In contain target vehicle, extract the target vehicle in the picture to be detected;Registration process is carried out to the target vehicle, so that Angle between the headstock direction of the target vehicle and the vertical direction of the target area is less than preset threshold;To described right Target vehicle after neat processing carries out feature extraction, and to obtain M feature, the M is the integer greater than 1;It is pre- using second If classifier classifies to the M feature;The vehicle of the target vehicle is determined according to the result of the classification.From certain It is said in meaning, which still falls within shallow neural network, is architecturally just difficult to realize unsupervised learning process.
Application No. is 201510071919.4 Chinese patent applications to propose a kind of vehicle based on convolutional neural networks Recognition methods is based on characteristic extracting module and vehicle cab recognition module, comprising the following steps: by design convolution and pond layer, entirely Articulamentum, classifier construct the neural network of vehicle cab recognition, and wherein convolution and pond layer and full articulamentum are for extracting vehicle Feature, classifier is for vehicle classification identification;Utilize the neural network of the database training comprising different automobile types feature, training side Formula is the study for having supervision that the data of tape label carry out, and carries out weight parameter matrix and offset with stochastic gradient descent method Adjustment;The weight parameter matrix and offset in trained each layer are obtained, they are accordingly assigned to the neural network In each layer, then the network has the function of vehicle feature extraction and identification.The invention lacks more detailed realization details, Only conceptually propose the data that the model recognizing method based on convolutional neural networks, especially training method are tape label The study for having supervision carried out.The shape image volume of vehicle belongs to mass data, and to be labeled to these image datas is one A extremely difficult thing;Furthermore the shape update of vehicle causes changing features fast very much fastly, so causing some in the invention Algorithm is difficult to meet practical application.In addition, the practical collected vehicle image in road surface is complicated;Include background complexity, vehicle it Between the interference such as block;If without processing is split to the image actually obtained, it will seriously affect final recognition result.
Application No. is 201510738852.5 and 201510738843.6 Chinese patent applications to propose a kind of be based on deeply The model recognizing method of Du Feisheer network constructs the 0th layer of Fei Sheer network first, to the database for having K kind vehicle image, Extract the SIFT feature of every kind of vehicle vehicle image;Then the 1st layer of Fei Sheer network is constructed, to the SIFT feature extracted Fei Sheer vector coding is carried out, the vector after coding is stacked in space, then carry out L2 normalization and PCA dimensionality reduction;1st layer is obtained The character representation arrived carries out Fei Sheer vector coding, by symbol square root and L2 normalized, forms Fei Sheer network 2nd layer;The global characteristics expression that different automobile types image obtains finally is used for linear SVM training, is obtained with K kind The identifying system of vehicle classification;It to vehicle to be identified, makes it through Fei Sheer network and obtains testing feature vector, import identification system System may recognize that vehicle vehicle to be identified.Also there are two deficiency, one is the absence of deep learning process for the invention, the second is lacking It is few that vehicles segmentation processing step is carried out to image.
Application No. is 201510738540.4 Chinese patent applications to propose one kind based on local feature Aggregation Descriptor Model recognizing method, first extraction model data library in vehicle image SIFT feature;Then to all vehicle images SIFT feature carries out Kmeans cluster, forms K cluster centre, obtains the dictionary with K vision word;Then for every Each SIFT feature is assigned to nearest vision word by vehicle image;Count SIFT feature around each vision word to The residual error cumulant of amount and Current vision word, obtains the local feature Aggregation Descriptor of current vehicle picture;Finally, will train The local feature Aggregation Descriptor of the n of module vehicle images obtains the n class vehicle class that can be indexed by quantization encoding Other coded image library;And to test vehicle image, its local feature Aggregation Descriptor is equally extracted, as query vector, is led Enter image library to be indexed, be matched by approximate KNN searching method, identifies test vehicle vehicle.The same invention Also there are two defect, one is the absence of deep learning, carries out vehicles segmentation processing step to image secondly being the absence of.
Convolutional neural networks are upper or relatively more successful in the identification of truck, buggy, bus, minibus, SUV and car , but classify in subclass, the remote precision less than major class classification if the precision in the different automobile types of identification vehicle.Usually The difficulty of subclass image recognition is generally placed at two o'clock:
(1) acquisition of subclass image labeling data is very difficult, it usually needs the expert of related fields is labeled.
(2) there are difference in big class for subclass image recognition, such as in vehicle cab recognition, the viewing angle of Che Butong, and Difference between small class.
In conclusion using convolutional neural networks even depth nerual network technique to vehicle cab recognition, still have at present as Several lower stubborn problems: 1) general image of tested vehicle how from complicated background is accurately partitioned into;2) how to the greatest extent The characteristic of vehicle vehicle may be accurately obtained using few label image data;3) how to know in vehicle vehicle major class Also can recognize that it is which kind of, which kind of color, which vehicle of age out on the basis of not;4) how to be obtained automatically by deep learning The feature for a vehicle of picking up the car;5) accuracy of identification and detection efficiency how have been taken into account, while having reduced trained and learning time as far as possible; 6) how in classifier design, which is the classificating requirement for being able to satisfy vehicle vehicle subclass, and can be in the shape of automobile It no longer needs to be trained study to whole network again after update;7) one CNN of use truly how is designed The frame of network implementations vehicle detection and identification end to end;8) influence for how reducing weather condition, increases the adaptive of system Ying Xing.
Summary of the invention
In order to overcome, automation and intelligent level in existing vehicle vehicle Visual identification technology are low, lack depth It practises, be difficult to adapting to ambient weather variation, be difficult to accurately extract vehicle general image for identification, be difficult to use visual manner Identification classification is carried out to vehicle vehicle subclass, is difficult to the deficiencies of taking into account accuracy of identification and time and detection efficiency, the present invention provides A kind of model recognizing method based on quick R-CNN deep neural network, can effectively improve vehicle visual identity automation and Intelligent level can preferably adapt to ambient weather variation and have extensive adaptivity, can guarantee in preferable detection identification essence There is real-time detection recognition capability on the basis of degree, the dependence to label vehicle data can be greatly reduced with automatic study and extract vehicle The contradiction of the ability of type feature, the complexity that can preferably solve vehicle cab recognition and generalization has preferable universality.
It realizes foregoing invention content, it is necessary to solve several key problems: (1) designing a kind of quick view of Vehicle Object Feel partitioning algorithm;(2) a kind of deep learning method is researched and developed, realizes unsupervised vehicle feature extraction;(3) one kind is designed to be suitable for The classifier of thousands of kinds of type subclass, and there is scalability;(4) one quick R- of use truly is designed The frame of CNN network implementations vehicle detection and identification end to end.
The technical solution adopted by the present invention to solve the technical problems is:
A kind of model recognizing method based on quick R-CNN deep neural network, including one be used for deep learning and instruction Practice the VGG network of identification, network is suggested in a region for extracting area-of-interest and one for vehicle classification Softmax classifier;
The VGG network, including 8 Ge Juan bases, 3 full articulamentums amount to 11 layers;There are 5 groups in 8 Ge Juan bases Convolutional layer, 2 classification layers extract characteristics of image, 1 classification layer characteristic of division;3 full articulamentums are separately connected classification layer 6, divide Class layer 7 and classification layer 8;
Network is suggested in the region, and the module of Classification Loss is calculated including 1 layer of classifying, 1 window recurrence layer, 1 The module of loss, p interested Suggestion box of output are returned with 1 calculation window;
The Softmax classifier, by the input data feature extracted and learning training obtain feature database data into Row compares, and calculates the probability of each classification results, the highest result of probability is then taken to be exported;
Quick R-CNN deep neural network has accessed the region at the 5th layer of end of the VGG network and has suggested Network, so that first 5 layers of low-level image feature extraction process and result of VGG network described in network share is suggested in the region;
The the 6th and the 7th layer of the VGG network suggests p interested suggestions of network output according to the region Characteristics of image in frame carries out convolution sum ReLU processing, obtains the p characteristic patterns containing 4096 vectors, then gives classification respectively Layer and window return layer and are handled, and realize the segmentation of vehicle image;On the other hand, the Softmax classifier contains p There is the characteristic pattern of 4096 vectors to carry out Classification and Identification, obtains the classification results of vehicle vehicle.
The Softmax classifier, using the learning outcome in quick R-CNN as softmax during learning training The input data of classifier;It is that the Logistic towards multicategory classification problem is returned that Softmax, which is returned, is that Logistic is returned General type, be suitable for classification between mutual exclusion the case where;Assuming that for training set { (x(1),y(1),…,x(m),y(m)), there is y(1) ∈ { 1,2 ..., k } inputs x for given sample, exports the vector of k dimension to indicate what each classification results occurred Probability is p (y=i | x), it is assumed that function h (x) is as follows:
θ12,…θkIt is the parameter of model, and all probability and be 1;Cost function after regularization term is added are as follows:
Partial derivative of the cost function to first of parameter of j-th of classification are as follows:
Finally, realizing that the classification of softmax returns by minimizing J (θ), classification regression result is saved in feature database In;
When identifying classification, the input data feature extracted is obtained into feature database data with learning training and is compared, The probability of each classification results is calculated, the highest result of probability is then taken to be exported.
Network is suggested in the region, for formation zone Suggestion box, is accessed at the 5th layer of end of the VGG network Network is suggested in the region, i.e., slides small network on the convolution Feature Mapping figure that the 5th layer of convolutional layer exports, this A network is connected to entirely in the spatial window of n × n of input convolution Feature Mapping;Each sliding window be mapped to a low-dimensional to In amount, low-dimensional vector is 256-d, the corresponding numerical value of a sliding window of each Feature Mapping;This vector is exported to two The layer of a peer connected entirely;- a window recurrence layer and a classification layer;Window returns layer and exports on each position, and 9 kinds Recommend region to correspond to window to need that there is translation scaling invariance, window returns layer and exports 4 translation scalings from 256 dimensional features Parameter has 4k output, i.e., the codes co-ordinates of k Suggestion box;Classification layer is exported from 256 dimensional features belongs to foreground and background Probability exports 2k Suggestion box score, i.e., is vehicle target/non-vehicle target estimated probability to each Suggestion box.
Whether the training of network is suggested in region, distribute a binary label to each candidate region, i.e., be vehicle pair As;Here distribute positive label to two class candidate regions: the enclosing region of (i) and some GT have the ratio between highest intersection union, IoU, the candidate region of overlapping;(ii) IoU with any GT enclosing region greater than 0.7 overlapping candidate region;It distributes simultaneously negative The IoU ratio that label gives all GT enclosing regions is below 0.3 candidate region;Leave out non-just non-negative candidate region;Tool Body algorithm is as follows:
STEP31: sequence reads every figure in training set;
STEP32: it to the true value candidate region of each calibration, overlaps the maximum candidate region of ratio and is denoted as prospect sample This;
STEP33:: to STEP32) remaining candidate region, if it is Chong Die with some calibration, IoU ratio is greater than 0.7, It is denoted as prospect sample;If the overlap proportion of itself and any one calibration is both less than 0.3, it is denoted as background sample;
STEP34: the remaining candidate region STEP32 and STEP33 is discarded;
STEP35: the candidate region across image boundary is discarded.
In order to carry out the screening and regional location refine of candidate region automatically, here using minimum objective function;To one Formula (14) expression of the cost function of a image,
In formula, i is the index of candidate region in a batch processing, NclsFor the normalization coefficient for layer of classifying, NregFor window The normalization coefficient of layer is returned, λ is balance weight, piFor the prediction probability of vehicle target,For GT label, if candidate region It is positiveIf candidate region is negativetiFor a vector, indicate that 4 parametrizations of the encirclement frame of prediction are sat Mark,The coordinate vector of frame, L are surrounded for GT corresponding with positive candidate regionclsFor the logarithm cost of classification, LregTo return logarithm Cost, L ({ pi},{ti) it is total logarithm cost;
The logarithm cost L of classificationclsIt is calculated by formula (15),
Window returns logarithm cost LregIt is calculated by formula (16),
In formula, R is the cost function of the robust of definition, belongs to Smooth L1 error, insensitive to outlier, uses formula (17) it calculates,
In formula (14)This means only positive candidate region, i.e.,Shi Caiyou returns generation Valence, other situations due toDo not return cost;Classification layer and window return the output of layer respectively by { piAnd { tiGroup At this two respectively by NclsAnd NregAnd a balance weight λ normalization, λ=10, N are selected herecls=256, Nreg= 2400, it is almost equal weight that selection sort layer and window in this way, which returns layer item,;
About position refine, here using 4 values, centre coordinate, width and height, calculation method is as follows,
In formula, x, y, w, h, which are respectively indicated, surrounds frame centre coordinate, width and height, xa、ya、wa、haRespectively indicate candidate Regional center coordinate, width and height, x*、y*、w*、h*Respectively indicate encirclement frame centre coordinate, width and the height of prediction;With The calculated result of formula (18) carries out position refine;In fact, using area completely there is no any candidate window is explicitly extracted Suggest that network itself completes judgement and position refine in domain.
The VGG network, the method that multilayer neural network is established in label vehicle image data are divided into two steps, and one It is one layer network of each training, second is that tuning, the advanced expression r for generating original representation X upwards and the advanced expression r give birth to downwards At X' it is as consistent as possible;
The propagated forward process of convolutional neural networks, upper one layer of output are the input of current layer, and pass through activation letter Number successively transmitting, therefore the practical calculating output of whole network is indicated with formula (4),
Op=Fn(…(F2(F1(XW1)W2)…)Wn) (4)
In formula, X expression is originally inputted, FlIndicate l layers of activation primitive, WlIndicate l layers of mapping weight matrix, Op Indicate the practical calculating output of whole network;
The output of current layer (5) expression,
Xl=fl(WlXl-1+bl) (5)
In formula, l represents the network number of plies, XlIndicate the output of current layer, Xl-1Indicate one layer of output, i.e. current layer Input, WlRepresent trained, current network layer mapping weight matrix, blBigoted, the f for the additivity of current networklIt is to work as The activation primitive of preceding network layer;The activation primitive f of uselTo correct linear unit, i.e. ReLU is indicated with formula (6),
In formula, l represents the network number of plies, WlRepresent trained, current network layer mapping weight matrix, flIt is to work as The activation primitive of preceding network layer;It is to allow it to be 0 if convolutional calculation result is less than 0 that it, which is acted on,;Otherwise keep its value constant.
The VGG network, first 5 layers are a typical depth convolutional neural networks, which is one Back-propagation process optimizes deconvolution parameter and biasing using stochastic gradient descent method by error function backpropagation Adjustment until network convergence or reaches maximum number of iterations stopping;
Backpropagation is needed by being compared to the training sample with label, right using square error cost function In c classification, the multi-class of N number of training sample is identified, network final output error function calculates mistake with formula (7) Difference,
In formula, ENFor square error cost function,It is tieed up for the kth of n-th of sample corresponding label,For n-th of sample K-th of output of corresponding network prediction;
When carrying out backpropagation to error function, using the similar calculation method of traditional BP algorithm, such as formula (8) institute Show,
In formula, δlRepresent the error function of current layer, δl+1Represent one layer of error function, Wl+1For upper one layer of mapping square Battle array, f' indicate the inverse function of activation primitive, that is, up-sample, ulIndicate upper one layer of the output for not passing through activation primitive, xl-1It indicates Next layer of input, WlWeight matrix is mapped for this layer;
After error back propagation, the error function δ of each network layer is obtainedl, then use stochastic gradient descent method To network weight WlIt modifies, then carries out next iteration, until network reaches the condition of convergence;It is needed when carrying out error propagation The up-sampling in formula (8) is first passed through so that two layers of front and back size is identical, then carries out error propagation;
Algorithm idea is: 1) successively building monolayer neuronal is first first, is one single layer network of training every time in this way;2) when After all layers have been trained, tuning is carried out using wake-sleep algorithm.
Deep learning training process is specific as follows:
STEP21: using unsupervised learning from bottom to top, i.e., since bottom, past top layer in layer is trained, and is learned It practises vehicle image feature: first with no label vehicle image data training first layer, first learning the parameter of first layer when training, due to The limitation of model capacity and sparsity constraints, the model enabled learns the structure to data itself, to obtain The feature of expression ability is had more than inputting;After study obtains l-1 layers, by l-1 layers of output as l layers of input, L layers of training, thus respectively obtains the parameter of each layer;It is specific to calculate as shown in formula (5), (6);
STEP22: top-down supervised learning goes to train by the vehicle image data of tape label, error from push up to Lower transmission, is finely adjusted network: specific to calculate as shown in formula (7), (8);
The parameter of entire multilayered model is further finely tuned based on the obtained each layer parameter of STEP21, this step, which is one, prison Superintend and direct training process;STEP21 similar to neural network random initializtion initial value process, due to the STEP21 of deep learning be not with Machine initialization, but obtained by the structure of study input data, thus this initial value is closer to global optimum, so as to Obtain better effect.
Preceding 5 layers of the model initialization of the VGG network: it is broadly divided into data preparation, calculates image mean value, network 5 steps such as definition, training and recovery data;
1) data preparation;The image data that all kinds of vehicles are had collected by crawler software, what is obtained substantially has mark The vehicle image data of label, as training image data;Another kind of data are the vehicle figures obtained by bayonet camera As data;
2) image mean value is calculated;Model needs to subtract mean value from every picture;
3) definition of network;Main definitions xml tag path, picture path, storage train.txt, val.txt, The path of test.txt and trainval.txt file;
4) training;Run training module;
5) restore data;The layer of ReLu5 before deleting, and the bottom of roi_pool5 is changed to data and rois;
The model initialization work of vehicle vehicle pre-training is completed by above-mentioned processing.
Suggest that preceding 5 layers of low-level image feature extraction of the VGG network is utilized as a result, i.e. two nets in network in the region Network has shared preceding 5 layers of low-level image feature of the VGG network, needs to learn by alternative optimization the shared feature of optimization;Tool Body algorithm is as follows:
STEP41: suggest that network is suggested in the optimization training region of network with region, with above-mentioned data preparation, calculate image 5 steps such as mean value, the definition of network, training and recovery data complete model initializations, and end-to-end fine tuning is built for region View task;
STEP42: suggesting the Suggestion box that network generates with the region of STEP1, by one individually inspection of quick R-CNN training Survey grid network, this detection network are equally by the model initialization of vehicle vehicle pre-training, and at this time two networks are not yet Shared convolutional layer;
STEP43: suggest network training, fixed shared convolutional layer with detection netinit region, and only finely tune area The exclusive layer of network is suggested in domain, at this moment two network share convolutional layers;
STEP44: keeping shared convolutional layer to fix, and finely tunes the classification layer of quick R-CNN;In this way, two network share phases Same convolutional layer, finally constitutes a unified network.
In the present invention, vehicle vehicle visual identity main flow is as follows;
STEP51: images to be recognized is read;
STEP52: being normalized images to be recognized, and it is normalized to obtain tri- different colours 224 × 224 of RGB Image data;
STEP53: the normalized image data of tri- different colours of RGB 224 × 224 is input to three channels CNN, through 5 Layer process of convolution obtains vehicle vehicle character image data;
STEP54: suggesting the Suggestion box of network generation to vehicle vehicle character image data by region, chooses one most The Suggestion box of high score is to get to an area-of-interest, RoI;This RoI is passed through maximum 5 layers of pondization to be handled to obtain The trellis diagram of one 6 × 6 × 256RoI;
STEP55: the trellis diagram of RoI is exported and obtains the feature of 4096 dimensions to after two layer processing connected entirely at the same level Vector, the input data as softmax classifier;
STEP56: with the classification regression analysis of softmax obtaining vehicle vehicle cab recognition to the feature vector of 4096 dimensions as a result, It identifies by which kind of vehicle the vehicle in altimetric image belongs to.
Beneficial effects of the present invention are mainly manifested in:
1) a kind of model recognizing method based on quick R-CNN deep neural network is provided;
2) a kind of deep learning method is researched and developed, realizes unsupervised vehicle feature extraction;
3) a kind of classifier of type subclass suitable for thousands of kinds is designed, and there is scalability;
4) one quick R-CNN network implementations vehicle detection and knowledge end to end of use truly is realized Other frame, and there is the Morphological Diversity suitable for vehicle target, illumination variation diversity is fast under the environment such as background diversity The vehicle subclass identification of speed, high-precision and robustness.
Detailed description of the invention
Fig. 1 is the detection algorithm process of marginal information candidate frame;
Fig. 2 is the process content in region suggestion network;
Fig. 3 is the sliding window schematic diagram of 3 kinds of scales and 3 kinds of length-width ratios;
Fig. 4 is the synoptic diagram that network is suggested in region;
Fig. 5 is that region Suggestion box generates explanatory diagram;
Fig. 6 is the explanatory diagram of the shared network in quick R-CNN network;
Fig. 7 is the candidate region for suggesting obtaining after network processes by region to vehicle on real road;
Fig. 8 is to realize that region suggests that network itself completes judgement and position refine schematic diagram with cost function;
Fig. 9 is the vehicle cab recognition flow diagram based on CNN;
Figure 10 is wake-sleep algorithmic descriptions figure;
Figure 11 is the vehicle cab recognition flow diagram of the CNN of VGG model;
Figure 12 is the training process figure of quick R-CNN network implementations vehicle detection and identification end to end;
Figure 13 is quick CNN network implementations vehicle detection and identification process synoptic diagram end to end;
Figure 14 is the flow chart of quick R-CNN network implementations vehicle detection and identification end to end.
Specific embodiment
Technical solution of the present invention is described in further detail with reference to the accompanying drawing.
Embodiment 1
Referring to Fig.1~14, the technical solution adopted by the present invention to solve the technical problems is:
(1) about the fast vision partitioning algorithm for designing a kind of Vehicle Object;
Firstly, designing a kind of fast vision partitioning algorithm of Vehicle Object, i.e., carry out regional choice to Vehicle Object and determining Position;
In order to which the position to vehicle target positions;Since vehicle target possibly is present at any position of image, and And the size of target, Aspect Ratio are not known yet, original technology be original adoption sliding window strategy to entire image into Row traversal, and need to be arranged different scales, different length-width ratios;Although the strategy of this exhaustion contains that target is all can The position that can occur, but disadvantage is also obvious: time complexity is too high, and generation redundancy window is too many, this is also serious Influence speed and performance that subsequent characteristics are extracted and classify;
For sliding window there are the problem of, the invention proposes a kind of solutions of candidate region;Find out in advance The position that vehicle target is likely to occur in figure;The information such as texture, edge, color in image, energy are utilized due to candidate region Guarantee to keep higher recall rate in the case where choosing less window;The time that subsequent operation can be effectively reduced in this way is complicated Degree, and the candidate window obtained is higher than the quality of sliding window;Available algorithm is selective search, i.e., Selective Search and marginal information candidate frame, i.e. edge Boxes;The core of these algorithms is that human vision is utilized " take a panoramic view of the situation " at a glance, direct " general position " of the discovery vehicle target in entire image;Since selective search is calculated Method time-consuming is bigger, is not suitable for and online vehicle cab recognition and detection;The present invention uses the detection algorithm of marginal information candidate frame.
The detection algorithm thought of marginal information candidate frame is: utilize marginal information, determine profile number in candidate frame and It with the profile number of candidate frame imbricate, and is scored candidate frame based on this, further according to the sequence of score Determine the candidate region information being made of size, length-width ratio, position;Detection algorithm process such as Fig. 1 institute of marginal information candidate frame Show;Algorithm steps are as follows:
STEP11: original image is handled with structure depth woods edge detection algorithm, obtained edge image, then with non- Maximum restrainable algorithms are further processed edge image to obtain a relatively sparse edge image;
STEP12: the marginal point in relatively sparse edge image almost point-blank puts together to form one A edge group, specific way are the marginal point of 8 connections ceaselessly to be found, until the orientation angle between marginal point two-by-two is poor Value and be greater than pi/2, just obtained the more a edge group s of N in this wayi∈S;
STEP13: calculating the similarity between two two edges groups with formula (1),
a(si,sj)=| cos (θiij)cos(θjij)|γ(1)
In formula, θiAnd θjThe average orientation of respectively two edge groups, siAnd sjRespectively indicate two edge groups, θijIt is two The mean place x of a edge groupiAnd xjBetween angle, γ be similar sensitivity coefficient, here select γ=2, a (si,sj) two Similarity between edge group;In order to improve computational efficiency, here by similarity a (si,sj) calculated value be more than threshold value, Ts≥ 0.05 edge group is stored, remaining is disposed as zero;
STEP14: assigning a weight to each edge group, and weight calculation method is provided by formula (2),
In formula, T is that s is reached since the edge of candidate frameiEdge group arrangement set path, Wb(si) it is edge si Weight, tjFor ..., (parameter interpretation);By W if not finding pathb(si) it is set as 1;
STEP15: calculating the scoring of candidate frame with formula (3),
In formula, miFor in edge group siIn all edge p size mpSummation, Wb(si) it is edge siWeight, bw And bhThe respectively width and height of candidate frame, k are size coefficient, define k=1.5 here;Calculation window inward flange number into Row marking, last Ordering and marking filter out low point of candidate frame;For present invention is mainly applied to the extraction of bayonet vehicle, this In just select the candidate frame of best result as tested vehicle object foreground image;
(2) about a kind of deep learning method is researched and developed, unsupervised vehicle feature extraction is realized;
Due to the Morphological Diversity of vehicle target, illumination variation diversity, the factors such as background diversity make design one The feature of robust is not so easy;However the quality for extracting feature directly influences the accuracy of classification;
Vehicle feature extraction that meet above three diversity requirement, robust must be by unsupervised deep learning come real It now completes, successively initialization is a highly useful solution;The essence of deep learning is by building with many hidden Layer machine learning model and magnanimity training data, to learn more useful feature, thus finally promoted classification or prediction Accuracy;Therefore, by following two points realize deep learning in the present invention: 1) depth of model structure is of five storeys~10 multilayers Hidden node;2) by layer-by-layer eigentransformation, the character representation by sample in former space transforms to a new feature space, from And makes to classify or predict to be more easier;
The present invention proposes one kind in non-supervisory data, i.e., without establishing multilayer nerve net in label vehicle image data The method of network is briefly divided into two steps, first is that one layer network of training every time, second is that tuning, generates original representation X upwards Advanced expression r and the advanced expression r X' generated downwards it is as consistent as possible;
The propagated forward process of convolutional neural networks, upper one layer of output are the input of current layer, and pass through activation letter Number successively transmitting, therefore the practical calculating output of whole network is indicated with formula (4),
Op=Fn(…(F2(F1(XW1)W2)…)Wn) (4)
In formula, X expression is originally inputted, FlIndicate l layers of activation primitive, WlIndicate l layers of mapping weight matrix, l= 1,2,3 ... represents the network number of plies, OpIndicate the practical calculating output of whole network;
The output of current layer formula (5) expression,
Xl=fl(WlXl-1+bl) (5)
In formula, l represents the network number of plies, XlIndicate the output of current layer, Xl-1Indicate one layer of output, i.e. current layer Input, WlRepresent trained, current network layer mapping weight matrix, blBigoted, the f for the additivity of current networklIt is to work as The activation primitive of preceding network layer;The activation primitive f that the present invention useslTo correct linear unit, i.e. Rectified Linear Units, ReLU, with formula (6) indicate,
In formula, l represents the network number of plies, WlRepresent trained, current network layer mapping weight matrix, flIt is to work as The activation primitive of preceding network layer;It is to allow it to be 0 if convolutional calculation result is less than 0 that it, which is acted on,;Otherwise keep its value constant;
Network parameter is further decreased using the shared method of local receptor field and weight in CNN.So-called part Receptive field refers to that every kind of convolution kernel is only connected with some specific region in image, i.e. every kind of convolution kernel convolved image A part, next again in other layers by these part convolution features link together, both met image pixel in space On relevance, and reduce deconvolution parameter quantity.And weight is shared to be passed through so that the weight of every kind of convolution kernel is all the same Increase the type of convolution kernel to extract the multi-party region feature of image;It is special in order to provide more details to the classification of vehicle vehicle subclass It levies, appropriately increases the type of convolution kernel in the present invention;
Convolutional neural networks training is a back-propagation process, similar with traditional BP algorithm, anti-by error function To propagation, deconvolution parameter and biasing are optimized and revised using stochastic gradient descent method, until network convergence or reached most Big the number of iterations stops.
Backpropagation needs to calculate error by being compared the training sample with label.For example, by using a square mistake Poor cost function, for c classification, the multi-class identification problem of N number of training sample, network final output error function formula (7) it indicates,
In formula, ENFor square error cost function,It is tieed up for the kth of n-th of sample corresponding label,For n-th of sample K-th of output of corresponding network prediction;
When carrying out backpropagation to error function, using the similar calculation method of traditional BP algorithm, such as formula (8) institute Show,
In formula, δlRepresent the error function of current layer, δl+1Represent one layer of error function, Wl+1For upper one layer of mapping square Battle array, f' indicate the inverse function of activation primitive, that is, up-sample, ulIndicate upper one layer of the output for not passing through activation primitive, xl-1It indicates Next layer of input, WlWeight matrix is mapped for this layer;
After error back propagation, the error function δ of each network layer is obtainedl, then use stochastic gradient descent method To network weight WlIt modifies, then carries out next iteration, until network reaches the condition of convergence;It should be noted that due to layer Size dimension between layer is different, needs to first pass through the up-sampling in formula (8) when carrying out error propagation so that front and back two Layer size is identical, then carries out error propagation;
Algorithm idea is: 1) successively building monolayer neuronal is first first, is one single layer network of training every time in this way;2) when After all layers have been trained, tuning is carried out using wake-sleep algorithm.
Deep learning training process is specific as follows:
STEP21: using unsupervised learning from bottom to top, i.e., since bottom, past top layer in layer is trained, and is learned It practises vehicle image feature: first with no label vehicle image data training first layer, first learning the parameter of first layer when training, due to The limitation of model capacity and sparsity constraints, the model enabled learns the structure to data itself, to obtain The feature of expression ability is had more than inputting;After study obtains l-1 layers, by l-1 layers of output as l layers of input, L layers of training, thus respectively obtains the parameter of each layer;It is specific to calculate as shown in formula (5), (6);
STEP22: top-down supervised learning goes to train by the vehicle image data of tape label, error from push up to Lower transmission, is finely adjusted network: specific to calculate as shown in formula (7), (8);
The parameter of entire multilayered model is further finely tuned based on the obtained each layer parameter of STEP21, this step, which is one, prison Superintend and direct training process;STEP21 similar to neural network random initializtion initial value process, due to the STEP21 of deep learning be not with Machine initialization, but obtained by the structure of study input data, thus this initial value is closer to global optimum, so as to Obtain better effect;So deep learning effect quality is largely attributed to the fact that the feature learning process of STEP21;
For the vehicle image data set of tape label, the present invention collects the vehicle figure of a variety of models using web crawlers technology Picture, vehicle image data of the vehicle image being collected into this way using manual confirmation as tape label;
About wake-sleep algorithm, see that attached drawing 8, main thought are to pass through in the wake stage to given generation weight Study obtains cognition weight;In the sleep stage to given cognition weight, obtain generating weight by study;
Wake stage, l layers of generation weight gl, it is updated with formula (9),
Δgl=ε sl+1(sl-pl) (9)
In formula, Δ glFor l layers of generation weight glUpdate changing value, ε is learning rate, sl+1For l+1 layers of neuron Liveness, slFor l layers of neuron liveness, plActivation probability when being driven for the current state of l layers of neuron;
Sleep stage, l layers of cognition weight wl, it is updated with formula (10),
Δwl=ε sl-1(sl-ql) (10)
In formula, Δ wlFor l layers of cognition weight wlUpdate changing value, ε is learning rate, sl-1For l-1 layers of neuron Liveness, slFor l layers of neuron liveness, qlThe preceding layer neuron of present cognitive weight is used for l layers of neuron Activation probability when current state drives;
(3) a kind of classifier of type subclass suitable for thousands of kinds is designed, and there is scalability;
The present invention classifies to vehicle using Softmax classifier;Softmax classifier is the feature robust the case where Under, there is preferable classifying quality, while this classifier has scalability, without to original training after new vehicle occurs Good network characterization carries out relearning training, increases the practicability of system;Softmax principle is the input number that will be extracted Be compared according to feature with feature database, calculate the probability of each classification results, then take the highest result of probability into Row output;
Using the learning outcome in CNN as the input data of softmax classifier;Softmax recurrence is towards multiclass point The case where Logistic of class problem is returned, and is the general type that Logistic is returned, is suitable for mutual exclusion between classification.Assuming that pair In training set { (x(1),y(1),…,x(m),y(m)), there is y(1)∈ { 1,2 ..., k } inputs x for given sample, exports one The vector of k dimension indicates that the probability that each classification results occurs is p (y=i | x), it is assumed that function h (x) is as follows:
θ12,…θkIt is the parameter of model, and all probability and be 1.Cost function after regularization term is added are as follows:
Partial derivative of the cost function to first of parameter of j-th of classification are as follows:
Finally, realizing that the classification of softmax returns by minimizing J (θ).
(4) one quick R-CNN network implementations vehicle detection and knowledge end to end of use truly is designed Other frame;
The present invention will focus on solving following three problems:
1) how design section suggests network;
2) region how to be trained to suggest network;
3) region how to be allowed to suggest that network and quick R-CNN network sharing features extract network;
Suggest that Web content specifically includes that network structure, feature extraction, the design of region suggestion network are suggested in region in region With training thinking, candidate region, window classification and position refine;
Suggest network structure, feature extraction, the design of region suggestion network in region:
Region suggests that network structure, feature extraction, region suggest that the design of network is as shown in Fig. 11, in order to multiple It is calculated on GPU, is divided into several groups in calculating in layer, the calculating in every group corresponds to GPU by it and completes, and can be promoted in this way Calculating speed;It is a kind of VGG network, i.e. visual geometry group shown in Figure 12,8 Ge Juan bases, 3 full articulamentums, It is 11 layers total;There are the convolutional layer of 5 groups, 2 classification layers to extract characteristics of image, 1 classification layer characteristic of division in 8 Ge Juan bases;3 A full articulamentum is separately connected classification layer 6, classification layer 7 and classification layer 8;
In Figure 11, normalized 224 × 224 image is sent directly into network, first five stage is the pond convolution+ReLU+ on basis The form of change, inputs p candidate region again at the ending of the 5th stage, and candidate region has 1 picture numbers and 4 geometry positions Confidence breath;Each candidate region is uniformly divided into M × N block by the pond the RoI- layer at the ending of the 5th stage, is carried out most to every piece Great Chiization operation;Candidate region not of uniform size on characteristic pattern is changed into the unified data of size, be sent into next layer of training with Identification;First five above-mentioned stage technique is all to be proved to mature technology in convolutional neural networks technology, it is important to how to make this A little candidate regions can be multiplexed the network characterization in these first five stages of picture;
Suggest that network and quick R-CNN network sharing features extract network in region:
Attached drawing 6 describes region and suggests the convolutional layer output how network and quick R-CNN share, first five stage in Figure 11 Belong to a primitive character and extract network, suggest network formation zone Suggestion box to region respectively later and is examined to quick R-CNN The characteristics of image in the Suggestion box of region is surveyed, two layers connected entirely at the same level, i.e. classification layer 6+ in Figure 11 are then output to ReLU6 and classification layer 7+ReLU7, obtains the p characteristic patterns containing 4096 vectors, then gives classification layer respectively and window returns Layer is handled;
Formation zone Suggestion box:
For formation zone Suggestion box, the present invention is sliding in the convolution Feature Mapping of the last one shared convolutional layer output Small network is moved, this network is connected to entirely in the spatial window of n × n of input convolution Feature Mapping, as shown in Fig. 5;Each Sliding window is mapped on a low-dimensional vector, and the low-dimensional vector in attached drawing 5 is 256-d, a sliding of each Feature Mapping Window corresponds to a numerical value;This vector is exported to two layers connected entirely at the same level;- a window recurrence layer and a classification Layer;Window returns layer and exports on each position, and 9 kinds of recommendation regions correspond to window and need to have translation scaling invariance, returns Layer exports 4 translation zooming parameters from 256 dimensional features;Classification layer is exported from 256 dimensional features belongs to the general of foreground and background Rate;
For example, in the position of each sliding window, while predicting that k region is suggested, therefore window returns layer and has 4k a Output, i.e., the codes co-ordinates of k Suggestion box;Classify layer export 2k Suggestion box score, i.e., to each Suggestion box be vehicle target/ The estimated probability of non-vehicle target;
Candidate region: k Suggestion box is parameterized by the corresponding k frame for being known as candidate regions, i.e., each candidate region with Centered on current sliding window mouth center, and a kind of corresponding scale and length-width ratio, the present invention use 3 kinds of scales and 3 kinds of length-width ratios, such as Shown in attached drawing 3;Just there is k=9 kind candidate region in each sliding position in this way;The convolution feature that size is W × H is reflected It penetrates, a total of W × H × k candidate region;
In entire quickly R-CNN algorithm, there are three types of graphical rules altogether: 1) original image scale: the size of original input picture, It is unrestricted, do not influence performance;2) Normalized Scale: input feature vector extracts the size of network, is arranged in test, candidate Region is set on this scale;This parameter and the relative size of candidate region determine the target zone for wanting detection;3) Network inputs scale: input feature vector detects the size of network, is arranged in training, is 224 × 224.
In conclusion region suggests that network is to export the set of rectangular target Suggestion box, often using an image as input A frame has the score of a Vehicle Object, as shown in Fig. 7;
Suggest that Web content specifically includes that training sample, cost function and hyper parameter in training region;
The training of region suggestion network:
The training of network is suggested in region, and whether the present invention distributes a binary label to each candidate region, i.e., be Vehicle Object;Here distribute positive label to two class candidate regions: (i) and some GT, the enclosing region of ground truth have most The ratio between high intersection union, IoU, Intersection-over-Union, the candidate region of overlapping;(ii) it is surrounded with any GT IoU of the region greater than 0.7 overlapping candidate region;Simultaneously distribute negative label give all GT enclosing regions IoU ratio it is all low In 0.3 candidate region;Leave out non-just non-negative candidate region;
Training sample algorithm:
STEP31: sequence reads every figure in training set;
STEP32: it to the true value candidate region of each calibration, overlaps the maximum candidate region of ratio and is denoted as prospect sample This;
STEP33:: to the remaining candidate region STEP32, if it is Chong Die with some calibration, IoU ratio is greater than 0.7, note For prospect sample;If the overlap proportion of itself and any one calibration is both less than 0.3, it is denoted as background sample;
STEP34: the remaining candidate region STEP32 and STEP33 is discarded;
STEP35: the candidate region across image boundary is discarded.
Cost function:
According to these definition, the multitask cost followed here, using minimum objective function;To the cost of an image Function formula (14) expression,
In formula, i is the index of candidate region in a batch processing, NclsFor the normalization coefficient for layer of classifying, NregTo return The normalization coefficient of layer, λ is balance weight, piFor the prediction probability of vehicle target,For GT label, if candidate region is positiveIf candidate region is negativetiFor a vector, 4 parametrization coordinates of the encirclement frame of prediction are indicated,For GT corresponding with positive candidate region surrounds the coordinate vector of frame, LclsFor the logarithm cost of classification, LregTo return logarithm cost, L ({pi},{ti) it is total logarithm cost;
The logarithm cost L of classificationclsIt is calculated by formula (15),
Window returns logarithm cost LregIt is calculated by formula (16),
In formula, R is the cost function of the robust of definition, belongs to Smooth L1 error, insensitive to outlier, uses formula (17) it calculates,
In formula (14)This means only positive candidate region, i.e.,Shi Caiyou returns generation Valence, other situations due toDo not return cost;Classification layer and window return the output of layer respectively by { piAnd { tiGroup At this two respectively by NclsAnd NregAnd a balance weight λ normalization, λ=10, N are selected herecls=256, Nreg= 2400, it is almost equal weight that selection sort layer and window in this way, which returns layer item,;
About position refine, here using 4 values, centre coordinate, width and height, calculation method is as follows,
In formula, x, y, w, h, which are respectively indicated, surrounds frame centre coordinate, width and height, xa、ya、wa、haRespectively indicate candidate Regional center coordinate, width and height, x*、y*、w*、h*Respectively indicate encirclement frame centre coordinate, width and the height of prediction;With The calculated result of formula (18) carries out position refine;In fact, using area completely there is no any candidate window is explicitly extracted Suggest that network itself completes judgement and position refine in domain;
The optimization of region suggestion network:
Region suggests that network can be embodied as full convolutional network naturally, end-to-end by backpropagation and stochastic gradient descent Training;Here using this network of the sampling policy training of picture centre, each batch processing is by containing many positive negative samples Single image composition;256 candidate regions are randomly sampled in one image, calculate the cost in batch processing with formula (14) Function, wherein the ratio of the positive and negative candidate region sampled is 1:1;As soon as if the positive sample number in image less than 128, we Other remaining candidate regions in this batch processing are filled up with negative sample;Here batch processing is dimensioned to 256;
By from the Gaussian Profile that zero-mean standard deviation is 0.01 weight that obtains come all mew layers of random initializtion, institute Meaning mew layer refers to suggesting the subsequent layer of network in region, such as the classification layer 6+ReLU6 and classification layer 7+ReLU7 in attached drawing 11; Every other layer, i.e. shared convolutional layer pass through the classification samples pre-training to vehicle vehicle such as first five layer in attached drawing 11 Model initializes;The present invention is 0.001 for the learning rate of 60k batch processing on vehicle model data collection, for The learning rate of next 20k batch processing is 0.0001;Momentum is 0.9, and weight decays to 0.0005;
The model initialization of vehicle vehicle pre-training: be broadly divided into data preparation, calculate image mean value, network definition, 5 steps such as training and recovery data;
1) data preparation;The new folder myself in data, we have collected all kinds of vehicles by crawler software Image data, due to band keyword scan for substantially with the vehicle image data of label, we by its As training data;Another kind of data are the vehicle image data that we are obtained by bayonet camera;
The input of training and test is described with train.txt and val.txt, lists the mark of All Files and they Label;The name of classification is the sequence of ASCII character, i.e. 0-999, and corresponding systematic name is mapped in synset_ with number In words.txt;Val.txt is unable to label, is all arranged to 0;Then the size of picture is unified into 256 × 256;Then exist Myself file is created in caffe-master/examples, then by caffe-maester/examples/imagenet Create_imagenet.sh copy under this document folder, its name is changed to create_animal.sh, modification training and is surveyed The setting for trying path, runs the sh;Finally obtain myself_train_lmdb and myself_val_lmdb;
2) image mean value is calculated;Model needs us to subtract mean value from every picture, thus we must obtain it is trained Mean value is realized with tools/compute_image_mean.cpp, equally replicates caffe-maester/examples/ In ./make_imagenet_mean to the examples/myself of imagenet, it is renamed as make_car_mean.sh, Modified path;
3) definition of network;All Files in caffe-master/models/bvlc_reference_caffenet It copies in caffe-master/examples/myself file, modifies train_val.prototxt, pay attention to modifying number According to the path of layer;
In training, we are with a softmax-loss layers of calculating loss function and initialize backpropagation, and are testing Card, our service precision layers detect our precision;There are one the agreement solver.prototxt run, copied, generals The first row path is changed to our path net: " examples/myself/train_val.prototxt ",
Test_iter:1000 refers to the batch of test;Test_interval:1000 refers to every 1000 iteration tests one It is secondary;Base_lr:0.01 is basic learning rate;Lr_policy: " step " learning rate variation;The variation of gamma:0.1 learning rate Ratio;Every 100000 iteration of stepsize:100000 reduce learning rate;The every 20 layers of display of display:20 is primary;max_ Iter:450000 maximum number of iterations;The parameter of momentum:0.9 study;The parameter of weight_decay:0.0005 study; 10000 display states of the every iteration of snapshot:10000;The end solver_mode:GPU adds a line, represents and uses GPU operation;
4) training;Train_caffenet.sh in caffe-master/examples/imagenet is copied next And entitled train_myself.sh operation is modified, modify the path of the inside;
5) restore data;The resume_training.sh duplication in caffe-master/examples/imagenet Come over and runs;
The model initialization work of vehicle vehicle pre-training is completed by above-mentioned processing;Further, the present invention proposes one 4 step training algorithms of kind, learn shared feature by alternative optimization;
STEP41: suggesting that network is suggested in the optimization training region of network with region, equal with above-mentioned data preparation, calculating image 5 steps such as value, the definition of network, training and recovery data complete model initializations, and end-to-end fine tuning is suggested for region Task;
STEP42: suggesting the Suggestion box that network generates with the region of STEP1, by one individually inspection of quick R-CNN training Survey grid network, this detection network are equally by the model initialization of vehicle vehicle pre-training, and at this time two networks are not yet Shared convolutional layer;
STEP43: suggest network training, fixed shared convolutional layer with detection netinit region, and only finely tune area The exclusive layer of network is suggested in domain, at this moment two network share convolutional layers;
STEP44: keeping shared convolutional layer to fix, and finely tunes the classification layer of quick R-CNN;In this way, two network share phases Same convolutional layer, finally constitutes a unified network;
The visual identity of vehicle vehicle:
We provide a vehicle vehicle visual identity main flow below, and entire process flow is as shown in Fig. 13;
STEP51: images to be recognized is read;
STEP52: being normalized images to be recognized, and it is normalized to obtain tri- different colours 224 × 224 of RGB Image data;
STEP53: the normalized image data of tri- different colours of RGB 224 × 224 is input to three channels CNN, through 5 Layer process of convolution obtains vehicle vehicle character image data;
STEP54: suggesting the Suggestion box of network generation to vehicle vehicle character image data by region, chooses one most The Suggestion box of high score is to get to an area-of-interest, RoI;This RoI is passed through maximum 5 layers of pondization to be handled to obtain The trellis diagram of one 6 × 6 × 256RoI;
STEP55: the trellis diagram of RoI is exported and obtains the feature of 4096 dimensions to after two layer processing connected entirely at the same level Vector, the input data as softmax classifier;
STEP56: with the classification regression analysis of softmax obtaining vehicle vehicle cab recognition to the feature vector of 4096 dimensions as a result, By vehicle vehicle classification at 1000 types in the present invention, to identify by which kind of vehicle the vehicle in altimetric image belongs to.
Embodiment 2
Visual identification technology of the invention has universality, and the subclass suitable for other objects identifies, as long as participating in training Data run learnt in the system developed of the present invention, can be achieved with the class object after obtaining the feature of the class object Subclass identification mission.
Embodiment 3
Visual identification technology of the invention has scalability, without to original trained network after new subclass occurs Feature carries out relearning training, as long as being trained study to new subclass, and softmax classifier in systems expands Open up the data of classification.
The foregoing is merely preferable implementation examples of the invention, are not intended to restrict the invention, it is all in spirit of that invention and Within principle, any modification, equivalent replacement, improvement and so on be should all be included in the protection scope of the present invention.

Claims (9)

1. a kind of model recognizing method based on quick R-CNN deep neural network, it is characterised in that: be used for depth including one The VGG network of study and training identification, a region suggestion network for being used to extract area-of-interest and one are for vehicle The Softmax classifier of classification;
The VGG network, including 8 Ge Juan bases, 3 full articulamentums amount to 11 layers;8 Ge Juan bases include the convolution of 5 groups Layer, 2 classification layers extract characteristics of image and 1 classification layer characteristic of division;3 full articulamentums are separately connected classification layer 6, classification layer 7 and classification layer 8;
Network is suggested in the region, including 1 layer of classifying, 1 window recurrence layer, 1 module for calculating Classification Loss and 1 Calculation window returns the module of loss, p interested Suggestion box of output;
The input data feature and the learning training that extract are obtained feature database data and compared by the Softmax classifier It is right, the probability of each classification results is calculated, the highest result of probability is then taken to be exported;
Quick R-CNN deep neural network has accessed the region at the 5th layer of end of the VGG network and has suggested network, So that first 5 layers of low-level image feature extraction process and result of VGG network described in network share is suggested in the region;
The 6th layer of the VGG network and the 7th layer p interested Suggestion box for suggesting that network is exported according to the region Interior characteristics of image carries out convolution sum ReLU processing, obtains the p characteristic patterns containing 4096 vectors, then gives classification layer respectively Layer is returned with window to be handled, and realizes the segmentation of vehicle image;The Softmax classifier contains 4096 vectors to p Characteristic pattern carry out Classification and Identification, obtain the classification results of vehicle vehicle.
2. model recognizing method as described in claim 1, it is characterised in that: network is suggested in the region, for generation area Domain Suggestion box has accessed the region at the 5th layer of end of the VGG network and has suggested network, i.e., in the 5th layer of convolutional layer A small network is slided on the convolution Feature Mapping figure of output, this network is connected to n × n's of input convolution Feature Mapping entirely In spatial window;Each sliding window is mapped on a low-dimensional vector, low-dimensional vector be 256-d, the one of each Feature Mapping The corresponding numerical value of a sliding window;This vector is exported to two layers connected entirely at the same level, and-a window returns layer and one A classification layer;Window returns layer and exports on each position, and 9 kinds of recommendation regions correspond to window translation and scale indeformable, window time Return layer to export 4 translation zooming parameters from 256 dimensional features, there is 4k output, i.e., the codes co-ordinates of k Suggestion box;Classification layer Output belongs to the probability of foreground and background from 256 dimensional features, exports 2k Suggestion box score, i.e., is vehicle to each Suggestion box The estimated probability of target or non-vehicle target.
3. model recognizing method as claimed in claim 1 or 2, it is characterised in that: the training of network is suggested in region, gives each time Whether favored area distributes a binary label, i.e., be Vehicle Object;Here positive label is distributed to two class candidate regions:
(i) there is the overlapping candidate region of the ratio between highest intersection union IoU with the enclosing region of some ground truth GT;
(ii) it is greater than 0.7 overlapping candidate region with the ratio between the intersection union of any ground truth GT enclosing region IoU;Simultaneously Distribute negative label give all GT enclosing regions IoU ratio be below 0.3 candidate region;
Leave out non-just non-negative candidate region;Specific algorithm is as follows:
STEP31: sequence reads every figure in training set;
STEP32: it to the true value candidate region of each calibration, overlaps the maximum candidate region of ratio and is denoted as prospect sample;
STEP33: to the remaining candidate region STEP32, if it is Chong Die with some calibration, and IoU ratio is greater than 0.7, is denoted as Prospect sample;If the overlap proportion of itself and any one calibration is both less than 0.3, it is denoted as background sample;
STEP34: the remaining candidate region STEP32 and STEP33 is discarded;
STEP35: the candidate region across image boundary is discarded.
4. model recognizing method as claimed in claim 3, it is characterised in that: in order to carry out screening and the area of candidate region automatically Domain position refine, here using minimum objective function;The cost function of one image is indicated with formula (14),
In formula, i is the index of candidate region in a batch processing, NclsFor the normalization coefficient for layer of classifying, NregFor window recurrence The normalization coefficient of layer, λ is balance weight, piFor the prediction probability of vehicle target,For GT label, if candidate region is positiveIf candidate region is negativetiFor a vector, 4 parametrization coordinates of the encirclement frame of prediction are indicated,For GT corresponding with positive candidate region surrounds the coordinate vector of frame, LclsFor the logarithm cost of classification, LregTo return logarithm cost, L ({pi},{ti) it is total logarithm cost;
The logarithm cost L of classificationclsIt is calculated by formula (15),
Window returns logarithm cost LregIt is calculated by formula (16),
In formula, R is the cost function of the robust of definition, belongs to Smooth L1 error, insensitive to outlier, with formula (17) It calculates,
In formula (14)This means only positive candidate region, i.e.,Shi Caiyou returns cost, His situation due toDo not return cost;Classification layer and window return the output of layer respectively by { piAnd { tiComposition, this two Item is respectively by NclsAnd NregAnd a balance weight λ normalization, λ=10, N are selected herecls=256, Nreg=2400, pass through It is equal weight that such selection sort layer and window, which return layer item,;
About position refine, here using 4 values, centre coordinate, width and height, calculation method is as follows,
In formula, x, y, w, h, which are respectively indicated, surrounds frame centre coordinate, width and height, xa、ya、wa、haIt respectively indicates in candidate region Heart coordinate, width and height, x*、y*、w*、h*Respectively indicate encirclement frame centre coordinate, width and the height of prediction, tx,ty、tw、th Respectively indicate encirclement frame centre coordinate, width and height after the refine of position;t* x,t* y、t* w、t* hIt respectively indicates through position essence Encirclement frame centre coordinate, width and the height of prediction after repairing;Position refine region is carried out with the calculated result of formula (18) to build Discuss network.
5. model recognizing method as described in claim 1, it is characterised in that: the VGG network, in label vehicle image number The method that multilayer neural network is established on is divided into two steps, first is that one layer network of training every time, second is that tuning, makes original representation X The advanced expression r generated upwards is consistent with the X' that the advanced expression r is generated downwards;
The propagated forward process of convolutional neural networks, upper one layer output be current layer input, and by activation primitive by Layer transmitting, therefore the practical calculating output of whole network is indicated with formula (4),
Op=Fn(…(F2(F1(XW1)W2)…)Wn) (4)
In formula, X expression is originally inputted, FnIndicate l layers of activation primitive, WnIndicate l layers of mapping weight matrix, OpIt indicates The practical calculating of whole network exports;
The output of current layer formula (5) expression,
Xl=fl(WlXl-1+bl) (5)
In formula, l represents the network number of plies, XlIndicate the output of current layer, Xl-1Indicate one layer of output, the i.e. input of current layer, WlRepresent trained, current network layer mapping weight matrix, blIt is biased for the additivity of current network, flIt is current net The activation primitive of network layers;The activation primitive f of uselTo correct linear unit, i.e. ReLU is indicated with formula (6),
In formula, l represents the network number of plies, WlRepresent trained, current network layer mapping weight matrix, flIt is current net The activation primitive of network layers;It is to allow it to be 0 if convolutional calculation result is less than 0 that it, which is acted on,;Otherwise keep its value constant.
6. model recognizing method as claimed in claim 5, it is characterised in that: the VGG network, first 5 layers are a depth Convolutional neural networks, the neural metwork training are that a back-propagation process by error function backpropagation utilizes boarding steps Degree descent method optimizes and revises deconvolution parameter and biasing, until network convergence or reaches maximum number of iterations stopping;
Backpropagation is needed by being compared to the training sample with label, using square error cost function, for c Classification, the multi-class of N number of training sample are identified that network final output error function calculates error with formula (7),
In formula, ENFor square error cost function,It is tieed up for the kth of n-th of sample corresponding label,Net is corresponded to for n-th of sample K-th of output of network prediction;
When carrying out backpropagation to error function, as shown in formula (8),
In formula, δlRepresent the error function of current layer, δl+1Represent one layer of error function, Wl+1For upper one layer of mapping matrix, f' The inverse function for indicating activation primitive, that is, up-sample, ulIndicate upper one layer of the output for not passing through activation primitive, xl-1Indicate next The input of layer, WlWeight matrix is mapped for this layer;
After error back propagation, the error function δ of each network layer is obtainedl, then using stochastic gradient descent method to this layer Map weight matrix WlIt modifies, then carries out next iteration, until network reaches the condition of convergence;When carrying out error propagation Need to first pass through the up-sampling in formula (8) so that two layers of front and back size is identical, then carries out error propagation;
Algorithm idea is: 1) successively building monolayer neuronal is first first, is one single layer network of training every time in this way;2) when all After layer has been trained, tuning is carried out using wake-sleep algorithm;
Deep learning training process is specific as follows:
STEP21: using unsupervised learning from bottom to top, i.e., since bottom, past top layer in layer is trained, and learns vehicle Characteristics of image: first with no label vehicle image data training first layer, when training, first learns the parameter of first layer, due to model The limitation of capacity and sparsity constraints, the model enabled learns the structure to data itself, to obtain than defeated Enter to have more the feature of expression ability;After study obtains l-1 layers, by l-1 layers of output as l layers of input, train L layers, thus respectively obtain the parameter of each layer;
STEP22: top-down supervised learning goes to train by the vehicle image data of tape label, the top-down biography of error It is defeated, network is finely adjusted.
7. model recognizing method as described in claim 1, it is characterised in that: preceding 5 layers of the model of the VGG network is initial Change: being divided into data preparation, calculates image mean value, the definition of network, training and restore 5 steps of data;
1) data preparation;The image data for collecting all kinds of vehicles is taken the photograph including the vehicle image data with label and by bayonet The vehicle image data that camera obtains, and using the vehicle image data with label as training image data;
2) image mean value is calculated;Mean value is subtracted from every picture;
3) definition of network;Define xml tag path, picture path, storage train.txt, val.txt, test.txt and The path of trainval.txt file;
4) training;Run training module;
5) restore data;The layer of ReLu5 before deleting, and the bottom of roi_pool5 is changed to data and rois;
The model initialization work of vehicle vehicle pre-training is completed by above-mentioned processing.
8. model recognizing method as claimed in claim 7, it is characterised in that: the region suggestion network is utilized described Preceding 5 layers of low-level image feature of VGG network extracts as a result, i.e. two network shares preceding 5 layers of low-level image feature of the VGG network, Need to learn by alternative optimization the shared feature of optimization;Specific algorithm is as follows:
STEP41: optimize training region and suggest network, with the data preparation, calculate image mean value, the definition of network, training Model initialization is completed with 5 steps of data are restored, and finely tunes region with mode end to end and suggests task;
STEP42: suggesting the Suggestion box that network generates with the region of STEP41, and by quick R-CNN training, one is individually detected net Network, this detection network are equally by the model initialization of vehicle vehicle pre-training, and at this time two networks are shared not yet Convolutional layer;
STEP43: suggest network training, fixed shared convolutional layer with detection netinit region, and only fine tuning region is built The exclusive layer of network is discussed, at this moment two network share convolutional layers;
STEP44: keeping shared convolutional layer to fix, and finely tunes the classification layer of quick R-CNN;In this way, two network shares are identical Convolutional layer finally constitutes a unified network.
9. model recognizing method as described in claim 1, it is characterised in that: vehicle vehicle visual identity process is as follows;
STEP51: images to be recognized is read;
STEP52: being normalized images to be recognized, obtains tri- normalized image datas of different colours of RGB;
STEP53: tri- normalized image datas of different colours of RGB are input to three channels CNN, are obtained through 5 layers of process of convolution To vehicle vehicle character image data;
STEP54: suggest the Suggestion box of network generation by region to vehicle vehicle character image data, choose a highest and obtain Point Suggestion box to get to an area-of-interest, RoI;This RoI is passed through maximum 5 layers of pondization to be handled to obtain one The trellis diagram of RoI;
STEP55: the trellis diagram of RoI being exported and obtains the feature vector of 4096 dimensions to after two layer processing connected entirely at the same level, Input data as softmax classifier;
STEP56: vehicle vehicle cab recognition is obtained as a result, identification with the classification regression analysis of softmax to the feature vector of 4096 dimensions Which kind of vehicle is belonged to by the vehicle in altimetric image out.
CN201610563184.1A 2016-07-15 2016-07-15 A kind of model recognizing method based on quick R-CNN deep neural network Active CN106250812B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610563184.1A CN106250812B (en) 2016-07-15 2016-07-15 A kind of model recognizing method based on quick R-CNN deep neural network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610563184.1A CN106250812B (en) 2016-07-15 2016-07-15 A kind of model recognizing method based on quick R-CNN deep neural network

Publications (2)

Publication Number Publication Date
CN106250812A CN106250812A (en) 2016-12-21
CN106250812B true CN106250812B (en) 2019-08-20

Family

ID=57613871

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610563184.1A Active CN106250812B (en) 2016-07-15 2016-07-15 A kind of model recognizing method based on quick R-CNN deep neural network

Country Status (1)

Country Link
CN (1) CN106250812B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111540203A (en) * 2020-04-30 2020-08-14 东华大学 Method for adjusting green light passing time based on fast-RCNN

Families Citing this family (196)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106647758A (en) * 2016-12-27 2017-05-10 深圳市盛世智能装备有限公司 Target object detection method and device and automatic guiding vehicle following method
CN106780558B (en) * 2016-12-27 2020-05-12 成都通甲优博科技有限责任公司 Method for generating unmanned aerial vehicle target initial tracking frame based on computer vision point
CN106682697B (en) * 2016-12-29 2020-04-14 华中科技大学 End-to-end object detection method based on convolutional neural network
CN106682696B (en) * 2016-12-29 2019-10-08 华中科技大学 The more example detection networks and its training method refined based on online example classification device
CN106599939A (en) * 2016-12-30 2017-04-26 深圳市唯特视科技有限公司 Real-time target detection method based on region convolutional neural network
CN107636727A (en) * 2016-12-30 2018-01-26 深圳前海达闼云端智能科技有限公司 Target detection method and device
CN106919978B (en) * 2017-01-18 2020-05-15 西南交通大学 Method for identifying and detecting parts of high-speed rail contact net supporting device
US10198655B2 (en) * 2017-01-24 2019-02-05 Ford Global Technologies, Llc Object detection using recurrent neural network and concatenated feature map
CN106845430A (en) * 2017-02-06 2017-06-13 东华大学 Pedestrian detection and tracking based on acceleration region convolutional neural networks
CN108446694B (en) * 2017-02-16 2020-11-27 杭州海康威视数字技术股份有限公司 Target detection method and device
CN106909924B (en) * 2017-02-18 2020-08-28 北京工业大学 Remote sensing image rapid retrieval method based on depth significance
CN106980858B (en) * 2017-02-28 2020-08-18 中国科学院信息工程研究所 Language text detection and positioning system and language text detection and positioning method using same
CN106910176B (en) * 2017-03-02 2019-09-13 中科视拓(北京)科技有限公司 A kind of facial image based on deep learning removes occlusion method
CN106846813A (en) * 2017-03-17 2017-06-13 西安电子科技大学 The method for building urban road vehicle image data base
CN107871136A (en) * 2017-03-22 2018-04-03 中山大学 The image-recognizing method of convolutional neural networks based on openness random pool
CN107016357B (en) * 2017-03-23 2020-06-16 北京工业大学 Video pedestrian detection method based on time domain convolutional neural network
CN107491720A (en) * 2017-04-01 2017-12-19 江苏移动信息系统集成有限公司 A kind of model recognizing method based on modified convolutional neural networks
CN107133616B (en) * 2017-04-02 2020-08-28 南京汇川图像视觉技术有限公司 Segmentation-free character positioning and identifying method based on deep learning
CN107067005A (en) * 2017-04-10 2017-08-18 深圳爱拼信息科技有限公司 A kind of method and device of Sino-British mixing OCR Character segmentations
CN107403424B (en) 2017-04-11 2020-09-18 阿里巴巴集团控股有限公司 Vehicle loss assessment method and device based on image and electronic equipment
CN107392218B (en) * 2017-04-11 2020-08-04 创新先进技术有限公司 Vehicle loss assessment method and device based on image and electronic equipment
CN106971187B (en) * 2017-04-12 2019-07-09 华中科技大学 A kind of vehicle part detection method and system based on vehicle characteristics point
CN107229929A (en) * 2017-04-12 2017-10-03 西安电子科技大学 A kind of license plate locating method based on R CNN
CN107239731B (en) * 2017-04-17 2020-10-30 浙江工业大学 Gesture detection and recognition method based on Faster R-CNN
CN107169421B (en) * 2017-04-20 2020-04-28 华南理工大学 Automobile driving scene target detection method based on deep convolutional neural network
CN107146237B (en) * 2017-04-24 2020-02-18 西南交通大学 Target tracking method based on online state learning and estimation
CN106971174B (en) * 2017-04-24 2020-05-22 华南理工大学 CNN model, CNN training method and CNN-based vein identification method
CN107424184B (en) * 2017-04-27 2019-10-11 厦门美图之家科技有限公司 A kind of image processing method based on convolutional neural networks, device and mobile terminal
US10572773B2 (en) * 2017-05-05 2020-02-25 Intel Corporation On the fly deep learning in machine learning for autonomous machines
CN107045642A (en) * 2017-05-05 2017-08-15 广东工业大学 A kind of logo image-recognizing method and device
CN107665351B (en) * 2017-05-06 2022-07-26 北京航空航天大学 Airport detection method based on difficult sample mining
JP7317717B2 (en) 2017-05-09 2023-07-31 ニューララ インコーポレイテッド Systems and methods that enable memory-bound continuous learning in artificial intelligence and deep learning, operating applications continuously across network computing edges
CN107610087B (en) * 2017-05-15 2020-04-28 华南理工大学 Tongue coating automatic segmentation method based on deep learning
CN107274451A (en) * 2017-05-17 2017-10-20 北京工业大学 Isolator detecting method and device based on shared convolutional neural networks
CN108960015A (en) * 2017-05-24 2018-12-07 优信拍(北京)信息科技有限公司 A kind of vehicle system automatic identifying method and device based on deep learning
CN107103308A (en) * 2017-05-24 2017-08-29 武汉大学 A kind of pedestrian's recognition methods again learnt based on depth dimension from coarse to fine
CN108229524A (en) * 2017-05-25 2018-06-29 北京航空航天大学 A kind of chimney and condensing tower detection method based on remote sensing images
CN107301376B (en) * 2017-05-26 2021-04-13 浙江大学 Pedestrian detection method based on deep learning multi-layer stimulation
CN107330446B (en) * 2017-06-05 2020-08-04 浙江工业大学 Image classification-oriented deep convolutional neural network optimization method
CN107247967B (en) * 2017-06-07 2020-09-18 浙江捷尚视觉科技股份有限公司 Vehicle window annual inspection mark detection method based on R-CNN
CN107730881A (en) * 2017-06-13 2018-02-23 银江股份有限公司 Traffic congestion vision detection system based on depth convolutional neural networks
CN107730903A (en) * 2017-06-13 2018-02-23 银江股份有限公司 Parking offense and the car vision detection system that casts anchor based on depth convolutional neural networks
CN107729799A (en) * 2017-06-13 2018-02-23 银江股份有限公司 Crowd's abnormal behaviour vision-based detection and analyzing and alarming system based on depth convolutional neural networks
CN107730905A (en) * 2017-06-13 2018-02-23 银江股份有限公司 Multitask fake license plate vehicle vision detection system and method based on depth convolutional neural networks
CN107730904A (en) * 2017-06-13 2018-02-23 银江股份有限公司 Multitask vehicle driving in reverse vision detection system based on depth convolutional neural networks
CN107316058A (en) * 2017-06-15 2017-11-03 国家新闻出版广电总局广播科学研究院 Improve the method for target detection performance by improving target classification and positional accuracy
CN107368845B (en) * 2017-06-15 2020-09-22 华南理工大学 Optimized candidate region-based Faster R-CNN target detection method
CN107247954A (en) * 2017-06-16 2017-10-13 山东省计算中心(国家超级计算济南中心) A kind of image outlier detection method based on deep neural network
CN107273502B (en) * 2017-06-19 2020-05-12 重庆邮电大学 Image geographic labeling method based on spatial cognitive learning
CN107301417A (en) * 2017-06-28 2017-10-27 广东工业大学 A kind of method and device of the vehicle brand identification of unsupervised multilayer neural network
CN107451602A (en) * 2017-07-06 2017-12-08 浙江工业大学 A kind of fruits and vegetables detection method based on deep learning
CN107341611A (en) * 2017-07-06 2017-11-10 浙江大学 A kind of operation flow based on convolutional neural networks recommends method
CN107292306A (en) * 2017-07-07 2017-10-24 北京小米移动软件有限公司 Object detection method and device
CN107688773A (en) * 2017-07-07 2018-02-13 北京联合大学 A kind of gesture identification method based on deep learning
CN107730906A (en) * 2017-07-11 2018-02-23 银江股份有限公司 Zebra stripes vehicle does not give precedence to the vision detection system of pedestrian behavior
CN107273872B (en) * 2017-07-13 2020-05-05 北京大学深圳研究生院 Depth discrimination network model method for re-identification of pedestrians in image or video
CN107369154B (en) * 2017-07-19 2020-05-05 电子科技大学 Image detection device
CN107239803A (en) * 2017-07-21 2017-10-10 国家海洋局第海洋研究所 Utilize the sediment automatic classification method of deep learning neutral net
CN110019896B (en) * 2017-07-28 2021-08-13 杭州海康威视数字技术股份有限公司 Image retrieval method and device and electronic equipment
CN107590178B (en) * 2017-07-31 2020-10-16 杭州大搜车汽车服务有限公司 Vehicle type matching method based on VIN code, electronic device and storage medium
CN107527068B (en) * 2017-08-07 2020-12-25 南京信息工程大学 Vehicle type identification method based on CNN and domain adaptive learning
CN107609483B (en) * 2017-08-15 2020-06-16 中国科学院自动化研究所 Dangerous target detection method and device for driving assistance system
CN107491764A (en) * 2017-08-25 2017-12-19 电子科技大学 A kind of violation based on depth convolutional neural networks drives detection method
CN107798335B (en) * 2017-08-28 2020-02-18 浙江工业大学 Vehicle logo identification method fusing sliding window and Faster R-CNN convolutional neural network
CN107679078B (en) * 2017-08-29 2020-01-10 银江股份有限公司 Bayonet image vehicle rapid retrieval method and system based on deep learning
CN107610113A (en) * 2017-09-13 2018-01-19 北京邮电大学 The detection method and device of Small object based on deep learning in a kind of image
CN107609522B (en) * 2017-09-19 2021-04-13 东华大学 Information fusion vehicle detection system based on laser radar and machine vision
CN107885764B (en) * 2017-09-21 2020-12-18 银江股份有限公司 Rapid Hash vehicle retrieval method based on multitask deep learning
CN107563454A (en) * 2017-09-25 2018-01-09 重庆邮电大学 A kind of related cascade of yardstick based on the analysis of 2D/3D automobiles suppresses sorting algorithm
CN107610224B (en) * 2017-09-25 2020-11-13 重庆邮电大学 3D automobile object class representation algorithm based on weak supervision and definite block modeling
CN107665355B (en) * 2017-09-27 2020-09-29 重庆邮电大学 Agricultural pest detection method based on regional convolutional neural network
CN108205580B (en) * 2017-09-27 2021-08-31 深圳市商汤科技有限公司 Image retrieval method and device and computer readable storage medium
CN107590489A (en) * 2017-09-28 2018-01-16 国家新闻出版广电总局广播科学研究院 Object detection method based on concatenated convolutional neutral net
CN107972662B (en) * 2017-10-16 2019-12-10 华南理工大学 Vehicle forward collision early warning method based on deep learning
CN107845116B (en) * 2017-10-16 2021-05-25 北京京东尚科信息技术有限公司 Method and apparatus for generating compression encoding of flat image
CN107909005A (en) * 2017-10-26 2018-04-13 西安电子科技大学 Personage's gesture recognition method under monitoring scene based on deep learning
CN107985189B (en) * 2017-10-26 2021-03-12 西安科技大学 Early warning method for lane changing depth of driver in high-speed driving environment
CN107680113A (en) * 2017-10-27 2018-02-09 武汉大学 The image partition method of multi-layer segmentation network based on Bayesian frame edge prior
CN107909082B (en) * 2017-10-30 2020-07-31 东南大学 Sonar image target identification method based on deep learning technology
CN107808138B (en) * 2017-10-31 2021-03-30 电子科技大学 Communication signal identification method based on FasterR-CNN
CN107844769B (en) * 2017-11-01 2021-06-01 浪潮集团有限公司 Vehicle detection method and system under complex scene
CN107679250B (en) * 2017-11-01 2020-12-01 浙江工业大学 Multi-task layered image retrieval method based on deep self-coding convolutional neural network
CN108052861A (en) * 2017-11-08 2018-05-18 北京卓视智通科技有限责任公司 A kind of nerve network system and the model recognizing method based on the nerve network system
CN107895367B (en) * 2017-11-14 2021-11-30 中国科学院深圳先进技术研究院 Bone age identification method and system and electronic equipment
CN109784131B (en) * 2017-11-15 2023-08-22 深圳光启合众科技有限公司 Object detection method, device, storage medium and processor
CN108038423B (en) * 2017-11-22 2022-03-04 广东数相智能科技有限公司 Automobile type identification method and device based on image identification
CN107871126A (en) * 2017-11-22 2018-04-03 西安翔迅科技有限责任公司 Model recognizing method and system based on deep-neural-network
CN108052881A (en) * 2017-11-30 2018-05-18 华中科技大学 The method and apparatus of multiclass entity object in a kind of real-time detection construction site image
CN108171112B (en) * 2017-12-01 2021-06-01 西安电子科技大学 Vehicle identification and tracking method based on convolutional neural network
CN108564097B (en) * 2017-12-05 2020-09-22 华南理工大学 Multi-scale target detection method based on deep convolutional neural network
CN108154149B (en) * 2017-12-08 2021-12-10 济南中维世纪科技有限公司 License plate recognition method based on deep learning network sharing
CN108052899A (en) * 2017-12-12 2018-05-18 成都睿码科技有限责任公司 A kind of method that electric bicycle and motorcycle are distinguished by video
CN108133186A (en) * 2017-12-21 2018-06-08 东北林业大学 A kind of plant leaf identification method based on deep learning
CN108171246B (en) * 2017-12-21 2022-02-08 北京科技大学 Clothing salient region detection method
CN108154504A (en) * 2017-12-25 2018-06-12 浙江工业大学 Method for detecting surface defects of steel plate based on convolutional neural network
CN108009526A (en) * 2017-12-25 2018-05-08 西北工业大学 A kind of vehicle identification and detection method based on convolutional neural networks
CN108021914B (en) * 2017-12-27 2020-07-28 清华大学 Method for extracting character area of printed matter based on convolutional neural network
CN108121986B (en) * 2017-12-29 2019-12-17 深圳云天励飞技术有限公司 Object detection method and device, computer device and computer readable storage medium
CN108364262A (en) * 2018-01-11 2018-08-03 深圳大学 A kind of restored method of blurred picture, device, equipment and storage medium
CN108460328A (en) * 2018-01-15 2018-08-28 浙江工业大学 A kind of fake-licensed car detection method based on multitask convolutional neural networks
CN108171203B (en) * 2018-01-17 2020-04-17 百度在线网络技术(北京)有限公司 Method and device for identifying vehicle
CN108171707A (en) * 2018-01-23 2018-06-15 武汉精测电子集团股份有限公司 A kind of Mura defects level evaluation method and device based on deep learning
CN108346145B (en) * 2018-01-31 2020-08-04 浙江大学 Identification method of unconventional cells in pathological section
CN108256498A (en) * 2018-02-01 2018-07-06 上海海事大学 A kind of non power driven vehicle object detection method based on EdgeBoxes and FastR-CNN
CN108509949B (en) * 2018-02-05 2020-05-15 杭州电子科技大学 Target detection method based on attention map
CN108460758A (en) * 2018-02-09 2018-08-28 河南工业大学 The construction method of Lung neoplasm detection model
JP7032536B2 (en) 2018-02-09 2022-03-08 ベイジン センスタイム テクノロジー デベロップメント カンパニー, リミテッド Instance segmentation methods and equipment, electronics, programs and media
CN108335305B (en) * 2018-02-09 2020-10-30 北京市商汤科技开发有限公司 Image segmentation method and apparatus, electronic device, program, and medium
CN108280490A (en) * 2018-02-28 2018-07-13 北京邮电大学 A kind of fine granularity model recognizing method based on convolutional neural networks
CN110210472A (en) * 2018-02-28 2019-09-06 佛山科学技术学院 A kind of method for checking object based on depth network
CN108334955A (en) * 2018-03-01 2018-07-27 福州大学 Copy of ID Card detection method based on Faster-RCNN
CN108549901A (en) * 2018-03-12 2018-09-18 佛山市顺德区中山大学研究院 A kind of iteratively faster object detection method based on deep learning
WO2019176877A1 (en) * 2018-03-15 2019-09-19 株式会社小糸製作所 Object identification system, automobile, vehicular lamp fitting, and method for identifying type of object
CN108320510B (en) * 2018-04-03 2020-12-04 深圳市智绘科技有限公司 Traffic information statistical method and system based on aerial video shot by unmanned aerial vehicle
CN108537732B (en) * 2018-04-10 2021-11-02 福州大学 PCA-SIFT-based rapid image splicing method
CN108921850B (en) * 2018-04-16 2022-05-17 博云视觉(北京)科技有限公司 Image local feature extraction method based on image segmentation technology
CN108537286B (en) * 2018-04-18 2020-11-24 北京航空航天大学 Complex target accurate identification method based on key area detection
CN108876849B (en) * 2018-04-24 2021-11-23 哈尔滨工程大学 Deep learning target identification and positioning method based on auxiliary identification
CN108830188B (en) * 2018-05-30 2022-03-04 西安理工大学 Vehicle detection method based on deep learning
CN109684906B (en) * 2018-05-31 2021-04-30 北京林业大学 Method for detecting red fat bark beetles based on deep learning
CN108776787B (en) * 2018-06-04 2020-09-29 京东数字科技控股有限公司 Image processing method and device, electronic device and storage medium
CN108871760B (en) * 2018-06-07 2020-07-17 广东石油化工学院 Efficient gear fault mode identification method
CN108830213A (en) * 2018-06-12 2018-11-16 北京理工大学 Car plate detection and recognition methods and device based on deep learning
CN108960079A (en) * 2018-06-14 2018-12-07 多伦科技股份有限公司 A kind of image-recognizing method and device
CN108830224B (en) * 2018-06-19 2021-04-02 武汉大学 High-resolution remote sensing image ship target detection method based on deep learning
CN110633717A (en) * 2018-06-21 2019-12-31 北京京东尚科信息技术有限公司 Training method and device for target detection model
CN108830254B (en) * 2018-06-27 2021-10-29 福州大学 Fine-grained vehicle type detection and identification method based on data balance strategy and intensive attention network
CN109166094B (en) * 2018-07-11 2022-03-25 华南理工大学 Insulator fault positioning and identifying method based on deep learning
CN109034245B (en) * 2018-07-27 2021-02-05 燕山大学 Target detection method using feature map fusion
CN109165582B (en) * 2018-08-09 2021-09-24 河海大学 Urban street garbage detection and cleanliness assessment method
CN109101934A (en) * 2018-08-20 2018-12-28 广东数相智能科技有限公司 Model recognizing method, device and computer readable storage medium
CN109131843B (en) * 2018-08-22 2022-04-26 王桥生 Long-term visual tracking active separation type undercarriage
CN109214441A (en) * 2018-08-23 2019-01-15 桂林电子科技大学 A kind of fine granularity model recognition system and method
CN109214505B (en) * 2018-08-29 2022-07-01 中山大学 Full convolution target detection method of densely connected convolution neural network
CN109376756B (en) * 2018-09-04 2020-08-18 青岛大学附属医院 System, computer device and storage medium for automatically identifying lymph node transferred from upper abdomen based on deep learning
CN109242516A (en) * 2018-09-06 2019-01-18 北京京东尚科信息技术有限公司 The single method and apparatus of processing service
CN109344825A (en) * 2018-09-14 2019-02-15 广州麦仑信息科技有限公司 A kind of licence plate recognition method based on convolutional neural networks
CN109543505B (en) * 2018-09-29 2023-03-21 江苏濠汉智能设备有限公司 Target detection system and method based on video image
CN109409518B (en) * 2018-10-11 2021-05-04 北京旷视科技有限公司 Neural network model processing method and device and terminal
CN109492761A (en) * 2018-10-30 2019-03-19 深圳灵图慧视科技有限公司 Realize FPGA accelerator, the method and system of neural network
CN109523015B (en) * 2018-11-09 2021-10-22 上海海事大学 Image processing method in neural network
CN109492586B (en) * 2018-11-12 2021-08-17 长讯通信服务有限公司 Mobile communication maintenance object detection method based on artificial intelligence and unmanned aerial vehicle
CN109614990A (en) * 2018-11-20 2019-04-12 成都通甲优博科技有限责任公司 A kind of object detecting device
CN109558902A (en) * 2018-11-20 2019-04-02 成都通甲优博科技有限责任公司 A kind of fast target detection method
CN109753581A (en) * 2018-11-30 2019-05-14 北京拓尔思信息技术股份有限公司 Image processing method, device, electronic equipment and storage medium
CN111260955B (en) * 2018-12-03 2021-12-28 魔门塔(苏州)科技有限公司 Parking space detection system and method adopting parking space frame lines and end points
CN109800778B (en) * 2018-12-03 2020-10-09 浙江工业大学 Faster RCNN target detection method based on difficultly-divided sample mining
CN109670501B (en) * 2018-12-10 2020-08-25 中国科学院自动化研究所 Object identification and grasping position detection method based on deep convolutional neural network
WO2020118616A1 (en) * 2018-12-13 2020-06-18 深圳先进技术研究院 Head and neck imaging method and device based on deep prior learning
CN109684956A (en) * 2018-12-14 2019-04-26 深源恒际科技有限公司 A kind of vehicle damage detection method and system based on deep neural network
CN109766775A (en) * 2018-12-18 2019-05-17 四川大学 A kind of vehicle detecting system based on depth convolutional neural networks
TWI706378B (en) * 2018-12-29 2020-10-01 鴻海精密工業股份有限公司 Cloud device, terminal device, and image classification method
CN109720275A (en) * 2018-12-29 2019-05-07 重庆集诚汽车电子有限责任公司 Multi-sensor Fusion vehicle environmental sensory perceptual system neural network based
CN111385598A (en) 2018-12-29 2020-07-07 富泰华工业(深圳)有限公司 Cloud device, terminal device and image classification method
CN109754071B (en) * 2018-12-29 2020-05-05 中科寒武纪科技股份有限公司 Activation operation method and device, electronic equipment and readable storage medium
CN109741318B (en) * 2018-12-30 2022-03-29 北京工业大学 Real-time detection method of single-stage multi-scale specific target based on effective receptive field
CN109934088A (en) * 2019-01-10 2019-06-25 海南大学 Sea ship discrimination method based on deep learning
CN109829491B (en) * 2019-01-22 2021-09-28 开易(北京)科技有限公司 Information processing method, apparatus and storage medium for image detection
CN109889525A (en) * 2019-02-26 2019-06-14 北京智芯微电子科技有限公司 Multi-communication protocol Intellisense method
CN110097534A (en) * 2019-03-04 2019-08-06 华北电力大学 A kind of nuclear fuel rod open defect detection method based on deep learning
CN110120047B (en) * 2019-04-04 2023-08-08 平安科技(深圳)有限公司 Image segmentation model training method, image segmentation method, device, equipment and medium
CN110110722A (en) * 2019-04-30 2019-08-09 广州华工邦元信息技术有限公司 A kind of region detection modification method based on deep learning model recognition result
CN110222593A (en) * 2019-05-18 2019-09-10 四川弘和通讯有限公司 A kind of vehicle real-time detection method based on small-scale neural network
CN110413825B (en) * 2019-06-21 2023-12-01 东华大学 Street-clapping recommendation system oriented to fashion electronic commerce
CN110348355A (en) * 2019-07-02 2019-10-18 南京信息工程大学 Model recognizing method based on intensified learning
CN110399816B (en) * 2019-07-15 2023-04-07 广西大学 High-speed train bottom foreign matter detection method based on Faster R-CNN
CN110397080A (en) * 2019-07-17 2019-11-01 深圳万海建筑工程科技有限公司 A kind of monitoring and warning system for pipe gallery
CN110414413A (en) * 2019-07-25 2019-11-05 北京麒麟智能科技有限公司 A kind of logistics trolley pedestrian detection method based on artificial intelligence
CN110532904B (en) * 2019-08-13 2022-08-05 桂林电子科技大学 Vehicle identification method
CN110472633A (en) * 2019-08-15 2019-11-19 南京拓控信息科技股份有限公司 A kind of detection of train license number and recognition methods based on deep learning
CN110570469B (en) * 2019-08-16 2020-08-25 广州威尔森信息科技有限公司 Intelligent identification method for angle position of automobile picture
CN110610210B (en) * 2019-09-18 2022-03-25 电子科技大学 Multi-target detection method
CN110807452A (en) * 2019-10-11 2020-02-18 上海上湖信息技术有限公司 Prediction model construction method, device and system and bank card number identification method
CN110942401B (en) * 2019-11-21 2023-12-19 黑龙江电力调度实业有限公司 Intelligent communication method for electric power Internet of things
CN111104942B (en) * 2019-12-09 2023-11-03 熵智科技(深圳)有限公司 Template matching network training method, recognition method and device
CN111145365A (en) * 2019-12-17 2020-05-12 北京明略软件系统有限公司 Method, device, computer storage medium and terminal for realizing classification processing
CN111368682B (en) * 2020-02-27 2023-12-12 上海电力大学 Method and system for detecting and identifying station caption based on master RCNN
CN111460909A (en) * 2020-03-09 2020-07-28 兰剑智能科技股份有限公司 Vision-based goods location management method and device
CN111524095A (en) * 2020-03-24 2020-08-11 西安交通大学 Target detection method for rotating object
CN111461128A (en) * 2020-03-31 2020-07-28 北京爱笔科技有限公司 License plate recognition method and device
CN111523579B (en) * 2020-04-14 2022-05-03 燕山大学 Vehicle type recognition method and system based on improved deep learning
WO2021218140A1 (en) * 2020-04-27 2021-11-04 平安科技(深圳)有限公司 Deformable convolution-based image recognition method and apparatus, and computer device
CN111652285A (en) * 2020-05-09 2020-09-11 济南浪潮高新科技投资发展有限公司 Tea cake category identification method, equipment and medium
CN111968127B (en) * 2020-07-06 2021-08-27 中国科学院计算技术研究所 Cancer focus area identification method and system based on full-section pathological image
CN112132222B (en) * 2020-09-27 2023-02-10 上海高德威智能交通系统有限公司 License plate category identification method and device and storage medium
CN112507247B (en) * 2020-12-15 2022-09-23 重庆邮电大学 Cross-social network user alignment method fusing user state information
US11971953B2 (en) 2021-02-02 2024-04-30 Inait Sa Machine annotation of photographic images
US11544914B2 (en) 2021-02-18 2023-01-03 Inait Sa Annotation of 3D models with signs of use visible in 2D images
WO2022175044A1 (en) 2021-02-18 2022-08-25 Inait Sa Annotation of 3d models with signs of use visible in 2d images
CN113076837A (en) * 2021-03-25 2021-07-06 高新兴科技集团股份有限公司 Convolutional neural network training method based on network image
CN112905213B (en) * 2021-03-26 2023-08-08 中国重汽集团济南动力有限公司 Method and system for realizing ECU (electronic control Unit) refreshing parameter optimization based on convolutional neural network
CN112949614B (en) * 2021-04-29 2021-09-10 成都市威虎科技有限公司 Face detection method and device for automatically allocating candidate areas and electronic equipment
CN113469190B (en) * 2021-06-10 2023-09-15 电子科技大学 Single-stage target detection algorithm based on domain adaptation
CN113392911B (en) * 2021-06-18 2023-04-18 电子科技大学 DW-ReSuMe algorithm-based image classification method
TWI830230B (en) * 2022-05-18 2024-01-21 逢甲大學 Object automatic tracking system and identification method thereof
CN117392179B (en) * 2023-12-11 2024-02-27 四川迪晟新达类脑智能技术有限公司 Target tracking method based on correlation filter and edge frame

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104657748A (en) * 2015-02-06 2015-05-27 中国石油大学(华东) Vehicle type recognition method based on convolutional neural network
CN105184271A (en) * 2015-09-18 2015-12-23 苏州派瑞雷尔智能科技有限公司 Automatic vehicle detection method based on deep learning
CN105404858A (en) * 2015-11-03 2016-03-16 电子科技大学 Vehicle type recognition method based on deep Fisher network

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9418319B2 (en) * 2014-11-21 2016-08-16 Adobe Systems Incorporated Object detection using cascaded convolutional neural networks

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104657748A (en) * 2015-02-06 2015-05-27 中国石油大学(华东) Vehicle type recognition method based on convolutional neural network
CN105184271A (en) * 2015-09-18 2015-12-23 苏州派瑞雷尔智能科技有限公司 Automatic vehicle detection method based on deep learning
CN105404858A (en) * 2015-11-03 2016-03-16 电子科技大学 Vehicle type recognition method based on deep Fisher network

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
基于深度卷积神经网络的车型识别研究;邓柳等;《计算机应用研究》;20160331;第33卷(第3期);全文
基于深度神经网络的汽车车型识别;王茜等;《图形图像》;20151231;全文

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111540203A (en) * 2020-04-30 2020-08-14 东华大学 Method for adjusting green light passing time based on fast-RCNN

Also Published As

Publication number Publication date
CN106250812A (en) 2016-12-21

Similar Documents

Publication Publication Date Title
CN106250812B (en) A kind of model recognizing method based on quick R-CNN deep neural network
CN108108657B (en) Method for correcting locality sensitive Hash vehicle retrieval based on multitask deep learning
Adarsh et al. YOLO v3-Tiny: Object Detection and Recognition using one stage improved model
CN107609601B (en) Ship target identification method based on multilayer convolutional neural network
CN109614985B (en) Target detection method based on densely connected feature pyramid network
CN111368896B (en) Hyperspectral remote sensing image classification method based on dense residual three-dimensional convolutional neural network
CN107885764B (en) Rapid Hash vehicle retrieval method based on multitask deep learning
CN111783831B (en) Complex image accurate classification method based on multi-source multi-label shared subspace learning
CN108830188A (en) Vehicle checking method based on deep learning
CN110889318B (en) Lane detection method and device using CNN
CN107066559A (en) A kind of method for searching three-dimension model based on deep learning
JP6980289B2 (en) Learning method and learning device that can detect lanes using a lane model, and test method and test device using this {LEARNING METHOD, LEARNING DEVICE FOR DETECTING LANE USING LANE MODEL AND TEST METHOD, TEST DEVICE
EP3938806A1 (en) Radar data collection and labeling for machine-learning
CN115861619A (en) Airborne LiDAR (light detection and ranging) urban point cloud semantic segmentation method and system of recursive residual double-attention kernel point convolution network
CN114241226A (en) Three-dimensional point cloud semantic segmentation method based on multi-neighborhood characteristics of hybrid model
Zhang et al. Lidar-based place recognition for autonomous driving: A survey
Hellert et al. Using algorithm selection for adaptive vehicle perception aboard UAV
Hwang et al. Object Detection for Cargo Unloading System Based on Fuzzy C Means.
CN108960005A (en) The foundation and display methods, system of subjects visual label in a kind of intelligent vision Internet of Things
Taha et al. Assessment of Approaches for the Extraction of Building Footprints from Plé iades Images
CN114297237A (en) Three-dimensional point cloud data retrieval method and device based on category fusion and computer equipment
Harras et al. Enhanced vehicle classification using transfer learning and a novel duplication-based data augmentation technique
Kamaleswari et al. An Assessment of Object Detection in Thermal (Infrared) Image Processing
Pavlove et al. Efficient Deep Learning Methods for Automated Visibility Estimation at Airports
Saravanakumar et al. An Enriched Model of Neutrosophical Fuzzy and Grasshopper Convolutional Neural Network Based Moving Object Detection and Classification to Improve Video Surveillance

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
TR01 Transfer of patent right

Effective date of registration: 20230925

Address after: Room 701, 7th Floor, Building 10, Jingshun Platinum Yuecheng, Xihu District, Hangzhou City, Zhejiang Province, 310023

Patentee after: Hangzhou Yixun Technology Service Co.,Ltd.

Address before: 71-3-501, Chaohui Sixth District, No. 64 Xinshi Street, Xiacheng District, Hangzhou City, Zhejiang Province, 310014

Patentee before: Tang Yiping

TR01 Transfer of patent right