CN109389136A - Classifier training method - Google Patents

Classifier training method Download PDF

Info

Publication number
CN109389136A
CN109389136A CN201710670289.1A CN201710670289A CN109389136A CN 109389136 A CN109389136 A CN 109389136A CN 201710670289 A CN201710670289 A CN 201710670289A CN 109389136 A CN109389136 A CN 109389136A
Authority
CN
China
Prior art keywords
training
classifier
scene type
neural network
new samples
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201710670289.1A
Other languages
Chinese (zh)
Inventor
胡振程
王艳明
孙皓
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shanghai Sen Sen Vehicle Sensor Technology Co Ltd
Original Assignee
Shanghai Sen Sen Vehicle Sensor Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shanghai Sen Sen Vehicle Sensor Technology Co Ltd filed Critical Shanghai Sen Sen Vehicle Sensor Technology Co Ltd
Priority to CN201711482401.5A priority Critical patent/CN109389142A/en
Priority to CN201710670289.1A priority patent/CN109389136A/en
Publication of CN109389136A publication Critical patent/CN109389136A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Computational Linguistics (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a kind of classifier training methods, comprising the following steps: generates data set by computer graphics model, data set includes the sample of several scenes classification;Training pattern is obtained using deep neural network training dataset;Training pattern is tested to obtain the test result of each scene type;The information of the corresponding scene type of accuracy minimum in test result is fed back into computer graphics model;Computer graphics model modifies parameter corresponding with the scene type, generates the new samples about the scene type;And new samples are added in data set.Classifier training method through the invention can be improved the accuracy of classifier training and reduce the time of classifier training.

Description

Classifier training method
Technical field
The present invention relates to a kind of classifier training methods, more particularly to one kind to be applied to ADAS (advanced driver Assistance system, driver assistance system) classifier training method.
Background technique
ADAS may be defined as in driving procedure by the perception to vehicle-surroundings environment and to the reason of driving behavior Solution judges danger that may be present, gives driver's warning, by the way of interactive so as to reducing traffic accident Incidence extends the safe driving time.ADAS system based on image has at low cost, easy for installation and high accuracy rate excellent Point.ADAS system needs to participate in object to the traffic in driving conditions to identify, to judge its shadow to current driving environment It rings.Usually, the traffic in driving conditions, which participates in object, to be divided into: static traffic participates in object, such as trees, fence, traffic mark Board, traffic lights, lane line, fixed obstacle etc.;Mobile traffic participate in object, as vehicle, pedestrian, bicycle, power assist vehicle, animal, Other objects with displacement characteristic.
ADAS system needs to participate in object to above-mentioned movement/static traffic to analyze with the motion conditions from vehicle, to reach The effect driven to auxiliary, primary goal is the detection and identification to above-mentioned target.Detection and identification for specific objective The machine learning algorithm that supervision can be used solves, and is broadly divided into two processes of training and test.It is needed in training process The sample marked is inputed into machine, machine is using specific algorithm and goes out optimal parameter according to sample training, provides point Class model;Machine examines the position of target and size according to trained model and the image of input during the test It surveys.Detection effect, which is directly undergone training, the influence of sample, the especially complexity of ADAS application environment and participates in the diversity of object, Influence of the Many times training sample to performance is even more than algorithm itself.
By taking vehicle as an example, the type of vehicle for needing to identify is numerous, comprising: car, SUV, MPV, sport car, microcaloire, pick up, big Bar, mini-bus, trailer, tank car, lorry etc..Vehicle running surface type can be divided into again: super expressway, urban road, suburb road Road and natural environment etc..Vehicle running environment is influenced by various illumination and weather condition again.Therefore, it collects in various weather And under natural environment, a variety of vehicle samples of various information of road surface have very big difficulty, in addition in view of collecting sample regards The sample size of the demand at angle, demand will be very huge, it will usually reach the sample of millions of or even more than one hundred million number of levels.Such as Fruit needs to collect the video data of substantial amounts by the way of manually marking, and therefrom calibrates the actual size of target by hand And position, it will bring huge workload and data integration time.
It is therefore desirable to be able to provide Massive Sample and using these sample training classifiers quickly and easily with effectively reliable Ground is applied to the method for ADAS.
Summary of the invention
The purpose of the present invention is to provide a kind of classifier training methods that can be applied to ADAS, use and pass through CG The magnanimity sample for the different scenes condition for selected target that (computer graphics, computer graphical) technology generates This.
According to an embodiment of the invention, providing a kind of classifier training method, comprising the following steps: pass through computer graphic Shape model generates data set, and data set includes the sample of several scenes classification;Come using deep neural network training dataset To training pattern;Training pattern is tested to obtain the test result of each scene type;By accuracy pair minimum in test result The information for the scene type answered feeds back to computer graphics model;Computer graphics model modifies ginseng corresponding with the scene type Number generates the new samples about the scene type;And new samples are added in data set.Note that in the training process, making It is trained with the sample under scenes all in training set to obtain training pattern, rather than is instructed with the sample under each scene Practice to obtain the training pattern under each scene.
Wherein, it includes: according to predetermined ratio by data that training pattern is obtained using deep neural network training dataset The sample of concentration is assigned randomly in training set, verifying collection and test set;Come using the deep neural network training training set Obtain multiple initial models;And multiple initial models are verified using verifying collection, obtain the training with optimized parameter Model.
In addition, repeating training, the training pattern of deep neural network after new samples are added in the training set Test, the feedback of test result, new samples generation processing, until the accuracy of the test result of each scene type is equal Until predetermined threshold.
Wherein, the training for repeating deep neural network includes any one of following two mode: utilizing depth nerve net Network re -training is added with the training set of new samples to obtain new training pattern;In deep neural network, using being added with The training set of new samples is finely adjusted the model parameter for the training pattern that previous training obtains to obtain new training pattern.
Preferably, the initial proportion of each scene type sample in training set is identical.
Preferably, it further includes authentic specimen that training set and verifying, which are concentrated,.
It preferably, further include authentic specimen in training set, verifying collection and test set.
It preferably, include the demarcating file of training objective in training set, verifying collection and test set.
Preferably, test training pattern includes: to test to instruct using test set come the test result for obtaining each scene type Practice model, respectively obtains the test result of each scene type.
Preferably, new samples are added to includes: that new samples are added in training set in data set.
Preferably, ratio of the new samples in training set is 20%.
Preferably, several scenes classification includes fine day, cloudy day, snowy day, rainy day, daytime, night, strong light etc..
According to the method for the present invention, it by generating sample using CG technology, effectively reduces and takes pictures sampling to actual environment Workload and working strength, and reduce manpower, improve work efficiency.By using deep neural network, Neng Gouti High training effect.By extraly increasing the verification processing of verifying collection in training, over-fitting can be prevented, so that training Model out has versatility, the image data being applicable to outside training sample.In addition, by by the lower information of accuracy CG is fed back to, CG is made to generate the new samples of corresponding scene type, re -training is added with the training set of new samples or to previous The model parameter of trained deep neural network is finely adjusted, to guarantee accuracy of the sample under more scene conditions.As a result, Classifier training method through the invention, can be improved the accuracy of classifier training and reduce classifier training when Between.
Detailed description of the invention
It in order to more clearly explain the embodiment of the invention or the technical proposal in the existing technology, below will be to institute in embodiment Attached drawing to be used is needed to be briefly described, it should be apparent that, the accompanying drawings in the following description is only some implementations of the invention Example, for those of ordinary skill in the art, without creative efforts, can also obtain according to these attached drawings Obtain other attached drawings.
Fig. 1 is the flow chart of classifier training method according to the present invention;And
Fig. 2 be deep neural network according to the present invention a kind of exemplary group at schematic diagram.
Specific embodiment
The description of this illustrated embodiment should be combined with corresponding attached drawing, and attached drawing should be used as the one of complete specification Part.In the accompanying drawings, the shape of embodiment or thickness can expand, and to simplify or facilitate mark.Furthermore it is respectively tied in attached drawing The part of structure will be to describe to be illustrated respectively, it is notable that the member for being not shown in the figure or not being illustrated by text Part is the form known to a person of ordinary skill in the art in technical field.
The description of embodiments herein, any reference in relation to direction and orientation, is merely for convenience of describing, and cannot manage Solution is any restrictions to the scope of the present invention.It can be related to the combination of feature below for the explanation of preferred embodiment, These features may be individually present or combine presence, and the present invention is not defined in preferred embodiment particularly.The present invention Range be defined by the claims.
The digital simulation that CG graphics can use magnanimity generates the driving scene with high realism, while generating target Accurate target position and size in the picture, mark letter big with data volume relative to the implementation manually demarcated Cease accurate, scene advantage abundant.The CG high realism true value mark database generated is used for target object in supervised learning Detection will be of great significance.Meanwhile there is the support of Massive Sample that can assess the various performances of sorting algorithm.
In the present invention, CG technology is combined with ADAS recognizer, being not only embodied in CG can provide for sorting algorithm The data of magnanimity are also embodied in sorting algorithm while CG can be instructed to generate what type of sample, to mention in maximum efficiency The performance of high algorithm.That is, one closed loop thinking generated based on CG data with classification algorithm training of building, CG can be used Good classifier of data " culture " of magnanimity, and classifier " can instruct " generation of CG sample database simultaneously, below into Row is described in detail.
Fig. 1 is the flow chart of classifier training method according to the present invention.
Referring to Fig. 1, classifier training method according to the present invention includes: to generate data set by computer graphics model, Data set includes the sample of several scenes classification;Training pattern is obtained using deep neural network training dataset;Test instruction Practice model to obtain the test result of each scene type;By the information of the corresponding scene type of accuracy minimum in test result Feed back to computer graphics model;Computer graphics model modifies parameter corresponding with the scene type, generates about the scene The new samples of classification;And new samples are added in data set.Note that in the training process, using fields all in training set Sample under scape is trained to obtain training pattern, rather than is trained with the sample under each scene to obtain each scene Under training pattern.
In classifier training method of the invention, data set is generated by CG technology, includes several scenes in data set Type, and saved by scene type, scene type is for example including fine day, cloudy day, snowy day, rainy day, daytime, night, strong light etc.. It should be understood that data set used in the present invention can also classify otherwise.
In classifier training method of the invention, deep neural network training sample is used.Fig. 2 is according to the present invention A kind of exemplary group of deep neural network at schematic diagram.Referring to Fig. 2, the deep neural network that the present invention uses is by convolution Layer, pond layer, full articulamentum and softmax layers of composition.Wherein, image characteristics extraction depends on convolution algorithm, is transported by convolution Calculation is filtered to obtain significant Expressive Features.
Specifically, input layer main function be by picture input be converted into output data required for model (data) and Label (label), input layer also convert data format required for platform, such as the lmdb of caffe etc. for picture simultaneously.
Convolutional layer main function is feature extraction, and the convolution operation of deep learning is similar to filter, by important information It extracts, main definitions convolution kernel and convolution operation.
Pond layer main function is characteristic pattern compression, and carrying out compressed characteristic pattern can be reduced calculation amount, reduce network Computation complexity, and characteristic pattern main information can be continued to extract.Pond layer operation includes average pond and maximum The modes such as pond.
Full articulamentum main function is to connect all features, and output valve is passed to softmax layers.Full connection Layer parameter is more, and full articulamentum is affected to parameter optimization.
Softmax layers of main function are classifiers, are last output.
It should be noted that Fig. 2 is only a kind of exemplary composed structure of deep neural network workable for the present invention, this hair The bright deep neural network that can also be applied to various composed structures, and can also be with using the deep neural network of other structures It realizes function of the invention and realizes identical effect.
Preferably, it includes: that will count according to predetermined ratio that training pattern is obtained using deep neural network training dataset It is assigned randomly in training set, verifying collection and test set according to the sample of concentration;Come using deep neural network training training set To multiple initial models;And multiple initial models are verified using verifying collection, obtain the training mould with optimized parameter Type.
In addition, repeating the training of deep neural network, the survey of training pattern after new samples are added in training set Examination, the generation of the feedback of test result, new samples processing, until the accuracy of the test result of each scene type is all larger than Until predetermined threshold.The predetermined threshold can according to circumstances use different values.Specifically, the sample in data set CG generated This is assigned randomly in training set, verifying collection and test set according to predetermined ratio, such as all data sets can be randomly divided into 50% training set, 25% verifying collection and 25% test set, or it is divided into 60% training set, 20% verifying collection and 20% test set, Certainly it will be understood by those skilled in the art that other ratios can also be used.
In an advantageous embodiment, when sample in the first sub-distribution data set, make each scene class in training set Very originally initial proportion is identical, that is, the quantity of the sample of each scene type is identical, to guarantee identification of the model to each scene Initialization weight having the same.
In addition, concentrating in training set and verifying includes authentic specimen, that is, training set and verifying collection are in addition to including what CG was generated It further include the various information of road surface under a small amount of various weather and natural environment collected by modes such as shootings except sample A variety of vehicle samples, ratio of the authentic specimen in data set is about 5% to 10%.
It include authentic specimen in training set, verifying collection and test set in another preferred embodiment.
In addition, including that (such as various static traffic participate in object and movement to training objective in training set, verifying collection and test set Traffic participates in object) demarcating file, demarcating file indicates the size and location of training objective in the picture.
After establishing data set according to each scene type with same ratio, the starting model instruction of deep neural network is carried out Practice, obtain multiple initial models, then multiple initial models are verified using verifying collection, obtain the instruction with optimized parameter Practice model.Note that in the training process, using multiple Artificial Neural Network Structures to the sample under scenes all in training set into Row training is to obtain multiple initial models.Wherein, training pattern can be known to the skilled in the art various training patterns. For example, the present invention can use multi-target detection model, it can also be using models such as semantic segmentations.
In classifier training method of the invention, other than including training set and test set, also it is additionally provided with and tests Card collection, verifying collection can carry out cross validation to the model that training set training obtains to select optimal models, and effect is to prevent Over-fitting because for training set train come model, for training set data, the parameter of model can reach best Effect, but when carrying out test set test, modelling effect may just have biggish change, and adding verifying collection can be well Solution such issues that, by verify collection obtain the model with optimized parameter, then the model is tested with test set again.
Deep neural network training is repetitive exercise, and when just having started to train, the effect of model has very big error rate.With The increase of neural metwork training number, model constantly update inherent parameters weight, the reversed biography of deep learning by backpropagation Broadcast and error be calculated by the output of true value and last softmax in true value file, by error it is reversed pass network back. Then, network punishes weighting parameter according to error, updates weight to the direction for reducing error, to guarantee network convergence And obtain further feature.By the verifying of verifying collection, the model for training deep neural network has versatility, can be applied to Image data outside sample.
As described above, the present invention carries out feature extraction and classification to the sample that CG is generated using the method for deep neural network Judgement, after obtaining preliminary depth sorting model, does not ensure that this model can meet verification and measurement ratio under any environment It is required that.Therefore, the present invention is also based on test result " guidance " CG and generates new samples carrying out supplementary training sample, to promote depth point The accuracy of class model.It after obtaining a depth sorting model, is analyzed by the accuracy rate to test set, changes CG sample The relevant parameter that library generates achievees the purpose that " instruct " generation of CG sample, and the scene ratio of sample database is updated, to update Training sample database used in next iteration.
In the present invention, training pattern is tested using test set, respectively obtains the test result of each scene type, tested As a result it can be indicated with modes such as accuracy, false detection rate, recall rates.That is, by evaluating model, available model Accuracy rate, false detection rate, recall rate and other evaluation results.For example, multi-target detection model can to for each classification into The evaluation of row accuracy rate and recall rate, while also carrying out the evaluation of consensus forecast rate.When scene type accuracy rate or call together When the rate of returning is relatively bad, CG can be fed back so that CG generates the new samples of the scene type and new samples are added to training set In, ratio of the scene type in training set is increased, makes the fine tuning training of depth network parameter to the effect of the scene type Rise.
Specifically, after Sample Refreshment, it can use the training set that deep neural network re -training is added with new samples To obtain new training pattern, that is, the process of training is identical with first time, and using deep neural network, training is new again completely Training set.Then, test set tests the training pattern newly obtained, instructs CG to generate new sample, Zhi Houzai according to accuracy New training set is trained using deep neural network, repeatedly iteration, until the test result of each scene type is all satisfied Until it is required that.
However, if completely re -training deep neural network will devote a tremendous amount of time, therefore the invention also provides Trimming scheme.That is, being carried out using a certain proportion of old sample and newly-increased sample to the old model parameter that last iteration obtains micro- It adjusts, to guarantee accuracy of the sample under more scenes.Wherein, ratio of the new samples in training set is 20%, certainly It can be other ratios.Specifically, it is finely adjusted, is learned with the model that the training set added with new samples obtains last training Practise new samples.The method of fine tuning is to extract model, and the value of reserving model weight and biasing is joined using the weight and biasing as starting Number, is then again trained new training set.Since model has had preferable testing result, so new samples start just Just there is good testing result in training stage.And due to increasing the general sample proportion of performance in new training set, thus In the case where guaranteeing original good scene type of performance, and the study to general scene type is showed is strengthened, thus To new model.Then, above-mentioned various processing are repeated, the evaluation result that can be got well until guaranteeing the data of each scene type.
Although the performance of entire training system is it should be noted that the mode efficiency of above two repetition training is different It is identical.
Wherein, CG generates the process of sample by targeted species and quantity, environment type and quantity, illumination, meteorology, object The influence of surface area object and camera parameter, such as the parameter that control CG sample generates are represented by Pcg={ VT1,…,n, VN, ET1,…,m, EN, SA, SI, SC, SVL, OVL, EL, RI, SI, FI, WP, SP, DP, CP, CD, CS, IQ }, wherein VT1,…,nIndicate vehicle Type, such as car, SUV, MPV, sport car, microcaloire, pick up, bus, mini-bus, trailer, tank car, lorry;VN indicates vehicle mould Type quantity represents the degree of crowding for generating vehicle in picture, can be divided into unimpeded and crowded two kinds of situations;ET1,…,mIndicate environment kind Class, such as urban road, expressway, tunnel, bridge road, hill path;EN indicates environmental model quantity, indicates the complexity of environment Degree, can be divided into two kinds of situations of single environment and complex environment;Illumination effect mostlys come from the influence of sunlight in daylight environment, SA indicates the natural environment sun angular of simulation, and being continuously can control value;SI indicates the natural environment sunlight intensity of simulation, is It continuously can control value;SC indicates the natural environment sun light color of simulation, and being continuously can control value;Illumination effect is main in night environment From the influence from vehicle, Ta Che and street lamp, wherein SVL indicates that being continuously can control value from the bright dark of car light;OVL indicates him Car light angular and bright dark, for continuously can control value;The angle and intensity of EL expression street lamp light are continuously can control value;Meteorologic parameter Middle RI indicates rainfall intensity, can be divided into weak, medium, strong three grades;SI indicates snowfall intensity, can be divided into weak, medium, strong three etc. Grade;FI indicates drop mist intensity, can be divided into weak, medium, strong three grades;Due to being influenced by meteorology, front vehicles surface can have been accumulated Different substances, WP indicate water accumulating volume, can be divided into weak, medium, strong three grades;SP indicates snowpack, can be divided into weak, medium, strong three A grade;DP indicates dust accumulation amount, can be divided into weak, medium, strong three grades;In the influence of camera parameter, CP indicates the direction of camera, For continuous controlled variable;CD indicates camera distortion and Fov, is commonly divided into pin-hole model and two kinds of flake model;CS indicates phase The parameters such as machine ISO are generally corresponding with camera model;IQ indicates picture quality, can be divided into three kinds basic, normal, high.It is above-mentioned by changing The value or type of parameter can simulate the samples pictures under various true environments, be supplied to classifier training disaggregated model.
It is not inventive point of the invention it should be noted that specifically how to generate sample about CG, therefore omits here detailed Description.
As described above, by generating sample using CG technology, being effectively reduced in classifier training method of the invention It takes pictures the workload and working strength of sampling to actual environment, and reduces manpower, improve work efficiency.By using depth Neural network is spent, can be improved training effect.By extraly increasing the verification processing of verifying collection in training, can prevent Fitting phenomenon, so that the model trained is applicable to the image data outside training sample.In addition, by the way that accuracy is lower Information feed back to CG, so that CG is generated the new samples of corresponding scene type, training set of the re -training added with new samples or The model parameter of previously trained deep neural network is finely adjusted, to guarantee that sample is accurate under more scene conditions Property.Classifier training method through the invention as a result, can be improved the accuracy of classifier training and reduce classifier instruction The experienced time.
The foregoing is merely illustrative of the preferred embodiments of the present invention, is not intended to limit the invention, all in essence of the invention Within mind and principle, any modification, equivalent replacement, improvement and so on be should all be included in the protection scope of the present invention.

Claims (12)

1. a kind of classifier training method, which is characterized in that the classifier training method the following steps are included:
Data set is generated by computer graphics model, the data set includes the sample of several scenes classification;
Training pattern is obtained using the deep neural network training data set;
The training pattern is tested to obtain the test result of each scene type;
The information of the corresponding scene type of accuracy minimum in test result is fed back into the computer graphics model;
The computer graphics model modifies parameter corresponding with the scene type, generates the new samples about the scene type; And
The new samples are added in the data set.
2. classifier training method according to claim 1, which is characterized in that use the deep neural network training number Include: to obtain training pattern according to collection
The sample in the data set is assigned randomly in training set, verifying collection and test set according to predetermined ratio;
Multiple initial models are obtained using the deep neural network training training set;And
The multiple initial model is verified using verifying collection, obtains the training pattern with optimized parameter.
3. classifier training method according to claim 2, which is characterized in that the new samples are being added to the instruction Practice after concentrating, repeat the training of the deep neural network, the test of the training pattern, the test result feedback, The processing of the generation of the new samples, until the accuracy of the test result of each scene type is all larger than predetermined threshold.
4. classifier training method according to claim 3, which is characterized in that repeat the training of the deep neural network Including any one of following two mode:
The training set of the new samples is added with using the deep neural network re -training to obtain new training mould Type;
In the deep neural network, using the training set added with the new samples to the previous training pattern trained and obtained Model parameter be finely adjusted to obtain new training pattern.
5. classifier training method according to claim 2, which is characterized in that each scene type sample in the training set This initial proportion is identical.
6. classifier training method according to claim 2, which is characterized in that
The training set and verifying concentration further include authentic specimen.
7. classifier training method according to claim 2, which is characterized in that
It further include authentic specimen in the training set, verifying collection and the test set.
8. classifier training method according to claim 2, which is characterized in that the training set, verifying collection and institute State the demarcating file in test set including training objective.
9. classifier training method according to claim 2, which is characterized in that it is each to obtain to test the training pattern The test result of scene type includes: to test the training pattern using the test set, respectively obtains each scene type Test result.
10. classifier training method according to claim 2, which is characterized in that the new samples are added to the number It include: that the new samples are added in the training set according to concentration.
11. classifier training method according to claim 10, which is characterized in that the new samples are in the training set Ratio be 20%.
12. classifier training method according to any one of claim 1 to 11, which is characterized in that the several scenes Classification includes fine day, cloudy day, snowy day, rainy day, daytime, night, strong light etc..
CN201710670289.1A 2017-08-08 2017-08-08 Classifier training method Pending CN109389136A (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN201711482401.5A CN109389142A (en) 2017-08-08 2017-08-08 Classifier training method
CN201710670289.1A CN109389136A (en) 2017-08-08 2017-08-08 Classifier training method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710670289.1A CN109389136A (en) 2017-08-08 2017-08-08 Classifier training method

Related Child Applications (1)

Application Number Title Priority Date Filing Date
CN201711482401.5A Division CN109389142A (en) 2017-08-08 2017-08-08 Classifier training method

Publications (1)

Publication Number Publication Date
CN109389136A true CN109389136A (en) 2019-02-26

Family

ID=65412973

Family Applications (2)

Application Number Title Priority Date Filing Date
CN201710670289.1A Pending CN109389136A (en) 2017-08-08 2017-08-08 Classifier training method
CN201711482401.5A Pending CN109389142A (en) 2017-08-08 2017-08-08 Classifier training method

Family Applications After (1)

Application Number Title Priority Date Filing Date
CN201711482401.5A Pending CN109389142A (en) 2017-08-08 2017-08-08 Classifier training method

Country Status (1)

Country Link
CN (2) CN109389136A (en)

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110135592A (en) * 2019-05-16 2019-08-16 腾讯科技(深圳)有限公司 Classifying quality determines method, apparatus, intelligent terminal and storage medium
CN110276443A (en) * 2019-05-29 2019-09-24 平安科技(深圳)有限公司 Model conversion consistency checking and analysis method, device and storage medium
CN110765876A (en) * 2019-09-20 2020-02-07 深圳码隆科技有限公司 Training method and device for commodity recognition model
CN110991538A (en) * 2019-12-05 2020-04-10 腾讯科技(深圳)有限公司 Sample classification method and device, storage medium and computer equipment
CN111292329A (en) * 2020-01-15 2020-06-16 北京字节跳动网络技术有限公司 Training method and device for video segmentation network and electronic equipment
CN111488939A (en) * 2020-04-15 2020-08-04 天津美腾科技股份有限公司 Model training method, classification method, device and equipment
CN112132796A (en) * 2020-09-15 2020-12-25 佛山读图科技有限公司 Visual detection method and system for improving detection precision by means of feedback data autonomous learning
CN112800233A (en) * 2021-04-13 2021-05-14 成都数联铭品科技有限公司 Text position detection method
CN114418036A (en) * 2022-03-28 2022-04-29 浙江所托瑞安科技集团有限公司 Method, device and storage medium for testing and training performance of neural network
WO2024060670A1 (en) * 2022-09-19 2024-03-28 北京沃东天骏信息技术有限公司 Method and apparatus for training classification model, and device and storage medium

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110110652B (en) * 2019-05-05 2021-10-22 达闼科技(北京)有限公司 Target detection method, electronic device and storage medium
CN110634036A (en) * 2019-09-27 2019-12-31 金瓜子科技发展(北京)有限公司 Method and device for distributing configuration objects, storage medium and electronic equipment

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2014049118A (en) * 2012-08-31 2014-03-17 Fujitsu Ltd Convolution neural network classifier system, training method for the same, classifying method, and usage
CN103971342A (en) * 2014-05-21 2014-08-06 厦门美图之家科技有限公司 Image noisy point detection method based on convolution neural network
CN104504376A (en) * 2014-12-22 2015-04-08 厦门美图之家科技有限公司 Age classification method and system for face images
CN104966097A (en) * 2015-06-12 2015-10-07 成都数联铭品科技有限公司 Complex character recognition method based on deep learning
CN105051755A (en) * 2013-01-18 2015-11-11 微软技术许可有限责任公司 Part and state detection for gesture recognition
CN105320957A (en) * 2014-07-10 2016-02-10 腾讯科技(深圳)有限公司 Classifier training method and device
CN106295637A (en) * 2016-07-29 2017-01-04 电子科技大学 A kind of vehicle identification method based on degree of depth study with intensified learning
CN106934319A (en) * 2015-12-29 2017-07-07 北京大唐高鸿软件技术有限公司 People's car objective classification method in monitor video based on convolutional neural networks

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101976461A (en) * 2010-10-25 2011-02-16 北京理工大学 Novel outdoor augmented reality label-free tracking registration algorithm
CN103679189B (en) * 2012-09-14 2017-02-01 华为技术有限公司 Method and device for recognizing scene

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2014049118A (en) * 2012-08-31 2014-03-17 Fujitsu Ltd Convolution neural network classifier system, training method for the same, classifying method, and usage
CN105051755A (en) * 2013-01-18 2015-11-11 微软技术许可有限责任公司 Part and state detection for gesture recognition
CN103971342A (en) * 2014-05-21 2014-08-06 厦门美图之家科技有限公司 Image noisy point detection method based on convolution neural network
CN105320957A (en) * 2014-07-10 2016-02-10 腾讯科技(深圳)有限公司 Classifier training method and device
CN104504376A (en) * 2014-12-22 2015-04-08 厦门美图之家科技有限公司 Age classification method and system for face images
CN104966097A (en) * 2015-06-12 2015-10-07 成都数联铭品科技有限公司 Complex character recognition method based on deep learning
CN106934319A (en) * 2015-12-29 2017-07-07 北京大唐高鸿软件技术有限公司 People's car objective classification method in monitor video based on convolutional neural networks
CN106295637A (en) * 2016-07-29 2017-01-04 电子科技大学 A kind of vehicle identification method based on degree of depth study with intensified learning

Cited By (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110135592A (en) * 2019-05-16 2019-08-16 腾讯科技(深圳)有限公司 Classifying quality determines method, apparatus, intelligent terminal and storage medium
CN110135592B (en) * 2019-05-16 2023-09-19 腾讯科技(深圳)有限公司 Classification effect determining method and device, intelligent terminal and storage medium
CN110276443A (en) * 2019-05-29 2019-09-24 平安科技(深圳)有限公司 Model conversion consistency checking and analysis method, device and storage medium
CN110765876A (en) * 2019-09-20 2020-02-07 深圳码隆科技有限公司 Training method and device for commodity recognition model
CN110991538B (en) * 2019-12-05 2021-12-03 腾讯科技(深圳)有限公司 Sample classification method and device, storage medium and computer equipment
CN110991538A (en) * 2019-12-05 2020-04-10 腾讯科技(深圳)有限公司 Sample classification method and device, storage medium and computer equipment
CN111292329A (en) * 2020-01-15 2020-06-16 北京字节跳动网络技术有限公司 Training method and device for video segmentation network and electronic equipment
CN111488939A (en) * 2020-04-15 2020-08-04 天津美腾科技股份有限公司 Model training method, classification method, device and equipment
CN112132796A (en) * 2020-09-15 2020-12-25 佛山读图科技有限公司 Visual detection method and system for improving detection precision by means of feedback data autonomous learning
CN112800233A (en) * 2021-04-13 2021-05-14 成都数联铭品科技有限公司 Text position detection method
CN114418036A (en) * 2022-03-28 2022-04-29 浙江所托瑞安科技集团有限公司 Method, device and storage medium for testing and training performance of neural network
CN114418036B (en) * 2022-03-28 2022-06-21 浙江所托瑞安科技集团有限公司 Method, device and storage medium for testing and training performance of neural network
WO2024060670A1 (en) * 2022-09-19 2024-03-28 北京沃东天骏信息技术有限公司 Method and apparatus for training classification model, and device and storage medium

Also Published As

Publication number Publication date
CN109389142A (en) 2019-02-26

Similar Documents

Publication Publication Date Title
CN109389136A (en) Classifier training method
CN109977812B (en) Vehicle-mounted video target detection method based on deep learning
CN111444821B (en) Automatic identification method for urban road signs
CN103279759B (en) A kind of vehicle front trafficability analytical procedure based on convolutional neural networks
CN108985194B (en) Intelligent vehicle travelable area identification method based on image semantic segmentation
CN110111335A (en) A kind of the urban transportation Scene Semantics dividing method and system of adaptive confrontation study
CN107133974A (en) The vehicle type classification method that Gaussian Background modeling is combined with Recognition with Recurrent Neural Network
CN108764298B (en) Electric power image environment influence identification method based on single classifier
CN105844257A (en) Early warning system based on machine vision driving-in-fog road denoter missing and early warning method
CN110716529A (en) Automatic generation method and device for automatic driving test case
Zhu et al. Extreme weather recognition using convolutional neural networks
CN114677507A (en) Street view image segmentation method and system based on bidirectional attention network
CN111027430A (en) Traffic scene complexity calculation method for intelligent evaluation of unmanned vehicles
CN110956146B (en) Road background modeling method and device, electronic equipment and storage medium
CN110851948B (en) Driving environment situation assessment method and device under unstructured road condition
CN103679205A (en) Preceding car detection method based on shadow hypothesis and layered HOG (histogram of oriented gradient) symmetric characteristic verification
CN106056102A (en) Video-image-analysis-based road vehicle type classification method
CN111369053B (en) Method for predicting tracks of pedestrians and vehicles in human-vehicle interaction scene
CN212009589U (en) Video identification driving vehicle track acquisition device based on deep learning
CN116205024A (en) Self-adaptive automatic driving dynamic scene general generation method for high-low dimension evaluation scene
Cheng et al. Modeling weather and illuminations in driving views based on big-video mining
CN114169444A (en) Driving style classification method considering risk potential field distribution under vehicle following working condition
CN109948419A (en) A kind of illegal parking automatic auditing method based on deep learning
Pan et al. Winter road surface condition recognition using a pretrained deep convolutional network
CN116246169A (en) SAH-Unet-based high-resolution remote sensing image impervious surface extraction method

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20190226

RJ01 Rejection of invention patent application after publication