CN107392125A - Training method/system, computer-readable recording medium and the terminal of model of mind - Google Patents

Training method/system, computer-readable recording medium and the terminal of model of mind Download PDF

Info

Publication number
CN107392125A
CN107392125A CN201710561654.5A CN201710561654A CN107392125A CN 107392125 A CN107392125 A CN 107392125A CN 201710561654 A CN201710561654 A CN 201710561654A CN 107392125 A CN107392125 A CN 107392125A
Authority
CN
China
Prior art keywords
data
model
data set
training
mind
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201710561654.5A
Other languages
Chinese (zh)
Inventor
汪宏
邵蔚元
郑莹斌
叶浩
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shanghai Information Technology Research Center
Shanghai Advanced Research Institute of CAS
Original Assignee
Shanghai Information Technology Research Center
Shanghai Advanced Research Institute of CAS
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shanghai Information Technology Research Center, Shanghai Advanced Research Institute of CAS filed Critical Shanghai Information Technology Research Center
Priority to CN201710561654.5A priority Critical patent/CN107392125A/en
Publication of CN107392125A publication Critical patent/CN107392125A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/172Classification, e.g. identification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/56Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
    • G06V20/58Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads
    • G06V20/584Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads of vehicle lights or traffic lights
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/08Detecting or categorising vehicles

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Physics & Mathematics (AREA)
  • Software Systems (AREA)
  • Multimedia (AREA)
  • Evolutionary Computation (AREA)
  • Medical Informatics (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Mathematical Physics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Human Computer Interaction (AREA)
  • Artificial Intelligence (AREA)
  • Image Analysis (AREA)

Abstract

The present invention provides a kind of training method/system, computer-readable recording medium and the terminal of model of mind, and training method includes:The first data set and the markup information related to training mission to input carry out initial model training, to obtain benchmark model;New data is added, is incorporated in the first data set, forms the second data set;To the data test in the second data set and value assessment, to pick out the data that mark value is more than default mark value, the 3rd data set is formed;The data for not marking markup information in 3rd data set are labeled, are incorporated in the 3rd data set;Retraining, the benchmark model after being updated are carried out to benchmark model;It is the first new data set by the 3rd data set definition, adds new data, circulation performs above step, until the accuracy of the model after repetitive exercise is more than default accuracy.The present invention reduces the quantity manually marked, it is not necessary to marks total data, so as to save mark cost, and improves model training efficiency.

Description

Training method/system, computer-readable recording medium and the terminal of model of mind
Technical field
The invention belongs to field of artificial intelligence, is related to a kind of training method and system, more particularly to a kind of intelligence Training method/system, computer-readable recording medium and the terminal of model.
Background technology
The development of artificial intelligence field is maked rapid progress, and in particular with the extensive use of depth learning technology, it is in object The fields such as detection, identification achieve breakthrough progress.The feature of traditional engineer is different from, depth learning technology passes through Substantial amounts of data are inputted, allow model itself to learn the character representation of object, the identification that can often match in excellence or beauty or even surmount the mankind is smart Degree.
In general, the learning process of a complete artificial intelligence model include preparing two steps of data set and model training Suddenly.Preparing data set includes data and the mark to data, for example, a recognition of face data set should include face picture with To the markup information of the piece identity of every pictures.Model training is then input input data and markup information simultaneously, according to not Same learning tasks carry out training pattern.Current most model learning flow is the learning process of one " static state ".I.e. once Property collect and labeled data, then train model according to these data, then learning process terminates.However, it is somebody's turn to do of " static state " Habit process has the disadvantage that in real application environment:
1) starting model training needs to wait all Data Collections and mark to finish.By taking depth learning technology as an example, model Training needs substantial amounts of data, can be from pictures up to ten thousand to hundreds of thousands pictures, and collecting data and labeled data needs Substantial amounts of man power and material is expended, cost is very high.
2) ability of model is shaped in an inquiry learning, and the data of disposable collecting are difficult complete adaptation practical application Scene, so as to cause the performance of model in actual applications to have larger gap compared with experimental situation.
Therefore, how training method/system, computer-readable recording medium and the terminal of a kind of model of mind are provided, with The problems such as solving Data Collection in the prior art and marking too high cost, model adaptability difference, the real this area that turned into are obtained employment Personnel's technical problem urgently to be resolved hurrily.
The content of the invention
In view of the above the shortcomings that prior art, it is an object of the invention to provide a kind of training side of model of mind Method/system, computer-readable recording medium and terminal, for solving Data Collection in the prior art and mark, cost is too high, mould The problem of type adaptability difference.
In order to achieve the above objects and other related objects, one aspect of the present invention provides a kind of training method of model of mind, The training method of the model of mind comprises the following steps:The first data set and the mark related to training mission letter to input Breath carries out initial model training, to obtain a benchmark model;Addition newly counts with data attribute identical in first data set According to, be incorporated in first data set, formed the second data set;The data tested in the second data set, to the second data set Middle data carry out value assessment, and to pick out the data that mark value is more than default mark value, the data that will be singled out are formed 3rd data set;The data for not marking markup information in 3rd data set are labeled, the data of markup information will be labeled with It is incorporated in the 3rd data set;Based on the 3rd data set after merging, retraining, the base after being updated are carried out to benchmark model Quasi-mode type;It is the first new data set by the 3rd data set definition, adds new data, circulation performs above step, to be changed Generation training, until the accuracy of the model after repetitive exercise is more than default accuracy.
In one embodiment of the invention, described pair of the first data set inputted and the markup information related to training mission Initial model training is carried out, to obtain the loss letter related to training mission for referring to optimization the step of a benchmark model and pre-setting Number, the functional value of the loss function is set constantly to reduce until preset function threshold value.
In one embodiment of the invention, after second data set is formed, the training method of the model of mind is also Including being pre-processed to the data in second data set;The pretreatment includes carrying out examination, image to the data Resolution ratio detection, image blur detection, and/or the adjustment of image balance degree.
In one embodiment of the invention, it is described test the second data set in data the step of include to described second count Convolution, Chi Hua and more classification processing are carried out according to intensive data.
It is described that value assessment is carried out to the second data intensive data in one embodiment of the invention, to pick out Mark value be more than the step of data of default mark value also include by it is predefined, for judging second data set The cost function of the mark value of middle data, pick out the data that mark value is more than default mark value.
In one embodiment of the invention, the markup information for being directed to selected data is by artificial notation methods The information of mark.
Another aspect of the present invention provides a kind of training system of model of mind, including:Initial training module, for input The first data set and the markup information related to training mission carry out initial model training, to obtain a benchmark model;Merge Module, for addition and data attribute identical new data in first data set, it is incorporated in first data set, shape Into the second data set;Processing module, for testing the data in the second data set, value is carried out to the second data intensive data and commented Estimate, to pick out the data that mark value is more than default mark value, the data that will be singled out form the 3rd data set;To the 3rd The data for not marking markup information in data set are labeled, and the data for being labeled with markup information are incorporated in into the 3rd data set In;Retraining module, for based on the 3rd data set after merging, retraining, the base after being updated to be carried out to benchmark model Quasi-mode type;Loop module, for being the first new data set by the 3rd data set definition, new data is added, described in circular flow Initial training module, merging module, processing module and retraining module, to be iterated training, until the mould after repetitive exercise The accuracy of type is more than default accuracy.
In one embodiment of the invention, after second data set is formed, the training system of the model of mind is also Including the pretreatment module coupled with the merging module, the pretreatment module is used for the data in second data set Pre-processed;It is described pretreatment include to the data carry out examination, image resolution ratio detection, image blur detection and/ Or image balance degree adjustment.
In one embodiment of the invention, the data that the processing module is tested in the second data set refer to described second Data intensive data carries out convolution, Chi Hua and more classification processing.
In one embodiment of the invention, the processing module be additionally operable to by it is predefined, for judging described second The cost function of the mark value of data intensive data, pick out the data that mark value is more than default mark value.
Another aspect of the invention provides a kind of computer-readable recording medium, is stored thereon with computer program, its feature It is, the program realizes the training method of the model of mind when being executed by processor.
Last aspect of the present invention provides a kind of terminal, including:Processor and memory;The memory is based on storing Calculation machine program, the processor is used for the computer program for performing the memory storage, so that the terminal performs the intelligence The training method of energy model.
As described above, training method/system, computer-readable recording medium and the terminal of the model of mind of the present invention, tool There is following beneficial effect:
Training method/system, computer-readable recording medium and the terminal of model of mind of the present invention reduce people The quantity of work mark, it is not necessary to mark total data, so as to save mark cost, and improve model training efficiency.
Brief description of the drawings
Fig. 1 is shown as schematic flow sheet of the training method of the model of mind of the present invention in an embodiment.
Fig. 2 is shown as theory structure schematic diagram of the training system of the model of mind of the present invention in an embodiment.
Component label instructions
The training system of 2 models of mind
21 initial training modules
22 merging modules
23 pretreatment modules
24 processing modules
25 retraining modules
26 loop modules
S11~S18 steps
Embodiment
Illustrate embodiments of the present invention below by way of specific instantiation, those skilled in the art can be by this specification Disclosed content understands other advantages and effect of the present invention easily.The present invention can also pass through specific realities different in addition The mode of applying is embodied or practiced, the various details in this specification can also be based on different viewpoints with application, without departing from Various modifications or alterations are carried out under the spirit of the present invention.It should be noted that in the case where not conflicting, following examples and implementation Feature in example can be mutually combined.
It should be noted that the diagram provided in following examples only illustrates the basic structure of the present invention in a schematic way Think, only show the component relevant with the present invention in schema then rather than according to component count, shape and the size during actual implement Draw, kenel, quantity and the ratio of each component can be a kind of random change during its actual implementation, and its assembly layout kenel It is likely more complexity.
Training method/system, computer-readable recording medium and the terminal of model of mind of the present invention include two Stage:Model initialization stage and model retraining stage.
The model initialization stage:In the starting stage of model training, prepare an initial data set and with training mission phase The markup information of pass.The structure of model is designed, using data set and its markup information as information is inputted, carries out model training, Obtain a benchmark model.Follow-up model carries out retraining on the basis of the model.
The model retraining stage:After getting out basic and training benchmark model.By retraining to model Parameter is updated.Here retraining refers to be trained again by inputting new data and its mark to benchmark model, right Model parameter is updated, and the model after renewal is typically higher than the model accuracy before renewal.Retraining needs the number inputted According to there is two kinds of sources:First, being tested using benchmark model on former training dataset, difficult example therein is selected, such as identify The relatively low data of the data and confidence level of mistake;Second, collect new data, by benchmark model and active Samples Estimates technology, Value marking is carried out to data, picks out the data for wherein most marking value, then manually this partial data is labeled. The data and its mark in two kinds of sources of merging collect to update the data.Retraining is carried out on new data set to benchmark model to obtain Model after renewal.Above step can constantly be repeated according to actual conditions or iteration.
Embodiment one
The present embodiment provides a kind of training method of model of mind, including:
The first data set and the markup information related to training mission to input carry out initial model training, to obtain one Benchmark model;
Addition and data attribute identical new data in first data set, are incorporated in first data set, shape Into the second data set;
The data tested in the second data set, value assessment is carried out to the second data intensive data, to pick out mark valency Value is more than the data of default mark value, and the data that will be singled out form the 3rd data set;
The data for not marking markup information in 3rd data set are labeled, the data for being labeled with markup information are merged In the 3rd data set;
Based on the 3rd data set after merging, retraining, the benchmark model after being updated are carried out to benchmark model;
It is the first new data set by the 3rd data set definition, adds new data, circulation performs above step, to be changed Generation training, until the accuracy of the model after repetitive exercise is more than default accuracy.
The training method of the model of mind provided below with reference to diagram the present embodiment is described in detail.Refer to Fig. 1, it is shown as schematic flow sheet of the training method of model of mind in an embodiment.As shown in figure 1, the model of mind Training method specifically includes following steps:
S11, the first data set and the markup information related to training mission to input carry out initial model training, to obtain Take a benchmark model.For example, first data set is the first LFW face public data collection and the identity related to training mission Markup information, or first data set are the positional information of the first vehicle detection public data collection and vehicle on picture.
Specific in practical application, to the first LFW face public data collection of input and the identity related to training mission Markup information carries out initial model training, to obtain recognition of face benchmark model.
Or initial model is carried out to the positional information of the first vehicle detection public data collection and vehicle of input on picture Training, to obtain vehicle detection benchmark model.
In the present embodiment, initial model training refers to the loss letter related to training mission that optimization is pre-set Number, the functional value of the loss function is set constantly to reduce until preset function value threshold value.The preset function value threshold value levels off to 0. In this process, the parameter of model is also continuously available renewal, so that the benchmark model preferably can appoint in adaptation training Business.
S12, addition and data attribute identical new data in first data set, are incorporated in first data set In, form the second data set.
In the present embodiment, the first LFW face public data collection is directed to, is added newly by modes such as web crawlers downloads Face picture, formed the 2nd LFW face public data collection.
Or for the first vehicle detection public data collection, mode is downloaded etc. by web crawlers and adds new vehicle pictures, Form the second vehicle detection public data collection.
S13, the data in second data set are pre-processed.The pretreatment includes sieving the data Look into, image resolution ratio detects, image blur detection, and/or image balance degree adjust.In the present embodiment, climbed by network The modes such as worm download collect the second data set.It is to be model retraining standard that data in second data set are carried out with pretreatment The high data of standby quality.
Specifically, such as face picture is concentrated to carry out just the 2nd LFW faces public data using Dlib, OpenCV instrument Walk the pretreatments such as face examination, image blur detection.
Or the vehicle figure concentrated using instruments such as Dlib, OpenCV to the second vehicle detection public data is carried out tentatively The pretreatments such as vehicle examination, image resolution ratio, fuzziness detection, the adjustment of image balance degree.
S14, using the benchmark model, test pretreated second data set.In the present embodiment, institute is tested The second data set is stated to refer to carry out the second data intensive data the processing such as convolution, Chi Hua and more classification processing.
Specifically, convolution, Chi Hua and more classification processing etc. are carried out to the 2nd LFW face public datas intensive data Processing.
Or the processing such as convolution, Chi Hua and more classification processing are carried out to the second vehicle detection public data intensive data.
S15, value assessment is carried out to second data set, to pick out the number that mark value is more than default mark value According to, and the costly data of mark that will be singled out form the 3rd data set.In the present embodiment, by it is predefined, be used for Judge the cost function of the mark value of the second data intensive data, pick out mark value and be more than default mark value Data.
Specifically, mark predefined, for judging the 2nd LFW faces public data concentration human body picture is passed through The cost function of value, to pick out the face picture that mark value is more than default mark value, and the mark valency that will be singled out It is worth high data and forms the 3rd LFW face public data collection.By it is predefined, for judging that second vehicle detection discloses The cost function of the mark value of vehicle pictures in data set, to pick out the vehicle figure that mark value is more than default mark value Piece, and the costly data of mark that will be singled out form the 3rd vehicle detection public data collection.
S16, the data for not marking markup information in the 3rd data set are labeled, the data of markup information will be labeled with It is incorporated in the 3rd data set.
In the present embodiment, the markup information for being directed to selected data is the letter marked by artificial notation methods Breath.For the face picture selected, face identity information is marked, and the face picture for being labeled with face identity information is merged In the 3rd LFW face public data collection;For the vehicle pictures selected, the positional information of vehicle is marked, and will be marked The vehicle pictures for having the positional information of vehicle are incorporated in the 3rd vehicle detection public data collection.The present embodiment is by selecting number Reduce the quantity manually marked, it is not necessary to total data is marked, so as to save according to the data of most worthy are concentrated to be labeled About mark cost.
S17, based on the 3rd data set after merging, retraining, the benchmark model after being updated are carried out to benchmark model. In the present embodiment, training technique is identical with the training technique in step S11 used by retraining, i.e. optimization is pre-set The loss function related to training mission, the functional value of the loss function is set constantly to reduce until preset function value threshold value.It is described Preset function value threshold value levels off to 0.
S18, it is the first new data set by the 3rd data set definition after merging, and in the first new data set New data is added, circulation performs above step S11 to S17, is trained to be iterated, until the model after repetitive exercise is accurate Degree is more than default accuracy.In the present embodiment, by the way of training-mark iteration, partial data once can be only marked, To partial data training pattern, renewal is then constantly iterated to model, training is more flexible, more rapidly.
Specifically, it is so that the accuracy of the model of recognition of face reaches more than 99% according to the demand of reality, works as model Accuracy when being less than the value, with regard to continuous repeat step S11 to S17, until the accuracy of model reaches default accuracy 99%.
It is so that the accuracy of the model of vehicle identification reaches more than 99% according to the demand of reality, when the accuracy of model During less than the value, with regard to continuous repeat step S11 to S17, until the accuracy of model reaches default accuracy 99%.
The present embodiment also provides a kind of computer-readable recording medium, is stored thereon with computer program, and the program is located Reason device realizes the training method of the model of mind when performing.One of ordinary skill in the art will appreciate that:Realize above-mentioned each side The all or part of step of method embodiment can be completed by the related hardware of computer program.Foregoing computer program can To be stored in a computer-readable recording medium.The program upon execution, execution the step of including above-mentioned each method embodiment; And foregoing storage medium includes:ROM, RAM, magnetic disc or CD etc. are various can be with the medium of store program codes.
The training method and computer-readable recording medium of model of mind described in the present embodiment reduce what is manually marked Quantity, it is not necessary to mark total data, so as to save mark cost, and improve model training efficiency.
Embodiment two
The present embodiment provides a kind of training system of model of mind, including:
Initial training module, carried out for the first data set to input and the markup information related to training mission initial Model training, to obtain a benchmark model;
Merging module, for add with data attribute identical new data in first data set, be incorporated in described the In one data set, the second data set is formed;
Processing module, for testing the data in the second data set, value assessment is carried out to the second data intensive data, with The data that mark value is more than default mark value are picked out, the data that will be singled out form the 3rd data set;To the 3rd data Concentrate the data for not marking markup information to be labeled, the data for being labeled with markup information are incorporated in the 3rd data set;
Retraining module, for based on the 3rd data set after merging, retraining being carried out to benchmark model, after obtaining renewal Benchmark model;
Loop module, for being the first new data set by the 3rd data set definition, new data is added, described in circular flow Initial training module, merging module, processing module and retraining module, to be iterated training, until the mould after repetitive exercise The accuracy of type is more than default accuracy.
The training system of the model of mind provided below with reference to diagram the present embodiment is described in detail.Need Bright is, it should be understood that the division of the modules of apparatus above is only a kind of division of logic function, can be with when actually realizing Completely or partially it is integrated on a physical entity, can also be physically separate.And these modules all can be passed through with software The form that treatment element calls is realized;All it can also realize in the form of hardware;Treatment element can be passed through with part of module The form of software is called to realize, part of module is realized by the form of hardware.For example, x modules can be the processing individually set up Element, it can also be integrated in some chip of said apparatus and realize, in addition it is also possible to be stored in the form of program code In the memory of said apparatus, called by some treatment element of said apparatus and perform the function of above x modules.Other moulds The realization of block is similar therewith.In addition these modules can completely or partially integrate, and can also independently realize.It is described here Treatment element can be a kind of integrated circuit, have signal disposal ability.In implementation process, each step of the above method Or more modules can be completed by the instruction of the integrated logic circuit of the hardware in processor elements or software form.
For example, the above module can be arranged to implement one or more integrated circuits of above method, such as: One or more specific integrated circuits (ApplicationSpecificIntegratedCircuit, abbreviation ASIC), or, one Or multi-microprocessor (digitalsingnalprocessor, abbreviation DSP), or, one or more field-programmable gate array Arrange (FieldProgrammableGateArray, abbreviation FPGA) etc..For another example, some module is dispatched by treatment element more than When the form of program code is realized, the treatment element can be general processor, such as central processing unit (CentralProcessingUnit, abbreviation CPU) or it is other can be with the processor of caller code.For another example, these modules can To integrate, realized in the form of on-chip system (system-on-a-chip, abbreviation SOC).
Referring to Fig. 2, it is shown as theory structure schematic diagram of the training system of model of mind in an embodiment.Such as Fig. 2 Shown, the training system 2 of the model of mind includes initial training module 21, merging module 22, pretreatment module 23, processing mould Block 24, retraining module 25 and loop module 26.
The initial training module 21 is used to enter the first data set of input and the markup information related to training mission Row initial model training, to obtain a benchmark model.
Marked for example, first data set is the first LFW face public data collection and the identity related to training mission Information, or first data set are the positional information of the first vehicle detection public data collection and vehicle on picture.
In specific practical application, the initial training module 21 to the first LFW face public data collection of input and with instruction Practice the related identity markup information of task and carry out initial model training, to obtain recognition of face benchmark model.
Position of the initial training module 21 to the first vehicle detection public data collection and vehicle of input on picture Information carries out initial model training, to obtain vehicle detection benchmark model.
In the present embodiment, initial model training refers to the loss letter related to training mission that optimization is pre-set Number, the functional value of the loss function is set constantly to reduce until preset function value threshold value.The preset function value threshold value levels off to 0. In this process, the parameter of model is also continuously available renewal, so that the benchmark model preferably can appoint in adaptation training Business.
It is used to add and number in first data set with the merging module 22 of the initial training module 21 coupling According to attribute identical new data, it is incorporated in first data set, forms the second data set.
In the present embodiment, the first LFW face public data collection is directed to, is added newly by modes such as web crawlers downloads Face picture, formed the 2nd LFW face public data collection.
For the first vehicle detection public data collection, new vehicle pictures, shape are added by modes such as web crawlers downloads Into the second vehicle detection public data collection.
It is used to carry out in advance the data in second data set with the pretreatment module 23 of the merging module 22 coupling Processing.The pretreatment includes carrying out the data examination, image resolution ratio detection, image blur detection, and/or image The degree of balance adjusts.In the present embodiment, the second data set is collected by modes such as web crawlers downloads.To in the second data set It is to prepare the high data of quality for model retraining that data, which carry out pretreatment,.
Specifically, the pretreatment module 23 is concentrated using instruments such as Dlib, OpenCV to the 2nd LFW faces public data Face picture carries out the pretreatments such as preliminary face examination, image blur detection.
Or the pretreatment module 23 is concentrated using instruments such as Dlib, OpenCV to the second vehicle detection public data Vehicle figure carry out the pretreatments such as preliminary vehicle examination, image resolution ratio, fuzziness detection, the adjustment of image balance degree.
It is used to utilize the benchmark model with the processing module 24 of the pretreatment module 23 coupling, tests pretreated Second data set.In the present embodiment, second data set is tested to refer to carry out the second data intensive data The processing such as convolution, Chi Hua and more classification processing.
Specifically, processing module 24 carries out convolution, Chi Hua and more to the 2nd LFW face public datas intensive data The processing such as classification processing.Or convolution, Chi Hua and more classification processing etc. are carried out to the second vehicle detection public data intensive data Reason.
The processing module 24 carries out value after the data in testing the second data set, to second data set and commented Estimate, be more than the data of default mark value to pick out mark value, and the costly data of mark that will be singled out form the Three data sets.In the present embodiment, the processing module 24 by it is predefined, for judging the second data intensive data Mark value cost function, pick out the data that mark value is more than default mark value.
Specifically, the processing module 24 by it is predefined, for judging that the 2nd LFW faces public data is concentrated The cost function of the mark value of human body picture, to pick out the face picture that mark value is more than default mark value, and will The costly data of mark picked out form the 3rd LFW face public data collection.
Or the processing module 24 by it is predefined, for judging that vehicle is concentrated in the second vehicle detection public data The cost function of the mark value of picture, it is more than the vehicle pictures of default mark value to pick out mark value, and will selects The costly data of mark gone out form the 3rd vehicle detection public data collection.
After the costly data of mark are picked out, the processing module 24 is additionally operable to not marked in the 3rd data set The data of note information are labeled, and the data for being labeled with markup information are incorporated in the 3rd data set.
In the present embodiment, the markup information for being directed to selected data is the information marked by artificial notation methods.
For the face picture selected, face identity information, and the face figure that face identity information will be labeled with are marked Piece is incorporated in the 3rd LFW face public data collection.
For the vehicle pictures selected, the positional information of vehicle is marked, and the car of the positional information of vehicle will be labeled with Picture is incorporated in the 3rd vehicle detection public data collection.The present embodiment is by selecting the data of most worthy in data set To be labeled, reduce the quantity manually marked, it is not necessary to total data is marked, so as to save mark cost.
It is used for the retraining module 25 of the processing module 24 coupling based on the 3rd data set after merging, to benchmark mould Type carries out retraining, the benchmark model after being updated.In the present embodiment, training technique is same used by retraining module 25 Training technique in initial training module 21 is identical, that is, optimizes the loss function related to training mission pre-set, make this The functional value of loss function is constantly reduced until preset function value threshold value.The preset function value threshold value levels off to 0.
With the initial training module 21, merging module 22, pretreatment module 23, processing module 24, retraining module 25 It is the first new data set that the loop module 26 of connection, which is used for the 3rd data set definition after merging, and described new first New data is added in data set, the initial training module 21, merging module 22, pretreatment module 23, place described in circular flow Reason module 24, retraining module 25 are to be iterated training, until the accuracy of the model after repetitive exercise is accurate more than default Degree.In the present embodiment, by the way of training-mark iteration, partial data once can be only marked, partial data is trained Model, renewal is then constantly iterated to model, training is more flexible, more rapidly.
Specifically, it is so that the accuracy of the model of recognition of face reaches more than 99% according to the demand of reality, works as model Accuracy when being less than the value, just constantly rerun the initial training module 21, merging module 22, pretreatment module 23rd, processing module 24, retraining module 25, until the accuracy of model reaches default accuracy 99%.
It is so that the accuracy of the model of vehicle identification reaches more than 99% according to the demand of reality, when the accuracy of model During less than the value, the initial training module 21 that just constantly reruns, merging module 22, pretreatment module 23, processing Module 24, retraining module 25, until the accuracy of model reaches default accuracy 99%.
Embodiment three
The present embodiment provides a kind of terminal, including:Processor, memory, transceiver, communication interface and system bus;Deposit Reservoir and communication interface are connected with processor and transceiver by system bus and complete mutual communication, and memory is used to deposit Computer program is stored up, communication interface is used for and other equipment is communicated, and processor and transceiver are used to run computer program, Make the step S11 to step S18 of the training system of model of mind of the terminal execution as described in embodiment one.
System bus mentioned above can be Peripheral Component Interconnect standard (PeripheralPomponentInterconnect, abbreviation PCI) bus or EISA (ExtendedIndustryStandardArchitecture, abbreviation EISA) bus etc..The system bus can be divided into address Bus, data/address bus, controlling bus etc..For ease of representing, only represented in figure with a thick line, it is not intended that only one total Line or a type of bus.Communication interface is used for accessing data base device and other equipment (such as client, read-write storehouse And read-only storehouse) between communication.Memory may include random access memory (RandomAccessMemory, abbreviation RAM), Nonvolatile memory (non-volatilememory), for example, at least a magnetic disk storage may also also be included.
Above-mentioned processor can be general processor, including central processing unit (CentralProcessingUnit, letter Claim CPU), network processing unit (NetworkProcessor, abbreviation NP) etc.;It can also be digital signal processor (DigitalSignalProcessing, abbreviation DSP), application specific integrated circuit (ApplicationSpecificIntegratedCircuit, abbreviation ASIC), field programmable gate array (Field- ProgrammableGateArray, abbreviation FPGA) either other PLDs, discrete gate or transistor logic device Part, discrete hardware components.
In summary, training method/system, computer-readable recording medium and the terminal of model of mind of the present invention Reduce the quantity manually marked, it is not necessary to mark total data, so as to save mark cost, and improve model training effect Rate.So the present invention effectively overcomes various shortcoming of the prior art and has high industrial utilization.
The above-described embodiments merely illustrate the principles and effects of the present invention, not for the limitation present invention.It is any ripe Know the personage of this technology all can carry out modifications and changes under the spirit and scope without prejudice to the present invention to above-described embodiment.Cause This, those of ordinary skill in the art is complete without departing from disclosed spirit and institute under technological thought such as Into all equivalent modifications or change, should by the present invention claim be covered.

Claims (12)

1. a kind of training method of model of mind, it is characterised in that the training method of the model of mind comprises the following steps:
The first data set and the markup information related to training mission to input carry out initial model training, to obtain a benchmark Model;
Addition and data attribute identical new data in first data set, are incorporated in first data set, form the Two data sets;
The data tested in the second data set, value assessment is carried out to the second data intensive data, it is big to pick out mark value In the data of default mark value, the data that will be singled out form the 3rd data set;To mark letter is not marked in the 3rd data set The data of breath are labeled, and the data for being labeled with markup information are incorporated in the 3rd data set;
Based on the 3rd data set after merging, retraining, the benchmark model after being updated are carried out to benchmark model;
It is the first new data set by the 3rd data set definition, adds new data, circulation performs above step, to be iterated instruction Practice, until the accuracy of the model after repetitive exercise is more than default accuracy.
2. the training method of model of mind according to claim 1, it is characterised in that the first data set of described pair of input And the markup information related to training mission carries out initial model training, set in advance with obtaining to refer to optimize the step of a benchmark model The loss function related to training mission put, the functional value of the loss function is set constantly to reduce until preset function threshold value.
3. the training method of model of mind according to claim 1, it is characterised in that forming second data set Afterwards, the training method of the model of mind also includes pre-processing the data in second data set;The pretreatment Adjusted including the data are carried out with examination, image resolution ratio detection, image blur detection, and/or image balance degree.
4. the training method of model of mind according to claim 1, it is characterised in that in the second data set of the test The step of data, includes carrying out the second data intensive data convolution, Chi Hua and more classification processing.
5. the training method of model of mind according to claim 1, it is characterised in that described in second data set Data carry out value assessment, are also included by predefined with picking out the step of mark value is more than the data of default mark value , the cost function of mark value for judging the second data intensive data, pick out mark value and be more than pre- bidding Note the data of value.
6. the training method of model of mind according to claim 1, it is characterised in that described to be directed to selected data Markup information is the information marked by artificial notation methods.
A kind of 7. training system of model of mind, it is characterised in that including:
Initial training module, initial model is carried out for the first data set to input and the markup information related to training mission Training, to obtain a benchmark model;
Merging module, for addition and data attribute identical new data in first data set, it is incorporated in first number According to concentration, the second data set is formed;
Processing module, for testing the data in the second data set, value assessment is carried out to the second data intensive data, to select Go out the data that mark value is more than default mark value, the data that will be singled out form the 3rd data set;To in the 3rd data set The data for not marking markup information are labeled, and the data for being labeled with markup information are incorporated in the 3rd data set;
Retraining module, for based on the 3rd data set after merging, retraining, the base after being updated to be carried out to benchmark model Quasi-mode type;
Loop module, for being the first new data set by the 3rd data set definition, add new data, it is initial described in circular flow Training module, merging module, processing module and retraining module, trained with being iterated, until the model after repetitive exercise Accuracy is more than default accuracy.
8. the training system of model of mind according to claim 7, it is characterised in that forming second data set Afterwards, the training system of the model of mind also includes the pretreatment module with merging module coupling, the pretreatment module For being pre-processed to the data in second data set;The pretreatment includes carrying out examination, image to the data Resolution ratio detection, image blur detection, and/or the adjustment of image balance degree.
9. the training system of model of mind according to claim 7, it is characterised in that the number of processing module test second Refer to carry out the second data intensive data convolution, Chi Hua and more classification processing according to the data of concentration.
10. the training system of model of mind according to claim 7, it is characterised in that the processing module is additionally operable to lead to The cost function of mark value predefined, for judging the second data intensive data is crossed, it is big to pick out mark value In the data of default mark value.
11. a kind of computer-readable recording medium, is stored thereon with computer program, it is characterised in that the program is by processor The training method of model of mind any one of claim 1 to 6 is realized during execution.
A kind of 12. terminal, it is characterised in that including:Processor and memory;
The memory is used to store computer program, and the processor is used for the computer journey for performing the memory storage Sequence, so that the terminal performs the training method of the model of mind as any one of claim 1 to 6.
CN201710561654.5A 2017-07-11 2017-07-11 Training method/system, computer-readable recording medium and the terminal of model of mind Pending CN107392125A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710561654.5A CN107392125A (en) 2017-07-11 2017-07-11 Training method/system, computer-readable recording medium and the terminal of model of mind

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710561654.5A CN107392125A (en) 2017-07-11 2017-07-11 Training method/system, computer-readable recording medium and the terminal of model of mind

Publications (1)

Publication Number Publication Date
CN107392125A true CN107392125A (en) 2017-11-24

Family

ID=60340400

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710561654.5A Pending CN107392125A (en) 2017-07-11 2017-07-11 Training method/system, computer-readable recording medium and the terminal of model of mind

Country Status (1)

Country Link
CN (1) CN107392125A (en)

Cited By (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108319938A (en) * 2017-12-31 2018-07-24 奥瞳系统科技有限公司 High quality training data preparation system for high-performance face identification system
CN108537289A (en) * 2018-04-24 2018-09-14 百度在线网络技术(北京)有限公司 Training method, device and the storage medium of data identification model
CN108537129A (en) * 2018-03-14 2018-09-14 北京影谱科技股份有限公司 The mask method of training sample, device and system
CN108564134A (en) * 2018-04-27 2018-09-21 网易(杭州)网络有限公司 Data processing method, device, computing device and medium
CN108664999A (en) * 2018-05-03 2018-10-16 北京图森未来科技有限公司 A kind of training method and its device, computer server of disaggregated model
CN108764372A (en) * 2018-06-08 2018-11-06 Oppo广东移动通信有限公司 Construction method and device, mobile terminal, the readable storage medium storing program for executing of data set
CN108875963A (en) * 2018-06-28 2018-11-23 北京字节跳动网络技术有限公司 Optimization method, device, terminal device and the storage medium of machine learning model
CN108875821A (en) * 2018-06-08 2018-11-23 Oppo广东移动通信有限公司 The training method and device of disaggregated model, mobile terminal, readable storage medium storing program for executing
CN108898162A (en) * 2018-06-08 2018-11-27 东软集团股份有限公司 A kind of data mask method, device, equipment and computer readable storage medium
CN108960409A (en) * 2018-06-13 2018-12-07 南昌黑鲨科技有限公司 Labeled data generation method, equipment and computer readable storage medium
CN109034397A (en) * 2018-08-10 2018-12-18 腾讯科技(深圳)有限公司 Model training method, device, computer equipment and storage medium
CN109359689A (en) * 2018-10-19 2019-02-19 科大讯飞股份有限公司 A kind of data identification method and device
CN109582793A (en) * 2018-11-23 2019-04-05 深圳前海微众银行股份有限公司 Model training method, customer service system and data labeling system, readable storage medium storing program for executing
CN109919317A (en) * 2018-01-11 2019-06-21 华为技术有限公司 A kind of machine learning model training method and device
CN109918444A (en) * 2019-02-01 2019-06-21 上海尚阵智能科技有限公司 Training/verifying/management method/system, medium and equipment of model result
CN110443141A (en) * 2019-07-08 2019-11-12 深圳中兴网信科技有限公司 Data set processing method, data set processing unit and storage medium
WO2020134533A1 (en) * 2018-12-29 2020-07-02 北京市商汤科技开发有限公司 Method and apparatus for training deep model, electronic device, and storage medium
WO2020206958A1 (en) * 2019-04-09 2020-10-15 苏宁云计算有限公司 Iterative learning-based smart aided labeling method and system
CN111815361A (en) * 2020-07-10 2020-10-23 北京思特奇信息技术股份有限公司 Region boundary calculation method and device, electronic equipment and storage medium
CN112036570A (en) * 2020-07-30 2020-12-04 第四范式(北京)技术有限公司 Method and device for scheduling annotation task and readable storage medium
CN112149701A (en) * 2019-06-28 2020-12-29 杭州海康威视数字技术股份有限公司 Image identification method, virtual sample data generation method and storage medium
CN112534367A (en) * 2018-10-01 2021-03-19 欧姆龙株式会社 Learning device, control device, learning method, and learning program

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090228411A1 (en) * 2008-03-06 2009-09-10 Kddi Corporation Reducing method for support vector
CN103793510A (en) * 2014-01-29 2014-05-14 苏州融希信息科技有限公司 Classifier construction method based on active learning
CN105117429A (en) * 2015-08-05 2015-12-02 广东工业大学 Scenario image annotation method based on active learning and multi-label multi-instance learning

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090228411A1 (en) * 2008-03-06 2009-09-10 Kddi Corporation Reducing method for support vector
CN103793510A (en) * 2014-01-29 2014-05-14 苏州融希信息科技有限公司 Classifier construction method based on active learning
CN105117429A (en) * 2015-08-05 2015-12-02 广东工业大学 Scenario image annotation method based on active learning and multi-label multi-instance learning

Cited By (38)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108319938A (en) * 2017-12-31 2018-07-24 奥瞳系统科技有限公司 High quality training data preparation system for high-performance face identification system
CN108319938B (en) * 2017-12-31 2022-05-17 奥瞳系统科技有限公司 High-quality training data preparation system for high-performance face recognition system
CN109919317A (en) * 2018-01-11 2019-06-21 华为技术有限公司 A kind of machine learning model training method and device
CN109919317B (en) * 2018-01-11 2024-06-04 华为技术有限公司 Machine learning model training method and device
CN108537129A (en) * 2018-03-14 2018-09-14 北京影谱科技股份有限公司 The mask method of training sample, device and system
CN108537129B (en) * 2018-03-14 2021-01-08 北京影谱科技股份有限公司 Method, device and system for marking training samples
CN108537289A (en) * 2018-04-24 2018-09-14 百度在线网络技术(北京)有限公司 Training method, device and the storage medium of data identification model
CN108537289B (en) * 2018-04-24 2023-04-07 百度在线网络技术(北京)有限公司 Training method and device of data recognition model and storage medium
CN108564134A (en) * 2018-04-27 2018-09-21 网易(杭州)网络有限公司 Data processing method, device, computing device and medium
CN108564134B (en) * 2018-04-27 2021-07-06 网易(杭州)网络有限公司 Data processing method, device, computing equipment and medium
CN108664999A (en) * 2018-05-03 2018-10-16 北京图森未来科技有限公司 A kind of training method and its device, computer server of disaggregated model
CN108664999B (en) * 2018-05-03 2021-02-12 北京图森智途科技有限公司 Training method and device of classification model and computer server
CN108898162A (en) * 2018-06-08 2018-11-27 东软集团股份有限公司 A kind of data mask method, device, equipment and computer readable storage medium
US11138478B2 (en) 2018-06-08 2021-10-05 Guangdong Oppo Mobile Telecommunications Corp., Ltd. Method and apparatus for training, classification model, mobile terminal, and readable storage medium
CN108875821A (en) * 2018-06-08 2018-11-23 Oppo广东移动通信有限公司 The training method and device of disaggregated model, mobile terminal, readable storage medium storing program for executing
CN108764372B (en) * 2018-06-08 2019-07-16 Oppo广东移动通信有限公司 Construction method and device, mobile terminal, the readable storage medium storing program for executing of data set
CN108898162B (en) * 2018-06-08 2021-03-30 东软集团股份有限公司 Data annotation method, device and equipment and computer readable storage medium
WO2019233297A1 (en) * 2018-06-08 2019-12-12 Oppo广东移动通信有限公司 Data set construction method, mobile terminal and readable storage medium
CN108764372A (en) * 2018-06-08 2018-11-06 Oppo广东移动通信有限公司 Construction method and device, mobile terminal, the readable storage medium storing program for executing of data set
CN108960409A (en) * 2018-06-13 2018-12-07 南昌黑鲨科技有限公司 Labeled data generation method, equipment and computer readable storage medium
CN108960409B (en) * 2018-06-13 2021-08-03 南昌黑鲨科技有限公司 Method and device for generating annotation data and computer-readable storage medium
CN108875963A (en) * 2018-06-28 2018-11-23 北京字节跳动网络技术有限公司 Optimization method, device, terminal device and the storage medium of machine learning model
CN109034397B (en) * 2018-08-10 2023-04-07 腾讯科技(深圳)有限公司 Model training method and device, computer equipment and storage medium
CN109034397A (en) * 2018-08-10 2018-12-18 腾讯科技(深圳)有限公司 Model training method, device, computer equipment and storage medium
US11971709B2 (en) 2018-10-01 2024-04-30 Omron Corporation Learning device, control device, learning method, and recording medium
CN112534367A (en) * 2018-10-01 2021-03-19 欧姆龙株式会社 Learning device, control device, learning method, and learning program
CN112534367B (en) * 2018-10-01 2023-09-19 欧姆龙株式会社 Learning device, control device, learning method, and computer-readable storage medium
CN109359689A (en) * 2018-10-19 2019-02-19 科大讯飞股份有限公司 A kind of data identification method and device
CN109582793A (en) * 2018-11-23 2019-04-05 深圳前海微众银行股份有限公司 Model training method, customer service system and data labeling system, readable storage medium storing program for executing
CN109582793B (en) * 2018-11-23 2023-05-23 深圳前海微众银行股份有限公司 Model training method, customer service system, data labeling system and readable storage medium
WO2020134533A1 (en) * 2018-12-29 2020-07-02 北京市商汤科技开发有限公司 Method and apparatus for training deep model, electronic device, and storage medium
CN109918444A (en) * 2019-02-01 2019-06-21 上海尚阵智能科技有限公司 Training/verifying/management method/system, medium and equipment of model result
WO2020206958A1 (en) * 2019-04-09 2020-10-15 苏宁云计算有限公司 Iterative learning-based smart aided labeling method and system
CN112149701A (en) * 2019-06-28 2020-12-29 杭州海康威视数字技术股份有限公司 Image identification method, virtual sample data generation method and storage medium
CN112149701B (en) * 2019-06-28 2024-05-10 杭州海康威视数字技术股份有限公司 Image recognition method, virtual sample data generation method and storage medium
CN110443141A (en) * 2019-07-08 2019-11-12 深圳中兴网信科技有限公司 Data set processing method, data set processing unit and storage medium
CN111815361A (en) * 2020-07-10 2020-10-23 北京思特奇信息技术股份有限公司 Region boundary calculation method and device, electronic equipment and storage medium
CN112036570A (en) * 2020-07-30 2020-12-04 第四范式(北京)技术有限公司 Method and device for scheduling annotation task and readable storage medium

Similar Documents

Publication Publication Date Title
CN107392125A (en) Training method/system, computer-readable recording medium and the terminal of model of mind
CN110349147B (en) Model training method, fundus macular region lesion recognition method, device and equipment
CN109034159A (en) image information extracting method and device
CN109176512A (en) A kind of method, robot and the control device of motion sensing control robot
CN108229341A (en) Sorting technique and device, electronic equipment, computer storage media, program
CN109544598A (en) Method for tracking target, device and readable storage medium storing program for executing
JP2023044669A (en) Graph model-based brain function registration method
CN109685077A (en) A kind of breast lump image-recognizing method and device
CN109934227A (en) System for recognizing characters from image and method
CN110781976B (en) Extension method of training image, training method and related device
CN107845092A (en) Clothes LOGO effect detection methods, device and electronic equipment
CN111897985B (en) Image multi-label classification method, system, equipment and readable storage medium
CN109902548A (en) A kind of object properties recognition methods, calculates equipment and system at device
CN106383912A (en) Picture retrieval method and apparatus
CN108537168A (en) Human facial expression recognition method based on transfer learning technology
CN111915618B (en) Peak response enhancement-based instance segmentation algorithm and computing device
Shu et al. LVC-Net: Medical image segmentation with noisy label based on local visual cues
CN110222330A (en) Method for recognizing semantics and device, storage medium, computer equipment
CN110264274A (en) Objective group's division methods, model generating method, device, equipment and storage medium
CN115345938B (en) Global-to-local-based head shadow mark point positioning method, equipment and medium
CN109740752A (en) Depth model training method and device, electronic equipment and storage medium
CN112132812B (en) Certificate verification method and device, electronic equipment and medium
CN108595815A (en) Artificial intelligence body training system and passive circuit Optimum Design System and method
CN106874913A (en) A kind of vegetable detection method
CN110033012A (en) A kind of production method for tracking target based on channel characteristics weighted convolution neural network

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20171124