CN110110861A - Determine method and apparatus, the storage medium of model hyper parameter and model training - Google Patents

Determine method and apparatus, the storage medium of model hyper parameter and model training Download PDF

Info

Publication number
CN110110861A
CN110110861A CN201910384551.5A CN201910384551A CN110110861A CN 110110861 A CN110110861 A CN 110110861A CN 201910384551 A CN201910384551 A CN 201910384551A CN 110110861 A CN110110861 A CN 110110861A
Authority
CN
China
Prior art keywords
parameter
path
machine learning
learning model
hyper parameter
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201910384551.5A
Other languages
Chinese (zh)
Other versions
CN110110861B (en
Inventor
林宸
李楚鸣
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Sensetime Technology Development Co Ltd
Original Assignee
Beijing Sensetime Technology Development Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Sensetime Technology Development Co Ltd filed Critical Beijing Sensetime Technology Development Co Ltd
Priority to CN201910384551.5A priority Critical patent/CN110110861B/en
Publication of CN110110861A publication Critical patent/CN110110861A/en
Application granted granted Critical
Publication of CN110110861B publication Critical patent/CN110110861B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Software Systems (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Medical Informatics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Artificial Intelligence (AREA)
  • Image Analysis (AREA)

Abstract

The embodiment of the present disclosure provides the technology and image processing techniques of a kind of determining model hyper parameter, is conducive to the image procossing performance of hoisting machine learning model, wherein the method for determining model hyper parameter comprises determining that the initial value of hyper parameter;According to the initial value of the hyper parameter and sample graph image set, M1 repetitive exercise is carried out to initial machine learning model by each path in parallel multiple paths, obtain each path first updates machine learning model, first based on each path in the multiple path updates the performance parameter of machine learning model, and the numerical value of the hyper parameter is updated to the first updated value;The first updated value and the sample graph image set based on the hyper parameter, the further numerical value update that machine learning model carries out M2 repetitive exercise and the hyper parameter is updated to the first of the multiple path, until reaching default cut-off condition, the final numerical value of the hyper parameter is obtained.

Description

Determine method and apparatus, the storage medium of model hyper parameter and model training
Technical field
This disclosure relates to machine learning techniques, and in particular to determine model hyper parameter and model training method and apparatus, Storage medium.
Background technique
In recent years, the machine learning models such as deep neural network various computer visions application in achieve significantly at Function.Under the driving of a large amount of flag datas, the performance of network has also reached surprising level.However, machine learning model is super Parameter is presently mainly to use engineer, after the good model hyper parameter of engineer, keeps the hyper parameter constant, to engineering It practises model to be trained, finally obtains the model parameter of machine learning model, it is finally obtained due to the limitation of engineer The performance of machine learning model need to be advanced optimized.
Summary of the invention
In view of this, the disclosure at least provides the technology and model training technology of a kind of determining model hyper parameter.
In a first aspect, providing a kind of method of determining model hyper parameter, which comprises determine the initial of hyper parameter Value;According to the initial value of the hyper parameter and sample graph image set, by each path in parallel multiple paths to initial machine Learning model carries out M1 repetitive exercise, and obtain each path first updates machine learning model, wherein the multiple The training parameter in different paths has the different numerical value sampled based on the hyper parameter in path, and M1 is greater than or equal to 1 and be less than or equal to the first numerical value;First based on each path in the multiple path updates the performance of machine learning model The numerical value of the hyper parameter is updated to the first updated value by parameter;The first updated value and the sample based on the hyper parameter Image set, to the first of the multiple path update machine learning model carry out M2 repetitive exercise and the hyper parameter into one Step Numerical updates, until reach default cut-off condition, obtains the final numerical value of the hyper parameter, wherein M2 be greater than or equal to 1 and Less than or equal to the first numerical value.
In one possible implementation, machine learning model is updated in first to the multiple path to carry out Before the further numerical value of M2 repetitive exercise and the hyper parameter updates, further includes: the first more new engine from multiple paths First object is chosen in learning model updates machine learning model;Machine learning model is updated by the first of the multiple path Model parameter is updated to the model parameter that the first object updates machine learning model.
It is in one possible implementation, described from the multiple road in conjunction with any embodiment that the disclosure provides The first of diameter, which updates, chooses first object update machine learning model in machine learning model, comprising: be based on the multiple path The first performance parameter for updating machine learning model, the is chosen from the first of the multiple path the update machine learning model One target update machine learning model.
It is in one possible implementation, described according to the super ginseng in conjunction with any embodiment that the disclosure provides Several initial values and sample graph image set carries out initial machine learning model M1 times by each path in parallel multiple paths Repetitive exercise, obtain each path first update machine learning model, comprising: initial value based on the hyper parameter and At least one first sample image in the sample image collection, by each path in multiple paths to the initial machine It practises model and carries out the first repetitive exercise, the first inner ring for obtaining each path updates machine learning model;Based on described super The second sample image of at least one of the initial value of parameter and the sample image collection, passes through each road in the multiple path Diameter updates machine learning model to first inner ring in each path and carries out secondary iteration training, obtains each path Second inner ring updates machine learning model;The second inner ring based on each path in the multiple path updates machine learning mould Type, obtain each path first update machine learning model.
It is in one possible implementation, described to be based on the super ginseng in conjunction with any embodiment that the disclosure provides At least one first sample image in several initial values and the sample image collection, by each path in multiple paths to institute It states initial machine learning model and carries out the first repetitive exercise, the first inner ring for obtaining each path updates machine learning mould Type, comprising: the initial value based on the hyper parameter carries out multiple repairing weld, obtains first instruction in each path in the multiple path Practice parameter;At least one of first training parameter and the sample image collection based on each path in the multiple path One sample image carries out the first repetitive exercise to the initial machine learning model, and the first inner ring for obtaining each path updates Machine learning model.
In conjunction with any embodiment that the disclosure provides, in one possible implementation, the of each path The training parameter used in one repetitive exercise and secondary iteration training is that the initial value based on the hyper parameter carries out different samplings It obtains.
It is in one possible implementation, described based on the multiple in conjunction with any embodiment that the disclosure provides The first of each path updates the performance parameter of machine learning model in path, and the numerical value of the hyper parameter is updated to first more New value, comprising: first based on each path in the multiple path updates the performance parameter of machine learning model, determine described in The model modification parameter in each path;The model modification parameter in the multiple path is averaging processing, average update is obtained Parameter;According to the average undated parameter, the numerical value of the hyper parameter is updated to the first updated value.
In conjunction with any embodiment that the disclosure provides, in one possible implementation, described based on described more The first of each path updates the performance parameter of machine learning model in a path, and the numerical value of the hyper parameter is updated to first Before updated value, the method also includes: the property of machine learning model is updated to first of each path in the multiple path Energy parameter is normalized;First based on each path in the multiple path updates the property of machine learning model Energy parameter, is updated to the first updated value for the numerical value of the hyper parameter, comprising: described in obtaining after the normalized The first of each path updates the performance parameter of machine learning model in multiple paths, and the numerical value of the hyper parameter is updated to the One updated value.
In conjunction with any embodiment that the disclosure provides, in one possible implementation, the performance parameter includes Accuracy rate.
It is in one possible implementation, described to be based on the super ginseng in conjunction with any embodiment that the disclosure provides The first several updated value and the sample graph image set update machine learning model to the first of the multiple path and carry out M2 times repeatedly In generation, the trained further numerical value with the hyper parameter updated, comprising: the first updated value and the sample image based on hyper parameter Collection updates machine learning model to first of each path in the multiple path and carries out M2 repetitive exercise, obtains described every The second of a path updates machine learning model;Second based on each path in the multiple path updates machine learning model Performance parameter, the numerical value of the hyper parameter is updated to the second updated value.
In conjunction with any embodiment that the disclosure provides, in one possible implementation, the hyper parameter includes using In the enhancing distribution parameter for carrying out image enhancement processing to the sample graph image set;The initial value according to the hyper parameter and Sample graph image set carries out M1 repetitive exercise, packet to initial machine learning model by each path in parallel multiple paths It includes: according to the enhancing distribution parameter, determining enhancing probability distribution, include multiple images enhancing behaviour in the enhancing probability distribution The probability of work;Based on the enhancing probability distribution, by sampling parallel multiple paths in the multiple data enhancement operations In the target data in each path enhance operation, at least one sample image in each path is carried out at image enhancement Reason obtains at least one enhancing image;Based at least one enhancing image in each path in the multiple path, to described first Beginning machine learning model carries out M1 repetitive exercise.
In conjunction with any embodiment that the disclosure provides, in one possible implementation, the described first more new engine The acquisition of the performance parameter of learning model, including handling as follows: updating machine by first of each path in the multiple path Device learning model handles at least one test image in test image collection, obtains processing result image;Based on described The corresponding described image processing result in each path in multiple paths, obtain each path first update machine learning model Performance parameter.
In conjunction with any embodiment that the disclosure provides, in one possible implementation, the default cut-off condition Including in following at least one of: default update times are reached to the update times of the hyper parameter;Alternatively, the multiple path The performance of obtained update machine learning model reaches target capabilities.
In conjunction with any embodiment that the disclosure provides, in one possible implementation, the method also includes: from Target machine is chosen in the final updated machine learning model in the multiple paths obtained in the case where reaching the default cut-off condition Device learning model, wherein the target machine learning model is the trained machine learning model for image procossing.
Second aspect provides a kind of device of determining model hyper parameter, and described device includes: initialization module, for true Determine the initial value of hyper parameter;Model training module, for the initial value and sample graph image set according to the hyper parameter, by parallel Multiple paths in each path to initial machine learning model carry out M1 repetitive exercise, obtain each path first update Machine learning model, wherein the training parameter in different paths has in the multiple path is sampled based on the hyper parameter Obtained different numerical value, M1 are greater than or equal to 1 and are less than or equal to the first numerical value;Super ginseng update module, for based on described more The first of each path updates the performance parameter of machine learning model in a path, and the numerical value of the hyper parameter is updated to first Updated value;Super ginseng obtains module, for based on the hyper parameter the first updated value and the sample graph image set, to the multiple The first of path updates the further numerical value update that machine learning model carries out M2 repetitive exercise and the hyper parameter, until reaching To default cut-off condition, the final numerical value of the hyper parameter is obtained, wherein M2 is greater than or equal to 1 and is less than or equal to the first number Value.
In conjunction with any embodiment that the disclosure provides, in one possible implementation, the super ginseng obtains module, It is also used to: updating machine learning model in first to the multiple path and carry out M2 repetitive exercise and the hyper parameter Further numerical value update before, updated from the first of the multiple path and choose first object in machine learning model and update machine Device learning model;The model parameter of the first update machine learning model in the multiple path is updated to the first object more The model parameter of new engine learning model.
In conjunction with any embodiment that the disclosure provides, in one possible implementation, the super ginseng obtains module, When choosing first object update machine learning model in machine learning model for updating from the first of the multiple path, packet Include: first based on the multiple path updates the performance parameter of machine learning model, the first more new engine from multiple paths First object is chosen in learning model updates machine learning model.
In conjunction with any embodiment that the disclosure provides, in one possible implementation, the model training module, Be specifically used for: at least one first sample image in initial value and the sample image collection based on the hyper parameter passes through Each path carries out the first repetitive exercise to the initial machine learning model in multiple paths, obtains the of each path One inner ring updates machine learning model;At least one of initial value and the sample image collection based on the hyper parameter second Sample image, by each path in the multiple path to first inner ring in each path update machine learning model into The training of row secondary iteration, the second inner ring for obtaining each path update machine learning model;Based in the multiple path Second inner ring in each path updates machine learning model, and obtain each path first updates machine learning model.
In conjunction with any embodiment that the disclosure provides, in one possible implementation, the model training module, When the first inner ring for obtaining each path updates machine learning model, comprising: based on the initial of the hyper parameter Value carries out multiple repairing weld, obtains first training parameter in each path in the multiple path;Based on every in the multiple path At least one first sample image in first training parameter in a path and the sample image collection, to the initial machine It practises model and carries out the first repetitive exercise, the first inner ring for obtaining each path updates machine learning model.
In conjunction with any embodiment that the disclosure provides, in one possible implementation, the of each path The training parameter used in one repetitive exercise and secondary iteration training is that the initial value based on the hyper parameter carries out different samplings It obtains.
In conjunction with any embodiment that the disclosure provides, in one possible implementation, the super ginseng update module, Be specifically used for: based in the multiple path each path first update machine learning model performance parameter, determine described in The model modification parameter in each path;The network undated parameter in the multiple path is averaging processing, average update is obtained Parameter;According to the average undated parameter, the numerical value of the hyper parameter is updated to the first updated value.
In conjunction with any embodiment that the disclosure provides, in one possible implementation, the super ginseng update module, It is specifically used for: place is normalized to the performance parameter of the first update machine learning model in each path in the multiple path Reason;First based on each path in the multiple path obtained after normalized updates the performance ginseng of machine learning model Number, is updated to the first updated value for the numerical value of the hyper parameter.
In conjunction with any embodiment that the disclosure provides, in one possible implementation, the performance parameter includes Accuracy rate.
In conjunction with any embodiment that the disclosure provides, in one possible implementation, the super ginseng obtains module, It is specifically used for: the first updated value and the sample graph image set based on the hyper parameter, to each path in the multiple path First update machine learning model carry out M2 repetitive exercise, obtain each path second update machine learning mould Type;Second based on each path in multiple paths updates the performance parameter of machine learning model, by the numerical value of the hyper parameter It is updated to the second updated value.
In conjunction with any embodiment that the disclosure provides, in one possible implementation, the hyper parameter includes using In the enhancing distribution parameter for carrying out image enhancement processing to the sample graph image set;The model training module, is specifically used for: root According to the enhancing distribution parameter, enhancing probability distribution is determined, include multiple images enhancing operation in the enhancing probability distribution Probability;It is every in parallel multiple paths by being sampled in the multiple data enhancement operations based on the enhancing probability distribution The target data in a path enhances operation, carries out image enhancement processing at least one sample image in each path, obtains Enhance image at least one;Based at least one enhancing image in each path in the multiple path, to the initial machine Device learning model carries out M1 repetitive exercise.
In conjunction with any embodiment that the disclosure provides, in one possible implementation, the super ginseng update module, It is also used to obtain the acquisition of the performance parameter of the first update machine learning model, including handles as follows: by the multiple path In each path first update machine learning model at least one test image in test image collection is handled, obtain Processing result image;Based on the corresponding described image processing result in path each in the multiple path, each path is obtained First updates the performance parameter of machine learning model.
In conjunction with any embodiment that the disclosure provides, in one possible implementation, the default cut-off condition Including in following at least one of: default update times are reached to the update times of the hyper parameter;Alternatively, the multiple path The performance of obtained update machine learning model reaches target capabilities.
In conjunction with any embodiment that the disclosure provides, in one possible implementation, the super ginseng obtains module, The final updated machine learning model in the multiple path for being also used to obtain in the case where reach the default cut-off condition Middle selection target machine learning model, wherein the target machine learning model is the trained machine for image procossing Learning model.
In conjunction with any embodiment that the disclosure provides, in one possible implementation, the super ginseng obtains module, It is also used to after the final numerical value for obtaining hyper parameter, the final numerical value based on the hyper parameter, training initialization model parameter Initial machine learning model, obtain training completion target machine learning model.
The third aspect additionally provides a kind of training method of machine learning model, comprising: is based on above-mentioned any embodiment In the method for hyper parameter of determination machine learning model obtain the final numerical value of hyper parameter, and based on the final of the hyper parameter Numerical value, training have the initial machine learning model of original model parameter, obtain target machine learning model.
The device that the embodiment of the present disclosure additionally provides for realizing the training method.
Fourth aspect provides a kind of image processing method, which comprises obtains image to be processed;Utilize engineering It practises model to handle the image to be processed, obtains processing result image, wherein the hyper parameter of the machine learning model What the hyper parameter training that the method for being based upon cover half type hyper parameter really described in disclosure any embodiment determines obtained.
5th aspect provides a kind of electronic equipment, and the electronic equipment includes memory, processor, and the memory is used In the computer instruction that storage can be run on a processor, the processor is used to realize this when executing the computer instruction Really the training method of the method or machine learning model of cover half type hyper parameter described in open any embodiment.
6th aspect, provides a kind of computer readable storage medium, is stored thereon with computer program, described program is located Manage the instruction that the method or machine learning model of cover half type hyper parameter really described in disclosure any embodiment are realized when device executes Practice method.
7th aspect, provides a kind of training system, comprising: parameter management server and controller, wherein parameter management clothes Business device is used to manage and update the numerical value of hyper parameter.Controller be used for the machine learning model based on hyper parameter carry out circulation or Iteration updates, and the performance parameter for updating foundation as hyper parameter is fed back to parameter management server, so that parameter management Server carries out the update of hyper parameter accordingly.
The embodiment of the present disclosure provides the technology of cover half type hyper parameter really, by being based on multiple paths to machine learning model Be iterated training, update the hyper parameter after having carried out M1 repetitive exercise to model, and based on updated hyper parameter after Continuous be iterated by multiple paths to model is trained and the further update of hyper parameter numerical value, by it is this with numerical value more Newly-service check is the mode of cycling element come the hyper parameter for determining machine learning model, accelerates the optimization efficiency of hyper parameter, And hyper parameter is updated by the performance parameter based on machine learning model, enables the search of hyper parameter with machine learning mould The performance parameter of type is optimization direction, to promote the performance of the machine learning model based on determining hyper parameter.
Detailed description of the invention
Technical solution in order to illustrate more clearly of disclosure one or more embodiment or in the related technology, below will be right Attached drawing needed in embodiment or description of Related Art is briefly described, it should be apparent that, be described below in it is attached Figure is only some embodiments recorded in disclosure one or more embodiment, for those of ordinary skill in the art, Without any creative labor, it is also possible to obtain other drawings based on these drawings.
Fig. 1 shows a kind of method of determining model hyper parameter of at least one embodiment of disclosure offer;
Fig. 2 shows the methods that the another kind that at least one embodiment of the disclosure provides determines model hyper parameter;
Fig. 3, which shows at least one embodiment of the disclosure and provides the application scenarios of the method for cover half type hyper parameter really, to be shown Example;
Fig. 4 shows another method for determining model hyper parameter of at least one embodiment of disclosure offer;
Fig. 5 shows a kind of model training process of at least one embodiment of disclosure offer;
Fig. 6 shows a kind of device of determining model hyper parameter of at least one embodiment of disclosure offer;
Fig. 7 shows a kind of process of image processing method of at least one embodiment of disclosure offer;
Fig. 8 shows a kind of training method of machine learning model of at least one embodiment of disclosure offer;
Fig. 9 shows a kind of training device of machine learning model of at least one embodiment of disclosure offer.
Specific embodiment
In order to make those skilled in the art more fully understand the technical solution in disclosure one or more embodiment, under Face will combine disclosure one or more embodiment in attached drawing, to the technical solution in disclosure one or more embodiment into Row clearly and completely describes, it is clear that described embodiment is only disclosure a part of the embodiment, rather than whole realities Apply example.Based on disclosure one or more embodiment, those of ordinary skill in the art are without creative efforts The range of disclosure protection all should belong in every other embodiment obtained.
The embodiment of the present disclosure provides the method and model training method of cover half type hyper parameter really, can be applied to engineering Practise the training platform of model, such as cloud training platform or end training platform, wherein the training platform may include one or more Equipment, correspondingly, the above method can be executed by cloud device, the network equipment or terminal device etc., and the embodiment of the present disclosure is to this Without limitation.In order to make it easy to understand, below in this way by being described for training equipment.
In the embodiment of the present disclosure, based on designing reasonable search space and using efficient way of search, automation is searched The reasonable hyper parameter of rope, can be improved the performance of the machine learning model based on the hyper parameter.
In some embodiments, constant current journey includes repeatedly circulation to hyper parameter really, wherein in each cycle, training is set The standby current value based on hyper parameter is sampled, and the training parameter that each path uses in multiple paths is obtained, and being based on should Training parameter carries out the model parameter adjustment of the initial machine learning model of this circulation, obtains updated machine learning mould Type, then, based on the performance parameter of updated machine learning model in multiple paths, the numerical value for carrying out hyper parameter updates, more Determination of the hyper parameter for the training parameter in multiple paths in subsequent cycle after new.
In the embodiments of the present disclosure, machine learning model can be neural network or other are trained based on hyper parameter Model, the embodiment of the present disclosure do not limit this.
In some embodiments, hyper parameter can be the parameter for obtaining loss function, or including increasing for data Strong parameter, etc., the embodiment of the present disclosure do not limit this.
The embodiment of the present disclosure provides a kind of method of determining model hyper parameter, and this method is intended to quickly search performance Preferable hyper parameter, and the training process of the search process of hyper parameter and model is combined together and carries out simultaneously by this method. Referring to Figure 1, Fig. 1 shows a kind of method of determining model hyper parameter of at least one embodiment of disclosure offer.
In step 100, the initial value of hyper parameter is determined.
Optionally, step 100~104 can correspond to the first time circulation process in all circulations, correspondingly, in step In 100, hyper parameter can be initialized, such as assigns the hyper parameter initial value.Alternatively, step 100~104 can correspond to all follow Certain intercycle process in ring correspondingly recycles the numerical value for the hyper parameter that last circulation obtains as this The initial value of hyper parameter, the embodiment of the present disclosure do not limit this.
In a step 102, according to the initial value of the hyper parameter and sample graph image set, by every in parallel multiple paths A path carries out M1 repetitive exercise to initial machine learning model, and obtain each path first updates machine learning mould Type.
It is alternatively possible to by multiple parallel paths repetitive exercise machine learning model, such as uniform machinery is learnt Model is distributed to multiple processing units or thread, obtains multiple machine learning models in multiple processing units or thread, each The process object of machine learning model can be at least one sample image in sample image collection.
Training equipment can be sampled based on the hyper parameter, and the training parameter of model is obtained, wherein the sampling can be with It carries out repeatedly, to obtain the training parameter that each path uses in multiple paths.The training parameter in different paths in multiple paths It can be and carry out what different samplings obtained based on the hyper parameter, correspondingly, different paths can have different training parameter numbers Value.
After obtaining training parameter, machine learning model is carried out in conjunction with the training parameter and at least one sample image Repetitive exercise.
In order to make it easy to understand, the model in this being recycled before repetitive exercise is known as initial machine learning model, And the model obtained after repetitive exercise is known as the first update machine learning model.The repetitive exercise number of model can be M1 Secondary, which may be greater than or be equal to 1 and be less than or equal to the first numerical value.Specifically, the number of the repetitive exercise can be one Secondary or multiple, for example, the numerical value of M1 is 50 or 30 or other numerical value, which can preset or based on machine learning model Current performance determine that or determine by other means, correspondingly, repetitive exercise number in difference circulation can be identical or not Together, the embodiment of the present disclosure does not limit this.
In addition, in one cycle, multiple paths can carry out the repetitive exercise of same number to machine learning model, or Person, different paths can also carry out the repetitive exercise of different numbers to machine learning model, and the embodiment of the present disclosure does not limit this It is fixed.
In some embodiments, it is assumed that M1 is the integer greater than 1, in each path, processing unit or at least one line Journey can carry out an iteration to initial machine learning model based on training parameter, obtain network parameter machine learning adjusted Model, and using network parameter machine learning model adjusted as the input of following iteration, wherein in following iteration, It can be sampled to obtain another training parameter based on hyper parameter, M1 iteration is carried out back and forth with this, obtains the first of the path Update machine learning model.
In addition, parallel multiple paths, the embodiment of the present disclosure are not intended to limit the quantity in path.Empirical experimentation is adopted Different number of paths is taken, influence might have on the performance of model training, in certain quantitative range, the increasing of number of paths Add the performance of meeting lift scheme;After number of paths reaches certain numerical value, the influence to the performance of model training may also can drop It is low.It can integrate and be considered according to the consumption of model performance and computing resource, determine selected number of paths.
At step 104, first based on each path in the multiple path updates the performance ginseng of machine learning model Number, is updated to the first updated value for the numerical value of the hyper parameter.
The update of hyper parameter can be the performance parameter that machine learning model is updated based on first in multiple paths, one In a little embodiments, the mode based on intensified learning carries out the update of hyper parameter, for example, the purpose for updating hyper parameter is so that being based on The hyper parameter can train to obtain performance parameter more preferably model.
The updated numerical value of hyper parameter can be known as the first updated value by training equipment.In this way, just completing hyper parameter more New one cycle.
In step 106, the first updated value based on the hyper parameter and the sample graph image set, to the multiple path First update machine learning model and carry out the further numerical value of M2 repetitive exercise and the hyper parameter and update, up to reaching pre- If cut-off condition, the final numerical value of the hyper parameter is obtained.
It optionally, can be based on the first updated value and sample image of hyper parameter after step 104 updates hyper parameter Collection, continues to be iterated update to the machine learning model in multiple paths.For example, changing to the first update machine learning model It can be M2 times for frequency of training, which may be greater than or be equal to 1 and be less than or equal to the first numerical value, and M2 and M1 can phase Together, it can also be different.
With the continuation repetitive exercise of machine learning model, hyper parameter also will further update numerical value, likewise, for example, It can continue the numerical value for updating hyper parameter based on the performance parameter of the machine learning model after continuation repetitive exercise.Optionally, when When reaching preset cut-off condition, the final numerical value of hyper parameter can be obtained, which is the preferably super ginseng searched Number numerical value.For example, the preset cut-off condition, which can be, reaches default update times to the update times of the hyper parameter, or The performance of person, the update machine learning model that the multiple path obtains reach target capabilities.
Really the method for cover half type hyper parameter in some embodiments, by based on multiple paths to machine learning model into Row iteration training is updated the hyper parameter after having carried out M1 repetitive exercise to model, and is continued based on updated hyper parameter Trained and hyper parameter numerical value further update is iterated to model by multiple paths, by this with numerical value update- Service check is the mode of cycling element come the hyper parameter for determining machine learning model, accelerates the optimization efficiency of hyper parameter, and And hyper parameter is updated by the performance parameter based on machine learning model, enable the search of hyper parameter with machine learning model Performance parameter be optimization direction, thus promoted based on determining hyper parameter machine learning model performance.
Fig. 2 shows the method that the another kind that at least one embodiment of the disclosure provides determines model hyper parameter, the implementations Example describes the processing mode of more detailed search model hyper parameter.As shown in Fig. 2, this method may include:
In step 100, the initial value of hyper parameter is determined.
In a step 102, according to the initial value of the hyper parameter and sample graph image set, by every in parallel multiple paths A path carries out M1 repetitive exercise to initial machine learning model, and obtain each path first updates machine learning mould Type.
Optionally, M1 repetitive exercise is carried out to the initial machine learning model in each path and obtains the first update engineering Model is practised, may include following process:
For example, can be by least one first sample image in initial value and sample image collection based on hyper parameter to first Beginning machine learning model carries out the model obtained after the first repetitive exercise, and referred to as the first inner ring updates machine learning model.It is obtaining During obtaining first inner ring update machine learning model, multiple repairing weld can be carried out based on the initial value of hyper parameter, obtained First training parameter in each path in multiple paths is then based on first training parameter and sample graph image set to initial machine Learning model carries out the first repetitive exercise, and the first inner ring for obtaining each path updates machine learning model.
Then, can the second sample image of at least one of initial value based on hyper parameter and sample image collection, pass through Each path updates machine learning model progress secondary iteration instruction to first inner ring in each path in the multiple path Practice, obtained more new model is properly termed as the second inner ring and updates machine learning model.
The second inner ring based on each path in the multiple path updates machine learning model, continues iteration and updates, directly Machine learning model is updated to obtain each path first.Wherein, first repetitive exercise in each path and The training parameter that uses is that the initial value based on the hyper parameter carries out different samplings and obtains in secondary iteration training.
At step 104, first based on each path in the multiple path updates the performance ginseng of machine learning model Number, is updated to the first updated value for the numerical value of the hyper parameter.
In some embodiments, the performance parameter based on model updates hyper parameter, may include: based on the multiple path In each path first update machine learning model performance parameter, determine the model modification parameter in each path.It is right The model modification parameter in the multiple path is averaging processing, and obtains average undated parameter.And joined according to the average update Number, is updated to the first updated value for the numerical value of the hyper parameter.
Wherein, the acquisition of the performance parameter of model can update engineering by first of each path in multiple paths It practises model to handle at least one test image in test image collection, obtains processing result image;And based on described The processing result image in multiple paths obtains the first of each path performance parameter for updating machine learning model.For example, this property Energy parameter can be the accuracy rate of model.
In addition, before the numerical value of hyper parameter is updated to the first updated value, it can also be to each road in multiple paths The performance parameter of first update machine learning model of diameter is normalized.And it is obtained after being based on the normalized Performance parameter, the numerical value of hyper parameter is updated to the first updated value.
In step s 106, it is updated from the first of the multiple path and chooses first object update machine in machine learning model The model parameter of the first update machine learning model in the multiple path is updated to the first object more by device learning model The model parameter of new engine learning model.
Optionally, before carrying out the repetitive exercise of model next time, the machine learning model in multiple paths can be unified Model parameter.For example, the model parameter of machine learning model can be updated based on the first of multiple path, from multiple first It updates and chooses a model in machine learning model as first object update machine learning model, and by the multiple path The model parameter of model is updated to the model parameter that the first object updates machine learning model.Illustratively, the mould Type performance parameter includes but is not limited to accuracy rate of the model on verifying collection.
In step 106, the first updated value based on the hyper parameter and the sample graph image set, to the multiple path First update machine learning model and carry out the further numerical value of M2 repetitive exercise and the hyper parameter and update, up to reaching pre- If cut-off condition, the final numerical value of the hyper parameter is obtained.
For example, when updating the further repetitive exercise of machine learning model to first, can based on hyper parameter first more New value and the sample graph image set update machine learning model to first of each path in the multiple path and carry out M2 times repeatedly Generation training, obtain each path second update machine learning model.Also, likewise, based on each in multiple paths The second of path updates the performance parameter of machine learning model, and the numerical value of the hyper parameter is updated to the second updated value.
As above, the update of hyper parameter and the update of model parameter carry out simultaneously, in the training process of model, execute The performance parameter of interim feedback model training, and hyper parameter is updated based on the performance parameter.
In addition, may exist two kinds of situations after the hyper parameter after being optimized:
For example, trained model can be obtained simultaneously.The institute obtained in the case where reach the default cut-off condition It states in the final updated machine learning model in multiple paths, chooses target machine learning model, wherein the target machine learns mould Type is the trained machine learning model for image procossing.
In another example can use the hyper parameter of final optimization pass, re -training obtains model as finally trained engineering Practise model.I.e. based on the final numerical value of the hyper parameter, the initial machine learning model of training initialization model parameter is instructed Practice the target machine learning model completed.
Really the method for cover half type hyper parameter in some embodiments, by the training process to machine learning model In, while hyper parameter is updated, accelerate the search efficiency of hyper parameter;Also, this method is with engineering when updating hyper parameter The performance parameter of model is practised as optimization foundation, to ensure that the effect of hyper parameter, realize both fast and sound search compared with Excellent hyper parameter.
Really the method for cover half type hyper parameter as above, the chess game optimization of the hyper parameter suitable for any machine learning.It is as follows Will be by taking one of hyper parameter as an example, which is the enhancing point for carrying out image enhancement processing to the sample graph image set Cloth parameter illustrates the implementation procedure of this method with this, but it is understood that, this method is not limited to the excellent of enhancing distribution parameter Change.
Data enhancing strategy can be applied to the training process to network, by using data enhancing strategy to the defeated of network Enter data and carry out data enhancing, the overfitting problem in network training process can be improved.For example, can be to the picture number of input According to the data enhancing such as being rotated, shearing, move, new training data is obtained for training network, helps to improve network Generalization ability improves the accuracy rate of neural network forecast.
And the data enhancing strategy used is different, the training effect of network is also different.For example, being enhanced using a certain data The generalization ability of strategy, trained network is slightly worse, and accuracy rate of the network on verifying collection is relatively low.And another data is used to increase Strong strategy, the accuracy rate of trained network can get a promotion.Therefore, a kind of preferably data enhancing strategy is searched for, for instruction Get the network important role of better performances.
The embodiment of the present disclosure provides a kind of optimization method of data enhancing strategy, this method be a kind of automation search for compared with The method of excellent data enhancing strategy, this method described below:
Firstly, illustrating some basic contents in order to which the description to this method is clearer:
Search space (Search Space) and data enhancement operations: in network training, network inputs data are carried out Which data enhancement operations can be using preset some data enhancement operations, be used by selecting in these preset operations, Preset multiple operations can be known as " search space ".
For example, image can be rotated when the input data of network is image, color adjustment, shear, is mobile etc. The single processing such as " rotation ", " movement " can be known as data enhancing element (an augmentation by a variety of processing Element), the group of two kinds of data enhancing elements can be collectively referred to as " data enhancement operations " (an by the disclosure augmentation operation).Assuming that data enhancing element quantity have 36, it would be possible that combination of two quantity That is the quantity K=36 of data enhancement operations2, i.e., include K data enhancement operations in search space.For each of network inputs A image can be applied to progress data enhancing in the image by selecting a data enhancement operations in the search space.
Wherein, in some embodiments, data enhancing element includes but is not limited to: horizontal shear (HorizontalShear), vertical shear (VerticalShear), move horizontally (HorizontalTranslate), be vertical Mobile (VerticalTranslate), rotation (Rotate), hue adjustment (ColorAdjust), tone separation (Posterize), exposure-processed (Solarize), comparison processing (Contrast), Edge contrast (Sharpness), at blast Manage (Brightness), automatic comparison (AutoContrast), hue balancing (Equalize), inversion processing (Invert) Deng.
Enhance distribution parameter (augmentation distribution parameter θ) and probability distribution (pθ): the increasing Strong distribution parameter can be a numerical value, each data enhancement operations can correspond to the numerical value of an enhancing distribution parameter. And probability distribution is converted to according to the enhancing distribution parameter, probability distribution is that each data enhancement operations are corresponding Enhancing distribution parameter numerical value be all converted to the numerical value between 0 to 1, that is, be converted to a probability.Also, the institute in search space The sum of corresponding probability of some data enhancement operations is 1.Illustratively, probability distribution can be { 0.1,0.08,0.32 ... }, K probability is shared, the sum of these probability are equal to 1, each of these probability can indicate that corresponding data enhancement operations are adopted The probability value that sample uses.
In embodiment of the disclosure, above-mentioned enhancing distribution parameter can be used as hyper parameter when network training (hyper-parameter), and can optimize simultaneously with the training process of network with network parameter.In addition, above-mentioned one kind Probability distribution can regard a kind of data enhancing strategy (an augmentation policy) as, because to the training data of network It carries out being to sample the data enhancement operations used based on the probability distribution when data enhancing, when probability distribution changes, sampling Data enhancement operations also change therewith.
Fig. 3, which shows at least one embodiment of the disclosure and provides the application scenarios of the method for cover half type hyper parameter really, to be shown Example.The application scenarios provide a kind of training system, which may include: parameter management server 11 and controller 12.Its In,
Parameter management server 11, for managing and updating the numerical value of hyper parameter.
Controller 12, including multiple machine learning models, the multiple machine learning model are based on hyper parameter and are recycled Or iteration updates, and obtains the performance parameter for the model that training obtains, which is fed back to parameter management by controller 12 Server, so that parameter management server carries out the update of hyper parameter accordingly.Controller 12 is also based on parameter management clothes The business updated hyper parameter of device 11 continues the training of model.
Fig. 3 is referred to, a kind of architectural form that the training system can use is illustrated, which exists using one kind The two-level architecture of line optimization can enable enhancing distribution parameter to carry out simultaneously with network parameter excellent by the two-level architecture Change.It wherein, include parameter management server 11 and controller 12 in the two-level architecture.Parameter management server 11 is responsible for depositing Storage and update enhancing distribution parameter, and the probability distribution of data enhancement operations is obtained based on the enhancing distribution parameter,.Controller It may include one group of network to be trained (Network) in 12, it is assumed that N number of network is shared, for example, the N can be in 4 to 16 Any value.The network structure disclosure of the network does not limit, such as can be CNN (Convolutional Neural Networks, convolutional neural networks).
It can be using the entirety of parameter management server 11 and controller 12 as an outer ring (Outer Loop), the outer ring It is run in a manner of intensified learning.Outer ring can carry out the dimension of undated parameter management server 11 with T time step of iteration (time step) The enhancing distribution parameter of shieldθ, for the enhancing distribution parameter as a hyper parameter, updating the enhancing distribution parameter can be used as parameter Behavior (action) of the management server 11 in intensified learning training process.And controller 12 is based on enhancing distribution parameter instruction Accuracy rate of the network got on verifying collection as the return reward value (Reward) in intensified learning training process, according to The Reward carries out the update of enhancing distribution parameter, and the optimization of Reward, i.e. accuracy rate are realized after T time step of iteration most Height is the target direction of intensified learning.The T=1,2 ... ... Tmax
And N number of network to be trained in controller 12, it can be used as inner ring (Inner Loop), which can be with Parallel operation.The training of each network, which can be, is adopted based on parameter management server 11 in the enhancing distribution parameter that outer ring updates The data enhancement operations that sample uses, and network training is carried out based on the enhanced data of data.The training of network parameter can make With stochastic gradient descent method (SGD, Stochastic gradient descent).Each network can with iteration i times (i=1, 2 ... .I), and using accuracy rate of the iteration i times network on verifying collection as above-mentioned return reward value Reward, for more New enhancing distribution parameter.
Fig. 4 shows another method for determining model hyper parameter of at least one embodiment of disclosure offer, as follows Embodiment description in, parallel multiple paths may include parallel multiple networks, which can be machine learning model. Also, in this example, hyper parameter is the enhancing distribution parameter for carrying out image enhancement processing to sample graph image set.In conjunction with figure Shown in 4, this method may include following processing step:
In step 400, parallel enhancing distribution parameter of multiple networks based on initialization respectively carries out enhancing operation and adopts Sample obtains the data enhancement operations of the Web vector graphic.
Optionally, the multiple networks for including in controller 12 can be with parallel training.
Each network has the network parameter of initialization, and the enhancing that can be safeguarded based on parameter management server 11 Probability distribution, sampling is applied to the data enhancement operations of the input data of network.The data of different network samples enhance behaviour It can be different.The data enhancement operations that sampling obtains can be known as to target data enhancing operation.
In step 402, each network in the multiple network is according to the data enhancement operations, to the network Input data carries out data enhancing, and carries out network training using enhanced data, obtains updated network parameter.
Optionally, each network carries out network training using enhanced data, can obtain after carrying out data enhancing Updated network parameter.The network training of this step can be with iteration preset times.
For example, data enhancement operations may include carrying out following processing to image: rotation, horizontal shear, is hung down at color adjustment Staight scissors cut, move horizontally, vertically moving, tone separation, exposure-processed etc..The enhancing operation of these target datas be can use to every At least one sample image in a path carries out image enhancement processing, obtains at least one enhancing image.And it is possible to based on institute At least one the enhancing image for stating each path in multiple paths, carries out M1 repetitive exercise to the initial network.
In step 404, the accuracy rate on each comfortable verifying collection of the multiple network that training obtains is obtained.
Optionally, the performance parameter of model is by taking accuracy rate as an example.
For example, can be by validation data set, the accuracy rate for the network that training obtains in verification step 404 (accuracy).In N number of network of controller 12, the data enhancement operations that heterogeneous networks sampling uses are different, correspondingly, training The effect of obtained network is also different, and the accuracy rate of each network may exist difference.
In a step 406, by nitrification enhancement, the accuracy rate based on the multiple network updates the enhancing distribution Parameter.
Optionally, the accuracy rate of the multiple networks obtained based on training is taken according to nitrification enhancement undated parameter management The enhancing distribution parameter that business device 11 is safeguarded.The optimization of accuracy rate can be used as the target direction of intensified learning, be based on accuracy rate Action (update for enhancing distribution parameter) is updated as Reward.
In a step 408, by the network parameter of the highest network of the accuracy rate, it is applied to the multiple network, is obtained The new network of next round iteration completes the update of a time step of intensified learning.
Collect the upper highest network of accuracy rate in verifying it is alternatively possible to determine in multiple networks of controller 12, and should The network parameter of the highest network of accuracy rate is synchronized to all networks, the i.e. network parameter of all-network in controller 12 all Using the network parameter of the synchronization.
Be synchronized the sum of network parameter, multiple networks in controller 12 are properly termed as new network, and can continue into Enter the update of next time step of intensified learning.
In step 410, multiple new networks are based on the updated enhancing distribution parameter, continue intensified learning The iteration of next time step updates.
The iteration of this step, which updates, repeats aforementioned step 400 to 408, for example, 11 base of parameter management server Updated probability distribution is obtained in updated enhancing distribution parameter, each network in controller 12 can be based on the update Probability distribution afterwards samples data enhancement operations to be used.And based on enhanced data training network, verifying accuracy rate, The operations such as enhancing distribution parameter are updated, are no longer described in detail.
In step 412, when reaching step of preset renewal time, verifying is collected into the upper highest network of accuracy rate and is determined as The network that final training obtains, and the enhancing distribution parameter of the last one time step update is obtained, to be distributed according to the enhancing Parameter determines the data enhancement operations taken.
Described reaches default renewal time step, is a kind of exemplary preset model cut-off condition.
The training of intensified learning can preset the quantity of renewal time step, for example, time step preferably at most updates Tmax, then When reaching TmaxWhen, it can be using the highest network of accuracy rate as final network, and the enhancing distribution parameter of final updating is made To optimize obtained hyper parameter.Probability distribution can be obtained based on the hyper parameter, and can be instructed according to the probability distribution in network Sampled data enhancing operation is trained the enhancing of data during white silk.
Really the method for cover half type hyper parameter in some embodiments, by the way that distribution parameter will be enhanced as network training Hyper parameter, and in the training process of network, the interim feedback for taking intensified learning updates the mode of the hyper parameter, makes The optimization of the optimization and network parameter that obtain hyper parameter carries out simultaneously, has been obviously improved the search efficiency of hyper parameter;Furthermore this method The mode training hyper parameter for the intensified learning taken, using the verifying collection accuracy rate of network as optimization aim, the update of hyper parameter Optimal for training direction with accuracy rate, so that obtained hyper parameter is also more accurate, effect is more preferable.
In addition, it should be noted that, the training of network (Network) is primary in traditional hyperparameter optimization mode Property training on earth, be primary training from the Final Network of the network of initialization to the end similar in the frame diagram of Fig. 3 It obtains, then based on the trained preferable data enhancing strategy of web search, entire search process is very time-consuming, and needs to instruct The tactful search result of final data enhancing could be obtained by practicing a fairly large number of network, and time cost and calculating cost are very It is high.And the optimization method of the disclosure, it is equivalent to and interrupts the disposable training process of above-mentioned network, from " the network of initialization The feedback of multiple stages is increased during Final Network " to the end, the certain number of network iteration just feeds back one It is secondary, distribution parameter is enhanced based on interim feedback updated in the way of intensified learning, this mode by hyperparameter optimization and The optimization of network parameter carries out simultaneously, can greatly speed up the optimization efficiency of hyper parameter, and time cost significantly reduces, moreover, super The optimization of parameter is using the accuracy rate of network as optimization aim direction, and effect is preferable, can both fast and sound be searched preferably Hyper parameter, trained the number networks also greatly reduce compared to traditional approach.
There is provided the data enhancing of one embodiment tactful optimization method again as follows, in the embodiment, to input picture For the data enhancement operations such as being rotated, being sheared, translated, searching for enhances strategy to the preferably data that input picture is taken, So that according to this, preferably the enhanced data of strategy progress train the network performance of obtained network more preferable.
It firstly, initializing in frame shown in Fig. 4, include N number of network (Network) N number of network tool in controller 12 There is identical initialization network parameter ω ', and initializes the enhancing distribution ginseng of each data enhancement operations in search space Number θ.Parameter management server 11 is according to enhancing distribution parameter, the probability distribution p θ of available data enhancement operations.
Further, it is also possible to set, for each network in inner ring, in network training, extensive chemical may be set in Each time step (the at each time step) iteration practised updates I times, then carries out the enhancing distribution parameter of an outer ring Update.The iteration of the intensified learning of outer ring can set total iteration TmaxIt is secondary, i.e. iteration TmaxFinal optimization net is obtained after secondary Network and optimization hyper parameter.
Fig. 5 illustrates the training process of one of network, wherein when the network training, each time network parameter repeatedly Generation update is the input data by one group of B input picture for network, and iteration can be used as a grouping, i.e. network each time It is to be iterated training by multiple groupings.
In step 500, the enhancing probability distribution based on data enhancing strategy, is made by sampling in multiple data enhancement operations Target data enhancing operation.
For example, including B input picture in this group of data when using wherein one group of input picture is trained.For Each input picture can sample a data enhancement operations based on enhancing probability distribution, which is properly termed as number of targets It is operated according to enhancing.Also, obtained these data enhancement operations of sampling can be regarded as and sampled based on hyper parameter The training parameter of model, the data enhancement operations that the network samples in different paths obtain can be different.
Illustratively, for one of input picture, the data enhancement operations of sampling may include that two data increase Strong element " rotation " and " shearing ".The input picture can be enhanced using the data enhancement operations, increased Strong image.
In step 502, network training is carried out using enhanced data, obtains updated network parameter.
Optionally, enhanced data can be enhancing image.The network training of this step can be multiple with iteration, final To updated network parameter.
When network parameter updates, gradient descent method can be used, network parameter is optimized with the minimum principle of loss function.Such as Under formula (1) illustrate a kind of update mode of network parameter:
Formula (1) as above, network parameter is when updating, and following relating to parameters:
ηwIt is the learning rate of network parameter;
Loss function value when being the grouping iteration, the loss function value enhance according to the data that current iteration samples OperationCurrent network parameterWith the input data x of current iterationB, yB, determine.
In step 504, judge whether the number of iterations of network reaches default the number of iterations.
As previously mentioned, may be set in each time step (at each time step) of intensified learning, in controller Each network iteration update I time, that is, pass through I grouping progress network parameter update.
Optionally, if judging that the number of iterations i of network is less than presets the number of iterations I, in return step 300, under B input picture of one grouping continues trained and iterative network parameter.
If judging, the number of iterations i of network is equal to default the number of iterations I, continues step 306.
For example, the default the number of iterations of this step can be M1 times.In addition, in different times in step, the iteration of network Number may be the same or different.
In addition, what training obtained for the first time is properly termed as the first inner ring more new engine in the successive ignition training of network Learning model is properly termed as the second inner ring again and updates machine learning model after iteration, and so on.
In step 506, the network that training obtains is obtained.
Each network N etwork is trained all in accordance with process shown in Fig. 3, and training here is referred in intensified learning Each time step in training.These N number of networks can be parallel training, finally obtain following parameter: { ωT, n}N=1:N, Obtain the network parameter of each network of N number of network.Wherein, ωT, nIndicate that the network parameter is the T in intensified learning The parameter of n-th of network in a time step.
Then, the network obtained based on N number of training can go the accuracy rate of verifying network by validation data set.It can be with Referring to the signal of Fig. 1, each network can obtain an accuracy rate Acc.In some embodiments, using accuracy rate as network Performance parameter for, certainly in other examples, expression of other parameters as performance parameter can also be used.
The acquisition of the accuracy rate can be the network obtained by training at least one test chart in test image collection As being handled, processing result image is obtained, and accuracy rate is obtained based on the processing result image.
Reward value reward of the above-mentioned accuracy rate as intensified learning can pass through reinforcing based on reward value reward Learning algorithm is updated enhancing distribution parameter.Intensified learning Policy-Gradient algorithm can be used in the embodiment of the present disclosure REINFORCE, the accuracy rate based on each network update enhancing distribution parameter.
Following formula (2) illustrates REINFORCE algorithm:
Formula (2) as above, acc (wT, n) be n-th of network of the T time step of intensified learning verifying collection it is quasi- True rate,It is the data enhancement operations of n-th of network, which sampled according to probability distribution.Therefore, according to public affairs Formula (2) is the data enhancement operations of the accuracy rate and the network samples according to each network, determines the multiple network pair The average gradient of the enhancing distribution parameter.
In formula (2) as above, still by taking performance parameter is accuracy rate as an example, according to the available each path of accuracy rate Model modification parameter, the model modification parameter can be the gradient of the network in the path, and the gradient in multiple paths be carried out flat It handles, obtains above-mentioned average gradient (be averaged undated parameter).
In addition, place first can be normalized in the accuracy rate in each path in multiple paths before executing formula (2) Reason, and calculated based on the accuracy rate after normalization.
Based on above-mentioned average gradient, enhancing distribution parameter can be updated, following formula (3):
Wherein, ηθIt is the learning rate of probability distribution.First is properly termed as according to the updated hyper parameter numerical value of formula (3) Updated value, certainly, after iteration updates repeatedly, the second updated value of available hyper parameter, third updated value etc..
As above, in a time step, based on network in the accuracy rate of verifying collection, enhancing distribution parameter, parameter are had updated Management server 11 can be according to the updated enhancing distribution parameter, update probability distribution, with next time step according to Updated probability distribution continues sampled data enhancing operation.
Continuing with combining referring to Fig. 3, dotted arrow 13 indicates the accuracy rate of each network feeding back to parameter management service Device 11, parameter management server 11, can be with update probability distribution ps according to above-mentioned REINFORCE algorithmθ.14 table of solid arrow Show the iteration that next time step is continued to execute based on updated probability distribution.In one embodiment, in next iteration Before beginning, the network parameter of the highest network of accuracy rate can be synchronized to the all-network of controller, as shown in figure 3, accurately The network parameter of the highest network of rate is parameter 15, indicates the parameter being applied to all-network by synchronous arrow 16.At it In his example, it is unitized that the network parameter before starting can also be iterated using other modes.
Updated network, which is based on updated probability distribution, to be continued to sample, and the iteration for starting next time step updates. In the iteration of next time step updates, obtained network is properly termed as the second update machine learning model, and equally can The performance parameter for updating machine learning model according to the second of path each in multiple path, more by the numerical value of the hyper parameter It is newly the second updated value.
Until the default renewal time for reaching intensified learning walks TmaxWhen, using the highest network of accuracy rate as Final The optimal network that Network, i.e. training are obtained, is referred to as target machine learning model.Meanwhile the last one time step makes The hyper parameter that the enhancing distribution parameter used is obtained as final optimization passThat is the final numerical value of hyper parameter.
The method flow of the above-mentioned Fig. 4 of the disclosure and determining model hyper parameter shown in fig. 5 can use the process such as the following table 1 It indicates, wherein obtaining hyper parameterWhile obtain Model Weight parameter.It is understood that other can also be used Process flow, as long as meeting the method for disclosure any embodiment cover half type hyper parameter really.
A kind of process of the method for determining model hyper parameter of following example:
For example, initialization hyper parameter is θ0, and the network parameter for initializing each network is w0
For each time step (total T=T of intensified learningmaxA time step), it is carried out following processing:
Each network of controller carries out i=I iteration altogether and updates, and iteration updates (if 0≤i≤I) according to public affairs each time Formula (1) updates network parameterAfter I iteration, network parameter w is obtainedT, n, the wT, nIt indicates the of intensified learning The network parameter of n-th of network in T time step, wherein share N number of network.
When obtaining N number of respective network parameter w of networkT, nAfterwards, multiple networks are calculated to the increasing according to formula (2) The average gradient of strong distribution parameterAnd hyper parameter is updated according to formula (3), make in the T time step of intensified learning Hyper parameter can be expressed as θT
Before the next time step for entering intensified learning carries out network training, verified by being selected in N number of network The network parameter w of the highest network of accuracy rate on collectionT, and by the network parameter synchronous applications to N number of network, so that this is N number of Network repetitive exercise starting point having the same.
When reaching preset intensified learning time step TmaxWhen, obtain final network parameter wTmaxAnd hyper parameter
In addition, being to reach the default renewal time of intensified learning and walk T in some embodimentsmaxAs hyperparameter optimization Cut-off condition in other examples can also be using other cut-off conditions.Also, the example is with hyperparameter optimization For obtaining trained network simultaneously, it is also possible to that trained network is finally not yet received, but is obtained using final optimization pass The initial machine learning model of one initialization model parameter of the hyper parameter re -training arrived obtains the target machine of training completion Learning model.
After obtaining hyper parameter, when a network will train, if the search for the data enhancement operations that the network uses Space is identical as search space when hyperparameter optimization, and data when can directly be carried out the network training using the hyper parameter are increased It by force, will be so that network obtains preferable performance.If the search space of the corresponding data enhancement operations of network of training is become Change, the optimization method that can also reuse disclosure offer searches for new hyper parameter, that is, re-searches for preferably data enhancing Strategy.
Following table 1, it may be said that bright disclosure any embodiment provides the experiment effect of the method for cover half type hyper parameter really Fruit.Refer to table 1, ResNet-18, the WideResNet-28-10 in left side etc. indicate different network structures, above table Baseline, Cutout etc. indicate that different data enhance tactful preparation method, wherein OHL-Auto-Aug indicates to utilize this public affairs The method opened.The data enhancing strategy that a variety of different methods are obtained is applied to the training of multiple network structure, demonstrate,proves through test Bright, disclosed method can reduce error rate compared with various other ways, wherein disclosed method relative to Baseline method trains the lower error rate of obtained network about 30%.The hyper parameter obtained using method of disclosure Training machine learning model, the performances such as accuracy rate for the model that can be improved.
1 method of disclosure of table is compared with the effect of other methods
Fig. 6 provides a kind of device of determining model hyper parameter, as shown in fig. 6, the apparatus may include: initialization mould Block 61, model training module 62, super ginseng update module 63 and super ginseng obtain module 64.
Initialization module 61, for determining the initial value of hyper parameter;
Model training module 62, for the initial value and sample graph image set according to the hyper parameter, by parallel multiple Each path carries out M1 repetitive exercise to initial machine learning model in path, and obtain each path first updates engineering Practise model, wherein the training parameter in different paths has the difference sampled based on the hyper parameter in multiple paths Numerical value, M1 are greater than or equal to 1 and are less than or equal to the first numerical value;
Super ginseng update module 63 updates machine learning model for first based on each path in the multiple path The numerical value of the hyper parameter is updated to the first updated value by performance parameter;;
Super ginseng obtains module 64, for based on the hyper parameter the first updated value and the sample graph image set, to described The first of multiple paths updates the further numerical value update that machine learning model carries out M2 repetitive exercise and the hyper parameter, directly To default cut-off condition is reached, the final numerical value of the hyper parameter is obtained, wherein M2 is greater than or equal to 1 and is less than or equal to the One numerical value.
In some alternative embodiments, surpass ginseng and obtain module 64, be also used to: described to the first of the multiple path Before updating the further numerical value update that machine learning model carries out M2 repetitive exercise and the hyper parameter, from the multiple road The first of diameter, which updates, chooses first object update machine learning model in machine learning model;More by the first of the multiple path The model parameter of new engine learning model is updated to the model parameter that the first object updates machine learning model.
In some alternative embodiments, surpass ginseng and obtain module 64, for updating machine from the first of the multiple path When choosing first object update machine learning model in device learning model, comprising: first based on the multiple path updates machine The performance parameter of device learning model updates from the first of the multiple path and chooses first object update machine in machine learning model Device learning model.
In some alternative embodiments, the model training module 62, is specifically used for: based on the initial of the hyper parameter At least one first sample image in value and the sample image collection, by each path in multiple paths to the initial machine Device learning model carries out the first repetitive exercise, and the first inner ring for obtaining each path updates machine learning model;Based on institute The second sample image of initial value and at least one of the sample image collection for stating hyper parameter, by every in the multiple path A path updates machine learning model to first inner ring in each path and carries out secondary iteration training, obtains each road Second inner ring of diameter updates machine learning model;The second inner ring based on each path in the multiple path updates machine learning Model, obtain each path first update machine learning model.
In some alternative embodiments, model training module 62, in the first inner ring for obtaining each path When updating machine learning model, comprising: the initial value based on the hyper parameter carries out multiple repairing weld, obtains in the multiple path First training parameter in each path;The first training parameter and the sample image based on each path in the multiple path At least one first sample image in collection carries out the first repetitive exercise to initial machine learning model, obtains each road First inner ring of diameter updates machine learning model.
In some alternative embodiments, it is used in first repetitive exercise in each path and secondary iteration training Training parameter be the initial value based on the hyper parameter carry out it is different sampling obtain.
In some alternative embodiments, surpass ginseng update module 63, be specifically used for: based on each road in the multiple path The first of diameter updates the performance parameter of machine learning model, determines the model modification parameter in each path;To the multiple The network undated parameter in path is averaging processing, and obtains average undated parameter;It, will be described super according to the average undated parameter The numerical value of parameter is updated to the first updated value.
In some alternative embodiments, surpass ginseng update module 63, be specifically used for: to each path in the multiple path First update machine learning model performance parameter be normalized;Described in being obtained after the normalized The first of each path updates the performance parameter of machine learning model in multiple paths, and the numerical value of the hyper parameter is updated to the One updated value.
In some alternative embodiments, the performance parameter includes accuracy rate.
In some alternative embodiments, surpass ginseng and obtain module 64, be specifically used for: first based on the hyper parameter updates Value and the sample graph image set update machine learning model to first of each path in the multiple path and carry out M2 iteration Training, obtain each path second update machine learning model;Second based on each path in multiple paths updates The numerical value of the hyper parameter is updated to the second updated value by the performance parameter of machine learning model.
In some alternative embodiments, the hyper parameter includes for carrying out at image enhancement to the sample graph image set The enhancing distribution parameter of reason;Model training module 62, is specifically used for: according to the enhancing distribution parameter, determining enhancing probability point Cloth, described enhance in probability distribution includes that multiple images enhance the probability operated;Based on the enhancing probability distribution, by described more The target data enhancing operation that each path in parallel multiple paths is sampled in a data enhancement operations, to described each At least one sample image in path carries out image enhancement processing, obtains at least one enhancing image;Based on the multiple path In each path at least one enhancing image, to the initial machine learning model carry out M1 repetitive exercise.
In some alternative embodiments, surpass ginseng update module 63, be also used to obtain the first update machine learning model The acquisition of performance parameter, including handling as follows: updating machine learning model by first of each path in the multiple path At least one test image in test image collection is handled, processing result image is obtained;Based in the multiple path The corresponding described image processing result in each path obtains the first of each path performance parameter for updating machine learning model.
In some alternative embodiments, the default cut-off condition includes at least one in following: to the super ginseng Several update times reach default update times;Alternatively, the performance for the update machine learning model that the multiple path obtains reaches To target capabilities.
In some alternative embodiments, surpass ginseng and obtain module 64, be also used to from the feelings for reaching the default cut-off condition Target machine learning model is chosen in the final updated machine learning model in the multiple path obtained under condition, wherein described Target machine learning model is the trained machine learning model for image procossing.
In some alternative embodiments, surpass ginseng and obtain module 64, be also used in the final numerical value for obtaining the hyper parameter Later, based on the final numerical value of the hyper parameter, the initial machine learning model of training initialization model parameter obtains having trained At target machine learning model.
The disclosure additionally provides a kind of electronic equipment, and the equipment includes memory, processor, and the memory is for depositing The computer instruction that can be run on a processor is stored up, the processor is for realizing the disclosure when executing the computer instruction The training method of any embodiment method or machine learning model of cover half type hyper parameter really.
The disclosure additionally provides a kind of computer readable storage medium, is stored thereon with computer program, described program quilt The training of the disclosure any embodiment method or machine learning model of cover half type hyper parameter really is realized when processor executes Method.
Fig. 7 provides the flow diagram of the image processing method in an embodiment of the present disclosure, as shown in fig. 7, the party Method may include:
In step 700, image to be processed is obtained.
Which kind of image the image to be processed that this step does not limit input model is.
In a step 702, the image to be processed is handled using machine learning model, obtains image procossing knot Fruit, wherein the hyper parameter of the machine learning model is by cover half type hyper parameter really described in disclosure any embodiment What method determined.
Image processing method in some embodiments, since the hyper parameter of the machine learning model of processing image is to pass through The method of the above-mentioned type hyper parameter of cover half really of the disclosure determines that the hyper parameter has preferable effect, so at using the model Managing obtained processing result image also has good performance.
Fig. 8 shows a kind of training method of machine learning model of at least one embodiment of disclosure offer, such as Fig. 8 institute Show, this method may include:
In step 800, the final numerical value of the hyper parameter is obtained.
Optionally, the method for cover half type hyper parameter really can be provided by disclosure any embodiment, determines the super ginseng Several final numerical value.
In step 802, based on the final numerical value of the hyper parameter, training has the initial machine of original model parameter Model is practised, target machine learning model is obtained.
Since hyper parameter is determined by the method for the determination hyper parameter of disclosure any embodiment, the effect of hyper parameter compared with It is good, therefore also there is preferable performance using the machine learning model of hyper parameter training.
Fig. 9 shows a kind of training device of machine learning model of at least one embodiment of disclosure offer, such as Fig. 9 institute Show, which includes: that super ginseng obtains module 91 and model training module 92.
Super ginseng obtains module 91, for obtaining by the method for cover half type hyper parameter really described in disclosure any embodiment Obtain the final numerical value of the hyper parameter;
Model training module 92, for the final numerical value based on the hyper parameter, training has the first of original model parameter Beginning machine learning model obtains target machine learning model.
The disclosure additionally provides a kind of computer readable storage medium, is stored thereon with computer program, described program quilt The method or machine learning model of cover half type hyper parameter really are realized described in the disclosure any embodiment when processor executes Training method.
The disclosure additionally provides a kind of electronic equipment, and the electronic equipment includes memory, processor, and the memory is used In the computer instruction that storage can be run on a processor, the processor is used to realize this when executing the computer instruction Really the training method of the method or machine learning model of cover half type hyper parameter described in open any embodiment.
It will be understood by those skilled in the art that disclosure one or more embodiment can provide as method, system or computer Program product.Therefore, complete hardware embodiment, complete software embodiment or combination can be used in disclosure one or more embodiment The form of embodiment in terms of software and hardware.Moreover, disclosure one or more embodiment can be used it is one or more its In include computer usable program code computer-usable storage medium (including but not limited to magnetic disk storage, CD-ROM, Optical memory etc.) on the form of computer program product implemented.
The embodiment of the present disclosure also provides a kind of computer readable storage medium, can store computer on the storage medium Program realizes the neural network for Text region of the disclosure any embodiment description when described program is executed by processor The step of training method, and/or, realize the disclosure any embodiment description character recognition method the step of.Wherein, described "and/or" indicates at least with one of them in the two, for example, " more and/or B " includes three kinds of schemes: more, B and " more And B ".
Various embodiments are described in a progressive manner in the disclosure, same and similar part between each embodiment It may refer to each other, each embodiment focuses on the differences from other embodiments.Especially for data processing For apparatus embodiments, since it is substantially similar to the method embodiment, so being described relatively simple, related place is referring to method The part of embodiment illustrates.
It is above-mentioned that disclosure specific embodiment is described.Other embodiments are within the scope of the appended claims. In some cases, the behavior recorded in detail in the claims or step can be executed according to the sequence being different from embodiment And desired result still may be implemented.In addition, process depicted in the drawing not necessarily require the particular order shown or Person's consecutive order is just able to achieve desired result.In some embodiments, multitasking and parallel processing are also possible Or it may be advantageous.
The embodiment of theme and feature operation described in the disclosure can be realized in the following: Fundamental Digital Circuit has The existing computer software of body or firmware, the computer hardware including structure disclosed in the disclosure and its structural equivalents, Or the combination of one or more of which.The embodiment of theme described in the disclosure can be implemented as one or more meters Calculation machine program, i.e. coding are on tangible non-transitory program carrier to execute or control data processing equipment by data processing equipment Operation computer program instructions in one or more modules.Alternatively, or in addition, program instruction can be encoded On manually generated transmitting signal, such as electricity, light or electromagnetic signal that machine generates, the signal are generated to encode information onto And suitable receiver apparatus is transferred to be executed by data processing equipment.Computer storage medium can be machine readable storage Equipment, machine readable storage substrate, random or serial access memory equipment or one or more of which combination.
Processing and logic flow described in the disclosure can be by one or more of the one or more computer programs of execution A programmable calculator executes, to execute corresponding function by the way that output is operated and generated according to input data.It is described Processing and logic flow can also be by dedicated logic circuit-such as FPG more (field programmable gate arrays) or more SIC (dedicated collection At circuit) Lai Zhihang, and device also can be implemented as dedicated logic circuit.
The computer for being suitable for carrying out computer program includes, for example, general and/or special microprocessor or it is any its The central processing unit of his type.In general, central processing unit will refer to from read-only memory and/or random access memory reception Order and data.The basic module of computer includes central processing unit for being practiced or carried out instruction and for storing instruction With one or more memory devices of data.In general, computer will also be including one or more great Rong for storing data Amount storage equipment, such as disk, magneto-optic disk or CD etc. or computer will be coupled operationally with this mass-memory unit To receive from it data or have both at the same time to its transmission data or two kinds of situations.However, computer is not required to have in this way Equipment.In addition, computer can be embedded in another equipment, such as mobile phone, personal digital assistant (PD is more), mobile sound Frequency or video player, game console, global positioning system (GPS) receiver or such as universal serial bus (USB) flash memory The portable memory apparatus of driver, names just a few.
It is suitable for storing computer program instructions and the computer-readable medium of data including the non-volatile of form of ownership Memory, medium and memory devices, for example including semiconductor memory devices (such as EPROM, EEPROM and flash memory device), Disk (such as internal hard drive or removable disk), magneto-optic disk and CD ROM and DVD-ROM disk.Processor and memory can be by special It is supplemented or is incorporated in dedicated logic circuit with logic circuit.
Although the disclosure includes many specific implementation details, these are not necessarily to be construed as limiting any scope of disclosure Or range claimed, and be primarily used for describing the feature of specifically disclosed specific embodiment.Multiple in the disclosure Certain features described in embodiment can also be combined implementation in a single embodiment.On the other hand, in a single embodiment The various features of description can also be performed separately in various embodiments or be implemented with any suitable sub-portfolio.Though in addition, Right feature can work in certain combinations as described above and even initially so be claimed, but come from required guarantor One or more features in the combination of shield can be removed from the combination in some cases, and combination claimed The modification of sub-portfolio or sub-portfolio can be directed toward.
Similarly, although depicting operation in the accompanying drawings with particular order, this is understood not to require these behaviour Make the particular order shown in execute or sequentially carry out or require the operation of all illustrations to be performed, to realize desired knot Fruit.In some cases, multitask and parallel processing may be advantageous.In addition, the various system modules in above-described embodiment Separation with component is understood not to be required to such separation in all embodiments, and it is to be understood that described Program assembly and system can be usually integrated in together in single software product, or be packaged into multiple software product.
The specific embodiment of theme has been described as a result,.Other embodiments are within the scope of the appended claims.? In some cases, the movement recorded in claims can be executed in different order and still realize desired result.This Outside, the processing described in attached drawing and it is nonessential shown in particular order or sequential order, to realize desired result.In certain realities In existing, multitask and parallel processing be may be advantageous.
The foregoing is merely the preferred embodiments of disclosure one or more embodiment, not to limit the disclosure One or more embodiments, all any modifications within the spirit and principle of disclosure one or more embodiment, made, etc. With replacement, improvement etc., should be included within the scope of the protection of disclosure one or more embodiment.

Claims (10)

1. a kind of method of determining model hyper parameter, which is characterized in that the described method includes:
Determine the initial value of hyper parameter;
According to the initial value of the hyper parameter and sample graph image set, by each path in parallel multiple paths to initial machine Learning model carries out M1 repetitive exercise, and obtain each path first updates machine learning model, wherein the multiple The training parameter in different paths has the different numerical value sampled based on the hyper parameter in path, and M1 is greater than or equal to 1 and be less than or equal to the first numerical value;
First based on each path in the multiple path updates the performance parameter of machine learning model, by the hyper parameter Numerical value is updated to the first updated value;
The first updated value and the sample graph image set based on the hyper parameter update engineering to the first of the multiple path It practises model and carries out the further numerical value update of M2 repetitive exercise and the hyper parameter, until reach default cut-off condition, acquisition The final numerical value of the hyper parameter, wherein M2 is greater than or equal to 1 and is less than or equal to the first numerical value.
2. the method according to claim 1, wherein first based on each path in the multiple path The performance parameter for updating machine learning model, is updated to the first updated value for the numerical value of the hyper parameter, comprising:
First based on each path in the multiple path updates the performance parameter of machine learning model, determines each road The model modification parameter of diameter;
The model modification parameter in the multiple path is averaging processing, average undated parameter is obtained;
According to the average undated parameter, the numerical value of the hyper parameter is updated to the first updated value.
3. method according to claim 1 or 2, which is characterized in that described based on each path in the multiple path First update machine learning model performance parameter, it is described before the numerical value of the hyper parameter is updated to the first updated value Method further include:
The performance parameter of the first update machine learning model in each path in the multiple path is normalized;
First based on each path in the multiple path updates the performance parameter of machine learning model, by the super ginseng Several numerical value is updated to the first updated value, comprising:
First based on each path in the multiple path obtained after the normalized updates machine learning model The numerical value of the hyper parameter is updated to the first updated value by performance parameter.
4. method according to any one of claims 1 to 3, which is characterized in that the hyper parameter includes for the sample The enhancing distribution parameter of image set progress image enhancement processing;
The initial value and sample graph image set according to the hyper parameter, by each path in parallel multiple paths to initial Machine learning model carries out M1 repetitive exercise, comprising:
According to the enhancing distribution parameter, enhancing probability distribution is determined, include that multiple images enhance in the enhancing probability distribution The probability of operation;
It is each in parallel multiple paths by being sampled in the multiple data enhancement operations based on the enhancing probability distribution The target data in path enhances operation, carries out image enhancement processing at least one sample image in each path, obtains At least one enhancing image;
Based at least one enhancing image in each path in the multiple path, M1 is carried out to the initial machine learning model Secondary repetitive exercise.
5. a kind of training method of machine learning model characterized by comprising
The final numerical value of the hyper parameter is obtained by method described in any one of Claims 1-4;
Final numerical value based on the hyper parameter, training have the initial machine learning model of original model parameter, obtain target Machine learning model.
6. a kind of device of determining model hyper parameter, which is characterized in that described device includes:
Initialization module, for determining the initial value of hyper parameter;
Model training module, for the initial value and sample graph image set according to the hyper parameter, by parallel multiple paths Each path carries out M1 repetitive exercise to initial machine learning model, and obtain each path first updates machine learning mould Type, wherein the training parameter in different paths has the difference sampled based on the hyper parameter in the multiple path Numerical value, M1 are greater than or equal to 1 and are less than or equal to the first numerical value;
Super ginseng update module updates the performance ginseng of machine learning model for first based on each path in the multiple path Number, is updated to the first updated value for the numerical value of the hyper parameter;
Super ginseng obtains module, for based on the hyper parameter the first updated value and the sample graph image set, to the multiple road The first of diameter updates the further numerical value update that machine learning model carries out M2 repetitive exercise and the hyper parameter, until reaching Default cut-off condition, obtains the final numerical value of the hyper parameter, wherein M2 is greater than or equal to 1 and is less than or equal to the first numerical value.
7. device according to claim 6, which is characterized in that the default cut-off condition includes at least one in following :
Default update times are reached to the update times of the hyper parameter;
Alternatively, the performance for the update machine learning model that the multiple path obtains reaches target capabilities.
8. a kind of training device of machine learning model characterized by comprising
Super ginseng obtains module, for obtaining the final number of the hyper parameter by method described in any one of Claims 1-4 Value;
Model training module, for the final numerical value based on the hyper parameter, training has the initial machine of original model parameter Learning model obtains target machine learning model.
9. a kind of electronic equipment, which is characterized in that the equipment includes memory, processor, and the memory can for storing The computer instruction run on a processor, the processor are used to realize claim 1 when executing the computer instruction To method described in 4 one, or method described in realization claim 5.
10. a kind of computer readable storage medium, is stored thereon with computer program, which is characterized in that described program is processed Claims 1-4 any method is realized when device executes, or realizes method described in claim 5.
CN201910384551.5A 2019-05-09 2019-05-09 Method and device for determining model hyper-parameters and training model and storage medium Active CN110110861B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910384551.5A CN110110861B (en) 2019-05-09 2019-05-09 Method and device for determining model hyper-parameters and training model and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910384551.5A CN110110861B (en) 2019-05-09 2019-05-09 Method and device for determining model hyper-parameters and training model and storage medium

Publications (2)

Publication Number Publication Date
CN110110861A true CN110110861A (en) 2019-08-09
CN110110861B CN110110861B (en) 2021-11-26

Family

ID=67489108

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910384551.5A Active CN110110861B (en) 2019-05-09 2019-05-09 Method and device for determining model hyper-parameters and training model and storage medium

Country Status (1)

Country Link
CN (1) CN110110861B (en)

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110889450A (en) * 2019-11-27 2020-03-17 腾讯科技(深圳)有限公司 Method and device for super-parameter tuning and model building
CN111061875A (en) * 2019-12-10 2020-04-24 深圳追一科技有限公司 Hyper-parameter determination method, device, computer equipment and storage medium
CN111260074A (en) * 2020-01-09 2020-06-09 腾讯科技(深圳)有限公司 Method for determining hyper-parameters, related device, equipment and storage medium
CN111275170A (en) * 2020-01-19 2020-06-12 腾讯科技(深圳)有限公司 Model training method and related device
CN111539177A (en) * 2020-04-22 2020-08-14 中国科学院微电子研究所 Method, device and medium for determining hyper-parameters of layout feature extraction
CN111613287A (en) * 2020-03-31 2020-09-01 武汉金域医学检验所有限公司 Report coding model generation method, system and equipment based on Glow network
CN111695624A (en) * 2020-06-09 2020-09-22 北京市商汤科技开发有限公司 Data enhancement strategy updating method, device, equipment and storage medium
CN112052942A (en) * 2020-09-18 2020-12-08 支付宝(杭州)信息技术有限公司 Neural network model training method, device and system
CN113555008A (en) * 2020-04-17 2021-10-26 阿里巴巴集团控股有限公司 Parameter adjusting method and device for model
CN113762327A (en) * 2020-06-05 2021-12-07 宏达国际电子股份有限公司 Machine learning method, machine learning system and non-transitory computer readable medium
CN113807397A (en) * 2021-08-13 2021-12-17 北京百度网讯科技有限公司 Training method, device, equipment and storage medium of semantic representation model

Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102591917A (en) * 2011-12-16 2012-07-18 华为技术有限公司 Data processing method and system and related device
US20140344193A1 (en) * 2013-05-15 2014-11-20 Microsoft Corporation Tuning hyper-parameters of a computer-executable learning algorithm
WO2017128961A1 (en) * 2016-01-30 2017-08-03 华为技术有限公司 Method and device for training model in distributed system
CN107018184A (en) * 2017-03-28 2017-08-04 华中科技大学 Distributed deep neural network cluster packet synchronization optimization method and system
CN107209873A (en) * 2015-01-29 2017-09-26 高通股份有限公司 Hyper parameter for depth convolutional network is selected
CN108021983A (en) * 2016-10-28 2018-05-11 谷歌有限责任公司 Neural framework search
CN108229647A (en) * 2017-08-18 2018-06-29 北京市商汤科技开发有限公司 The generation method and device of neural network structure, electronic equipment, storage medium
US20180225391A1 (en) * 2017-02-06 2018-08-09 Neural Algorithms Ltd. System and method for automatic data modelling
CN109272118A (en) * 2018-08-10 2019-01-25 北京达佳互联信息技术有限公司 Data training method, device, equipment and storage medium
CN109299142A (en) * 2018-11-14 2019-02-01 中山大学 A kind of convolutional neural networks search structure method and system based on evolution algorithm
US20190095785A1 (en) * 2017-09-26 2019-03-28 Amazon Technologies, Inc. Dynamic tuning of training parameters for machine learning algorithms
CN109657805A (en) * 2018-12-07 2019-04-19 泰康保险集团股份有限公司 Hyper parameter determines method, apparatus, electronic equipment and computer-readable medium

Patent Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102591917A (en) * 2011-12-16 2012-07-18 华为技术有限公司 Data processing method and system and related device
US20140344193A1 (en) * 2013-05-15 2014-11-20 Microsoft Corporation Tuning hyper-parameters of a computer-executable learning algorithm
CN107209873A (en) * 2015-01-29 2017-09-26 高通股份有限公司 Hyper parameter for depth convolutional network is selected
WO2017128961A1 (en) * 2016-01-30 2017-08-03 华为技术有限公司 Method and device for training model in distributed system
CN108021983A (en) * 2016-10-28 2018-05-11 谷歌有限责任公司 Neural framework search
US20180225391A1 (en) * 2017-02-06 2018-08-09 Neural Algorithms Ltd. System and method for automatic data modelling
CN107018184A (en) * 2017-03-28 2017-08-04 华中科技大学 Distributed deep neural network cluster packet synchronization optimization method and system
CN108229647A (en) * 2017-08-18 2018-06-29 北京市商汤科技开发有限公司 The generation method and device of neural network structure, electronic equipment, storage medium
US20190095785A1 (en) * 2017-09-26 2019-03-28 Amazon Technologies, Inc. Dynamic tuning of training parameters for machine learning algorithms
CN109272118A (en) * 2018-08-10 2019-01-25 北京达佳互联信息技术有限公司 Data training method, device, equipment and storage medium
CN109299142A (en) * 2018-11-14 2019-02-01 中山大学 A kind of convolutional neural networks search structure method and system based on evolution algorithm
CN109657805A (en) * 2018-12-07 2019-04-19 泰康保险集团股份有限公司 Hyper parameter determines method, apparatus, electronic equipment and computer-readable medium

Non-Patent Citations (6)

* Cited by examiner, † Cited by third party
Title
BARRET ZOPH ET AL.: "Neural Architecture Search with Reinforcement Learning", 《HTTPS://ARXIV.ORG/ABS/1611.01578V2》 *
EKIN D.CUBUK ET AL.: "autoAugment:learning augmentation policies from data", 《HTTPS://ARXIV.ORG/ABS/1805.09501V2》 *
JAMES BERGSTRA ET AL.: "Random Search for Hyper-Parameter Optimization", 《JOURNAL OF MACHINE LEARNING RESEARCH》 *
VOLODYMYR MNIH ET AL.: "Asynchronous methods for deep reinforcement learning", 《HTTPS://ARXIV.ORG/ABS/1602.01783V2》 *
朱汇龙 等: "基于人群的神经网络超参数优化的研究", 《信息技术》 *
陆高: "基于智能计算的超参数优化及其应用研究", 《中国优秀硕士学位论文全文数据库信息科技辑》 *

Cited By (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110889450B (en) * 2019-11-27 2023-08-11 腾讯科技(深圳)有限公司 Super-parameter tuning and model construction method and device
CN110889450A (en) * 2019-11-27 2020-03-17 腾讯科技(深圳)有限公司 Method and device for super-parameter tuning and model building
CN111061875B (en) * 2019-12-10 2023-09-15 深圳追一科技有限公司 Super parameter determination method, device, computer equipment and storage medium
CN111061875A (en) * 2019-12-10 2020-04-24 深圳追一科技有限公司 Hyper-parameter determination method, device, computer equipment and storage medium
CN111260074A (en) * 2020-01-09 2020-06-09 腾讯科技(深圳)有限公司 Method for determining hyper-parameters, related device, equipment and storage medium
CN111260074B (en) * 2020-01-09 2022-07-19 腾讯科技(深圳)有限公司 Method for determining hyper-parameters, related device, equipment and storage medium
CN111275170A (en) * 2020-01-19 2020-06-12 腾讯科技(深圳)有限公司 Model training method and related device
CN111275170B (en) * 2020-01-19 2023-11-24 腾讯科技(深圳)有限公司 Model training method and related device
CN111613287A (en) * 2020-03-31 2020-09-01 武汉金域医学检验所有限公司 Report coding model generation method, system and equipment based on Glow network
CN113555008A (en) * 2020-04-17 2021-10-26 阿里巴巴集团控股有限公司 Parameter adjusting method and device for model
CN111539177A (en) * 2020-04-22 2020-08-14 中国科学院微电子研究所 Method, device and medium for determining hyper-parameters of layout feature extraction
TWI831016B (en) * 2020-06-05 2024-02-01 宏達國際電子股份有限公司 Machine learning method, machine learning system and non-transitory computer-readable storage medium
CN113762327A (en) * 2020-06-05 2021-12-07 宏达国际电子股份有限公司 Machine learning method, machine learning system and non-transitory computer readable medium
WO2021248791A1 (en) * 2020-06-09 2021-12-16 北京市商汤科技开发有限公司 Method and apparatus for updating data enhancement strategy, and device and storage medium
JP2022541370A (en) * 2020-06-09 2022-09-26 ベイジン・センスタイム・テクノロジー・デベロップメント・カンパニー・リミテッド Data enrichment policy update method, apparatus, device and storage medium
TWI781576B (en) * 2020-06-09 2022-10-21 大陸商北京市商湯科技開發有限公司 Method, equipment and storage medium for updating data enhancement strategy
CN111695624A (en) * 2020-06-09 2020-09-22 北京市商汤科技开发有限公司 Data enhancement strategy updating method, device, equipment and storage medium
CN111695624B (en) * 2020-06-09 2024-04-16 北京市商汤科技开发有限公司 Updating method, device, equipment and storage medium of data enhancement strategy
CN112052942B (en) * 2020-09-18 2022-04-12 支付宝(杭州)信息技术有限公司 Neural network model training method, device and system
CN112052942A (en) * 2020-09-18 2020-12-08 支付宝(杭州)信息技术有限公司 Neural network model training method, device and system
CN113807397A (en) * 2021-08-13 2021-12-17 北京百度网讯科技有限公司 Training method, device, equipment and storage medium of semantic representation model
CN113807397B (en) * 2021-08-13 2024-01-23 北京百度网讯科技有限公司 Training method, training device, training equipment and training storage medium for semantic representation model

Also Published As

Publication number Publication date
CN110110861B (en) 2021-11-26

Similar Documents

Publication Publication Date Title
CN110110861A (en) Determine method and apparatus, the storage medium of model hyper parameter and model training
CN107590603B (en) Based on the dispatching method and system for improving change neighborhood search and differential evolution algorithm
CN108053119B (en) Improved particle swarm optimization method for solving scheduling problem of zero-waiting line shop
US7249142B2 (en) Automatic machine for production of sequences based on profiles as well as method for automatic production of sequences
CN110084378A (en) A kind of distributed machines learning method based on local learning strategy
CN108399450A (en) Improvement particle cluster algorithm based on biological evolution principle
CN110348571A (en) A kind of neural network model training method, device, chip and system
CN106779372A (en) Based on the agricultural machinery dispatching method for improving immune Tabu search algorithm
Gajpal et al. An ant colony algorithm for scheduling in flowshops with sequence-dependent setup times of jobs
CN110428046A (en) Acquisition methods and device, the storage medium of neural network structure
Alanazi et al. Runtime analysis of selection hyper-heuristics with classical learning mechanisms
CN110599068A (en) Cloud resource scheduling method based on particle swarm optimization algorithm
CN107229966A (en) A kind of model data update method, apparatus and system
CN113220356A (en) User computing task unloading method in mobile edge computing
US20210124860A1 (en) High-throughput computational material simulation optimisation method and apparatus based on time prediction
CN105678401A (en) Global optimization method based on strategy adaptability differential evolution
Agrawal et al. Acceleration based particle swarm optimization for graph coloring problem
CN106919504B (en) Test data evolution generation method based on GA algorithm
CN106251031A (en) A kind of improved Particle Swarm Optimization inspired based on biology
CN105589896B (en) Data digging method and device
CN110378464A (en) The management method and device of the configuration parameter of artificial intelligence platform
Yan et al. A fast evolutionary algorithm for combinatorial optimization problems
Katayama et al. Iterated k-opt local search for the maximum clique problem
CN113220311A (en) Mobile-aware cloud-edge-side collaborative application unloading method and system and storage medium thereof
CN106156366A (en) A kind of pinning control node selecting method based on cluster

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
CB02 Change of applicant information
CB02 Change of applicant information

Address after: Room 1101-1117, floor 11, No. 58, Beisihuan West Road, Haidian District, Beijing 100080

Applicant after: BEIJING SENSETIME TECHNOLOGY DEVELOPMENT Co.,Ltd.

Address before: 100084, room 7, floor 3, building 1, No. 710-712, Zhongguancun East Road, Beijing, Haidian District

Applicant before: BEIJING SENSETIME TECHNOLOGY DEVELOPMENT Co.,Ltd.

GR01 Patent grant
GR01 Patent grant