CN111860630A - Model establishing method and system based on feature importance - Google Patents

Model establishing method and system based on feature importance Download PDF

Info

Publication number
CN111860630A
CN111860630A CN202010661710.4A CN202010661710A CN111860630A CN 111860630 A CN111860630 A CN 111860630A CN 202010661710 A CN202010661710 A CN 202010661710A CN 111860630 A CN111860630 A CN 111860630A
Authority
CN
China
Prior art keywords
feature
importance
model
feature importance
sub
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202010661710.4A
Other languages
Chinese (zh)
Other versions
CN111860630B (en
Inventor
林建明
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Wuyu Technology Co ltd
Original Assignee
Shenzhen Wuyu Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen Wuyu Technology Co ltd filed Critical Shenzhen Wuyu Technology Co ltd
Priority to CN202010661710.4A priority Critical patent/CN111860630B/en
Publication of CN111860630A publication Critical patent/CN111860630A/en
Application granted granted Critical
Publication of CN111860630B publication Critical patent/CN111860630B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting

Landscapes

  • Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The invention discloses a model building method and a system based on feature importance, wherein the model building method comprises the following steps: step S1, initializing feature data; step S2, sampling the feature data to form a plurality of groups of feature data combinations, wherein each group of feature data combination is used as a sub-model; s3, setting model parameters of each group of characteristic models; each group of feature combinations is provided with the same model parameter range; s4, training a model, and calculating the importance of each sub-model feature; step S5, calculating the importance of the comprehensive weighting characteristics by using all the sub models; s6, sorting the feature importance to obtain a new importance sorting; and step S7, modeling again according to the characteristic sequence. The model establishing method and system based on the feature importance can reduce the relative sorting fluctuation of the feature importance obtained by calculation and improve the reliability.

Description

Model establishing method and system based on feature importance
Technical Field
The invention belongs to the technical field of big data processing, relates to a model building method, and particularly relates to a model building method and system based on feature importance.
Background
In the field of big data wind control modeling, the commonly used data sources are various, and due to the fact that the data obtaining rate is different and the access time is sequential, the quality of usable data of the wind control modeling is low finally. In the face of the data situation, the traditional algorithm has limited performance, a large amount of data preprocessing work is needed, and the risk control requirement of the business cannot be met, so that the selection of a model with higher performance is a primary target, XGBOST is one of the commonly used algorithms, and the content of the patent is based on XGBOST algorithm expansion.
The XGB OST algorithm adopts an integrated learning scheme to construct trees, N trees are finally learned, the characteristic importance of the N trees is the proportion of times appearing on tree nodes, parameters such as maximum tree depth, node splitting conditions and the like are limited due to the fact that other factors such as overfitting and stability of a test set of a model need to be considered in the tree construction process, and finally the characteristics used by the whole forest are limited, so that most of characteristic importance is zero in a large number of molded characteristic variables, the condition that most of characteristic importance is zero is not beneficial to subsequent characteristic screening work, and the characteristic importance is zero and does not represent the characteristics.
In view of the above, there is an urgent need to design a new model building method to overcome at least some of the above-mentioned disadvantages of the existing model building methods.
Disclosure of Invention
The invention provides a model establishing method and system based on feature importance, which can reduce the relative sorting fluctuation of the feature importance obtained by calculation and improve the reliability.
In order to solve the technical problem, according to one aspect of the present invention, the following technical solutions are adopted:
a model building method based on feature importance comprises the following steps:
step S1, initializing feature data;
step S2, sampling the feature data to form a plurality of groups of feature data combinations, wherein each group of feature data combination is used as a sub-model;
s3, setting model parameters of each group of characteristic models; each group of feature combinations is provided with the same model parameter range;
s4, training a model, and calculating the importance of each sub-model feature;
step S5, calculating the importance of the comprehensive weighting characteristics by using all the sub models;
s6, sorting the feature importance to obtain a new importance sorting;
and step S7, modeling again according to the characteristic sequence.
As an embodiment of the present invention, the step S5 includes:
step S51, traversing each sub-model;
step S52, modifying the feature importance calculation mode of a single sub-model: test set ks of the submodeli+ksiImportance of this feature fimpi
Step S53, aggregating the feature importance of all sub-models, and solving the mean value of the new feature importance of each feature to obtain the comprehensive weighted feature importance;
Figure BDA0002578791920000021
where m represents the number of models, ks, that are trained in totaliKs value, fimp representing the ith modeliRepresenting the feature importance of all the modelled features of the ith model.
In one embodiment of the present invention, in step S2, the feature data is randomly sampled without being put back.
In one embodiment of the present invention, in step S2, the replacement equal percentage sampling is performed on the feature data.
In one embodiment of the present invention, in step S2, the random non-recurrent combined features are used, and the non-recurrent sampling equal ratio features are sampled for a plurality of times, so as to ensure that each feature can participate in the model training with the same weight, i.e., the number of models participating in the training for each feature is the same.
As an embodiment of the present invention, in the step S2, a random combination method is introduced, and a final model test set ks and sub-model feature importance weighting and averaging method is combined, so that the obtained feature importance is more representative, and a smoother quantitative evaluation index is provided for the feature, which provides an excellent basis for subsequent modeling feature screening.
In an embodiment of the invention, in step S4, the XGBOOST model is used as the training model.
In an embodiment of the present invention, in step S4, each group of models trains multiple groups of parameters, resulting in multiple groups of basic XGBOOST models.
In an embodiment of the present invention, in step S4, the obtained feature importance is combined with the feature IV value to form an evaluation index again, and the XGBOOST tree evaluation method and the linear IV evaluation method are compatible. Setting a weight combination scheme with different feature importance for iv and the feature, and deriving additional feature evaluation indexes again; given the iv weight a and the feature importance weight b, the derived new feature evaluation index is f _ index _ new ═ iv a + f _ imp _ new ×.b.
As an embodiment of the present invention, the step S6 further includes: and modeling in an increasing, decreasing or step wise mode according to the obtained characteristic sequence rows.
According to another aspect of the invention, the following technical scheme is adopted: a feature importance-based modeling system, the modeling system comprising:
the characteristic data initialization module is used for initializing the characteristic data;
the characteristic data sampling module is used for sampling the characteristic data to form a plurality of groups of characteristic data combinations, and each group of characteristic data combination is used as a sub-model;
The model parameter setting module is used for setting model parameters of each group of characteristic models; each group of feature combinations is provided with the same model parameter range;
the model training module is used for training the model and calculating the importance of the characteristics of each sub-model;
the importance calculating module is used for calculating the importance of the comprehensive weighting characteristics by using all the sub models;
the importance ranking module is used for ranking the feature importance to obtain a new importance ranking; and
and the training model modeling module is used for setting the model parameter range again by taking the characteristic subset with the top rank according to the characteristic sequence and retraining the model modeling.
As an embodiment of the present invention, the importance calculating module includes:
the submodel traversing unit is used for traversing each submodel;
the feature importance calculating unit is used for modifying the feature importance calculating mode of a single sub-model as follows: test set ks of the submodeli+ksiImportance of this feature fimpi(ii) a And
the feature importance aggregation unit is used for aggregating the feature importance of all the submodels and solving the mean value of the new feature importance of each feature to obtain the comprehensive weighted feature importance;
Figure BDA0002578791920000031
where m represents the number of models, ks, that are trained in total iKs value, fimp representing the ith modeliRepresenting the feature importance of all the modelled features of the ith model.
The invention integrates the weighted feature importance core calculation thought: (1) random no-put-back combination features: sampling is carried out for multiple times without the returned sampling equal ratio features, and each feature can participate in the model training with the same weight (namely the number of models participating in the training of each feature is consistent). (2) And (3) training multiple groups of parameters of each group of models to obtain multiple groups of basic XGBOOST models, and calculating the importance of the comprehensive characteristics of the multiple models.
The invention has the overall improved effect that: (1) the random non-playback combination sampling mode disorganizes and combines the features, the more the combination is, the more easily the good feature combination and the poor feature combination appear, the larger the difference of the corresponding model effect is, namely, the orderliness of the model effect is related to the use feature thereof. (2) The native XGBOST feature importance may fluctuate due to different model parameters, and is not consistent every time, but the feature importance calculated by the scheme of the invention has small fluctuation in relative sequencing and high reliability. (3) After the importance of the features of the multiple groups of models is calculated through ks weighting, the importance of each feature is not zero, namely the problem that the importance of a large number of features of a single XGBOOST model is zero is solved. (4) Compared with the primary feature importance of the XGB OST, the calculated comprehensive feature importance and the numerical value are smoother.
The invention has the beneficial effects that: the model establishing method based on the feature importance can reduce the relative sorting fluctuation of the feature importance obtained by calculation and improve the reliability.
Drawings
Fig. 1 is a flowchart of a feature importance-based model building method according to an embodiment of the present invention.
Fig. 2 is a schematic composition diagram of a model building system based on feature importance according to an embodiment of the present invention.
FIG. 3 is a schematic diagram illustrating an importance calculation module of the model building system according to an embodiment of the present invention.
Detailed Description
Preferred embodiments of the present invention will be described in detail below with reference to the accompanying drawings.
For a further understanding of the invention, reference will now be made to the preferred embodiments of the invention by way of example, and it is to be understood that the description is intended to further illustrate features and advantages of the invention, and not to limit the scope of the claims.
The description in this section is for several exemplary embodiments only, and the present invention is not limited only to the scope of the embodiments described. It is within the scope of the present disclosure and protection that the same or similar prior art means and some features of the embodiments may be interchanged.
The invention discloses a model building method based on feature importance, and FIG. 1 is a flow chart of the model building method based on feature importance in an embodiment of the invention; referring to fig. 1, the model building method includes:
Step S1 initializes the feature data.
Step S2, the feature data are sampled to form a plurality of sets of feature data combinations, and each set of feature data combination is used as a sub-model.
In one embodiment of the invention, the feature data is randomly sampled without playback. And sampling for multiple times by adopting random non-playback combination characteristics and non-playback sampling equal ratio characteristics to ensure that each characteristic can participate in model training and the participating weights are the same, namely the number of models participating in training of each characteristic is consistent. By introducing a random combination method and combining a final model test set ks and a sub-model feature importance weighting and averaging mode, the obtained feature importance is more representative, and meanwhile, a smoother quantitative evaluation index is provided for the features, so that an excellent basis is provided for subsequent modeling feature screening.
In another embodiment of the present invention, the replacement equal percentage sampling is performed on the feature data.
Step S3, setting model parameters of each group of feature models; each set of feature combinations sets the same model parameter ranges.
Step S4, the model is trained, and the importance of each sub-model feature is calculated.
In an embodiment of the invention, the XGBOST model is adopted as the training model. In one embodiment, each model set trains multiple sets of parameters, resulting in multiple sets of basic XGBOOST models.
In an embodiment of the invention, the feature importance is obtained based on the XGBOOST scheme, the core of the evaluation mostly depends on the tree node segmentation calculation mode, the obtained feature importance is combined with the feature IV value, an evaluation index can be formed again, and the evaluation mode of the XGBOOST tree and the evaluation mode of the linear IV are considered. Setting a weight combination scheme with different feature importance for iv and the feature, and deriving additional feature evaluation indexes again; given the iv weight a and the feature importance weight b, the derived new feature evaluation index is f _ index _ new ═ iv a + f _ imp _ new ×.b. For example, given an iv weight of 50% and a feature importance weight of 50%, the derived new feature evaluation index is f _ index _ new ═ iv 0.5+ f _ imp _ new ×. 0.5.
Step S5, the importance of the integrated weighted feature is calculated using all sub models.
In an embodiment of the present invention, the step S5 includes:
step S51, traversing each sub-model;
step S52, modifying the feature importance of the individual submodelsThe calculation method is as follows: test set ks of the submodel i+ksiImportance of this feature fimpi
Step S53, aggregating the feature importance of all sub-models, and solving the mean value of the new feature importance of each feature to obtain the comprehensive weighted feature importance;
Figure BDA0002578791920000051
where m represents the number of models, ks, that are trained in totaliKs value, fimp representing the ith modeliRepresenting the feature importance of all the modelled features of the ith model.
In one embodiment of the invention, ks is used in the calculation mode of the weighted importance of the comprehensive characteristics; in another embodiment of the present invention, other models such as AUC and LIFT may be used.
Step S6, the feature importance is ranked to obtain a new importance ranking.
In an embodiment of the present invention, the step S6 further includes: modeling in an increasing, decreasing or step wise mode according to the obtained characteristic sequencing rows; the stability and the upper performance limit of the model are improved, and the use characteristics are reduced, so that the deployment difficulty of the online model is greatly reduced.
Step S7, modeling the features again according to the feature sequence;
step S8 ends.
The invention also discloses a model building system based on the feature importance, and FIG. 2 is a schematic composition diagram of the model building system based on the feature importance in an embodiment of the invention; referring to fig. 2, in an embodiment of the present invention, the model building system includes: the system comprises a characteristic data initialization module 1, a characteristic data sampling module 2, a model parameter setting module 3, a model training module 4, an importance calculating module 5, an importance sequencing module 6 and a training model modeling module 7.
The characteristic data initialization module 1 is used for initializing characteristic data; the characteristic data sampling module 2 is used for sampling the characteristic data to form a plurality of groups of characteristic data combinations, and each group of characteristic data combination is used as a sub-model; the model parameter setting module 3 is used for setting model parameters of each group of feature models; each group of feature combinations is provided with the same model parameter range; the model training module 4 is used for training the model and calculating the importance of the characteristics of each sub-model; the importance calculating module 5 is used for calculating the importance of the comprehensive weighting characteristic by using all the sub models; the importance ranking module 6 is used for ranking the feature importance to obtain a new importance ranking; the training model modeling module 7 is used for setting the model parameter range again by taking the characteristic subset with the top rank according to the characteristic sequence and retraining the model modeling.
FIG. 3 is a schematic diagram illustrating an importance calculation module of the model building system according to an embodiment of the present invention; referring to fig. 3, in an embodiment of the present invention, the importance calculating module 5 includes: a submodel traversing unit 51, a feature importance calculating unit 52, and a feature importance aggregating unit 53.
The sub-model traversal unit 51 is used to traverse the sub-models. The feature importance calculating means 52 is used to modify the feature importance calculation of the individual submodels in the following manner: test set ks of the submodel i+ksiImportance of this feature fimpi. The feature importance aggregating unit 53 is configured to aggregate feature importance of all sub-models, and calculate an average of new feature importance of each feature, i.e., obtain an integrated weighted feature importance;
Figure BDA0002578791920000061
where m represents the number of models, ks, that are trained in totaliKs value, fimp representing the ith modeliRepresenting the feature importance of all the modelled features of the ith model.
In summary, the feature importance-based model building method and system provided by the invention can reduce the relative sorting fluctuation of the calculated feature importance and improve the reliability.
The technical features of the embodiments described above may be arbitrarily combined, and for the sake of brevity, all possible combinations of the technical features in the embodiments described above are not described, but should be considered as being within the scope of the present specification as long as there is no contradiction between the combinations of the technical features.
The description and applications of the invention herein are illustrative and are not intended to limit the scope of the invention to the embodiments described above. Effects or advantages referred to in the embodiments may not be reflected in the embodiments due to interference of various factors, and the description of the effects or advantages is not intended to limit the embodiments. Variations and modifications of the embodiments disclosed herein are possible, and alternative and equivalent various components of the embodiments will be apparent to those skilled in the art. It will be clear to those skilled in the art that the present invention may be embodied in other forms, structures, arrangements, proportions, and with other components, materials, and parts, without departing from the spirit or essential characteristics thereof. Other variations and modifications of the embodiments disclosed herein may be made without departing from the scope and spirit of the invention.

Claims (10)

1. A model building method based on feature importance is characterized by comprising the following steps:
step S1, initializing feature data;
step S2, sampling the feature data to form a plurality of groups of feature data combinations, wherein each group of feature data combination is used as a sub-model;
s3, setting model parameters of each group of characteristic models; each group of feature combinations is provided with the same model parameter range;
s4, training a model, and calculating the importance of each sub-model feature;
step S5, calculating the importance of the comprehensive weighting characteristics by using all the sub models;
s6, sorting the feature importance to obtain a new importance sorting;
and step S7, setting the parameter range of the model again by taking the characteristic subset with the top rank according to the characteristic sequence, and retraining the model for modeling.
2. The feature importance-based model building method according to claim 1, wherein:
the step S5 includes:
step S51, traversing each sub-model;
step S52, modifying the feature importance calculation mode of a single sub-model: test set ks of the submodeli+ksiImportance of this feature fimpi
Step S53, aggregating the feature importance of all sub-models, and solving the mean value of the new feature importance of each feature to obtain the comprehensive weighted feature importance;
Figure FDA0002578791910000011
Where m represents the number of models, ks, that are trained in totaliKs value, fimp representing the ith modeliRepresenting the feature importance of all the modelled features of the ith model.
3. The feature importance-based model building method according to claim 1, wherein:
in step S2, performing random non-playback sampling on the feature data; alternatively, the feature data is sampled with equal percentage put back.
4. The feature importance-based model building method according to claim 1, wherein:
in step S2, the random non-playback combination features are used, and sampling is performed for multiple times through the non-playback sampling equal ratio features, so that each feature can participate in the model training, and the participating weights are the same, that is, the number of models that each feature participates in the training is the same.
5. The feature importance-based model building method according to claim 1, wherein:
in the step S2, by introducing a random combination method and combining a final model test set ks and a sub-model feature importance weighting and averaging manner, the obtained feature importance is more representative, and a smoother quantitative evaluation index is provided for the feature, which provides an excellent basis for subsequent modeling feature screening.
6. The feature importance-based model building method according to claim 1, wherein:
in the step S4, the training model adopts an XGBOOST model; and training multiple groups of parameters by each group of models to obtain multiple groups of basic XGBOST models.
7. The feature importance-based model building method according to claim 1, wherein:
in the step S4, the obtained feature importance and the feature IV value are combined together to form an evaluation index again, and the evaluation mode of the XGBOOST tree and the evaluation mode of the linear IV are considered;
setting weight combination schemes with different IV and feature importance, and deriving additional feature evaluation indexes again; given the IV weight a and the feature importance weight b, the derived new feature evaluation index is f _ index _ new ═ IV × a + f _ imp _ new × b.
8. The feature importance-based model building method according to claim 1, wherein:
the step S6 further includes: and modeling in an increasing, decreasing or step wise mode according to the obtained characteristic sequence rows.
9. A model building system based on feature importance, the model building system comprising:
the characteristic data initialization module is used for initializing the characteristic data;
The characteristic data sampling module is used for sampling the characteristic data to form a plurality of groups of characteristic data combinations, and each group of characteristic data combination is used as a sub-model;
the model parameter setting module is used for setting model parameters of each group of characteristic models; each group of feature combinations is provided with the same model parameter range;
the model training module is used for training the model and calculating the importance of the characteristics of each sub-model;
the importance calculating module is used for calculating the importance of the comprehensive weighting characteristics by using all the sub models;
the importance ranking module is used for ranking the feature importance to obtain a new importance ranking; and
and the training model modeling module is used for setting the model parameter range again by taking the characteristic subset with the top rank according to the characteristic sequence and retraining the model modeling.
10. The feature importance based model building system according to claim 9, wherein:
the importance calculation module comprises:
the submodel traversing unit is used for traversing each submodel;
the feature importance calculating unit is used for modifying the feature importance calculating mode of a single sub-model as follows: test set ks of the submodeli+ksiImportance of this feature fimp i(ii) a And
the feature importance aggregation unit is used for aggregating the feature importance of all the submodels and solving the mean value of the new feature importance of each feature to obtain the comprehensive weighted feature importance;
Figure FDA0002578791910000031
where m represents the number of models, ks, that are trained in totaliKs value, fimp representing the ith modeliRepresenting the feature importance of all the modelled features of the ith model.
CN202010661710.4A 2020-07-10 2020-07-10 Model building method and system based on feature importance Active CN111860630B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010661710.4A CN111860630B (en) 2020-07-10 2020-07-10 Model building method and system based on feature importance

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010661710.4A CN111860630B (en) 2020-07-10 2020-07-10 Model building method and system based on feature importance

Publications (2)

Publication Number Publication Date
CN111860630A true CN111860630A (en) 2020-10-30
CN111860630B CN111860630B (en) 2023-10-13

Family

ID=73153137

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010661710.4A Active CN111860630B (en) 2020-07-10 2020-07-10 Model building method and system based on feature importance

Country Status (1)

Country Link
CN (1) CN111860630B (en)

Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE102013225768A1 (en) * 2013-12-12 2015-06-18 Robert Bosch Gmbh Method and apparatus for determining a LOLIMOT model
CN105589683A (en) * 2014-10-22 2016-05-18 腾讯科技(深圳)有限公司 Sample extraction method and apparatus
CN107316082A (en) * 2017-06-15 2017-11-03 第四范式(北京)技术有限公司 For the method and system for the feature importance for determining machine learning sample
CN107730154A (en) * 2017-11-23 2018-02-23 安趣盈(上海)投资咨询有限公司 Based on the parallel air control application method of more machine learning models and system
WO2018145596A1 (en) * 2017-02-13 2018-08-16 腾讯科技(深圳)有限公司 Method and device for extracting feature information, server cluster, and storage medium
CN108764597A (en) * 2018-04-02 2018-11-06 华南理工大学 A kind of product quality control method based on integrated study
CN109035003A (en) * 2018-07-04 2018-12-18 北京玖富普惠信息技术有限公司 Anti- fraud model modelling approach and anti-fraud monitoring method based on machine learning
CN109460825A (en) * 2018-10-24 2019-03-12 阿里巴巴集团控股有限公司 For constructing the Feature Selection Algorithms, device and equipment of machine learning model
CN110334773A (en) * 2019-07-12 2019-10-15 四川新网银行股份有限公司 Model based on machine learning enters the screening technique of modular character
CN110908908A (en) * 2019-11-21 2020-03-24 深圳无域科技技术有限公司 Method and device for testing micro-service Dubbo interface
CN110991474A (en) * 2019-10-12 2020-04-10 未鲲(上海)科技服务有限公司 Machine learning modeling platform
US20220012639A1 (en) * 2020-07-08 2022-01-13 Vmware, Inc. Quantizing training data sets using ml model metadata

Patent Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE102013225768A1 (en) * 2013-12-12 2015-06-18 Robert Bosch Gmbh Method and apparatus for determining a LOLIMOT model
CN105589683A (en) * 2014-10-22 2016-05-18 腾讯科技(深圳)有限公司 Sample extraction method and apparatus
WO2018145596A1 (en) * 2017-02-13 2018-08-16 腾讯科技(深圳)有限公司 Method and device for extracting feature information, server cluster, and storage medium
CN107316082A (en) * 2017-06-15 2017-11-03 第四范式(北京)技术有限公司 For the method and system for the feature importance for determining machine learning sample
CN107730154A (en) * 2017-11-23 2018-02-23 安趣盈(上海)投资咨询有限公司 Based on the parallel air control application method of more machine learning models and system
CN108764597A (en) * 2018-04-02 2018-11-06 华南理工大学 A kind of product quality control method based on integrated study
CN109035003A (en) * 2018-07-04 2018-12-18 北京玖富普惠信息技术有限公司 Anti- fraud model modelling approach and anti-fraud monitoring method based on machine learning
CN109460825A (en) * 2018-10-24 2019-03-12 阿里巴巴集团控股有限公司 For constructing the Feature Selection Algorithms, device and equipment of machine learning model
CN110334773A (en) * 2019-07-12 2019-10-15 四川新网银行股份有限公司 Model based on machine learning enters the screening technique of modular character
CN110991474A (en) * 2019-10-12 2020-04-10 未鲲(上海)科技服务有限公司 Machine learning modeling platform
CN110908908A (en) * 2019-11-21 2020-03-24 深圳无域科技技术有限公司 Method and device for testing micro-service Dubbo interface
US20220012639A1 (en) * 2020-07-08 2022-01-13 Vmware, Inc. Quantizing training data sets using ml model metadata

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
崔晓旭: "基于数据挖掘的疾病预测组合模型研究", 《中国优秀硕士学位论文全文数据库》, no. 12, pages 062 - 154 *

Also Published As

Publication number Publication date
CN111860630B (en) 2023-10-13

Similar Documents

Publication Publication Date Title
CN107633301A (en) The training method of testing and its application system of a kind of BP neural network regression model
CN107330902A (en) Chaos-Genetic BP neural network image partition method based on Arnold conversion
CN108805268A (en) Deeply learning strategy network training method based on evolution algorithm
CN109816150A (en) One kind adopting rear Table Grape forecasting shelf life method and device
CN109670655B (en) Multi-target particle swarm optimization scheduling method for electric power system
CN113722980A (en) Ocean wave height prediction method, system, computer equipment, storage medium and terminal
CN113240263A (en) Comprehensive energy system planning combination evaluation method based on entropy weight fuzzy
CN111461284A (en) Data discretization method, device, equipment and medium
CN110110447B (en) Method for predicting thickness of strip steel of mixed frog leaping feedback extreme learning machine
CN108197280A (en) Mining ability evaluation method based on industrial equipment data
CN115775026A (en) Federated learning method based on organization similarity
CN105426959B (en) Aluminium electroloysis energy-saving and emission-reduction method based on BP neural network Yu adaptive M BFO algorithms
CN113705098A (en) Air duct heater modeling method based on PCA and GA-BP network
CN111860630A (en) Model establishing method and system based on feature importance
CN108256623A (en) Particle swarm optimization on multiple populations based on period interaction mechanism and knowledge plate synergistic mechanism
CN112132259B (en) Neural network model input parameter dimension reduction method and computer readable storage medium
CN115936317A (en) Suspension cable-stayed cooperative system scheme evaluation method and system
CN116014764A (en) Distributed energy storage optimization processing method and device
CN115438842A (en) Load prediction method based on adaptive improved dayflies and BP neural network
CN107590540A (en) A kind of forest hat width evaluation method dependent on neighboring trees feature
CN114781244A (en) Grouping and parameter optimization method in wind power plant
CN111553398B (en) Wind power scene uncertain continuous interval obtaining method based on multidimensional normal distribution
CN112817959A (en) Construction method of ancient biomorphic phylogenetic tree based on multi-metric index weight
CN113255883A (en) Weight initialization method based on power law distribution
Lou An intelligent teaching test paper generation system based on ant colony hybrid genetic algorithms

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant