CN111860630B - Model building method and system based on feature importance - Google Patents

Model building method and system based on feature importance Download PDF

Info

Publication number
CN111860630B
CN111860630B CN202010661710.4A CN202010661710A CN111860630B CN 111860630 B CN111860630 B CN 111860630B CN 202010661710 A CN202010661710 A CN 202010661710A CN 111860630 B CN111860630 B CN 111860630B
Authority
CN
China
Prior art keywords
feature
model
importance
models
sub
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202010661710.4A
Other languages
Chinese (zh)
Other versions
CN111860630A (en
Inventor
林建明
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Wuyu Technology Co ltd
Original Assignee
Shenzhen Wuyu Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen Wuyu Technology Co ltd filed Critical Shenzhen Wuyu Technology Co ltd
Priority to CN202010661710.4A priority Critical patent/CN111860630B/en
Publication of CN111860630A publication Critical patent/CN111860630A/en
Application granted granted Critical
Publication of CN111860630B publication Critical patent/CN111860630B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting

Landscapes

  • Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The invention discloses a model building method and a system based on feature importance, wherein the model building method comprises the following steps: step S1, initializing characteristic data; step S2, sampling the characteristic data to form a plurality of groups of characteristic data combinations, wherein each group of characteristic data combination is used as a sub-model; s3, setting model parameters of each group of feature models; each group of feature combinations is provided with the same model parameter range; s4, training a model, and calculating the importance of each sub-model feature; s5, calculating importance of the comprehensive weighting characteristics by using all sub-models; s6, sorting the feature importance to obtain a new importance sorting; and S7, modeling again according to the characteristic sequence. The model establishing method and system based on the feature importance can reduce the relative sorting fluctuation of the feature importance obtained by calculation and improve the credibility.

Description

Model building method and system based on feature importance
Technical Field
The invention belongs to the technical field of big data processing, relates to a model building method, and particularly relates to a model building method and system based on feature importance.
Background
In the field of big data wind control modeling, commonly used data sources are numerous, and due to the fact that the data searching yields are different and the access time is sequential, the quality of the data which can be used for wind control modeling is low. In the face of the data situation, the traditional algorithm has limited performance, and needs to do a large amount of data preprocessing work, and cannot meet the control demands of business on risks, so that a model with higher performance is selected as a primary target, XGBOOST is one of the common algorithms, and the patent is based on XGBOOST algorithm development.
The XGBOOST algorithm adopts an integrated learning scheme to construct the tree, finally learns N trees, and is characterized in that the number of times of the N trees appearing on the nodes of the tree is proportional, parameters such as maximum tree depth, splitting conditions of the nodes and the like are required to be limited because other factors such as fitting, stability and the like of a test set of the model are required to be considered in the tree construction process, and finally the characteristics used by the whole tree forest are limited, so that most of characteristic variables are zero in the model entering characteristic variables, the condition that most of characteristic importance is zero is unfavorable for subsequent characteristic screening work, and the characteristic importance is zero and does not represent useless characteristics.
In view of this, there is an urgent need to design a new model building method to overcome at least some of the above-mentioned drawbacks of the existing model building methods.
Disclosure of Invention
The invention provides a model building method and a system based on feature importance, which can reduce the relative sorting fluctuation of the feature importance obtained by calculation and improve the credibility.
In order to solve the technical problems, according to one aspect of the present invention, the following technical scheme is adopted:
a model building method based on feature importance, the model building method comprising:
step S1, initializing characteristic data;
step S2, sampling the characteristic data to form a plurality of groups of characteristic data combinations, wherein each group of characteristic data combination is used as a sub-model;
s3, setting model parameters of each group of feature models; each group of feature combinations is provided with the same model parameter range;
s4, training a model, and calculating the importance of each sub-model feature;
s5, calculating importance of the comprehensive weighting characteristics by using all sub-models;
s6, sorting the feature importance to obtain a new importance sorting;
and S7, modeling again according to the characteristic sequence.
As an embodiment of the present invention, the step S5 includes:
step S51, traversing each sub-model;
step S52, modifying the feature importance calculation mode of the single submodel to be: test set ks of the submodel i +ks i * Importance fimp of the feature i
Step S53, feature importance of all sub-models is aggregated, and the average value of new feature importance of each feature is calculated, so that comprehensive weighted feature importance is obtained;wherein m represents the number of models trained overall, ks i Ks value, fimp, representing the ith model i Representing the feature importance of all the in-mold features of the ith model.
In one embodiment of the present invention, in the step S2, the feature data is randomly sampled without substitution.
In the step S2, the feature data is sampled with a percentage of the same value as the feature data.
In step S2, the random non-replacement combined feature is adopted, and multiple sampling is performed through the feature with the non-replacement sampling equal ratio, so that each feature can participate in model training, and the participating weights are the same, i.e. the number of models of each feature participating in training is the same.
In step S2, a random combination method is introduced, and a final model test set ks and a sub-model feature importance weighting and averaging method are combined, so that the obtained feature importance is more representative, and a smoother quantitative evaluation index is provided for the feature, which provides an excellent basis for subsequent modeling feature screening.
In step S4, the XGBOOST model is used as the training model.
In step S4, each set of models trains multiple sets of parameters to obtain multiple sets of basic XGBOOST models.
In the step S4, the obtained feature importance and the feature IV value are combined to form one evaluation index again, and the evaluation method of XGBOOST tree and the evaluation method of linear IV are compatible. Setting a weight combination scheme with iv and different feature importance, and deriving additional feature evaluation indexes again; given iv weight a, feature importance weight b, the new feature evaluation index after derivation is f_index_new=iv_a+f_imp_new_new.
As an embodiment of the present invention, the step S6 further includes: the progressive, decremental or step wise modeling is performed according to the resulting feature ordering lines.
According to another aspect of the invention, the following technical scheme is adopted: a feature importance based model building system, the model building system comprising:
the feature data initializing module is used for initializing feature data;
the characteristic data sampling module is used for sampling characteristic data to form a plurality of groups of characteristic data combinations, and each group of characteristic data combination is used as a sub-model;
the model parameter setting module is used for setting model parameters of each group of characteristic models; each group of feature combinations is provided with the same model parameter range;
the model training module is used for training a model and calculating the importance of each sub-model characteristic;
an importance calculation module for calculating the importance of the comprehensive weighted features using all sub-models;
the importance ranking module is used for ranking the feature importance to obtain a new importance ranking; and
and the training model modeling module is used for taking the feature subset which is ranked at the front according to the feature sequence to set the model parameter range again and retrain the model modeling again.
As one embodiment of the present invention, the importance calculating module includes:
the sub-model traversing unit is used for traversing each sub-model;
the feature importance calculating unit is used for modifying the feature importance calculating mode of the single submodel to be as follows: test set ks of the submodel i +ks i * Importance fimp of the feature i The method comprises the steps of carrying out a first treatment on the surface of the And
the feature importance aggregation unit is used for aggregating the feature importance of all the submodels and solving the mean value of the new feature importance of each feature to obtain the comprehensive weighted feature importance;wherein m represents the number of models trained overall, ks i Ks value, fimp, representing the ith model i Representing the feature importance of all the in-mold features of the ith model.
The invention synthesizes the core calculation thought of the weighted feature importance: (1) random non-put back combination feature: sampling is carried out for a plurality of times without replaced sampling equal ratio characteristics, so that each characteristic can participate in model training, and the participating weights are the same (namely, the quantity of models of each characteristic participating in training is consistent). (2) Each group of models trains a plurality of groups of parameters to obtain a plurality of groups of basic XGBOOST models, and the comprehensive feature importance of the plurality of models is calculated.
The invention has the overall improvement effect that: (1) The random sampling mode without the put-back combination is used for combining the characteristics in disorder, the more the combination is, the more good characteristic combination and poor characteristic combination are easy to appear, and the larger the corresponding model effect is, namely the ordering of the model effect is related to the use characteristics. (2) The original XGBOOST feature importance may have fluctuation due to different model parameters and is not consistent every time, but the feature importance calculated by the scheme of the invention has small fluctuation of relative ordering and higher reliability. (3) After the feature importance of the multiple groups of models is calculated through the ks weighting, the importance of each feature is not zero, namely the problem that a large number of features of a single XGBOOST model are zero is solved. (4) Compared with the original characteristic importance of XGBOOST, the calculated comprehensive characteristic importance is smoother in value.
The invention has the beneficial effects that: the model establishment method based on the feature importance can reduce the relative sequencing fluctuation of the feature importance obtained by calculation and improve the credibility.
Drawings
FIG. 1 is a flow chart of a method for modeling based on feature importance in an embodiment of the invention.
FIG. 2 is a schematic diagram of a feature importance based modeling system according to an embodiment of the present invention.
FIG. 3 is a schematic diagram showing the components of a model building system importance calculating module according to an embodiment of the present invention.
Detailed Description
Preferred embodiments of the present invention will be described in detail below with reference to the accompanying drawings.
For a further understanding of the present invention, preferred embodiments of the invention are described below in conjunction with the examples, but it should be understood that these descriptions are merely intended to illustrate further features and advantages of the invention, and are not limiting of the claims of the invention.
The description of this section is intended to be illustrative of only a few exemplary embodiments and the invention is not to be limited in scope by the description of the embodiments. It is also within the scope of the description and claims of the invention to interchange some of the technical features of the embodiments with other technical features of the same or similar prior art.
The invention discloses a model building method based on feature importance, and FIG. 1 is a flow chart of the model building method based on feature importance in an embodiment of the invention; referring to fig. 1, the method for establishing the model includes:
initializing characteristic data (step S1).
And (S2) sampling the characteristic data to form a plurality of groups of characteristic data combinations, wherein each group of characteristic data combination is used as a submodel.
In one embodiment of the invention, the feature data is randomly non-subsampled. The random non-replacement combined features are adopted, and sampling is carried out for multiple times through the features of non-replacement sampling equal ratio, so that each feature can participate in model training, the participating weights are the same, and the number of models of each feature participating in training is consistent. By introducing a random combination method and combining a final model test set ks and sub-model feature importance weighting average method, the obtained feature importance is more representative, and a smoother quantitative evaluation index is provided for the feature, so that an excellent basis is provided for subsequent modeling feature screening.
In another embodiment of the invention, the feature data is sampled with a put back equal percentage.
Setting model parameters of each group of feature models; the same model parameter range is set for each set of feature combinations.
Training the model and calculating the importance of each sub-model feature (step S4).
In one embodiment of the invention, the training model employs the XGBOOST model. In one embodiment, each set of models trains multiple sets of parameters, resulting in multiple sets of basic XGBOOST models.
In an embodiment of the present invention, the feature importance is obtained based on XGBOOST scheme, the core of the evaluation is mostly dependent on the segmentation calculation mode of the tree node, the obtained feature importance and the feature IV value are combined together, an evaluation index can be formed again, and the evaluation mode of XGBOOST tree and the evaluation mode of linear IV are considered. Setting a weight combination scheme with iv and different feature importance, and deriving additional feature evaluation indexes again; given iv weight a, feature importance weight b, the new feature evaluation index after derivation is f_index_new=iv_a+f_imp_new_new. For example, given that iv weights are 50%, feature importance weights are 50%, the new feature evaluation index after derivation is f_index_new=iv×0.5+f_imp_new×0.5.
Step S5 the importance of the comprehensive weighting characteristics is calculated using all sub-models.
In an embodiment of the present invention, the step S5 includes:
step S51, traversing each sub-model;
step S52, modifying the feature importance calculation mode of the single submodel to be: test set ks of the submodel i +ks i * Importance fimp of the feature i
Step S53, feature importance of all sub-models is aggregated, and the average value of new feature importance of each feature is calculated, so that comprehensive weighted feature importance is obtained;wherein m represents the number of models trained overall, ks i Ks value, fimp, representing the ith model i Representing the feature importance of all the in-mold features of the ith model.
In one embodiment of the present invention, ks is used as the integrated feature weighted importance calculation; in another embodiment of the present invention, evaluation indexes of other models such as AUC, LIFT, etc. may be used.
And (6) sorting the feature importance to obtain a new importance sort.
In an embodiment of the present invention, the step S6 further includes: modeling in an incremental, decremental or step wise manner according to the obtained feature ordering row; the stability and the performance upper limit of the model are improved, and meanwhile, the used characteristics are reduced, so that the deployment difficulty of the online model is greatly reduced.
Step S7, performing feature re-modeling according to the feature sequence;
the step S8 ends.
The invention also discloses a model building system based on the feature importance, and FIG. 2 is a schematic diagram of the model building system based on the feature importance in an embodiment of the invention; referring to fig. 2, in an embodiment of the present invention, the modeling system includes: the system comprises a feature data initialization module 1, a feature data sampling module 2, a model parameter setting module 3, a model training module 4, an importance calculating module 5, an importance sorting module 6 and a training model modeling module 7.
The characteristic data initializing module 1 is used for initializing characteristic data; the feature data sampling module 2 is used for sampling feature data to form a plurality of groups of feature data combinations, and each group of feature data combinations is used as a sub-model; the model parameter setting module 3 is used for setting the model parameters of each group of feature models; each group of feature combinations is provided with the same model parameter range; the model training module 4 is used for training a model and calculating the importance of each sub-model feature; the importance calculating module 5 is used for calculating the importance of the comprehensive weighting characteristics by using all sub-models; the importance ranking module 6 is used for ranking the feature importance to obtain a new importance ranking; the training model modeling module 7 is configured to take the feature subset with the top rank according to the feature order to set the model parameter range again, and retrain the model modeling.
FIG. 3 is a schematic diagram showing the components of a model building system importance calculating module according to an embodiment of the present invention; referring to fig. 3, in an embodiment of the present invention, the importance calculating module 5 includes: a sub-model traversing unit 51, a feature importance calculating unit 52, and a feature importance aggregating unit 53.
The sub-model traversing unit 51 is used for traversing each sub-model. The feature importance calculating unit 52 is configured to modify the feature importance calculating manner of the single submodel as follows: test set ks of the submodel i +ks i * Importance fimp of the feature i . The feature importance aggregation unit 53 is configured to aggregate feature importance of all sub-models, and calculate a mean value of new feature importance of each feature, thereby obtaining a comprehensive weighted feature importance;wherein m represents the number of models trained overall, ks i Ks value, fimp, representing the ith model i Representing the feature importance of all the in-mold features of the ith model.
In summary, the model building method and system based on the feature importance provided by the invention can reduce the relative sorting fluctuation of the feature importance obtained by calculation and improve the reliability.
The technical features of the above-described embodiments may be arbitrarily combined, and all possible combinations of the technical features in the above-described embodiments are not described for brevity of description, however, as long as there is no contradiction between the combinations of the technical features, they should be considered as the scope of the description.
The description and applications of the present invention herein are illustrative and are not intended to limit the scope of the invention to the embodiments described above. Effects or advantages referred to in the embodiments may not be embodied in the embodiments due to interference of various factors, and description of the effects or advantages is not intended to limit the embodiments. Variations and modifications of the embodiments disclosed herein are possible, and alternatives and equivalents of the various components of the embodiments are known to those of ordinary skill in the art. It will be clear to those skilled in the art that the present invention may be embodied in other forms, structures, arrangements, proportions, and with other assemblies, materials, and components, without departing from the spirit or essential characteristics thereof. Other variations and modifications of the embodiments disclosed herein may be made without departing from the scope and spirit of the invention.

Claims (6)

1. The model building method based on the feature importance is characterized by being used in the field of big data wind control modeling, and comprises the following steps of:
step S1, initializing characteristic data;
step S2, sampling the characteristic data to form a plurality of groups of characteristic data combinations, wherein each group of characteristic data combination is used as a sub-model;
s3, setting model parameters of each group of feature models; each group of feature combinations is provided with the same model parameter range;
s4, training a model, and calculating the importance of each sub-model feature;
s5, calculating importance of the comprehensive weighting characteristics by using all sub-models;
s6, sorting the feature importance to obtain a new importance sorting;
s7, according to the feature sequence, the feature subset which is ranked at the front is taken to set the parameter range of the model again, and model modeling is retrained;
in the step S2, random non-replacement combined features are adopted, and multiple sampling is carried out through non-replacement sampling equal ratio features, so that each feature can participate in model training, the participating weights are the same, namely the number of models of each feature participating in training is consistent;
in the step S4, the training model adopts an XGBOOST model; training multiple groups of parameters by each group of models to obtain multiple groups of basic XGBOOST models;
the step S5 includes:
step S51, traversing each sub-model;
step S52, modifying the feature weight of the single submodelThe significance calculation mode is as follows: test set ks of the submodel i +ks i * Importance fimp of the feature i
Step S53, feature importance of all sub-models is aggregated, and the average value of new feature importance of each feature is calculated, so that comprehensive weighted feature importance is obtained;wherein m represents the number of models trained overall, ks i Ks value, fimp, representing the ith model i Representing the feature importance of all the in-mold features of the ith model.
2. The feature importance-based model building method according to claim 1, characterized in that:
in the step S2, random unreplaced sampling is carried out on the characteristic data; alternatively, the feature data is sampled with a put back equal percentage.
3. The feature importance-based model building method according to claim 1, characterized in that:
in the step S2, a method of random combination is introduced, and a final model test set ks and sub-model feature importance weighting averaging method is combined, so that the obtained feature importance is more representative, and a smoother quantitative evaluation index is provided for the feature, so that an excellent basis is provided for subsequent modeling feature screening.
4. The feature importance-based model building method according to claim 1, characterized in that:
in the step S4, the obtained feature importance and the feature IV value are combined together to form an evaluation index again, and the evaluation mode of the XGBOOST tree and the evaluation mode of the linear IV are considered;
setting weight combination schemes with different IV and feature importance, and deriving additional feature evaluation indexes again; given IV weight a, feature importance weight b, and the new feature evaluation index after derivation is f_index new W
iv*a+f_imp_new*b。
5. The feature importance-based model building method according to claim 1, characterized in that:
the step S6 further includes: the progressive, decremental or step wise modeling is performed according to the resulting feature ordering lines.
6. A model building system based on feature importance, wherein the model building system is used in the big data wind control modeling field, the model building system comprises:
the feature data initializing module is used for initializing feature data;
the characteristic data sampling module is used for sampling characteristic data to form a plurality of groups of characteristic data combinations, and each group of characteristic data combination is used as a sub-model; the feature data sampling module adopts random non-replacement combined features, and performs multiple sampling through the features of non-replacement sampling equal ratio, so that each feature can participate in model training, and the weights of the participation are the same, namely the number of models of each feature participating in training is consistent;
the model parameter setting module is used for setting model parameters of each group of characteristic models; each group of feature combinations is provided with the same model parameter range;
the model training module is used for training a model and calculating the importance of each sub-model characteristic; the training model adopts an XGBOOST model; training multiple groups of parameters by each group of models to obtain multiple groups of basic XGBOOST models;
an importance calculation module for calculating the importance of the comprehensive weighted features using all sub-models;
the importance ranking module is used for ranking the feature importance to obtain a new importance ranking; and
the training model modeling module is used for taking the feature subset which is ranked at the front according to the feature sequence to set the model parameter range again, and retraining model modeling;
the importance calculating module includes:
the sub-model traversing unit is used for traversing each sub-model;
the feature importance calculating unit is used for modifying the feature importance calculating mode of the single submodel to be as follows: test set ks of the submodel i +ks i * Importance fimp of the feature i The method comprises the steps of carrying out a first treatment on the surface of the And
the feature importance aggregation unit is used for aggregating the feature importance of all the submodels and solving the mean value of the new feature importance of each feature to obtain the comprehensive weighted feature importance;wherein m represents the number of models trained overall, ks i Ks value, fimp, representing the ith model i Representing the feature importance of all the in-mold features of the ith model.
CN202010661710.4A 2020-07-10 2020-07-10 Model building method and system based on feature importance Active CN111860630B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010661710.4A CN111860630B (en) 2020-07-10 2020-07-10 Model building method and system based on feature importance

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010661710.4A CN111860630B (en) 2020-07-10 2020-07-10 Model building method and system based on feature importance

Publications (2)

Publication Number Publication Date
CN111860630A CN111860630A (en) 2020-10-30
CN111860630B true CN111860630B (en) 2023-10-13

Family

ID=73153137

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010661710.4A Active CN111860630B (en) 2020-07-10 2020-07-10 Model building method and system based on feature importance

Country Status (1)

Country Link
CN (1) CN111860630B (en)

Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE102013225768A1 (en) * 2013-12-12 2015-06-18 Robert Bosch Gmbh Method and apparatus for determining a LOLIMOT model
CN105589683A (en) * 2014-10-22 2016-05-18 腾讯科技(深圳)有限公司 Sample extraction method and apparatus
CN107316082A (en) * 2017-06-15 2017-11-03 第四范式(北京)技术有限公司 For the method and system for the feature importance for determining machine learning sample
CN107730154A (en) * 2017-11-23 2018-02-23 安趣盈(上海)投资咨询有限公司 Based on the parallel air control application method of more machine learning models and system
WO2018145596A1 (en) * 2017-02-13 2018-08-16 腾讯科技(深圳)有限公司 Method and device for extracting feature information, server cluster, and storage medium
CN108764597A (en) * 2018-04-02 2018-11-06 华南理工大学 A kind of product quality control method based on integrated study
CN109035003A (en) * 2018-07-04 2018-12-18 北京玖富普惠信息技术有限公司 Anti- fraud model modelling approach and anti-fraud monitoring method based on machine learning
CN109460825A (en) * 2018-10-24 2019-03-12 阿里巴巴集团控股有限公司 For constructing the Feature Selection Algorithms, device and equipment of machine learning model
CN110334773A (en) * 2019-07-12 2019-10-15 四川新网银行股份有限公司 Model based on machine learning enters the screening technique of modular character
CN110908908A (en) * 2019-11-21 2020-03-24 深圳无域科技技术有限公司 Method and device for testing micro-service Dubbo interface
CN110991474A (en) * 2019-10-12 2020-04-10 未鲲(上海)科技服务有限公司 Machine learning modeling platform

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11645587B2 (en) * 2020-07-08 2023-05-09 Vmware, Inc. Quantizing training data sets using ML model metadata

Patent Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE102013225768A1 (en) * 2013-12-12 2015-06-18 Robert Bosch Gmbh Method and apparatus for determining a LOLIMOT model
CN105589683A (en) * 2014-10-22 2016-05-18 腾讯科技(深圳)有限公司 Sample extraction method and apparatus
WO2018145596A1 (en) * 2017-02-13 2018-08-16 腾讯科技(深圳)有限公司 Method and device for extracting feature information, server cluster, and storage medium
CN107316082A (en) * 2017-06-15 2017-11-03 第四范式(北京)技术有限公司 For the method and system for the feature importance for determining machine learning sample
CN107730154A (en) * 2017-11-23 2018-02-23 安趣盈(上海)投资咨询有限公司 Based on the parallel air control application method of more machine learning models and system
CN108764597A (en) * 2018-04-02 2018-11-06 华南理工大学 A kind of product quality control method based on integrated study
CN109035003A (en) * 2018-07-04 2018-12-18 北京玖富普惠信息技术有限公司 Anti- fraud model modelling approach and anti-fraud monitoring method based on machine learning
CN109460825A (en) * 2018-10-24 2019-03-12 阿里巴巴集团控股有限公司 For constructing the Feature Selection Algorithms, device and equipment of machine learning model
CN110334773A (en) * 2019-07-12 2019-10-15 四川新网银行股份有限公司 Model based on machine learning enters the screening technique of modular character
CN110991474A (en) * 2019-10-12 2020-04-10 未鲲(上海)科技服务有限公司 Machine learning modeling platform
CN110908908A (en) * 2019-11-21 2020-03-24 深圳无域科技技术有限公司 Method and device for testing micro-service Dubbo interface

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
基于数据挖掘的疾病预测组合模型研究;崔晓旭;《中国优秀硕士学位论文全文数据库》(第12期);E062-154 *

Also Published As

Publication number Publication date
CN111860630A (en) 2020-10-30

Similar Documents

Publication Publication Date Title
CN113191484B (en) Federal learning client intelligent selection method and system based on deep reinforcement learning
Zhu et al. Grouped network vector autoregression
CN110070116A (en) Segmented based on the tree-shaped Training strategy of depth selects integrated image classification method
CN108805193A (en) A kind of power loss data filling method based on mixed strategy
CN106951471A (en) A kind of construction method of the label prediction of the development trend model based on SVM
Xiong et al. A large-scale consensus model to manage non-cooperative behaviors in group decision making: A perspective based on historical data
CN108830478A (en) A kind of team's recommended method towards the processing of crowdsourcing task
CN111709523A (en) Width learning method based on internal integration
CN113722980A (en) Ocean wave height prediction method, system, computer equipment, storage medium and terminal
CN113240263A (en) Comprehensive energy system planning combination evaluation method based on entropy weight fuzzy
CN115775026A (en) Federated learning method based on organization similarity
CN115861671A (en) Double-layer self-adaptive clustering method considering load characteristics and adjustable potential
CN113361928B (en) Crowd-sourced task recommendation method based on heterogram attention network
CN114679372A (en) Node similarity-based attention network link prediction method
CN110110962A (en) A kind of task gunz executes the preferred method of team
CN111860630B (en) Model building method and system based on feature importance
CN113705098A (en) Air duct heater modeling method based on PCA and GA-BP network
Xie et al. The study of methods for post-pruning decision trees based on comprehensive evaluation standard
CN113129188A (en) Provincial education teaching evaluation system based on artificial intelligence big data
CN113361776A (en) Power load probability prediction method based on user power consumption behavior clustering
CN111292062A (en) Crowdsourcing garbage worker detection method and system based on network embedding and storage medium
CN111353525A (en) Modeling and missing value filling method for unbalanced incomplete data set
CN115695429A (en) Non-IID scene-oriented federal learning client selection method
CN114841501A (en) Large-group satellite emergency scheme decision method and system in social network environment
Li et al. Grey-incidence clustering decision-making method with three-parameter interval grey number based on regret theory

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant