CN113345581B - Cerebral apoplexy post thrombolysis bleeding probability prediction method based on ensemble learning - Google Patents

Cerebral apoplexy post thrombolysis bleeding probability prediction method based on ensemble learning Download PDF

Info

Publication number
CN113345581B
CN113345581B CN202110525660.1A CN202110525660A CN113345581B CN 113345581 B CN113345581 B CN 113345581B CN 202110525660 A CN202110525660 A CN 202110525660A CN 113345581 B CN113345581 B CN 113345581B
Authority
CN
China
Prior art keywords
model
training
folds
features
prediction
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202110525660.1A
Other languages
Chinese (zh)
Other versions
CN113345581A (en
Inventor
梁浩然
倪文田
李峰
施超
孙阳
梁荣华
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang University of Technology ZJUT
Original Assignee
Zhejiang University of Technology ZJUT
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang University of Technology ZJUT filed Critical Zhejiang University of Technology ZJUT
Priority to CN202110525660.1A priority Critical patent/CN113345581B/en
Publication of CN113345581A publication Critical patent/CN113345581A/en
Application granted granted Critical
Publication of CN113345581B publication Critical patent/CN113345581B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H50/00ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
    • G16H50/30ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for calculating health indices; for individual health risk assessment
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • G06N20/20Ensemble learning
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H50/00ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
    • G16H50/70ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for mining of medical data, e.g. analysing previous cases of other patients
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02ATECHNOLOGIES FOR ADAPTATION TO CLIMATE CHANGE
    • Y02A90/00Technologies having an indirect contribution to adaptation to climate change
    • Y02A90/10Information and communication technologies [ICT] supporting adaptation to climate change, e.g. for weather forecasting or climate simulation

Landscapes

  • Engineering & Computer Science (AREA)
  • Medical Informatics (AREA)
  • Health & Medical Sciences (AREA)
  • Data Mining & Analysis (AREA)
  • Public Health (AREA)
  • Primary Health Care (AREA)
  • Software Systems (AREA)
  • Databases & Information Systems (AREA)
  • Epidemiology (AREA)
  • General Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Theoretical Computer Science (AREA)
  • Pathology (AREA)
  • Evolutionary Computation (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Artificial Intelligence (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The cerebral apoplexy post-thrombolysis bleeding probability prediction method based on ensemble learning comprises the steps of preprocessing original data to ensure that the data for model training is sufficiently simplified, and simultaneously, the original data can be completely represented; then inputting the preprocessed data into four sub learners, respectively training to generate a preliminary result, and summarizing the result; and finally, inputting the summarized result obtained in the last step into a Catboost model, and obtaining a final probability prediction model through training. According to the invention, theoretical and practical analysis is performed on a plurality of factors related to the post-thrombolytic hemorrhage probability of cerebral apoplexy, key factors are extracted, and a prediction model with higher accuracy is trained by means of the performance advantage of integrated learning. By virtue of deep research on the correlation among various features and reasonable utilization of ensemble learning, the model finally presented by the invention can give reliable theoretical guidance to doctors clinically.

Description

Cerebral apoplexy post thrombolysis bleeding probability prediction method based on ensemble learning
Technical Field
The invention relates to a cerebral apoplexy post-thrombolysis bleeding probability prediction method based on ensemble learning.
Background
Cerebral stroke, i.e. the blockage of cerebral vessels by thrombus, once it occurs, will cause great harm to the life health of the patient. The common clinical treatment method is to use thrombolytic drugs for thrombolysis, but after the thrombolytic drugs are used, serious bleeding complications are caused with a certain probability, and the probability of serious disability or death of patients suffering from bleeding exceeds 90 percent. Under the general condition, the thrombolytic bleeding probability of the doctor informing the family members of the patient is mostly obtained according to the experience of the doctor, the judgment basis for the individual patient is lacking, and each doctor has own judgment standard and has larger randomness. For this situation, an effective prediction method is needed in clinic.
In recent years, with the advent of products such as AlphaGo, the heat of machine learning is higher and higher, and more fields such as natural language processing and image style migration are involved, and considerable roles are played in these fields. The integrated learning is a large class in machine learning, and is mainly characterized in that a plurality of learners are built to complete learning tasks: 1) The sub-learners are not limited in type, and the integrated total learners can be homogenous (the sub-learners are the same in type) or heterogeneous (the self-learning Xi Qi is different in type), so that the performance and compatibility of the integrated learning are improved; 2) Typically, a generalization performance significantly superior to that of a single learner can be achieved through integration, which is more pronounced if the sub learner is weaker.
Disclosure of Invention
In order to overcome the defects of the prior art, in order to provide a reliable enough basis for a doctor to judge the bleeding probability of a cerebral apoplexy patient after thrombolysis, the invention provides an integrated learning-based cerebral apoplexy post-thrombolysis bleeding probability prediction method, which integrates a learning model, integrates factors of a plurality of aspects of the patient to consider and outputs the final judgment.
In order to solve the technical problems, the invention provides the following technical scheme:
an ensemble learning-based post-thrombolytic hemorrhage probability prediction method for cerebral apoplexy, comprising the following steps:
(1) Input feature processing: after the original data is obtained, irrelevant features and the like are required to be removed, so that the performance of a final model is improved;
(2) Training a single model: training on the preprocessed data set by using four sub learners respectively to obtain respective prediction results, and waiting for the next processing; the process is as follows:
firstly, dividing a data set into a plurality of folds, and setting the total number of the folds as (N+1), wherein the number of samples contained in the first N folds is the same in the (N+1) folds, and the (N+1) th folds have no requirement, but the number of the contained samples is as close to the first F folds as possible;
for each sub-learner, training is required N times in total; in the T-th training, the training set is the first N folds with folds removed T The remaining (N-1) folds of the model is used to predict folds T Fold N+1 Is provided with fold T And fold N+1 The number of samples contained is M and M respectively, and after one round of training, two vectors are obtained, namely the model is related to fold respectively T Is the prediction result of (2)
Figure BDA0003065625590000021
Regarding fold N+1 Predicted outcome of->
Figure BDA0003065625590000022
After training by N rounds, all
Figure BDA0003065625590000023
Splicing to obtain the prediction result of the sub learner about the first N folds +.>
Figure BDA0003065625590000024
For all->
Figure BDA0003065625590000025
Averaging to obtain the child learner about fold N+1 Final prediction result->
Figure BDA0003065625590000026
Due to adoption ofFour sub-learners, thus after a single model training step, two prediction matrices are obtained, i.e. prediction result matrices for the first N folds
Figure BDA0003065625590000027
Regarding fold N+1 Is>
Figure BDA0003065625590000028
(3) Multi-model fusion: fusing the results output in the last step according to a specific method, and outputting final probability prediction; the operation of the multimodal fusion is as follows:
construction of a new learner
Figure BDA0003065625590000029
To learn the outputs of the respective sub-learners, < >>
Figure BDA00030656255900000210
Is a Catboost model whose training set is +.>
Figure BDA00030656255900000211
Matrix, verification set is +.>
Figure BDA00030656255900000212
Matrix, training completed->
Figure BDA00030656255900000213
I.e., the final model, for predicting the new sample; when predicting a new sample, the new sample is processed in the steps (1) and (2) and finally +.>
Figure BDA00030656255900000214
The received input is a preliminary prediction of the single model training output of step (2), after which a probability prediction of post-thrombolytic hemorrhage can be output.
Further, in the step (1), the method for processing the input features includes the following steps:
(1-1) for conventional features, it is possible to input directly into the model without any processing;
(1-2) directly rejecting useless features so as not to pollute the trained model;
(1-3) merging the relevant features;
(1-4) continuous feature normalization, compressing numerical features into the same range;
(1-5) discrete feature one-time thermal encoding.
The integrated learning model provided by the invention comprises four sub learners, namely CatBoost, XGBoost, lightGBM and Factorizationachine respectively. It should be noted that, in addition to the factor learning, three other sub learners are actually integrated with a plurality of sub learners, and thus, from this point of view, the present invention can be regarded as "integrated integration". In order to avoid ambiguity, the sub-learner refers to all four frameworks unless otherwise specified.
The ensemble learning can be divided into two main categories according to the generation mode of the sub-learner: serial and parallel. There is a strong dependency between the serially generated sub-learners, and there is no parallel. The three integrated learning frameworks (except the factor learning) related by the invention are all serial generation sub learners, and in particular belong to Boosting one family of algorithms.
The earliest Boosting algorithm was AdaBoost, which appeared in 1997. When training, adaBoost firstly gives the same initial weight to each sample, each round of learner adjusts the weight of each sample according to the performance of each round of learner after training, and the weights of the erroneous classifying samples are increased, so that the learner can get more attention in the subsequent learners, a plurality of learners are trained according to the process, and finally weighted combination is carried out. Gradient Boosting, which is generally similar to AdaBoost, was developed in 2001 with the greatest difference that the learner was more concerned with the method of classifying erroneous samples in the previous round in each round. AdaBoost is the weight of the error sample added, and Gradient Boosting is the error of the last round corrected by fitting the negative gradient of the last round.
XGBoost, which was shown in 2015 and is known as Extreme Gradient Boosting, is a new implementation of Gradient Boosting, with the main differences in loss function: 1) The regularization term is added on the basis of the original loss function, so that a new objective function is generated; 2) The objective function is second-order taylor-expanded and optimized in a manner similar to newton's method. Equation (1) is the objective function of the mth learner in XGBoost, where N is the total number of samples, Ω (f) m ) Is a newly added regularization term.
Figure BDA0003065625590000031
The LightGBM was proposed by microsoft in 2017, and is known in full as Light Gradient Boosting Machine, whose name reveals its greatest features, light. Compared with XGBoost, lightGBM, the training device has the advantages of faster training speed, better accuracy and lower memory consumption. In addition, the method supports distributed computing, and can rapidly process mass data. Overall, the LightGBM is in principle just as practical and XGBoost is to fit a negative gradient, which is optimized mainly in terms of multithreading, decision trees, etc.
CatBOOST was open-sourced by Russian search giant Yandex in month 4 of 2017, which is the most mainstream three of the Boosting algorithm family with XGBoost, lightGBM. CatBoost, categorical Boost, which is based on a symmetrical tree, the developer claiming accuracy is better than XGBoost and LightGBM, and the pain point which is mainly solved is to efficiently and reasonably process category characteristics. The biggest innovation of Catboost, compared to the former two, is the fact that a symmetrical tree is used.
FactorizationMachine (FM), factorization machine. In a traditional linear model, each feature is independent, if interaction between the features needs to be considered, the features may need to be manually combined in a crossing way, and when the feature dimension is too high, the method is conceivable to be feasible; the nonlinear SVM can perform kernel mapping on the features, but at feature heightIn the case of sparsity, learning is not performed well; other decomposition models, such as matrix decomposition MF, svd++, etc., do learn interactive hidden relationships between features, but basically each model is only used for a particular scene. For this reason, FM occurs under highly sparse data scenes such as recommendation systems. Equation (2) is the objective function of FM, the most critical part of which is<V i ,V j >x i x j ,<V i ,V j >Is the dot product, x, of the ith and jth rows of matrix V i And x j Then two features are adopted, and through the operation, all the features of the sample are connected with each other to different degrees, so that hidden connection among the features can be conveniently found out.
Figure BDA0003065625590000041
The beneficial effects of the invention are as follows: the theory and practice analysis is carried out on a plurality of factors related to the post-thrombolytic hemorrhage probability of cerebral apoplexy, key factors are extracted, and a prediction model with higher accuracy is trained by means of the performance advantage of ensemble learning; by virtue of deep research on the correlation among various features and reasonable utilization of ensemble learning, the model finally presented by the invention can give reliable theoretical guidance to doctors clinically.
Drawings
FIG. 1 is a flow chart of the present invention.
Fig. 2 is a view of data preprocessing according to the present invention.
FIG. 3 is a single model training view of the present invention.
Detailed description of the preferred embodiments
The invention is further described below with reference to the accompanying drawings.
Referring to fig. 1 to 3, a method for predicting post-thrombolytic hemorrhage probability of cerebral apoplexy based on ensemble learning includes the following steps:
1) The input feature processing, as shown in fig. 1, in the original feature space, the subsequent processing performed according to the need can divide the features into five categories: a) Conventional features, which do not require any processing and can be directly input into the model; b) Useless features, such as blood type and the like, obviously do not play any positive role on the target according to experience, and for the features, the method of the invention directly eliminates the model trained by the training model so as not to be polluted; c) In many cases, the original data contains some features with similar implicit meanings, such as blood sugar level and diabetes, if the features exist as they are, the same implicit features of the features may occupy excessive weight in the final model, so that the correct judgment of the model on the result is affected, and in this case, the method of the invention combines the features; d) The continuous feature normalization, a large class of features common in raw data are continuous numerical features, but there are often magnitude differences between different numerical features, such as systolic pressure and BMI, which are essentially unproblematic, but when used to train a model, larger values tend to take up more weight. Therefore, for the situation, the invention selects to perform feature standardization and compresses numerical type features into the same range; e) Discrete features are thermally encoded uniquely, the discrete features being categorical features such as gender. For class-type features, one approach is to assign a value to each class in natural number order to represent that class. However, as in d), although the difference between the representative numbers of each category is not particularly large, the representative numbers with different sizes always have a certain influence on the result, so that the invention selects to treat the category type characteristics by using the single-heat coding, and the memory burden is increased, but the rationality of the model is greatly improved;
2) Single model training FIG. 2 illustrates the learning process of one of the sub-learners during a single model training process. Training has N rounds, N can be set by oneself according to specific situations, or a plurality of N values are selected and then an optimal N value is selected according to a final model effect, as shown in fig. 2, a data set after the previous feature pretreatment is divided into n+1fold, wherein the first N fold is used as a training set, and the n+1fold is used as a verification set to verify the performance of the model. Training deviceThe number of samples of each fold in the training set is M, the number of samples of the n+1st fold is M, in fig. 2, the right is the total single model training process, training is performed by four sub learners respectively, and then the results of all the sub learners are integrated to generate the final prediction matrix. The left is the detailed training process of a single sub-learner, as shown, a total of N rounds of training, in the T-th round of training, fold T (Tth fold) is a verification set, the rest N-1 folds are training sets, and a preliminary model obtained by training the sub-learner on the training sets which are further divided is used for correcting the folds T Fold N+1 Making predictions to generate two prediction vectors, e.g. obtained by first training
Figure BDA0003065625590000051
Is->
Figure BDA0003065625590000052
According to the operation of the step, after N rounds of training, the sub learner is used for merging the prediction results of the previous N folds to obtain the corresponding folds 1 To fold N For the predictive vector of (2) for fold N+1 The prediction vector of (2) can be defined by N +.>
Figure BDA0003065625590000053
The vectors are averaged, and after all the sub learners perform N rounds of training, the obtained results are combined to obtain corresponding prediction matrixes;
3) The multi-model fusion is carried out, the result output in the last step is fused according to a specific method, and then final probability prediction can be output; the operation of the multimodal fusion is as follows:
construction of a new learner
Figure BDA0003065625590000061
To learn the outputs of the respective sub-learners, < >>
Figure BDA0003065625590000062
Is a Catboost model whose training set is +.>
Figure BDA0003065625590000063
Matrix, verification set is +.>
Figure BDA0003065625590000064
Matrix, training completed->
Figure BDA0003065625590000065
I.e., the final model, for predicting the new sample; when predicting a new sample, the new sample is processed in the steps (1) and (2) and finally +.>
Figure BDA0003065625590000066
The received input is a preliminary prediction of the single model training output of step (2), after which a probability prediction of post-thrombolytic hemorrhage can be output.
According to the embodiment, theoretical and practical analysis is carried out on a plurality of factors related to the post-thrombolytic hemorrhage probability of cerebral apoplexy, key factors are extracted, and a prediction model with higher accuracy is trained by means of the performance advantage of integrated learning; by virtue of deep research on the correlation among various features and reasonable utilization of ensemble learning, the model finally presented by the invention can give reliable theoretical guidance to doctors clinically.

Claims (2)

1. The cerebral apoplexy post-thrombolysis bleeding probability prediction method based on ensemble learning is characterized by comprising the following steps of:
(1) Input feature processing: after the original data is obtained, irrelevant feature processing is needed to be removed, and the performance of a final model is improved;
in the original feature space, features are divided into five main categories: a) Conventional features, which do not need to be processed and are directly input into the model; b) Directly removing useless features including blood types; c) Relevant characteristics including the height of blood sugar and the existence of diabetes are combined; d) Continuous feature standardization, including systolic pressure and BMI, is performed, and numerical features are compressed into the same range; e) Single-heat encoding of discrete features;
(2) Training a single model: training on the preprocessed data set by using four sub learners respectively to obtain respective prediction results, and waiting for the next processing; the process is as follows:
firstly, dividing a data set into a plurality of folds, and setting the total number of the folds as (N+1), wherein the number of samples contained in the first N folds is the same in the (N+1) folds, and the (N+1) th folds have no requirement, but the number of the contained samples is as close to the first N folds as possible;
for each sub-learner, training is required N times in total; in the T-th training, the training set is the first N folds with folds removed T The remaining (N-1) folds of the model is used to predict folds T Fold N+1 Is provided with fold T And fold N+1 The number of samples contained is M and M respectively, and after one round of training, two vectors are obtained, namely the model is related to fold respectively T Is the prediction result of (2)
Figure QLYQS_1
Regarding fold N+1 Predicted outcome of->
Figure QLYQS_2
After N-round training, all +.>
Figure QLYQS_3
Splicing to obtain the prediction result of the sub learner about the first N folds +.>
Figure QLYQS_4
For all->
Figure QLYQS_5
Averaging to obtain the child learner about fold N+1 Final prediction result->
Figure QLYQS_6
Due to the collection ofFour sub-learners are used, so after a single model training step, two prediction matrices are obtained, namely a prediction result matrix for the first N folds and a prediction result matrix for the folds N+1 Is a prediction result matrix of (a);
(3) Multi-model fusion: fusing the results output in the last step according to a specific method, and outputting final probability prediction; the operation of the multimodal fusion is as follows:
constructing a new learner I to learn the output of each sub learner, wherein I is a Catboost model, the training set of the new learner is the matrix in the step (2), the verification set is the matrix, and the trained I is a final model for predicting new samples; when a new sample is predicted, the new sample is processed in the step (1) and the step (2), and finally the input received by I is preliminary prediction output by the single model training in the step (2), and then the probability prediction of bleeding after thrombolysis can be output.
2. The method for predicting post-thrombolytic hemorrhage based on ensemble learning as set forth in claim 1, wherein in said step (1), the method for inputting feature processing comprises the steps of:
(1-1) for conventional features, it is possible to input directly into the model without any processing;
(1-2) directly rejecting useless features so as not to pollute the trained model;
(1-3) merging the relevant features;
(1-4) continuous feature normalization, compressing numerical features into the same range;
(1-5) discrete feature one-time thermal encoding.
CN202110525660.1A 2021-05-14 2021-05-14 Cerebral apoplexy post thrombolysis bleeding probability prediction method based on ensemble learning Active CN113345581B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110525660.1A CN113345581B (en) 2021-05-14 2021-05-14 Cerebral apoplexy post thrombolysis bleeding probability prediction method based on ensemble learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110525660.1A CN113345581B (en) 2021-05-14 2021-05-14 Cerebral apoplexy post thrombolysis bleeding probability prediction method based on ensemble learning

Publications (2)

Publication Number Publication Date
CN113345581A CN113345581A (en) 2021-09-03
CN113345581B true CN113345581B (en) 2023-06-27

Family

ID=77469741

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110525660.1A Active CN113345581B (en) 2021-05-14 2021-05-14 Cerebral apoplexy post thrombolysis bleeding probability prediction method based on ensemble learning

Country Status (1)

Country Link
CN (1) CN113345581B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113951845B (en) * 2021-12-01 2022-08-05 中国人民解放军总医院第一医学中心 Method and system for predicting severe blood loss and injury condition of wound
CN115064255A (en) * 2022-06-27 2022-09-16 上海梅斯医药科技有限公司 Medical expense prediction method, system, equipment and storage medium

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110033860A (en) * 2019-02-27 2019-07-19 杭州贝安云科技有限公司 A kind of Inherited Metabolic Disorders recall rate method for improving based on machine learning
CN110472778A (en) * 2019-07-29 2019-11-19 上海电力大学 A kind of short-term load forecasting method based on Blending integrated study
CN111199343A (en) * 2019-12-24 2020-05-26 上海大学 Multi-model fusion tobacco market supervision abnormal data mining method
CN111968741A (en) * 2020-07-15 2020-11-20 华南理工大学 Diabetes complication high-risk early warning system based on deep learning and integrated learning
CN112700325A (en) * 2021-01-08 2021-04-23 北京工业大学 Method for predicting online credit return customers based on Stacking ensemble learning

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2492691A1 (en) * 2011-02-22 2012-08-29 Institut de Recerca Hospital Universitari Vall d'Hebron Method of predicting the evolution of a patient suffering of a neurovascular disease
US20210125207A1 (en) * 2019-10-29 2021-04-29 Somnath Banerjee Multi-layered market forecast framework for hotel revenue management by continuously learning market dynamics

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110033860A (en) * 2019-02-27 2019-07-19 杭州贝安云科技有限公司 A kind of Inherited Metabolic Disorders recall rate method for improving based on machine learning
CN110472778A (en) * 2019-07-29 2019-11-19 上海电力大学 A kind of short-term load forecasting method based on Blending integrated study
CN111199343A (en) * 2019-12-24 2020-05-26 上海大学 Multi-model fusion tobacco market supervision abnormal data mining method
CN111968741A (en) * 2020-07-15 2020-11-20 华南理工大学 Diabetes complication high-risk early warning system based on deep learning and integrated learning
CN112700325A (en) * 2021-01-08 2021-04-23 北京工业大学 Method for predicting online credit return customers based on Stacking ensemble learning

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
Predict Onset Age of Hypertension Using Catboost and Medical Big Data;Huanhuan Zhao等;《2020 International Conference on Networking and Network Applications》;全文 *
基于多源特征分析的冠心病预测模型研究;暨思铭;中国优秀硕士论文 信息科技;全文 *
基于多维特征和模型融合的血尿酸预测模型;安永利;中国优秀硕士论文 信息科技;全文 *
基于机器学习算法的脑出血相关肺炎预测模型研究;王孟等;中国卒中杂志;全文 *

Also Published As

Publication number Publication date
CN113345581A (en) 2021-09-03

Similar Documents

Publication Publication Date Title
CN108733742B (en) Global normalized reader system and method
Chowdhury et al. An artificial neural network model for neonatal disease diagnosis
CN113345581B (en) Cerebral apoplexy post thrombolysis bleeding probability prediction method based on ensemble learning
US11328125B2 (en) Method and server for text classification using multi-task learning
CN109801705A (en) Treat recommended method, system, device and storage medium
CN111145912B (en) Machine learning-based prediction device for personalized ovulation promotion scheme
US20230154627A1 (en) Semi-Supervised Machine Learning Method and System Suitable for Identification of Patient Subgroups in Electronic Healthcare Records
CN116779091B (en) Automatic generation method of multi-mode network interconnection and fusion chest image diagnosis report
CN114023412A (en) ICD code prediction method and system based on joint learning and denoising mechanism
CN114036934A (en) Chinese medical entity relation joint extraction method and system
Jamshidi et al. Symptom prediction and mortality risk calculation for COVID-19 using machine learning
Juraev et al. Multilayer dynamic ensemble model for intensive care unit mortality prediction of neonate patients
CN117153393A (en) Cardiovascular disease risk prediction method based on multi-mode fusion
Zheng et al. Chinese medical named entity recognition using crf-mt-adapt and ner-mrc
CN112052874A (en) Physiological data classification method and system based on generation countermeasure network
Togunwa et al. Deep hybrid model for maternal health risk classification in pregnancy: synergy of ANN and random forest
Sampath et al. Ensemble Nonlinear Machine Learning Model for Chronic Kidney Diseases Prediction
CN116072265B (en) Sleep stage analysis system and method based on convolution of time self-attention and dynamic diagram
Wang et al. DeepTriager: a neural attention model for emergency triage with electronic health records
CN116469534A (en) Hospital number calling management system and method thereof
EP4141744A1 (en) Semi-supervised machine learning method and system suitable for identification of patient subgroups in electronic healthcare records
CN113488165B (en) Text matching method, device, equipment and storage medium based on knowledge graph
Chraibi et al. A Deep Learning Framework for Automated ICD-10 Coding.
Liu et al. Using Artificial Neural Network Condensation to Facilitate Adaptation of Machine Learning in Medical Settings by Reducing Computational Burden: Model Design and Evaluation Study
Le et al. KNN loss and deep KNN

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant