CN111028913A - Hemodialysis treatment scheme aid decision-making method - Google Patents

Hemodialysis treatment scheme aid decision-making method Download PDF

Info

Publication number
CN111028913A
CN111028913A CN201911197143.5A CN201911197143A CN111028913A CN 111028913 A CN111028913 A CN 111028913A CN 201911197143 A CN201911197143 A CN 201911197143A CN 111028913 A CN111028913 A CN 111028913A
Authority
CN
China
Prior art keywords
model
time
loss
medical record
task
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201911197143.5A
Other languages
Chinese (zh)
Inventor
王冠
宋胜男
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing University of Technology
Original Assignee
Beijing University of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing University of Technology filed Critical Beijing University of Technology
Priority to CN201911197143.5A priority Critical patent/CN111028913A/en
Publication of CN111028913A publication Critical patent/CN111028913A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H20/00ICT specially adapted for therapies or health-improving plans, e.g. for handling prescriptions, for steering therapy or for monitoring patient compliance
    • G16H20/40ICT specially adapted for therapies or health-improving plans, e.g. for handling prescriptions, for steering therapy or for monitoring patient compliance relating to mechanical, radiation or invasive therapies, e.g. surgery, laser therapy, dialysis or acupuncture
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H50/00ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
    • G16H50/70ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for mining of medical data, e.g. analysing previous cases of other patients

Landscapes

  • Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Public Health (AREA)
  • Medical Informatics (AREA)
  • Biomedical Technology (AREA)
  • Computational Linguistics (AREA)
  • Epidemiology (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Biophysics (AREA)
  • Primary Health Care (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Pathology (AREA)
  • Databases & Information Systems (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Surgery (AREA)
  • Urology & Nephrology (AREA)
  • External Artificial Organs (AREA)

Abstract

The invention discloses a hemodialysis treatment scheme assistant decision method based on an Attention-LSTM model, which comprises the following steps of 1, preprocessing hemodialysis historical medical record data marked by a professional dialysis doctor to obtain a total sample set, 2, establishing a long-time and short-time memory network based on an Attention mechanism, and distributing Attention weights α to time-series medical records of each time-series medical record input into the modeltThereby calculating a medical record information global feature vector c; inputting the global feature vector c into a multi-task sharing layer to learn scheme labels; 4, distributing weight lambda to loss of each learning tasknWeighting and summing the losses to obtain the total loss; updating the network parameters of the model by using an Adam optimization algorithm to obtain the final bloodThe hemodialysis treatment protocol assists the decision making model. The model in the invention can assist dialysis clinicians to set up individualized hemodialysis treatment schemes for patients, reduce the manual workload of the clinicians and improve the hemodialysis efficiency of a hemodialysis center.

Description

Hemodialysis treatment scheme aid decision-making method
Technical Field
The invention relates to the field of intelligent analysis, in particular to an Attention-LSTM-based hemodialysis treatment scheme assistant decision-making method.
Background
Hemodialysis is one of the kidney replacement treatment modes of patients with acute and chronic renal failure. It slowly leads blood out of the blood vessel passage through a hemodialysis machine, exchanges substances with dialysate through the principles of dispersion, ultrafiltration, adsorption and convection, removes metabolic wastes in the body, maintains the balance of electrolytes and acid and alkali, and returns the purified blood to the body. Traditionally, hemodialysis physicians have developed individualized hemodialysis protocols for patients based on their own clinical experience. The doctor consults the age, weight, renal function, blood routine, electrolyte, cardiac function and other indexes of the dialysis patient to make an individualized hemodialysis scheme, wherein the scheme comprises dialysis frequency, single dialysis time, dialysate concentration, anticoagulation mode, dosage and the like. Because hemodialysis patient treatment cycle is generally longer, so can produce a series of historical dialysis case histories along with the patient that carries out the treatment, and the clinician often adjusts the dialysis scheme according to historical dialysis case history, and manual work volume is great. With the improvement of medical electronic medical record information systems, the field of health and medical care has gradually entered the information era. A large amount of electronic medical record information data provides a solid data base for the development of AI + medical treatment, and provides great potential for improving the level of the medical industry. For the field of hemodialysis, there is also the application of "AI + medical" mode. The method for predicting the hemodialysis protocol is disclosed in patent CN 109686446 a filed by Jiangsu Kogyo health science and technology Limited. The method comprises the steps of preprocessing training samples of a hemodialysis center, screening important features by adopting Lasso sparse constraint and RandomForest random forest respectively, and combining two groups of screened training samples into a final training sample. And finally, loading the training sample into a multiple linear regression model, and continuously iterating to obtain a dialysis scheme prediction model. The method does not consider the time sequence characteristic of hemodialysis medical record data, and meanwhile, a prediction model of the method is a simple multiple linear regression model, and the fitting degree of a complex model is possibly poor.
Disclosure of Invention
In order to solve the above problems and reduce the manual workload of dialysis physicians in making dialysis solutions, the present invention provides an Attention-LSTM-based hemodialysis treatment solution decision-making assisting method. The hemodialysis system is based on a deep learning algorithm, can assist a dialysis doctor to provide a personalized dialysis scheme for a patient, reduces heavy manual analysis work of the dialysis doctor, and improves hemodialysis efficiency of a hemodialysis center.
The technical scheme adopted by the invention is that the hemodialysis treatment scheme assistant decision method based on Attention-LSTM comprises the following steps:
step 1, preprocessing the hemodialysis history medical record data labeled by a professional dialysis doctor to obtain a total sample set;
step 2, establishing a long-time memory network based on a self-attention mechanism, and distributing attention weights α to each time sequence medical record of the input modeltThereby calculating a medical record information global feature vector c;
step 3, inputting the global feature vector c into a multitask sharing layer to learn scheme labels;
step 4, distributing weight lambda to the loss of each learning tasknWeighting and summing the losses to obtain the total loss; and updating the network parameters of the model by using an Adam optimization algorithm to obtain a final hemodialysis treatment scheme assistant decision model.
The data preprocessing of step 1 comprises the following substeps:
step 1.1, sensitive fields related to personal information of a patient in hemodialysis historical data, such as the real name, the telephone, the home address and the like of the patient, are erased;
and 1.2, determining the characteristic data of the input model and the dialysis scheme label data output by the model. The characteristic data includes age, weight, renal function, blood routine, electrolyte, cardiac function, etc. of the patient, and the model label data includes dialysis frequency, single dialysis time, dialysate concentration, anticoagulation mode, dosage, etc. of the patient. In particular, the time interval between each history medical record (including one medical record corresponding to the new dialysis scheme) and the new dialysis scheme of the patient is taken as the characteristic input;
step 1.3, filling up the missing model input characteristic data by using a mean filling method;
step 1.4, taking the patient ID as a unique mark for distinguishing patient data; sequencing the dialysis medical record data according to time to form serialized data by taking the ID of the patient as a basis;
and 1.5, randomly dividing the preprocessed sample set into a model training set, a verification set and a test set according to the proportion of 7:1: 2.
Establishing a long-time memory network based on a self-attention mechanism in the step 2, and distributing attention weights α to each time sequence medical record of the input modeltTherefore, the medical record information global feature vector c is calculated, and the method comprises the following substeps:
and 2.1, constructing a long-term and short-term memory network. The training algorithm of the long-time memory network is a back propagation algorithm, and when the error reaches an expected value, the network learning is finished. The long-and-short-time memory network introduces a gate control unit to endow the network with the capability of controlling the accumulation of internal information, so that the network can master long-distance dependence and selectively forget information during learning to prevent overload; input door itFor updating the state of the cells; forget door ftFor deciding which information should be discarded or retained; output gate otFor determining the next hidden state htA value of (d); x is the number oftInput for the current time step network; h ist-1Hidden state for last time step; c. CtMemory cells at the current time step; c. Ct-1Memory cells of the last time step;
Figure BDA0002294952320000041
is a candidate memory cell; wxi、Whi、Wxf、Whf、Wxo、Who、Wxc、WhcAll are weight matrices of the network model; bi、bf、bo、bcAre the offsets of the network model weights; sigma is a sigmoid activation function and ranges from 0 to 1; the tanh is a hyperbolic tangent activation function, and the long-time and short-time memory network formula is as follows:
it=σ(Wxixt+Whiht-1+bi) (1)
ft=σ(Wxfxt+Whfht-1+bf) (2)
ot=σ(Wxoxt+Whoht-1+bo) (3)
Figure BDA0002294952320000042
Figure BDA0002294952320000043
ht=ot*tanh(ct) (6)
the historical medical record data of the dialysis patients has time series, the medical records with longer time have smaller influence on the new dialysis scheme, and the medical records with closer time have larger influence on the new dialysis scheme. The serialized dialysis medical record data is input into a long-time memory network, certain historical information is removed or added through a well-designed gate structure, and the physiological change characteristics and the future treatment trend of the patient are mined.
And 2.2, adding a self-attention mechanism on the basis of the long-time and short-time memory network. The attention mechanism is to obtain the parts needing important attention from all information, thereby suppressing other useless information. The self-attention mechanism is a special case in the attention mechanism, and is hidden layer information h of each medical record obtained after long-time and short-time memory networks input patient serialized medical record datatDistributing weight and integrating hidden layer information h of each medical recordtAs a final representation of the patient's historical medical record information. Hidden layer information h of each case historytMultiplying the output u by a parameter matrix W to obtain an output u through calculationtAnd a trainable parameter matrix u dot product representing the context information to obtain the similarity score of the two, and using softmax to normalize the score of all time steps to obtain the corresponding attention weight αtFinally, the hidden layer information h of each case history is usedtAnd its corresponding weight αtThe global feature vector c used for representing the historical medical record information of the patient is obtained by multiplication, and the formula is as follows:
ut=tanh(Wht) (7)
Figure BDA0002294952320000051
Figure BDA0002294952320000053
Figure BDA0002294952320000052
the step 3 of inputting the global feature vector c into the multitask sharing layer for the scheme label learning includes the following steps:
the method of hard parameter sharing in multi-task learning is applied to construct independent learning structures aiming at different learning tasks by taking a label contained in a patient dialysis scheme as a target of the multi-task learning, namely, a hidden layer is shared among all tasks and a task related layer of a specific task is reserved at the same time. And the global feature vector c learned by the Attention-LSTM network is input into the multitask shared hidden layer. The scheme label comprises: dialysis frequency, single dialysis time, dialysate concentration, anticoagulation mode and dosage, etc.
Step 4, distributing weight lambda to loss of each learning tasknAnd weighting and summing the losses to obtain the total loss. Updating network parameters of the model by using an Adam optimization algorithm to obtain a final hemodialysis treatment scheme assistant decision model, and comprising the following substeps:
step 4.1, inputting the patient time-series medical record data into a model for training after preprocessing and normalization processing; different tasks adopt corresponding Loss functions Loss when multi-task training is carried outnAnd presetting a loss weight lambda for itnAnd a weight parameter λnThe importance of each task in the total loss is determined, a larger weight is set for the important task, and the learning sensitivity of the model to the task is improved; for the classification task, its loss function is the cross-entropy loss function CE, where i denotes the first sample,
Figure BDA0002294952320000061
indicating whether j is the true label of the ith sample,
Figure BDA0002294952320000062
Figure BDA0002294952320000063
indicating the probability that j is the true label of the ith sample,
Figure BDA0002294952320000064
for the regression task, the loss function is the mean square error loss function MSE, yiIn order to be the true value of the value,
Figure BDA0002294952320000065
is a predicted value; the model total Loss is equal to the weighted sum of the Loss functions of all tasks, and the formula is as follows:
Figure BDA0002294952320000066
Figure BDA0002294952320000067
Loss=λ1Loss12Loss2+…+λnLossn(13)
and 4.2, updating the network parameters of the model by using an Adam optimization algorithm. And (4) repeatedly updating the model, continuously reducing the total loss value until the termination condition is met, and stopping training the model. And verifying the model effect by using the verification set, adjusting the model parameters according to the actual effect, and finally verifying by using the test set.
The invention has the following advantages:
(1) the invention further excavates patient serialized medical record data information by utilizing a long-time memory network based on a self-attention mechanism, allocates attention weight to each time sequence medical record input into the model, integrates the medical record data information of each time sequence into a global feature vector for multi-task learning, and improves the final learning effect of the model. Attention-driven mechanisms stem from the study of human vision, where humans selectively focus on a portion of all information while ignoring other visible information. The attention mechanism mimics the response of the human brain in observing things, i.e., more attention is assigned to the more important parts. The self-attention mechanism is an improvement of the attention mechanism, reduces the dependence on external information, and is better at capturing the internal correlation of data. Because the dialysis medical record data of the dialysis patient have time sequence and the reference values of the medical record data at different times are different, the importance of the medical record data of the patient at each time is mined by using a self-attention mechanism, and the information with larger attention weight can bring more attention to the model, thereby being beneficial to improving the final learning effect of the model.
(2) The method utilizes a multi-task learning mechanism to train a plurality of scheme labels simultaneously, reduces the risk of model overfitting and improves the generalization performance of the model. The method for sharing the hard parameters in the multi-task learning is applied to construct independent learning structures aiming at different learning tasks based on the sharing layer, namely, the hidden layer is shared among all the tasks, and meanwhile, the task related layer of a specific task is reserved, so that the risk of overfitting is reduced. When a certain task is learned, the part irrelevant to the task is equivalent to noise in the learning process, and the generalization effect of model learning can be improved by introducing the noise. During single-task learning, the reverse propagation of the gradient tends to fall into a local minimum value, while the local minimum values of different tasks are in different positions in multi-task learning, and the model can be helped to escape from the local minimum value through interaction, so that the performance of the model is improved.
Drawings
FIG. 1 is a flow chart of an aided decision method for an Attention-LSTM-based hemodialysis treatment protocol according to the present invention.
FIG. 2 is a schematic diagram of a long-term memory network of an aided decision-making method of an Attention-LSTM-based hemodialysis treatment protocol according to the present invention.
FIG. 3 is a schematic diagram of a self-Attention mechanism-based long-term memory network of an Attention-LSTM-based hemodialysis treatment protocol assistant decision method provided by the invention.
FIG. 4 is a schematic diagram of a hard parameter sharing method in multitask learning of an Attention-LSTM-based hemodialysis treatment protocol aided decision method provided by the present invention.
Detailed Description
The present invention will be described in detail with reference to fig. 1 to 4, and the technical solutions in the embodiments of the present invention will be clearly and completely described below, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
Referring to fig. 1, the present invention herein provides an Attention-LSTM-based hemodialysis treatment protocol aided decision method, comprising the steps of:
step 1, preprocessing the hemodialysis history medical record data labeled by a professional dialysis doctor to obtain a total sample set;
step 2, establishing a long-time memory network based on a self-attention mechanism, and distributing attention weights α to each time sequence medical record of the input modeltThereby calculating a medical record information global feature vector c;
step 3, inputting the global feature vector c into a multitask sharing layer to learn scheme labels;
step 4, distributing weight lambda to the loss of each learning tasknAnd weighting and summing the losses to obtain the total loss. And updating the network parameters of the model by using an Adam optimization algorithm to obtain a final hemodialysis treatment scheme assistant decision model.
The data preprocessing of step 1 comprises the following substeps:
step 1.1: removing sensitive fields related to personal information of the patient in hemodialysis historical data, such as the real name, telephone, home address and the like of the patient;
step 1.2: characteristic data input into the model and dialysis scheme label data output by the model are determined. The characteristic data includes age, weight, renal function, blood routine, electrolyte, cardiac function, etc. of the patient, and the model label data includes dialysis frequency, single dialysis time, dialysate concentration, anticoagulation mode, dosage, etc. of the patient. In particular, the time interval between each history medical record (including one medical record corresponding to the new dialysis scheme) and the new dialysis scheme of the patient is taken as the characteristic input;
step 1.3: filling up the missing model input characteristic data by using a mean filling method;
step 1.4, taking the patient ID as a unique mark for distinguishing patient data; sequencing the dialysis medical record data according to time to form serialized data by taking the ID of the patient as a basis;
and 1.5, randomly dividing the preprocessed sample set into a model training set, a verification set and a test set according to the proportion of 7:1: 2.
Step 2, establishing a long-time and short-time memory network based on a self-attention mechanism, and allocating attention weights α to each time sequence medical record of the input modeltTherefore, the medical record information global feature vector c is calculated, and the method comprises the following substeps:
and 2.1, constructing a long-term memory network by referring to the figure 2. The training algorithm of the long-time memory network is a back propagation algorithm, and when the error reaches an expected value, the network learning is finished. The long-and-short-time memory network introduces a gate control unit to endow the network with the capability of controlling the accumulation of internal information, so that the network can master long-distance dependence and selectively forget information during learning to prevent overload; input door itFor updating the state of the cells; forget door ftFor deciding which information should be discarded or retained; output gate otFor determining the next hidden state htA value of (d); x is the number oftInput for the current time step network; h ist-1Hidden state for last time step; c. CtMemory cells at the current time step; c. Ct-1Memory cells of the last time step;
Figure BDA0002294952320000101
is a candidate memory cell; wxi、Whi、Wxf、Whf、Wxo、Who、Wxc、WhcAll are weight matrices of the network model; bi、bf、bo、bcAre the offsets of the network model weights; sigma is a sigmoid activation function and ranges from 0 to 1; the tanh is a hyperbolic tangent activation function, and the long-time and short-time memory network formula is as follows:
it=σ(Wxixt+Whiht-1+bi) (1)
ft=σ(Wxfxt+Whfht-1+bf) (2)
ot=σ(Wxoxt+Whoht-1+bo) (3)
Figure BDA0002294952320000102
Figure BDA0002294952320000103
ht=ot*tanh(ct) (6)
the historical medical record data of the dialysis patients has time series, the medical records with longer time have smaller influence on the new dialysis scheme, and the medical records with closer time have larger influence on the new dialysis scheme. The serialized dialysis medical record data is input into a long-time memory network, certain historical information is removed or added through a well-designed gate structure, and the physiological change characteristics and the future treatment trend of the patient are mined.
And 2.2, referring to fig. 3, adding a self-attention mechanism on the basis of the long-time and short-time memory network. The attention mechanism is to obtain the parts needing important attention from all information, thereby suppressing other useless information. The self-attention mechanism is a special case in attention mechanism, and is a long-time and short-time memory netHidden layer information h of each medical record obtained after patient serialized medical record data is input in a network modetDistributing weight and integrating hidden layer information h of each medical recordtAs a final representation of the patient's historical medical record information. Hidden layer information h of each case historytMultiplying the output u by a parameter matrix W to obtain an output u through calculationtAnd a trainable parameter matrix u dot product representing the context information to obtain the similarity score of the two, and using softmax to normalize the score of all time steps to obtain the corresponding attention weight αtFinally, the hidden layer information h of each case history is usedtAnd its corresponding weight αtThe global feature vector c used for representing the historical medical record information of the patient is obtained by multiplication, and the formula is as follows:
ut=tanh(Wht) (7)
Figure BDA0002294952320000111
Figure BDA0002294952320000113
Figure BDA0002294952320000112
step 3, inputting the global feature vector c into the multitask sharing layer for learning the scheme label, which comprises the following steps:
referring to fig. 4, the method of hard parameter sharing in multitask learning is to build independent learning structures for different learning tasks based on the shared layer, i.e. a hidden layer is shared among all tasks while preserving the task-dependent layer of a specific task. The label included in the dialysis scheme of the patient is taken as the target of multi-task learning, and the global feature vector c learned by the Attention-LSTM network is input into the shared hidden layer of the multi-task. The scheme label comprises: dialysis frequency, single dialysis time, dialysate concentration, anticoagulation mode and dosage, etc.
The multi-task learning is an induction migration mechanism, and aims to use a shared representation layer to train a plurality of different tasks in parallel, reduce the risk of model overfitting and improve the generalization performance of the model. The multi-task learning method can acquire internal relation among different learning tasks while learning, and fully utilizes data information in the input model. Hard parameter sharing is the most common way in multi-task learning of neural networks, and can be applied to all hidden layers of all tasks, and the output layer related to the tasks is reserved. Because different tasks have different noise modes, the simultaneous learning of a plurality of tasks can enable the model to obtain a more generalized representation, thereby reducing the risk of overfitting.
Step 4, distributing weight lambda to loss of each learning tasknWeighting and summing the losses to obtain the total loss; updating network parameters of the model by using an Adam optimization algorithm to obtain a final hemodialysis treatment scheme assistant decision model, and comprising the following substeps:
step 4.1, inputting the patient time-series medical record data into a model for training after preprocessing and normalization processing; different tasks adopt corresponding Loss functions Loss when multi-task training is carried outnAnd presetting a loss weight lambda for itnAnd a weight parameter λnThe importance of each task in the total loss is determined, a larger weight is set for the important task, and the learning sensitivity of the model to the task is improved; for the classification task, its loss function is the cross-entropy loss function CE, where i denotes the first sample,
Figure BDA0002294952320000131
indicating whether j is the true label of the ith sample,
Figure BDA0002294952320000132
Figure BDA0002294952320000133
indicating the probability that j is the true label of the ith sample,
Figure BDA0002294952320000134
for the regression task, itThe loss function is mean square error loss function MSE, yiIn order to be the true value of the value,
Figure BDA0002294952320000135
is a predicted value; the model total Loss is equal to the weighted sum of the Loss functions of all tasks, and the formula is as follows:
Figure BDA0002294952320000136
Figure BDA0002294952320000137
Loss=λ1Loss12Loss2+…+λnLossn(13)
and 4.2, updating the network parameters of the model by using an Adam optimization algorithm. And (4) repeatedly updating the model, continuously reducing the total loss value until the termination condition is met, and stopping training the model. And verifying the model effect by using the verification set, adjusting the model parameters according to the actual effect, and finally verifying by using the test set.
The above-mentioned details are only one embodiment of the present invention, and do not limit the scope of the present invention, therefore, those skilled in the art can make various changes and modifications without departing from the spirit and scope of the present invention, and these changes and modifications should also be construed as the protection scope of the present invention. The protection scope of the present invention is defined by the claims, and all structural changes equivalent to the content of the description and drawings of the present invention should be considered to be included in the protection scope of the present invention.

Claims (5)

1. An Attention-LSTM-based hemodialysis treatment scheme decision-making assisting method is characterized in that: the method comprises the following steps:
step 1, preprocessing the hemodialysis history medical record data labeled by a professional dialysis doctor to obtain a total sample set;
step 2, establishing the duration based on the self-attention mechanismA memory network for assigning attention weights α to each time-series medical record of the input modeltThereby calculating a medical record information global feature vector c;
step 3, inputting the global feature vector c into a multitask sharing layer to learn scheme labels;
step 4, distributing weight lambda to the loss of each learning tasknWeighting and summing the losses to obtain the total loss; and updating the network parameters of the model by using an Adam optimization algorithm to obtain a final hemodialysis treatment scheme assistant decision model.
2. An Attention-LSTM based hemodialysis treatment protocol aid decision method according to claim 1, wherein: the step 1 comprises the following substeps:
step 1.1, sensitive fields related to personal information of a patient, such as the real name, the telephone, the home address and the like of the patient, in the hemodialysis historical medical record data are erased;
step 1.2, determining characteristic data of an input model and dialysis scheme label data output by the model;
step 1.3, filling up the missing characteristic data by using a mean filling method;
step 1.4, taking the patient ID as a unique mark for distinguishing patient data; sequencing the dialysis medical record data according to time to form serialized data by taking the ID of the patient as a basis;
and 1.5, randomly dividing the preprocessed sample set into a model training set, a verification set and a test set according to the proportion of 7:1: 2.
3. An Attention-LSTM based hemodialysis treatment protocol aid decision method according to claim 1, wherein: the step 2 comprises the following substeps:
step 2.1, constructing a long-time and short-time memory network; the long-and-short-time memory network introduces a gate control unit to endow the network with the capability of controlling the accumulation of internal information, so that the network can master long-distance dependence and selectively forget information during learning to prevent overload; input door itFor updating the state of the cells; forget door ftFor deciding which information should be discarded or retained; output gate otFor determining the next hidden state htA value of (d); x is the number oftInput for the current time step network; h ist-1Hidden state for last time step; c. CtMemory cells at the current time step; c. Ct-1Memory cells of the last time step;
Figure FDA0002294952310000023
is a candidate memory cell; wxi、Whi、Wxf、Whf、Wxo、Who、Wxc、WhcAll are weight matrices of the network model; bi、bf、bo、bcAre the offsets of the network model weights; sigma is a sigmoid activation function and ranges from 0 to 1; the tanh is a hyperbolic tangent activation function, and the long-time and short-time memory network formula is as follows:
it=σ(Wxixt+Whiht-1+bi) (1)
ft=σ(Wxfxt+Whfht-1+bf) (2)
ot=σ(Wxoxt+Whoht-1+bo) (3)
Figure FDA0002294952310000021
Figure FDA0002294952310000022
ht=ot*tanh(ct) (6)
step 2.2, adding a self-attention mechanism on the basis of the long-time and short-time memory network; it is a hidden layer information h of each medical record obtained after long-time memory network inputs patient serialized medical record datatDistributing weight and integrating hidden layer information h of each medical recordtAs a final representation of patient historical medical record information; hidden layer information h of each case historytMultiplying the output u by a parameter matrix W to obtain an output u through calculationtAnd a trainable parameter matrix u dot product representing the context information to obtain the similarity score of the two, and using softmax to normalize the score of all time steps to obtain the corresponding attention weight αtFinally, the hidden layer information h of each case history is usedtAnd its corresponding weight αtThe global feature vector c used for representing the historical medical record information of the patient is obtained by multiplication, and the formula is as follows:
ut=tanh(Wht) (7)
Figure FDA0002294952310000031
score(ut,u)=ut Tu (9)
Figure FDA0002294952310000032
4. an Attention-LSTM based hemodialysis treatment protocol aid decision method according to claim 1, wherein: the step 3 comprises the following steps:
the method of hard parameter sharing in multi-task learning is applied to construct independent learning structures for different learning tasks based on a sharing layer by taking a label contained in a patient dialysis scheme as a target of the multi-task learning, namely, a hidden layer is shared among all tasks and a task related layer of a specific task is reserved at the same time. And the global feature vector c learned by the Attention-LSTM network is input into the multitask shared hidden layer.
5. An Attention-LSTM based hemodialysis treatment protocol aid decision-making approach in accordance with claim 1, wherein: the step 4 comprises the following substeps:
step 4.1, inputting the patient time-series medical record data into a model for training after preprocessing and normalization processing; different tasks adopt corresponding Loss functions Loss when multi-task training is carried outnAnd presetting a loss weight lambda for itnAnd a weight parameter λnThe importance of each task in the total loss is determined, a larger weight is set for the important task, and the learning sensitivity of the task is improved; for the classification task, its loss function is the cross-entropy loss function CE, where i denotes the first sample,
Figure FDA0002294952310000033
indicating whether j is the true label of the ith sample,
Figure FDA0002294952310000034
Figure FDA0002294952310000035
indicating the probability that j is the true label of the ith sample,
Figure FDA0002294952310000036
for the regression task, the loss function is the mean square error loss function MSE, yiIn order to be the true value of the value,
Figure FDA0002294952310000037
is a predicted value; the model total Loss is equal to the weighted sum of the Loss functions of all tasks, and the formula is as follows:
Figure FDA0002294952310000041
Figure FDA0002294952310000042
Loss=λ1Loss12Loss2+…+λnLossn(13)
step 4.2, updating the network parameters of the model by using an Adam optimization algorithm; repeatedly updating the model in an iterative manner, continuously reducing the total loss value until the termination condition is met, and stopping training the model; and verifying the model effect by using the verification set, adjusting the model parameters according to the actual effect, and finally verifying by using the test set.
CN201911197143.5A 2019-11-29 2019-11-29 Hemodialysis treatment scheme aid decision-making method Pending CN111028913A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911197143.5A CN111028913A (en) 2019-11-29 2019-11-29 Hemodialysis treatment scheme aid decision-making method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911197143.5A CN111028913A (en) 2019-11-29 2019-11-29 Hemodialysis treatment scheme aid decision-making method

Publications (1)

Publication Number Publication Date
CN111028913A true CN111028913A (en) 2020-04-17

Family

ID=70203275

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911197143.5A Pending CN111028913A (en) 2019-11-29 2019-11-29 Hemodialysis treatment scheme aid decision-making method

Country Status (1)

Country Link
CN (1) CN111028913A (en)

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111640512A (en) * 2020-06-02 2020-09-08 戴松世 Kidney replacement therapy starting strategy evaluation method and device and electronic equipment
CN111681767A (en) * 2020-06-12 2020-09-18 电子科技大学 Electronic medical record data processing method and system
CN112782762A (en) * 2021-01-29 2021-05-11 东北大学 Earthquake magnitude determination method based on deep learning
CN112957553A (en) * 2021-02-01 2021-06-15 南通市第二人民医院 Intelligent monitoring system and method for complications of hemodialysis patients
CN113436733A (en) * 2021-08-26 2021-09-24 肾泰网健康科技(南京)有限公司 Characteristic construction method of hemodialysis scheme generation model based on fusion experience
CN113488159A (en) * 2021-08-11 2021-10-08 中国医学科学院阜外医院 Medical department recommendation method and device based on neural network
CN114582463A (en) * 2022-05-05 2022-06-03 成都尚医信息科技有限公司 Personalized motion guidance system and method based on machine learning
CN114937486A (en) * 2022-06-22 2022-08-23 肾泰网健康科技(南京)有限公司 Construction method and application of IDH prediction and intervention measure recommendation multitask model
CN116313127A (en) * 2023-03-23 2023-06-23 珠海市安克电子技术有限公司 Decision support system based on pre-hospital first-aid big data
CN116453706A (en) * 2023-06-14 2023-07-18 之江实验室 Hemodialysis scheme making method and system based on reinforcement learning

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2019025601A1 (en) * 2017-08-03 2019-02-07 Koninklijke Philips N.V. Hierarchical neural networks with granularized attention
CN109493933A (en) * 2018-08-08 2019-03-19 浙江大学 A kind of prediction meanss of the adverse cardiac events based on attention mechanism
CN110136828A (en) * 2019-05-16 2019-08-16 杭州健培科技有限公司 A method of medical image multitask auxiliary diagnosis is realized based on deep learning
CN110334843A (en) * 2019-04-22 2019-10-15 山东大学 A kind of time-varying attention improves be hospitalized medial demand prediction technique and the device of Bi-LSTM

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2019025601A1 (en) * 2017-08-03 2019-02-07 Koninklijke Philips N.V. Hierarchical neural networks with granularized attention
CN109493933A (en) * 2018-08-08 2019-03-19 浙江大学 A kind of prediction meanss of the adverse cardiac events based on attention mechanism
CN110334843A (en) * 2019-04-22 2019-10-15 山东大学 A kind of time-varying attention improves be hospitalized medial demand prediction technique and the device of Bi-LSTM
CN110136828A (en) * 2019-05-16 2019-08-16 杭州健培科技有限公司 A method of medical image multitask auxiliary diagnosis is realized based on deep learning

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111640512B (en) * 2020-06-02 2023-08-15 戴松世 Kidney substitution therapy starting strategy evaluation method and device and electronic equipment
CN111640512A (en) * 2020-06-02 2020-09-08 戴松世 Kidney replacement therapy starting strategy evaluation method and device and electronic equipment
CN111681767B (en) * 2020-06-12 2022-07-05 电子科技大学 Electronic medical record data processing method and system
CN111681767A (en) * 2020-06-12 2020-09-18 电子科技大学 Electronic medical record data processing method and system
CN112782762A (en) * 2021-01-29 2021-05-11 东北大学 Earthquake magnitude determination method based on deep learning
CN112957553A (en) * 2021-02-01 2021-06-15 南通市第二人民医院 Intelligent monitoring system and method for complications of hemodialysis patients
CN113488159A (en) * 2021-08-11 2021-10-08 中国医学科学院阜外医院 Medical department recommendation method and device based on neural network
CN113436733B (en) * 2021-08-26 2021-11-30 肾泰网健康科技(南京)有限公司 Characteristic construction method of hemodialysis scheme generation model based on fusion experience
CN113436733A (en) * 2021-08-26 2021-09-24 肾泰网健康科技(南京)有限公司 Characteristic construction method of hemodialysis scheme generation model based on fusion experience
CN114582463A (en) * 2022-05-05 2022-06-03 成都尚医信息科技有限公司 Personalized motion guidance system and method based on machine learning
CN114937486A (en) * 2022-06-22 2022-08-23 肾泰网健康科技(南京)有限公司 Construction method and application of IDH prediction and intervention measure recommendation multitask model
CN114937486B (en) * 2022-06-22 2023-09-26 肾泰网健康科技(南京)有限公司 IDH prediction and intervention measure recommendation multi-task model construction method and application
CN116313127A (en) * 2023-03-23 2023-06-23 珠海市安克电子技术有限公司 Decision support system based on pre-hospital first-aid big data
CN116453706A (en) * 2023-06-14 2023-07-18 之江实验室 Hemodialysis scheme making method and system based on reinforcement learning
CN116453706B (en) * 2023-06-14 2023-09-08 之江实验室 Hemodialysis scheme making method and system based on reinforcement learning

Similar Documents

Publication Publication Date Title
CN111028913A (en) Hemodialysis treatment scheme aid decision-making method
Zhang ATTAIN: Attention-based time-aware LSTM networks for disease progression modeling.
Ambekar et al. Disease risk prediction by using convolutional neural network
Barbiero et al. Graph representation forecasting of patient's medical conditions: Toward a digital twin
Akter et al. Comprehensive performance assessment of deep learning models in early prediction and risk identification of chronic kidney disease
Zohora et al. Forecasting the risk of type ii diabetes using reinforcement learning
CN111798954A (en) Drug combination recommendation method based on time attention mechanism and graph convolution network
EP1534122A2 (en) Medical decision support systems utilizing gene expression and clinical information and method for use
Pujianto et al. Comparison of Naïve Bayes Algorithm and Decision Tree C4. 5 for Hospital Readmission Diabetes Patients using HbA1c Measurement.
CN116453706B (en) Hemodialysis scheme making method and system based on reinforcement learning
Pan et al. A self-correcting deep learning approach to predict acute conditions in critical care
CN115579141A (en) Interpretable disease risk prediction model construction method and disease risk prediction device
Sheikhalishahi et al. Benchmarking machine learning models on eICU critical care dataset
Srivastava et al. Hybrid diabetes disease prediction framework based on data imputation and outlier detection techniques
Diware et al. Severity-based hierarchical ECG classification using neural networks
Zhang et al. Time-aware adversarial networks for adapting disease progression modeling
CN115985513B (en) Data processing method, device and equipment based on multiple groups of chemical cancer typing
CN117038096A (en) Chronic disease prediction method based on low-resource medical data and knowledge mining
CN113077901A (en) Electronic medical record analysis device and method
Liu et al. Combination of neuro-fuzzy network models with biological knowledge for reconstructing gene regulatory networks
Usha et al. Revived ant colony optimization-based adaboost algorithm for heart disease and diabetes (hdd) prediction
CN114694841B (en) Adverse event risk prediction method based on patient electronic health record
CN113780394B (en) Training method, device and equipment for strong classifier model
Miazi et al. A cloud-based app for early detection of type II diabetes with the aid of deep learning
Xiao et al. Predicting ICU mortality from heterogeneous clinical events with prior medical knowledge

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination