CN117421678B - Single-lead atrial fibrillation recognition system based on knowledge distillation - Google Patents

Single-lead atrial fibrillation recognition system based on knowledge distillation Download PDF

Info

Publication number
CN117421678B
CN117421678B CN202311750008.5A CN202311750008A CN117421678B CN 117421678 B CN117421678 B CN 117421678B CN 202311750008 A CN202311750008 A CN 202311750008A CN 117421678 B CN117421678 B CN 117421678B
Authority
CN
China
Prior art keywords
lead
training module
student
identification network
network training
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202311750008.5A
Other languages
Chinese (zh)
Other versions
CN117421678A (en
Inventor
彭博
蔡家骏
王俞松
周秘
谭静
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Southwest Petroleum University
Chengdu Wenjiang District Peoples Hospital
Original Assignee
Southwest Petroleum University
Chengdu Wenjiang District Peoples Hospital
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Southwest Petroleum University, Chengdu Wenjiang District Peoples Hospital filed Critical Southwest Petroleum University
Priority to CN202311750008.5A priority Critical patent/CN117421678B/en
Publication of CN117421678A publication Critical patent/CN117421678A/en
Application granted granted Critical
Publication of CN117421678B publication Critical patent/CN117421678B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/24Detecting, measuring or recording bioelectric or biomagnetic signals of the body or parts thereof
    • A61B5/316Modalities, i.e. specific diagnostic methods
    • A61B5/318Heart-related electrical modalities, e.g. electrocardiography [ECG]
    • A61B5/346Analysis of electrocardiograms
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/24Detecting, measuring or recording bioelectric or biomagnetic signals of the body or parts thereof
    • A61B5/316Modalities, i.e. specific diagnostic methods
    • A61B5/318Heart-related electrical modalities, e.g. electrocardiography [ECG]
    • A61B5/346Analysis of electrocardiograms
    • A61B5/349Detecting specific parameters of the electrocardiograph cycle
    • A61B5/361Detecting fibrillation
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/72Signal processing specially adapted for physiological signals or for diagnostic purposes
    • A61B5/7235Details of waveform analysis
    • A61B5/7264Classification of physiological signals or data, e.g. using neural networks, statistical classifiers, expert systems or fuzzy systems
    • A61B5/7267Classification of physiological signals or data, e.g. using neural networks, statistical classifiers, expert systems or fuzzy systems involving training the classification device
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/10Pre-processing; Data cleansing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/213Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/217Validation; Performance evaluation; Active pattern learning techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/243Classification techniques relating to the number of classes
    • G06F18/2431Multiple classes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/253Fusion techniques of extracted features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/042Knowledge-based neural networks; Logical representations of neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/0464Convolutional networks [CNN, ConvNet]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/0495Quantised networks; Sparse networks; Compressed networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/096Transfer learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2218/00Aspects of pattern recognition specially adapted for signal processing
    • G06F2218/02Preprocessing
    • G06F2218/04Denoising
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2218/00Aspects of pattern recognition specially adapted for signal processing
    • G06F2218/08Feature extraction
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2218/00Aspects of pattern recognition specially adapted for signal processing
    • G06F2218/12Classification; Matching

Abstract

The invention relates to the field of artificial intelligence, in particular to a knowledge distillation-based single-lead atrial fibrillation recognition system, which comprises: the twelve-lead preprocessing module is used for preprocessing original twelve-lead data; the II lead extraction module is used for extracting the II leads from the data processed by the twelve-lead preprocessing module; the teacher recognizes the network training module, utilize ResNet34 network as teacher's network to train twelve lead data; the student identification network training module is used for carrying out knowledge distillation by taking a ResNet10 network as a student network and utilizing knowledge of the teacher identification network training module; and the middle layer knowledge transfer module based on the attention feature fusion utilizes an attention mechanism to perform feature fitting on two layers adjacent to the student identification network training module to form a new middle layer. The invention uses the knowledge distillation single-lead electrocardiograph recognition model, can reduce the network memory requirement while maintaining high performance, so that the method can be deployed in an environment with limited resources and has higher prediction speed.

Description

Single-lead atrial fibrillation recognition system based on knowledge distillation
Technical Field
The invention belongs to the field of artificial intelligence, and particularly relates to a knowledge distillation-based single-lead atrial fibrillation recognition system.
Background
Electrocardiography (ECG) is one of the most commonly used noninvasive techniques for recording fluctuations in cardiac bioelectric activity in medical diagnosis, plays a significant role in detecting arrhythmias, and can help doctors to perform targeted therapy on patients; however, in practical clinical diagnostics, experienced doctors often spend a lot of time accurately identifying the electrocardiogram, which brings inefficiency and difficulty to long-term monitoring, and as the population ages, the number of arrhythmic patients also presents explosive growth, which forces us to seek more efficient, accurate and cost-effective automated electrocardiographic diagnostic tools, the development of which is crucial for timely finding and treating heart-related problems, improving the quality of life of patients and reducing adverse consequences of cardiovascular disease.
Therefore, since the 60 s of the 20 th century, computer-aided electrocardiographic interpretation systems have emerged for heart rhythm diagnosis, and with the upgrading of software and hardware and the gradual development of artificial intelligence, the degree of automation of electrocardiographic recognition has been increasing, especially in the field of deep learning.
The deep learning model has great potential in treating arrhythmia, because the deep learning model can automatically learn characteristics and can process a large amount of data so as to improve accuracy and robustness, but the traditional electrocardiographic data identification models are all multi-lead or single-lead-based models, and the researches show that ECG (electrocardiogram) shows excellent performance on the multi-lead model and the single-lead model, however, the memory parameters of the multi-lead model are huge, and the single lead lacks multi-lead electrocardiographic information, so that the models are difficult to deploy on edge equipment due to the defect of poor performance; therefore, the invention provides a knowledge distillation-based single-lead atrial fibrillation recognition system, which can effectively solve or alleviate the problems.
Disclosure of Invention
Based on the problems, the invention provides a knowledge distillation-based single-lead atrial fibrillation recognition system, which solves the difficulty that the application of a deep learning multi-lead model on edge equipment is limited by huge model parameters and effectively improves the performance of the single-lead model, and comprises the following steps:
the twelve-lead preprocessing module is used for preprocessing each electrocardiographic data lead based on the original twelve-lead electrocardiographic database;
the II lead extraction module is used for extracting II leads of the twelve lead modules and is used for training of the student identification network training module;
the teacher identification network training module is used for training the preprocessed twelve-lead electrocardiograph data through a ResNet34 deep neural network to obtain a teacher identification network;
the middle layer knowledge transfer module is used for reconstructing the middle layer of the student identification network training module, and performing feature fusion on the middle layers of the adjacent student identification network training modules by using an attention mechanism to form a middle layer of a new student identification network training module;
the student identification network training module is used for training the preprocessed II lead electrocardiographic data through an intermediate layer knowledge transfer module based on attention feature fusion and a ResNet10 deep neural network under the intermediate layer and output guidance of the teacher identification network training module to obtain a student identification network;
in an embodiment, the twelve-lead preprocessing module is specifically configured to use a butterworth band-pass filtering method and smooth convolution for twelve-lead data in the original twelve-lead electrocardiograph database, and perform standardized processing on all data, so as to remove noise and burrs of the data and improve generalization of the data;
in an embodiment, the II-lead extraction module is specifically configured to extract, according to the position index, all the II-lead data of the twelve leads from the data processed by the twelve-lead preprocessing module, and only retain the data shape of a single lead;
in one embodiment, the teacher recognition network training module uses the twelve-lead data set processed by the twelve-lead preprocessing module to train and predict, in the data set, 8384 pieces of data and corresponding labels marked by the professional doctor are used as training sets, and 2099 pieces of data and corresponding labels marked by the professional doctor are used as test sets; the teacher identification network training module uses a ResNet34 deep neural network, all network layers are used for extracting characteristics of electrocardiographic data, each layer except a convolution layer and a feedforward neural network consists of a plurality of residual blocks so as to avoid the problem of gradient disappearance, and multi-classification cross entropy loss is used as a loss function during model training;
in an embodiment, the middle layer knowledge transfer module based on attention feature fusion is configured to perform feature fusion on each pair of adjacent residual layers in the student identification network training module by using an attention mechanism-based method to form a middle layer of a new student identification network training module, and perform co-location matching with the middle layer of the teacher identification network training module, so as to enrich feature tensor information of the middle layer of the student identification network;
in one embodiment, the student identification network training module uses all the II lead data sets extracted by the II lead extraction module to train and predict, in the data sets, 8384 pieces of single lead data and corresponding labels marked by the specialist doctor are used as training sets, and 2099 pieces of single lead data and corresponding labels marked by the specialist doctor are used as test sets; the student identification network training module uses a ResNet10 deep neural network, all network layers are used for extracting characteristics of electrocardiographic data, wherein each layer except a convolution layer and a feedforward neural network consists of a plurality of residual blocks so as to avoid the problem of gradient disappearance, and the residual layers are modified by an intermediate layer knowledge transfer module based on attention characteristic fusion; the model training process adopts multi-classification cross entropy loss, information tensor transmitted by the middle layer of the teacher recognition network training module and information tensor transmitted by the middle layer of the student recognition network training module under the middle layer knowledge transmission module based on attention feature fusion to perform loss calculation by using a feature pyramid formula, and performs KL divergence calculation based on the output probability of the teacher recognition network training module and the output probability of the student recognition network training module, wherein the three are combined to serve as a loss function.
Compared with the prior art, the system firstly uses the twelve-lead preprocessing module to preprocess the original twelve-lead data set, which eliminates noise and burrs in the twelve leads and improves generalization; dividing the preprocessed twelve-lead data into a training set and a testing set, inputting the training set into a teacher identification network training module for training, and verifying by the testing set to obtain an offline teacher identification model; and then the II leads in the preprocessed twelve leads are extracted by the II lead extraction module, and finally the II leads are input into the student identification network training module for final feature classification according to the same method, and the residual layer in the student identification network training module adopts an intermediate layer knowledge transfer module based on attention feature fusion, so that the technology can effectively reduce the size of the multi-lead data model while improving the accuracy.
Drawings
In order to more clearly illustrate the technical solution of the embodiments of the present invention, the following will briefly introduce the drawings that are needed in the description of the embodiments of the present invention or the prior art:
fig. 1 is a schematic structural diagram of a knowledge distillation-based single-lead atrial fibrillation recognition system, in which numbers are marked as serial numbers of modules, and the execution sequence of the system is indicated by combining line segments, in which 101 is the first operation of data processing, i.e. twelve-lead data after preprocessing is obtained, 102 is the second operation of data processing, II leads in the twelve-lead after preprocessing are extracted, 103 is the twelve-lead data obtained by using 101, the twelve-lead data are sent to a teacher recognition network training module for training, 105 is the II-lead data obtained by using 102 and sent to a student recognition network training module for training, 104 is the knowledge distillation of the student recognition network obtained by using 103 through an intermediate knowledge transfer module based on attention feature fusion, and finally the knowledge distillation-based single-lead atrial fibrillation recognition network is obtained;
FIG. 2 is a graph of the results of data processed by the twelve lead preprocessing module in one embodiment, wherein the labels in the graph represent 6 limb leads (I, II, III, aVR, aVL, aVF) and 6 chest leads (V1, V2, V3, V4, V5, V6);
FIG. 3 is a diagram of the extraction results of the II lead extraction module in one embodiment;
FIG. 4 is a schematic diagram of a knowledge distillation network in one embodiment;
FIG. 5 is a schematic diagram of a network structure of an intermediate layer knowledge transfer module based on attention-based feature fusion in an embodiment.
Description of the embodiments
In order to more clearly illustrate the technical solutions of the embodiments of the present invention, the drawings used in the description of the embodiments will be briefly described below; it is to be understood that the drawings in the following description are merely exemplary or embodiments of the invention and that it will be apparent to those skilled in the art that the invention may be applied to other similar situations in light of the accompanying drawings without undue effort, unless otherwise apparent from the language environment or otherwise indicated, in which like reference numerals refer to like structures or operations.
As used in the specification and claims, the terms "a," "an," "the," and/or "the" are not specific to a singular, but may include a plural number, unless the context clearly dictates otherwise; generally, the terms "comprises" and "comprising" merely indicate that the steps and elements are explicitly identified, and they do not constitute an exclusive list, as other steps or elements may be included in a method or apparatus.
It will be understood that when an element or module is referred to as being "connected," "coupled" to another element, module, or block, it can be directly connected or coupled or in communication with the other element, module, or block, or intervening elements, modules, or blocks may be present unless the context clearly dictates otherwise; the term "and/or" as used herein may include any and all combinations of one or more of the associated listed items.
As shown in fig. 1, an embodiment of the present invention provides a knowledge distillation-based single-lead atrial fibrillation recognition system, which includes a twelve-lead preprocessing module 101 for preprocessing each electrocardiographic data lead based on an original twelve-lead electrocardiographic database; the II lead extraction module 102 is used for extracting II leads of the twelve lead modules and training the student identification network training module; the teacher identification network training module 103 is used for training the preprocessed twelve-lead electrocardiographic data through a ResNet34 deep neural network to obtain a teacher identification network; the middle layer knowledge transfer module 104 is used for reconstructing the middle layer of the student identification network training module, and performing feature fusion on the adjacent middle layer of the student network by using an attention mechanism to form a middle layer of a new student identification network training module; the student identification network training module 105 is used for training the preprocessed II-lead electrocardiographic data through the intermediate layer knowledge transfer module based on attention feature fusion and the ResNet10 deep neural network under the intermediate layer and output guidance of the teacher identification network training module to obtain the student identification network.
In the embodiment, firstly, an original twelve-lead data set is subjected to data preprocessing by using a twelve-lead preprocessing module; this process eliminates noise and burrs in the twelve leads and improves generalization; next, dividing the preprocessed twelve-lead data into a training set and a testing set, inputting the training set into a teacher identification network training module for training, verifying by the testing set to obtain an offline teacher identification model, extracting II leads in the preprocessed twelve-lead data by using an II lead extraction module, inputting the II leads into a student identification network training module for final feature classification according to the same method, wherein a residual layer in the student identification network training module adopts an intermediate layer knowledge transfer module based on feature fusion of attention; therefore, the technology can effectively reduce the size of the multi-lead data model while improving the accuracy;
specifically, the original twelve lead data set includes twelve pieces of information data, I, II, III, AVR, AVL, AVF, V, V2, V3, V4, V5, and V6.
In an embodiment, the twelve-lead preprocessing module is specifically configured to use a butterworth band-pass filtering method and smooth convolution for twelve-lead data in the original twelve-lead electrocardiograph database, and perform standardized processing on all data, so as to remove noise and burrs of the data and improve generalization of the data;
as shown in fig. 2, fig. 2 is a graph of the result of data processed by the twelve-lead preprocessing module in the present invention, except for the above processing technique, on the premise that the PQRST wave of the II leads is the most obvious known information, all the leads are uniformly divided according to the R-wave index position of the II leads, so as to ensure that the data waveform can be matched with the R-wave index position, thereby providing a reliable data prediction result;
the specific pretreatment process comprises the following steps:
inputting each sample according to the twelve-lead sequence, and intercepting signal frequencies beyond 50 Hz-1 Hz by using a Butterworth band-pass filter because electrocardiograph data need to exclude myoelectric interference, power frequency interference and baseline drift;
removing burrs from each lead by using a convolution smoothing window, wherein the size of the window is 5, so that a relatively smooth image can be obtained, and the burr is prevented from excessively interfering with model identification;
finally, standardization is adopted for each twelve-lead sample, so that the generalization degree of the sample is improved.
In an embodiment, the II-lead extraction module is specifically configured to extract the data processed by the twelve-lead preprocessing module according to the position index of the II-leads, extract all II-lead data of the twelve-leads, and only preserve the data shape of a single lead;
as shown in fig. 3, fig. 3 is an extraction result diagram using the II lead extraction module after passing through the twelve lead preprocessing module; the extraction steps are as follows:
directly storing II data obtained by sequentially indexing twelve lead data into a one-dimensional data tensor, but not clearing other lead data;
and superposing all the extracted II leads according to the same dimension to form a new II lead data set.
In one embodiment, the teacher recognition network training module uses the twelve-lead data set processed by the twelve-lead preprocessing module to train and predict, in the data set, 8384 pieces of data and corresponding labels marked by the professional doctor are used as training sets, and 2099 pieces of data and corresponding labels marked by the professional doctor are used as test sets; the teacher identification network training module uses a ResNet34 deep neural network, all network layers are used for extracting characteristics of electrocardiographic data, wherein each layer except a convolution layer and a feedforward neural network consists of a plurality of residual blocks so as to avoid the problem of gradient disappearance; the multi-classification cross entropy loss is adopted as a loss function during model training;
the multi-class cross entropy loss function is:
wherein,representing the number of categories, the output of the model is a model containing +.>A vector of individual category scores; />Is a loss function; />Is the +.f. in the true class label vector>Element, if the sample belongs to +.>The number of categories is 1, otherwise 0; />Is the +.o in the output vector of the model>Element, representing model pair +>Prediction probabilities for individual categories;
as shown in res net34 (teacher recognition network) above fig. 4, the network receives data with tensor (12,2500), firstly, the data passes through a convolution layer with a convolution kernel of 7 and a step length of 2, then passes through a maximum pooling layer with a convolution kernel of 3 and a step length of 2, then passes through residual layers formed by 3,4,6,3 residual blocks respectively, and finally, the result is obtained by combining the average pooling layer with a full connection layer; the residual block consists of an input characteristic tensor and a characteristic tensor obtained by the characteristic tensor through a convolution layer with the convolution kernel of 3, and the output of one residual layer is set as an intermediate layer output tensor of the teacher identification network training module.
In an embodiment, the middle layer knowledge transfer module based on attention feature fusion is configured to perform feature fusion on each pair of adjacent residual layers in the student identification network training module by using an attention mechanism-based method to form a middle layer of a new student identification network training module, and perform co-location matching with the middle layer of the teacher identification network training module, so as to enrich feature tensor information of the middle layer of the student identification network training module;
as shown in fig. 5, the middle layer knowledge transfer module based on attention feature fusion fuses adjacent middle layer feature tensors of the residual layer output of the student identification network training module by using an attention mechanism-based method, and re-distributes the fused middle layer feature tensors as weights to two middle layers, and then forms a middle layer output of a new student identification network training module, as shown in a block of a res net10 (student identification network) below in fig. 4, the formula steps are as follows:
firstly, carrying out feature matching on two adjacent layers of a middle layer required to be output by a student identification network training module to unify tensors, then fusing the two tensors by using a concentration method, and then respectively transmitting fused results back to the original two layers as weights to finally obtain a new integrated output layer, wherein the method comprises the following steps:
wherein interpolate is used to match adjacent intermediate layer feature tensors, f j And f j+1 Front and rear feature tensors of adjacent intermediate layers respectively;
the AT is used for carrying out attention-based fusion on the feature layers after tensor matching, and the unsqueeze is used for increasing feature dimensions so as to facilitate the fusion;
and finally, respectively combining the new front and back layer images with the original front and back layers according to the sequence, and then combining to obtain better characteristic tensors.
In one embodiment, the student identification network training module uses all the II lead data sets extracted by the II lead extraction module to train and predict, in the data sets, 8384 pieces of single lead data and corresponding labels marked by the specialist doctor are used as training sets, and 2099 pieces of single lead data and corresponding labels marked by the specialist doctor are used as test sets; the student identification network training module uses a ResNet10 deep neural network, all network layers are used for extracting characteristics of electrocardiographic data, wherein each layer except a convolution layer and a feedforward neural network consists of a plurality of residual blocks so as to avoid the problem of gradient disappearance, and the residual layers are modified by an intermediate layer knowledge transfer module based on attention characteristic fusion; when the model is trained, the multi-classification cross entropy loss, the information tensor transmitted by the middle layer of the teacher recognition network training module and the information tensor of the middle layer of the student recognition network training module under the middle layer knowledge transmission module based on attention feature fusion are adopted to perform loss calculation by using a feature pyramid formula, the output probability of the teacher recognition network training module and the output probability of the student recognition network training module are used for performing KL divergence calculation, and the three are combined to be used as a loss function, and the formula steps are as follows:
matching the middle layer output of the new student identification network with the corresponding middle layer output of the offline teacher identification network, then fitting by using a feature pyramid, and finally calculating a loss function, wherein the formula is as follows:
wherein S is z Intermediate layer knowledge representing student identification network training module using attention-based feature fusionThe new student obtained by the transmission module identifies the middle layer of the network training module, wherein n is 5;
T z similarly, each middle layer output representing an offline teacher recognition network training module corresponding to a new middle layer tensor size of the student recognition network training module; performing loss calculation on the hierarchical structure context in the form of a feature pyramid for the layers of the two corresponding networks;
finally, calculating the total loss of the distillation network:
firstly, we extend the loss calculation of the hierarchical structure context of a certain corresponding layer in the teacher identification network training module and the student identification network training module after the intermediate layer knowledge transfer module based on the attention feature fusion to all required layers, and the total loss is as follows:
and then, using a knowledge distillation calculation method based on probability response to transmit knowledge to the student network from the result output by the offline teacher identification network, and calculating the loss of output probability:
wherein T is temperature, alpha is proportion, both variables are hyper-parameters, KLDivLoss means using KL divergence calculation, f student And f teacher Output tensors respectively representing output layers of the student network and the teacher network, stu_Loss representing output prediction Loss of the student network;
finally, the total loss is calculated:
wherein gamma is a superparameter, L KD And Loss of review The loss of the knowledge distillation based on the output probability and the loss of the knowledge distillation based on the intermediate layer characteristic tensor are respectively used as optimization targets, the loss function is minimized, and the distillation network model is optimized.
In one embodiment, 1 ten thousand samples collected from the original twelve lead dataset include 3889 sinus bradycardia, 1826 sinus rhythms, 1780 atrial fibrillation, 1568 sinus tachycardia, 445 atrial flutter, 399 sinus arrhythmia, 587 supraventricular tachycardia, all samples of which are used according to a 8:2 segmentation ratio to obtain a training set and a test set, the training set is sent to a teacher identification network training module and a student identification network training module of an intermediate knowledge transfer module based on attention feature fusion through a knowledge distillation method, the teacher identification network training module carries out knowledge transfer on the student identification network training module of the intermediate knowledge transfer module based on attention feature fusion, finally obtains a single lead atrial fibrillation identification network based on knowledge distillation, and the single lead atrial fibrillation identification network is tested by using the test set, and performance tests are carried out in the case of atrial fibrillation classification and seven classification respectively as shown in the following table 1;
table one: performance of single-lead atrial fibrillation recognition system based on knowledge distillation under two-classification and seven-classification standards of atrial fibrillation
In summary, the invention provides a knowledge distillation-based single-lead atrial fibrillation recognition system with better classification performance.
The technical features of the above embodiments may be arbitrarily combined, and all possible combinations of the technical features in the above embodiments are not described for brevity of description, however, as long as there is no contradiction between the combinations of the technical features, they should be considered as the scope of the description.
The foregoing examples merely represent several embodiments of the present application, which are described in more detail and are not to be construed as limiting the scope of the invention; it should be noted that it is possible for a person skilled in the art to make several variations and modifications without departing from the concept of the present application, which are all within the scope of protection of the present application; accordingly, the scope of protection of the present application is to be determined by the claims appended hereto.

Claims (1)

1. A knowledge distillation-based single-lead atrial fibrillation recognition system, comprising at least the following steps:
the twelve-lead preprocessing module is used for realizing noise reduction preprocessing on each electrocardiographic data lead based on an original twelve-lead electrocardiographic database;
the II lead extraction module is used for extracting II leads of the twelve lead modules according to the position index, and superposing the II leads according to the same dimension to form a new II lead data set for training of the student identification network training module;
the teacher identification network training module is used for training the preprocessed twelve-lead electrocardiograph data through a ResNet34 deep neural network formed by two-dimensional convolution blocks to obtain a teacher identification network;
the middle layer knowledge transfer module based on the attention feature fusion is used for reconstructing the middle layer of the student identification network training module, performing the feature fusion on the middle layers of the adjacent student identification network training modules by using an attention mechanism, then, re-distributing the middle layers as weights, and finally, forming a middle layer output of a new student identification network training module;
the student identification network training module is transformed by the middle layer knowledge transfer module based on attention feature fusion, under the middle layer and output guidance of the teacher identification network training module, the preprocessed II lead electrocardiographic data is sent into the ResNet10 deep neural network formed by one-dimensional convolution blocks to be trained to obtain a student identification network, and finally, the student identification network is used for atrial fibrillation identification of two classifications and seven classifications, and the specific details are as follows:
the residual layer of the student identification network training module is transformed by the middle layer knowledge transfer module based on attention feature fusion to obtain a new middle layer of the student identification network training module, then multi-classification cross entropy loss is adopted during model training, the information tensor transferred by the middle layer of the teacher identification network training module and the information tensor transferred by the middle layer of the student identification network training module under the middle layer knowledge transfer module based on attention feature fusion are used for carrying out loss calculation by using a feature pyramid formula, and KL divergence calculation is carried out based on the output probability of the teacher identification network training module and the output probability of the student identification network training module, and the three are combined to serve as a loss function, wherein the steps and calculation formulas are as follows:
firstly, carrying out feature matching on two adjacent layers of a middle layer required to be output by a student identification network training module to unify tensors, then fusing the two tensors by using a concentration method, and then respectively transmitting fused results back to the original two layers as weights to finally obtain a new integrated output layer, wherein the method comprises the following steps:
f i =interpolate(f j );
f i+1 =interpolate(f j+1 );
wherein interpolate is used to match adjacent intermediate layer feature tensors, f j And f j+1 Front and rear feature tensors of adjacent intermediate layers respectively;
F j =AT(unsqueeze(f i ),unsqueeze(f i+1 ));
the AT is used for carrying out attention-based fusion on the feature layers after tensor matching, and the unsqueeze is used for increasing feature dimensions so as to facilitate the fusion;
finally, combining the new front and back layer images with the original front and back layers respectively according to the sequence, and then combining to obtain better characteristic tensors;
and matching the middle layer output of the new student identification network with the corresponding middle layer output of the offline teacher identification network, fitting by using a characteristic pyramid, and finally calculating a loss function, wherein the formula is as follows:
S z =s 1 ,s 2 ,s 3 ,...,s n
T z =t 1 ,t 2 ,t 3 ,...,t n
wherein S is z Representing a middle layer of a new student identification network training module obtained by the student identification network training module by using a middle layer knowledge transfer module based on attention feature fusion, wherein n is 5;
T z similarly, each middle layer output representing an offline teacher recognition network training module corresponding to a new middle layer tensor size of the student recognition network training module; performing loss calculation on the hierarchical structure context in the form of a feature pyramid for the layers of the two corresponding networks;
finally, calculating the total loss of the distillation network:
firstly, we extend the loss calculation of the hierarchical structure context of a certain corresponding layer in the teacher identification network training module and the student identification network training module after the intermediate layer knowledge transfer module based on the attention feature fusion to all required layers, and the total loss is as follows:
and then, using a knowledge distillation calculation method based on probability response to transmit knowledge to the student network from the result output by the offline teacher identification network, and calculating the loss of output probability:
Distillation_Loss=KLDivLoss(f student ,f teacher *(T^2)/f student .shape[0];
L KD =α*Stu_Loss+(1-α)*Distillation_Loss;
wherein T is temperature, alpha is proportion, both variables are hyper-parameters, KLDivLoss means using KL divergence calculation, f student And f teacher Output tensors respectively representing output layers of the student network and the teacher network, stu_Loss representing output prediction Loss of the student network;
finally, the total loss is calculated:
Loss_all=γ*L KD +(1-γ)*Loss review
wherein gamma is a superparameter, L KD And Loss of review The loss of the knowledge distillation based on the output probability and the loss of the knowledge distillation based on the intermediate layer characteristic tensor are respectively used as optimization targets, the loss function is minimized, and the distillation network model is optimized.
CN202311750008.5A 2023-12-19 2023-12-19 Single-lead atrial fibrillation recognition system based on knowledge distillation Active CN117421678B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202311750008.5A CN117421678B (en) 2023-12-19 2023-12-19 Single-lead atrial fibrillation recognition system based on knowledge distillation

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202311750008.5A CN117421678B (en) 2023-12-19 2023-12-19 Single-lead atrial fibrillation recognition system based on knowledge distillation

Publications (2)

Publication Number Publication Date
CN117421678A CN117421678A (en) 2024-01-19
CN117421678B true CN117421678B (en) 2024-03-22

Family

ID=89525216

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202311750008.5A Active CN117421678B (en) 2023-12-19 2023-12-19 Single-lead atrial fibrillation recognition system based on knowledge distillation

Country Status (1)

Country Link
CN (1) CN117421678B (en)

Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110379506A (en) * 2019-06-14 2019-10-25 杭州电子科技大学 The cardiac arrhythmia detection method of binaryzation neural network is used for ECG data
CN111488793A (en) * 2020-02-21 2020-08-04 广州视源电子科技股份有限公司 Electrocardiosignal classification method and device, electronic equipment and storage medium
CN112418343A (en) * 2020-12-08 2021-02-26 中山大学 Multi-teacher self-adaptive joint knowledge distillation
CN113288162A (en) * 2021-06-03 2021-08-24 北京航空航天大学 Short-term electrocardiosignal atrial fibrillation automatic detection system based on self-adaptive attention mechanism
CN116071635A (en) * 2023-03-06 2023-05-05 之江实验室 Image recognition method and device based on structural knowledge propagation
CN116090503A (en) * 2022-12-23 2023-05-09 北京鹰瞳科技发展股份有限公司 Method for training neural network model based on knowledge distillation and related products
CN116257751A (en) * 2023-02-23 2023-06-13 安徽理工大学 Distillation method and device based on online cooperation and feature fusion
CN116260642A (en) * 2023-02-27 2023-06-13 南京邮电大学 Knowledge distillation space-time neural network-based lightweight Internet of things malicious traffic identification method
CN116796810A (en) * 2023-06-28 2023-09-22 河海大学 Deep neural network model compression method and device based on knowledge distillation
CN116824334A (en) * 2023-06-25 2023-09-29 中国科学院软件研究所 Model back door attack countermeasure method based on frequency domain feature fusion reconstruction
CN116913504A (en) * 2023-07-13 2023-10-20 重庆理工大学 Self-supervision multi-view knowledge distillation method for single-lead arrhythmia diagnosis
WO2023212997A1 (en) * 2022-05-05 2023-11-09 五邑大学 Knowledge distillation based neural network training method, device, and storage medium
CN117132870A (en) * 2023-10-25 2023-11-28 西南石油大学 Wing icing detection method combining CenterNet and mixed attention

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB201910657D0 (en) * 2019-07-25 2019-09-11 Univ Oxford Innovation Ltd Deep end-to-end classification of electrocardiogram data

Patent Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110379506A (en) * 2019-06-14 2019-10-25 杭州电子科技大学 The cardiac arrhythmia detection method of binaryzation neural network is used for ECG data
CN111488793A (en) * 2020-02-21 2020-08-04 广州视源电子科技股份有限公司 Electrocardiosignal classification method and device, electronic equipment and storage medium
CN112418343A (en) * 2020-12-08 2021-02-26 中山大学 Multi-teacher self-adaptive joint knowledge distillation
CN113288162A (en) * 2021-06-03 2021-08-24 北京航空航天大学 Short-term electrocardiosignal atrial fibrillation automatic detection system based on self-adaptive attention mechanism
WO2023212997A1 (en) * 2022-05-05 2023-11-09 五邑大学 Knowledge distillation based neural network training method, device, and storage medium
CN116090503A (en) * 2022-12-23 2023-05-09 北京鹰瞳科技发展股份有限公司 Method for training neural network model based on knowledge distillation and related products
CN116257751A (en) * 2023-02-23 2023-06-13 安徽理工大学 Distillation method and device based on online cooperation and feature fusion
CN116260642A (en) * 2023-02-27 2023-06-13 南京邮电大学 Knowledge distillation space-time neural network-based lightweight Internet of things malicious traffic identification method
CN116071635A (en) * 2023-03-06 2023-05-05 之江实验室 Image recognition method and device based on structural knowledge propagation
CN116824334A (en) * 2023-06-25 2023-09-29 中国科学院软件研究所 Model back door attack countermeasure method based on frequency domain feature fusion reconstruction
CN116796810A (en) * 2023-06-28 2023-09-22 河海大学 Deep neural network model compression method and device based on knowledge distillation
CN116913504A (en) * 2023-07-13 2023-10-20 重庆理工大学 Self-supervision multi-view knowledge distillation method for single-lead arrhythmia diagnosis
CN117132870A (en) * 2023-10-25 2023-11-28 西南石油大学 Wing icing detection method combining CenterNet and mixed attention

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
Distilling Knowledge via Knowledge Review;Pengguang Chen等;《arXiv:2104.09044v1》;20210419;第1-10页第1-3节 *
面向房颤信号提取网络的知识蒸馏方法研究;谢卓延;《中国优秀硕士学位论文全文数据库 医药卫生科技辑》;20230115;E062-475 *

Also Published As

Publication number Publication date
CN117421678A (en) 2024-01-19

Similar Documents

Publication Publication Date Title
Kachuee et al. Ecg heartbeat classification: A deep transferable representation
JP7091451B2 (en) Automatic ECG analysis method based on artificial intelligence self-learning, equipment used to execute the analysis method, computer program products and computer-readable storage media
JP6986724B2 (en) ECG interference identification method based on artificial intelligence
US11617528B2 (en) Systems and methods for reduced lead electrocardiogram diagnosis using deep neural networks and rule-based systems
Alquran et al. ECG classification using higher order spectral estimation and deep learning techniques
CN111990989A (en) Electrocardiosignal identification method based on generation countermeasure and convolution cyclic network
Musa et al. A systematic review and Meta-data analysis on the applications of Deep Learning in Electrocardiogram
CN112906748A (en) 12-lead ECG arrhythmia detection classification model construction method based on residual error network
Zhou et al. ECG quality assessment using 1D-convolutional neural network
Zhang et al. Semi-supervised learning for automatic atrial fibrillation detection in 24-hour Holter monitoring
Deevi et al. HeartNetEC: a deep representation learning approach for ECG beat classification
Wang et al. Multiscale residual network based on channel spatial attention mechanism for multilabel ECG classification
Al-Huseiny et al. Diagnosis of arrhythmia based on ECG analysis using CNN
CN114041800A (en) Electrocardiosignal real-time classification method and device and readable storage medium
Xu et al. Arrhythmia detection using gated recurrent unit network with ECG signals
Tung et al. Multi-lead ECG classification via an information-based attention convolutional neural network
CN117421678B (en) Single-lead atrial fibrillation recognition system based on knowledge distillation
JP2023104885A (en) Electrocardiographic heart rate multi-type prediction method based on graph convolution
CN113171102B (en) ECG data classification method based on continuous deep learning
Lu et al. A New Multichannel Parallel Network Framework for the Special Structure of Multilead ECG
Zhang et al. Multi-scale and multi-channel information fusion for exercise electrocardiogram feature extraction and classification
Bengherbia et al. Real-time smart system for ecg monitoring using a one-dimensional convolutional neural network
Yin et al. An algorithm for locating pvc and spb in wearable ecgs
Zhang et al. Multi-scale and attention based ResNet for heartbeat classification
Xie et al. Arrhythmia Detection Based on Semantic Segmentation for Multi-lead ECG

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant