CN112837741B - Protein secondary structure prediction method based on cyclic neural network - Google Patents

Protein secondary structure prediction method based on cyclic neural network Download PDF

Info

Publication number
CN112837741B
CN112837741B CN202110097155.1A CN202110097155A CN112837741B CN 112837741 B CN112837741 B CN 112837741B CN 202110097155 A CN202110097155 A CN 202110097155A CN 112837741 B CN112837741 B CN 112837741B
Authority
CN
China
Prior art keywords
protein sequence
secondary structure
residue
protein
matrix
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202110097155.1A
Other languages
Chinese (zh)
Other versions
CN112837741A (en
Inventor
胡俊
殷文杰
贾宁欣
曾文武
董明
张贵军
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang University of Technology ZJUT
Original Assignee
Zhejiang University of Technology ZJUT
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang University of Technology ZJUT filed Critical Zhejiang University of Technology ZJUT
Priority to CN202110097155.1A priority Critical patent/CN112837741B/en
Publication of CN112837741A publication Critical patent/CN112837741A/en
Application granted granted Critical
Publication of CN112837741B publication Critical patent/CN112837741B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16BBIOINFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR GENETIC OR PROTEIN-RELATED DATA PROCESSING IN COMPUTATIONAL MOLECULAR BIOLOGY
    • G16B15/00ICT specially adapted for analysing two-dimensional or three-dimensional molecular structures, e.g. structural or functional relations or structure alignment
    • G16B15/20Protein or domain folding
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16BBIOINFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR GENETIC OR PROTEIN-RELATED DATA PROCESSING IN COMPUTATIONAL MOLECULAR BIOLOGY
    • G16B40/00ICT specially adapted for biostatistics; ICT specially adapted for bioinformatics-related machine learning or data mining, e.g. knowledge discovery or pattern finding
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02ATECHNOLOGIES FOR ADAPTATION TO CLIMATE CHANGE
    • Y02A90/00Technologies having an indirect contribution to adaptation to climate change
    • Y02A90/10Information and communication technologies [ICT] supporting adaptation to climate change, e.g. for weather forecasting or climate simulation

Landscapes

  • Life Sciences & Earth Sciences (AREA)
  • Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • Medical Informatics (AREA)
  • Engineering & Computer Science (AREA)
  • Spectroscopy & Molecular Physics (AREA)
  • General Health & Medical Sciences (AREA)
  • Theoretical Computer Science (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Biotechnology (AREA)
  • Evolutionary Biology (AREA)
  • Biophysics (AREA)
  • Artificial Intelligence (AREA)
  • Chemical & Material Sciences (AREA)
  • Crystallography & Structural Chemistry (AREA)
  • Bioethics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • Epidemiology (AREA)
  • Evolutionary Computation (AREA)
  • Public Health (AREA)
  • Software Systems (AREA)
  • Investigating Or Analysing Biological Materials (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

Protein based on cyclic neural networkAccording to the input protein sequence to be subjected to secondary structure prediction, the protein sequence is encoded by a single thermal encoding mode to obtain an L multiplied by 20 characteristic matrix M 1 Generating a protein position-specific scoring matrix PSSM using a PSI-BLAST program; then, M is 1 Performing matrix addition operation with PSSM to obtain matrix M 2 The method comprises the steps of carrying out a first treatment on the surface of the Secondly, obtaining a characteristic vector of the residue; thirdly, constructing a circulating neural network framework, constructing a sample set by acquiring protein sequences of known secondary structure labels from a PDB database, and training the constructed circulating neural network; and finally, inputting the feature vector of the residue in the protein sequence to be predicted into a trained model, and predicting the secondary structure category of the residue in the protein sequence according to the output probability value. The method has low calculation cost and high prediction accuracy.

Description

Protein secondary structure prediction method based on cyclic neural network
Technical Field
The invention relates to the fields of bioinformatics, deep learning and computer application, in particular to a protein secondary structure prediction method based on a circulating neural network.
Background
Prediction of the secondary structure of a protein is an important intermediate step in the prediction of tertiary structure, and is also a bridge linking the protein sequence and tertiary structure. Accurate identification of the secondary structure of the protein can help people to know complex dependency relationship between protein sequences and tertiary structures, and can also promote functional analysis of the protein and drug design.
Currently, the methods for predicting the secondary structure of proteins by deep learning are as follows: SSpro8 (Gianluca, pollastri, darisz, et al, improving the prediction of protein secondary structure in three and eight classes using recurrent neural networks and profiles [ J ]. Proteins Structure Function & Genetics, 2002: gianluca et al SSpro 8. Prediction of class 3 and class 8 protein secondary structures using cyclic neural networks and spectrum files [ J ]. Protein structure, function and Genetics,2002 ], BLSTM (Snderby S K, winther O.protein Secondary Structure Prediction with Long Short Term Memory Networks [ J ]. Computer Science,2014. I.e., snderby S K, winther O.BLSTM: prediction of class 3 and class 8 protein secondary structures based on long and short time memory networks [ J ]. Computer Science,2014 ], GSN (Zhou J, troynsky O.Deep Supervised and Convolutional Generative Stochastic Network for Protein Secondary Structure Prediction [ J ]. Computer Science,2014:745-753. I.e., zhou J ]. Troynsky O G. Prediction of class III secondary structures based on deep-supervision convolution and CRNs [ J ]. 2014:745-753. ]) are performed with (Buzzoz, jnyan [ L ]. 35 of class III.e., 35) of new biological structures of Buzzo [ J ]. 35, 35 mg [ J ]. 1-92 ] of Buzzo.e.g. biological structures of Buzzo [ J ]. Three classes of Bungan [ J ]. 35 ] were studied. Compared with the traditional machine learning method, the deep learning-based method can more fully extract the amino acid characteristics and the hidden modes in the protein sequence. Although existing protein secondary structure prediction methods have achieved good results, these methods only extract features along the amino acid residue dimension. Thus, these methods may ignore some important features hidden in the protein sequence feature vector, which may be useful for predicting secondary structure.
In summary, the existing protein secondary structure prediction methods have great differences from the actual application requirements in terms of calculation cost and prediction accuracy, and urgent improvement is needed.
Disclosure of Invention
In order to overcome the defects of the existing protein secondary structure prediction method in terms of calculation cost and prediction accuracy, the invention provides a protein secondary structure prediction method based on a cyclic neural network, which is low in calculation cost and high in prediction accuracy.
The technical scheme adopted for solving the technical problems is as follows:
a method for predicting protein secondary structure based on a recurrent neural network, the method comprising the steps of:
1) Inputting a protein sequence with L residues and to be subjected to protein secondary structure prediction, and marking the protein sequence as S;
2) The protein sequence S is encoded by a single thermal encoding mode to obtain a characteristic matrix with the size of L multiplied by 20, which is marked as M 1 ,M 1 The ith row and jth column elements of (a) are expressed as:
wherein A is 1 ,A 2 ,...,A 20 Residues representing 20 common amino acid types, T i Indicating the type of the i-th residue in sequence S;
3) Protein sequence S was scored using the PSI-BLAST (https:// www.ebi.ac.uk/Tools/sss/psiblast /) program to generate a protein position-specific scoring matrix, designated PSSM;
4) For M 1 Performing matrix addition operation with PSSM to obtain a matrix with size of Lx20, and recording as M 2
5) Calculating matrix M 2 The average value of each row in the row (C) is obtained, and a vector with the size of L multiplied by 1 is obtained and is recorded as H;
6) Calculating matrix M 2 Obtaining a vector with the size of L multiplied by 1, and recording the vector as V;
7) For the protein sequence S, a feature matrix F with the size of L multiplied by 20 is obtained through the following formula, and the elements in the ith row and the jth column in F are expressed as:
wherein,represents M 2 The ith row and the jth column elements of H i Represents the i-th element in H, V i Represents the i-th element in V;
8) Any row F in matrix F i I=1, 2, L, a feature vector representing the i-th residue in the protein sequence S;
9) Obtaining a protein sequence of a known residue secondary structure class from a PDB (https:// www.rcsb.org /) database as a training set, wherein the residue secondary structure class comprises burial, intermediate state and exposure, and generating feature vectors of all residues in the training set by using the steps 2) -8); combining the residue secondary structure class labels to construct a training sample set;
10 A second-level structure class of residues in the predicted protein sequence S of the cyclic neural network is built, the network consists of three parts, wherein the first part is a convolution layer part which consists of a convolution layer with a convolution kernel size of 1 multiplied by 20, a convolution layer with a convolution kernel size of 3 multiplied by 1, a normalization layer and a pooling layer; the second part is a recycled layer part, which consists of two LSTM layers; the last part is a fully connected layer part which consists of two fully connected layers, wherein the output of each layer is used as the input of the next layer, and the sigmoid activation function is used to enable the output value of the network to be in the range of (0, 1);
11 Training the cyclic neural network constructed in step 10) using the training sample set constructed in step 9), the training stage adjusting parameters in the network using a cross entropy loss function, the cross entropy loss function being noted as:
wherein u represents the true tag of the residue to be detected in the protein sequence,representing an output value corresponding to the type of the predicted residue of the network model, wherein Y represents the difference between the predicted output and the real label;
12 Inputting the feature vector of the residue in the sequence S into the model trained in the step 11), outputting the probability value of each residue secondary structure category according to the model, wherein the category corresponding to the maximum probability value is the prediction category of the residue secondary structure.
The technical conception of the invention is as follows: firstly, according to an input protein sequence to be subjected to secondary structure prediction, encoding the protein sequence by using a single thermal encoding mode to obtain an L multiplied by 20 characteristic matrix M 1 Generating a protein position-specific scoring matrix PSSM using a PSI-BLAST program; then, M is 1 Performing matrix addition operation with PSSM to obtain matrix M 2 The method comprises the steps of carrying out a first treatment on the surface of the Next, the eigenvectors of the residues are obtained by: calculating matrix M 2 Mean and variance of each row, matrix M 2 Subtracting the average value of the current row from any element, dividing the average value by the variance of the current row, and generating a corresponding element in a feature matrix F, wherein any row in F represents a feature vector of a corresponding residue; thirdly, constructing a circulating neural network framework, constructing a sample set by acquiring protein sequences of known secondary structure labels from a PDB database, and training the constructed circulating neural network; and finally, inputting the feature vector of the residue in the protein sequence to be predicted into a trained model, and predicting the secondary structure category of the residue in the protein sequence according to the output probability value. The invention provides a protein secondary structure prediction method based on a cyclic neural network, which has low calculation cost and high prediction accuracy.
The beneficial effects of the invention are as follows: on one hand, the convolution part of the network uses asymmetric convolution for extracting the connection between residue parts, so that preparation is made for further improving the accuracy of protein secondary structure prediction; on the other hand, the BLSTM part of the network is used for capturing long-distance interdependence relation among residues, so that the accuracy of protein secondary structure prediction is better ensured.
Drawings
FIG. 1 is a schematic diagram of a method for predicting a protein secondary structure based on a recurrent neural network.
FIG. 2 is a graph showing the results of secondary structure prediction of protein sequence 1F2C using a cyclic neural network-based protein secondary structure prediction method.
Detailed Description
The invention is further described below with reference to the accompanying drawings.
Referring to fig. 1 and 2, a protein secondary structure prediction method based on a recurrent neural network includes the steps of:
1) Inputting a protein sequence with L residues and to be subjected to protein secondary structure prediction, and marking the protein sequence as S;
2) The protein sequence S is encoded by a single thermal encoding mode to obtain a characteristic matrix with the size of L multiplied by 20, which is marked as M 1 ,M 1 The ith row and jth column elements of (a) are expressed as:
wherein A is 1 ,A 2 ,...,A 20 Residues representing 20 common amino acid types, T i Indicating the type of the i-th residue in sequence S;
3) Protein sequence S was scored using the PSI-BLAST (https:// www.ebi.ac.uk/Tools/sss/psiblast /) program to generate a protein position-specific scoring matrix, designated PSSM;
4) For M 1 Performing matrix addition operation with PSSM to obtain a matrix with size of Lx20, and recording as M 2
5) Calculating matrix M 2 The average value of each row in the row (C) is obtained, and a vector with the size of L multiplied by 1 is obtained and is recorded as H;
6) Calculating matrix M 2 Obtaining a vector with the size of L multiplied by 1, and recording the vector as V;
7) For the protein sequence S, a feature matrix F with the size of L multiplied by 20 is obtained through the following formula, and the elements in the ith row and the jth column in F are expressed as:
wherein,represents M 2 The ith row and the jth column elements of H i Represents the i-th element in H, V i Represents the i-th element in V;
8) Any row F in matrix F i I=1, 2, L, a feature vector representing the i-th residue in the protein sequence S;
9) Obtaining a protein sequence of a known residue secondary structure class from a PDB (https:// www.rcsb.org /) database as a training set, wherein the residue secondary structure class comprises burial, intermediate state and exposure, and generating feature vectors of all residues in the training set by using the steps 2) -8); combining the residue secondary structure class labels to construct a training sample set;
10 A second-level structure class of residues in the predicted protein sequence S of the cyclic neural network is built, the network consists of three parts, wherein the first part is a convolution layer part which consists of a convolution layer with a convolution kernel size of 1 multiplied by 20, a convolution layer with a convolution kernel size of 3 multiplied by 1, a normalization layer and a pooling layer; the second part is a recycled layer part, which consists of two LSTM layers; the last part is a fully connected layer part which consists of two fully connected layers, wherein the output of each layer is used as the input of the next layer, and the sigmoid activation function is used to enable the output value of the network to be in the range of (0, 1);
11 Training the cyclic neural network constructed in step 10) using the training sample set constructed in step 9), the training stage adjusting parameters in the network using a cross entropy loss function, the cross entropy loss function being noted as:
wherein u represents the true tag of the residue to be detected in the protein sequence,representing an output value corresponding to the type of the predicted residue of the network model, wherein Y represents the difference between the predicted output and the real label;
12 Inputting the feature vector of the residue in the sequence S into the model trained in the step 11), outputting the probability value of each residue secondary structure category according to the model, wherein the category corresponding to the maximum probability value is the prediction category of the residue secondary structure.
The embodiment takes the prediction of the secondary structure of the protein 1F2C as an embodiment, and the prediction method of the secondary structure of the protein based on the circulating neural network comprises the following steps:
1) Inputting a protein sequence 1F2C with the number of residues 215 to be subjected to protein secondary structure prediction, and marking the protein sequence as S;
2) The protein sequence S is encoded by a single thermal encoding method to obtain a matrix with the size of 215 multiplied by 20, which is denoted as M 1 ,M 1 The ith row and jth column elements of (b) are represented by the following equations:
wherein A is 1 ,A 2 ,...,A 20 Representing 20 common amino acid types, T i Represents the type of the i-th residue in the protein sequence 1F 2C;
3) Protein sequence S was scored using the PSI-BLAST (https:// www.ebi.ac.uk/Tools/sss/psiblast /) program to generate a protein position-specific scoring matrix, designated PSSM;
4) For M 1 Performing matrix addition operation with PSSM to obtain a matrix with size of Lx20, and recording as M 2
5) Calculating matrix M 2 The average value of each row in the row (C) is obtained, and a vector with the size of L multiplied by 1 is obtained and is recorded as H;
6) Calculating matrix M 2 Obtaining a vector with the size of L multiplied by 1, and recording the vector as V;
7) For the protein sequence S, a feature matrix F with the size of L multiplied by 20 is obtained through the following formula, and the elements in the ith row and the jth column in F are expressed as:
wherein,represents M 2 The ith row and the jth column elements of H i Represents the i-th element in H, V i Represents the i-th element in V;
8) Any row F in matrix F i I=1, 2, L, a feature vector representing the i-th residue in the protein sequence S;
9) Obtaining a protein sequence of a known residue secondary structure class from a PDB (https:// www.rcsb.org /) database as a training set, wherein the residue secondary structure class comprises burial, intermediate state and exposure, and generating feature vectors of all residues in the training set by using the steps 2) -8); combining the residue secondary structure class labels to construct a training sample set;
10 A second-level structure class of residues in the predicted protein sequence S of the cyclic neural network is built, the network consists of three parts, wherein the first part is a convolution layer part which consists of a convolution layer with a convolution kernel size of 1 multiplied by 20, a convolution layer with a convolution kernel size of 3 multiplied by 1, a normalization layer and a pooling layer; the second part is a recycled layer part, which consists of two LSTM layers; the last part is a fully connected layer part which consists of two fully connected layers, wherein the output of each layer is used as the input of the next layer, and the sigmoid activation function is used to enable the output value of the network to be in the range of (0, 1);
11 Training the cyclic neural network constructed in step 10) using the training sample set constructed in step 9), the training stage adjusting parameters in the network using a cross entropy loss function, the cross entropy loss function being noted as:
wherein u represents the true tag of the residue to be detected in the protein sequence,representing an output value corresponding to the type of the predicted residue of the network model, wherein Y represents the difference between the predicted output and the real label;
12 Inputting the feature vector of the residue in the protein sequence S into the model trained in the step 11), outputting the probability value of each residue secondary structure category according to the model, wherein the category corresponding to the maximum probability value is the prediction category of the residue secondary structure.
The prediction of the protein sequence 1F2C using the above method for partitioning was performed using the secondary structure prediction of the protein sequence 1F2C as shown in FIG. 2.
The above description is that the predicted result obtained by taking the secondary structure of the protein sequence 1F2C as an example of the present invention is not limited to the implementation scope of the present invention, and various modifications and improvements can be made thereto without departing from the scope of the present invention, which is not to be excluded from the protection scope of the present invention.

Claims (1)

1. A method for predicting a protein secondary structure based on a recurrent neural network, the method comprising the steps of:
1) Inputting a protein sequence with L residues and to be subjected to protein secondary structure prediction, and marking the protein sequence as S;
2) The protein sequence S is encoded by a single thermal encoding mode to obtain a characteristic matrix with the size of L multiplied by 20, which is marked as M 1 ,M 1 The ith row and jth column elements of (a) are expressed as:
wherein A is 1 ,A 2 ,...,A 20 Residues representing 20 common amino acid types, T i Indicating the type of the i-th residue in sequence S;
3) Protein sequence S was scored using the PSI-BLAST program to generate a protein position-specific scoring matrix, designated PSSM;
4) For M 1 And PSSM performs matrix addition to obtain a matrix with L×20, denoted as M 2
5) Calculating matrix M 2 The average value of each row in the row (C) is obtained, and a vector with the size of L multiplied by 1 is obtained and is recorded as H;
6) Calculating matrix M 2 Obtaining a vector with the size of L multiplied by 1, and recording the vector as V;
7) For the protein sequence S, a feature matrix F with the size of L multiplied by 20 is obtained through the following formula, and the elements in the ith row and the jth column in F are expressed as:
wherein,represents M 2 The ith row and the jth column elements of H i Represents the i-th element in H, V i Represents the i-th element in V;
8) Any row F in matrix F i I=1, 2, L, a feature vector representing the i-th residue in the protein sequence S;
9) Obtaining a protein sequence of a secondary structure class of known residues from a PDB database as a training set, wherein the secondary structure class of the residues comprises burial, intermediate state and exposure, and generating feature vectors of all residues in the training set by using the steps 2) -8); combining the residue secondary structure class labels to construct a training sample set;
10 A second-level structure class of residues in the predicted protein sequence S of the cyclic neural network is built, the network consists of three parts, wherein the first part is a convolution layer part which consists of a convolution layer with a convolution kernel size of 1 multiplied by 20, a convolution layer with a convolution kernel size of 3 multiplied by 1, a normalization layer and a pooling layer; the second part is a recycled layer part, which consists of two LSTM layers; the last part is a fully connected layer part which consists of two fully connected layers, wherein the output of each layer is used as the input of the next layer, and the sigmoid activation function is used to enable the output value of the network to be in the range of (0, 1);
11 Training the network constructed in step 10) using the training sample set constructed in step 9), the training stage adjusting parameters in the network using a cross entropy loss function, the cross entropy loss function being noted as:
wherein u represents the true tag of the residue to be detected in the protein sequence,representing an output value corresponding to the type of the predicted residue of the network model, wherein Y represents the difference between the predicted output and the real label;
12 Inputting the feature vector of the residue in the protein sequence S into the model trained in the step 11), outputting the probability value of each residue secondary structure category according to the model, wherein the category corresponding to the maximum probability value is the prediction category of the residue secondary structure.
CN202110097155.1A 2021-01-25 2021-01-25 Protein secondary structure prediction method based on cyclic neural network Active CN112837741B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110097155.1A CN112837741B (en) 2021-01-25 2021-01-25 Protein secondary structure prediction method based on cyclic neural network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110097155.1A CN112837741B (en) 2021-01-25 2021-01-25 Protein secondary structure prediction method based on cyclic neural network

Publications (2)

Publication Number Publication Date
CN112837741A CN112837741A (en) 2021-05-25
CN112837741B true CN112837741B (en) 2024-04-16

Family

ID=75931348

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110097155.1A Active CN112837741B (en) 2021-01-25 2021-01-25 Protein secondary structure prediction method based on cyclic neural network

Country Status (1)

Country Link
CN (1) CN112837741B (en)

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113851192B (en) * 2021-09-15 2023-06-30 安庆师范大学 Training method and device for amino acid one-dimensional attribute prediction model and attribute prediction method
CN114496068A (en) * 2022-01-27 2022-05-13 中国农业银行股份有限公司 Protein secondary structure prediction method, device, equipment and storage medium
CN115312119B (en) * 2022-10-09 2023-04-07 之江实验室 Method and system for identifying protein structural domain based on protein three-dimensional structure image
CN116312754B (en) * 2023-03-16 2023-10-03 安庆师范大学 Protein structure prediction method based on mixed deep learning model
CN116705150A (en) * 2023-06-05 2023-09-05 国家超级计算天津中心 Method, device, equipment and medium for determining gene expression efficiency

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110689920A (en) * 2019-09-18 2020-01-14 上海交通大学 Protein-ligand binding site prediction algorithm based on deep learning
CN111667880A (en) * 2020-05-27 2020-09-15 浙江工业大学 Protein residue contact map prediction method based on depth residual error neural network
CN112085245A (en) * 2020-07-21 2020-12-15 浙江工业大学 Protein residue contact prediction method based on deep residual error neural network
CN112149885A (en) * 2020-09-07 2020-12-29 浙江工业大学 Ligand binding residue prediction method based on sequence template

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110689920A (en) * 2019-09-18 2020-01-14 上海交通大学 Protein-ligand binding site prediction algorithm based on deep learning
CN111667880A (en) * 2020-05-27 2020-09-15 浙江工业大学 Protein residue contact map prediction method based on depth residual error neural network
CN112085245A (en) * 2020-07-21 2020-12-15 浙江工业大学 Protein residue contact prediction method based on deep residual error neural network
CN112149885A (en) * 2020-09-07 2020-12-29 浙江工业大学 Ligand binding residue prediction method based on sequence template

Also Published As

Publication number Publication date
CN112837741A (en) 2021-05-25

Similar Documents

Publication Publication Date Title
CN112837741B (en) Protein secondary structure prediction method based on cyclic neural network
CN110532900B (en) Facial expression recognition method based on U-Net and LS-CNN
CN110135459B (en) Zero sample classification method based on double-triple depth measurement learning network
WO2016192612A1 (en) Method for analysing medical treatment data based on deep learning, and intelligent analyser thereof
CN116417093A (en) Drug target interaction prediction method combining transducer and graph neural network
CN109192316B (en) Disease subtype prediction system based on gene network analysis
CN117122303B (en) Brain network prediction method, system, equipment and storage medium
CN113096722A (en) Protein phosphorylation site prediction method based on inner product self-attention neural network
CN112364708A (en) Multi-mode human body action recognition method based on knowledge distillation and antagonistic learning
CN106951509B (en) Multi-tag coring canonical correlation analysis search method
CN112434718A (en) New coronary pneumonia multi-modal feature extraction fusion method and system based on depth map
CN114511759A (en) Method and system for identifying categories and determining characteristics of skin state images
CN110781828A (en) Fatigue state detection method based on micro-expression
CN112085245B (en) Protein residue contact prediction method based on depth residual neural network
CN106778252B (en) Intrusion detection method based on rough set theory and WAODE algorithm
CN111695570A (en) Variational prototype reasoning-based semantic segmentation method under small sample
Vauterin et al. Integrated databasing and analysis
CN114299342B (en) Unknown mark classification method in multi-mark picture classification based on deep learning
CN115757855A (en) Image retrieval method based on graph structure matching
CN113571126B (en) Protein residue contact prediction method based on multi-loss training
CN111259176B (en) Cross-modal Hash retrieval method based on matrix decomposition and integrated with supervision information
Sen et al. DeepGlut: A deep learning framework for prediction of glutarylation sites in proteins
US20150302171A1 (en) Bioimaging grid
Narra et al. Use of extended phylogenetic profiles with E-values and support vector machines for protein family classification
Howe et al. Protein residue contact prediction using support vector machine

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant