CN112837741B - Protein secondary structure prediction method based on cyclic neural network - Google Patents
Protein secondary structure prediction method based on cyclic neural network Download PDFInfo
- Publication number
- CN112837741B CN112837741B CN202110097155.1A CN202110097155A CN112837741B CN 112837741 B CN112837741 B CN 112837741B CN 202110097155 A CN202110097155 A CN 202110097155A CN 112837741 B CN112837741 B CN 112837741B
- Authority
- CN
- China
- Prior art keywords
- protein sequence
- secondary structure
- residue
- protein
- matrix
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 102000004169 proteins and genes Human genes 0.000 title claims abstract description 96
- 108090000623 proteins and genes Proteins 0.000 title claims abstract description 96
- 238000000034 method Methods 0.000 title claims abstract description 25
- 238000013528 artificial neural network Methods 0.000 title claims abstract description 24
- 125000004122 cyclic group Chemical group 0.000 title claims abstract description 14
- 239000011159 matrix material Substances 0.000 claims abstract description 43
- 239000013598 vector Substances 0.000 claims abstract description 29
- 150000001413 amino acids Chemical class 0.000 claims description 5
- 230000000306 recurrent effect Effects 0.000 claims description 5
- 230000004913 activation Effects 0.000 claims description 4
- 238000009933 burial Methods 0.000 claims description 4
- 238000010606 normalization Methods 0.000 claims description 4
- 238000011176 pooling Methods 0.000 claims description 4
- 238000004364 calculation method Methods 0.000 abstract description 5
- 230000001537 neural effect Effects 0.000 abstract 1
- 230000006870 function Effects 0.000 description 11
- 238000013135 deep learning Methods 0.000 description 3
- 125000000539 amino acid group Chemical group 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000004883 computer application Methods 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 238000009510 drug design Methods 0.000 description 1
- 238000010230 functional analysis Methods 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 230000015654 memory Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000002360 preparation method Methods 0.000 description 1
- 230000006403 short-term memory Effects 0.000 description 1
- 238000000638 solvent extraction Methods 0.000 description 1
- 238000001228 spectrum Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16B—BIOINFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR GENETIC OR PROTEIN-RELATED DATA PROCESSING IN COMPUTATIONAL MOLECULAR BIOLOGY
- G16B15/00—ICT specially adapted for analysing two-dimensional or three-dimensional molecular structures, e.g. structural or functional relations or structure alignment
- G16B15/20—Protein or domain folding
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16B—BIOINFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR GENETIC OR PROTEIN-RELATED DATA PROCESSING IN COMPUTATIONAL MOLECULAR BIOLOGY
- G16B40/00—ICT specially adapted for biostatistics; ICT specially adapted for bioinformatics-related machine learning or data mining, e.g. knowledge discovery or pattern finding
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02A—TECHNOLOGIES FOR ADAPTATION TO CLIMATE CHANGE
- Y02A90/00—Technologies having an indirect contribution to adaptation to climate change
- Y02A90/10—Information and communication technologies [ICT] supporting adaptation to climate change, e.g. for weather forecasting or climate simulation
Landscapes
- Life Sciences & Earth Sciences (AREA)
- Physics & Mathematics (AREA)
- Health & Medical Sciences (AREA)
- Medical Informatics (AREA)
- Engineering & Computer Science (AREA)
- Spectroscopy & Molecular Physics (AREA)
- General Health & Medical Sciences (AREA)
- Theoretical Computer Science (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Biotechnology (AREA)
- Evolutionary Biology (AREA)
- Biophysics (AREA)
- Artificial Intelligence (AREA)
- Chemical & Material Sciences (AREA)
- Crystallography & Structural Chemistry (AREA)
- Bioethics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Data Mining & Analysis (AREA)
- Databases & Information Systems (AREA)
- Epidemiology (AREA)
- Evolutionary Computation (AREA)
- Public Health (AREA)
- Software Systems (AREA)
- Investigating Or Analysing Biological Materials (AREA)
- Management, Administration, Business Operations System, And Electronic Commerce (AREA)
Abstract
Protein based on cyclic neural networkAccording to the input protein sequence to be subjected to secondary structure prediction, the protein sequence is encoded by a single thermal encoding mode to obtain an L multiplied by 20 characteristic matrix M 1 Generating a protein position-specific scoring matrix PSSM using a PSI-BLAST program; then, M is 1 Performing matrix addition operation with PSSM to obtain matrix M 2 The method comprises the steps of carrying out a first treatment on the surface of the Secondly, obtaining a characteristic vector of the residue; thirdly, constructing a circulating neural network framework, constructing a sample set by acquiring protein sequences of known secondary structure labels from a PDB database, and training the constructed circulating neural network; and finally, inputting the feature vector of the residue in the protein sequence to be predicted into a trained model, and predicting the secondary structure category of the residue in the protein sequence according to the output probability value. The method has low calculation cost and high prediction accuracy.
Description
Technical Field
The invention relates to the fields of bioinformatics, deep learning and computer application, in particular to a protein secondary structure prediction method based on a circulating neural network.
Background
Prediction of the secondary structure of a protein is an important intermediate step in the prediction of tertiary structure, and is also a bridge linking the protein sequence and tertiary structure. Accurate identification of the secondary structure of the protein can help people to know complex dependency relationship between protein sequences and tertiary structures, and can also promote functional analysis of the protein and drug design.
Currently, the methods for predicting the secondary structure of proteins by deep learning are as follows: SSpro8 (Gianluca, pollastri, darisz, et al, improving the prediction of protein secondary structure in three and eight classes using recurrent neural networks and profiles [ J ]. Proteins Structure Function & Genetics, 2002: gianluca et al SSpro 8. Prediction of class 3 and class 8 protein secondary structures using cyclic neural networks and spectrum files [ J ]. Protein structure, function and Genetics,2002 ], BLSTM (Snderby S K, winther O.protein Secondary Structure Prediction with Long Short Term Memory Networks [ J ]. Computer Science,2014. I.e., snderby S K, winther O.BLSTM: prediction of class 3 and class 8 protein secondary structures based on long and short time memory networks [ J ]. Computer Science,2014 ], GSN (Zhou J, troynsky O.Deep Supervised and Convolutional Generative Stochastic Network for Protein Secondary Structure Prediction [ J ]. Computer Science,2014:745-753. I.e., zhou J ]. Troynsky O G. Prediction of class III secondary structures based on deep-supervision convolution and CRNs [ J ]. 2014:745-753. ]) are performed with (Buzzoz, jnyan [ L ]. 35 of class III.e., 35) of new biological structures of Buzzo [ J ]. 35, 35 mg [ J ]. 1-92 ] of Buzzo.e.g. biological structures of Buzzo [ J ]. Three classes of Bungan [ J ]. 35 ] were studied. Compared with the traditional machine learning method, the deep learning-based method can more fully extract the amino acid characteristics and the hidden modes in the protein sequence. Although existing protein secondary structure prediction methods have achieved good results, these methods only extract features along the amino acid residue dimension. Thus, these methods may ignore some important features hidden in the protein sequence feature vector, which may be useful for predicting secondary structure.
In summary, the existing protein secondary structure prediction methods have great differences from the actual application requirements in terms of calculation cost and prediction accuracy, and urgent improvement is needed.
Disclosure of Invention
In order to overcome the defects of the existing protein secondary structure prediction method in terms of calculation cost and prediction accuracy, the invention provides a protein secondary structure prediction method based on a cyclic neural network, which is low in calculation cost and high in prediction accuracy.
The technical scheme adopted for solving the technical problems is as follows:
a method for predicting protein secondary structure based on a recurrent neural network, the method comprising the steps of:
1) Inputting a protein sequence with L residues and to be subjected to protein secondary structure prediction, and marking the protein sequence as S;
2) The protein sequence S is encoded by a single thermal encoding mode to obtain a characteristic matrix with the size of L multiplied by 20, which is marked as M 1 ,M 1 The ith row and jth column elements of (a) are expressed as:
wherein A is 1 ,A 2 ,...,A 20 Residues representing 20 common amino acid types, T i Indicating the type of the i-th residue in sequence S;
3) Protein sequence S was scored using the PSI-BLAST (https:// www.ebi.ac.uk/Tools/sss/psiblast /) program to generate a protein position-specific scoring matrix, designated PSSM;
4) For M 1 Performing matrix addition operation with PSSM to obtain a matrix with size of Lx20, and recording as M 2 ;
5) Calculating matrix M 2 The average value of each row in the row (C) is obtained, and a vector with the size of L multiplied by 1 is obtained and is recorded as H;
6) Calculating matrix M 2 Obtaining a vector with the size of L multiplied by 1, and recording the vector as V;
7) For the protein sequence S, a feature matrix F with the size of L multiplied by 20 is obtained through the following formula, and the elements in the ith row and the jth column in F are expressed as:
wherein,represents M 2 The ith row and the jth column elements of H i Represents the i-th element in H, V i Represents the i-th element in V;
8) Any row F in matrix F i I=1, 2, L, a feature vector representing the i-th residue in the protein sequence S;
9) Obtaining a protein sequence of a known residue secondary structure class from a PDB (https:// www.rcsb.org /) database as a training set, wherein the residue secondary structure class comprises burial, intermediate state and exposure, and generating feature vectors of all residues in the training set by using the steps 2) -8); combining the residue secondary structure class labels to construct a training sample set;
10 A second-level structure class of residues in the predicted protein sequence S of the cyclic neural network is built, the network consists of three parts, wherein the first part is a convolution layer part which consists of a convolution layer with a convolution kernel size of 1 multiplied by 20, a convolution layer with a convolution kernel size of 3 multiplied by 1, a normalization layer and a pooling layer; the second part is a recycled layer part, which consists of two LSTM layers; the last part is a fully connected layer part which consists of two fully connected layers, wherein the output of each layer is used as the input of the next layer, and the sigmoid activation function is used to enable the output value of the network to be in the range of (0, 1);
11 Training the cyclic neural network constructed in step 10) using the training sample set constructed in step 9), the training stage adjusting parameters in the network using a cross entropy loss function, the cross entropy loss function being noted as:
wherein u represents the true tag of the residue to be detected in the protein sequence,representing an output value corresponding to the type of the predicted residue of the network model, wherein Y represents the difference between the predicted output and the real label;
12 Inputting the feature vector of the residue in the sequence S into the model trained in the step 11), outputting the probability value of each residue secondary structure category according to the model, wherein the category corresponding to the maximum probability value is the prediction category of the residue secondary structure.
The technical conception of the invention is as follows: firstly, according to an input protein sequence to be subjected to secondary structure prediction, encoding the protein sequence by using a single thermal encoding mode to obtain an L multiplied by 20 characteristic matrix M 1 Generating a protein position-specific scoring matrix PSSM using a PSI-BLAST program; then, M is 1 Performing matrix addition operation with PSSM to obtain matrix M 2 The method comprises the steps of carrying out a first treatment on the surface of the Next, the eigenvectors of the residues are obtained by: calculating matrix M 2 Mean and variance of each row, matrix M 2 Subtracting the average value of the current row from any element, dividing the average value by the variance of the current row, and generating a corresponding element in a feature matrix F, wherein any row in F represents a feature vector of a corresponding residue; thirdly, constructing a circulating neural network framework, constructing a sample set by acquiring protein sequences of known secondary structure labels from a PDB database, and training the constructed circulating neural network; and finally, inputting the feature vector of the residue in the protein sequence to be predicted into a trained model, and predicting the secondary structure category of the residue in the protein sequence according to the output probability value. The invention provides a protein secondary structure prediction method based on a cyclic neural network, which has low calculation cost and high prediction accuracy.
The beneficial effects of the invention are as follows: on one hand, the convolution part of the network uses asymmetric convolution for extracting the connection between residue parts, so that preparation is made for further improving the accuracy of protein secondary structure prediction; on the other hand, the BLSTM part of the network is used for capturing long-distance interdependence relation among residues, so that the accuracy of protein secondary structure prediction is better ensured.
Drawings
FIG. 1 is a schematic diagram of a method for predicting a protein secondary structure based on a recurrent neural network.
FIG. 2 is a graph showing the results of secondary structure prediction of protein sequence 1F2C using a cyclic neural network-based protein secondary structure prediction method.
Detailed Description
The invention is further described below with reference to the accompanying drawings.
Referring to fig. 1 and 2, a protein secondary structure prediction method based on a recurrent neural network includes the steps of:
1) Inputting a protein sequence with L residues and to be subjected to protein secondary structure prediction, and marking the protein sequence as S;
2) The protein sequence S is encoded by a single thermal encoding mode to obtain a characteristic matrix with the size of L multiplied by 20, which is marked as M 1 ,M 1 The ith row and jth column elements of (a) are expressed as:
wherein A is 1 ,A 2 ,...,A 20 Residues representing 20 common amino acid types, T i Indicating the type of the i-th residue in sequence S;
3) Protein sequence S was scored using the PSI-BLAST (https:// www.ebi.ac.uk/Tools/sss/psiblast /) program to generate a protein position-specific scoring matrix, designated PSSM;
4) For M 1 Performing matrix addition operation with PSSM to obtain a matrix with size of Lx20, and recording as M 2 ;
5) Calculating matrix M 2 The average value of each row in the row (C) is obtained, and a vector with the size of L multiplied by 1 is obtained and is recorded as H;
6) Calculating matrix M 2 Obtaining a vector with the size of L multiplied by 1, and recording the vector as V;
7) For the protein sequence S, a feature matrix F with the size of L multiplied by 20 is obtained through the following formula, and the elements in the ith row and the jth column in F are expressed as:
wherein,represents M 2 The ith row and the jth column elements of H i Represents the i-th element in H, V i Represents the i-th element in V;
8) Any row F in matrix F i I=1, 2, L, a feature vector representing the i-th residue in the protein sequence S;
9) Obtaining a protein sequence of a known residue secondary structure class from a PDB (https:// www.rcsb.org /) database as a training set, wherein the residue secondary structure class comprises burial, intermediate state and exposure, and generating feature vectors of all residues in the training set by using the steps 2) -8); combining the residue secondary structure class labels to construct a training sample set;
10 A second-level structure class of residues in the predicted protein sequence S of the cyclic neural network is built, the network consists of three parts, wherein the first part is a convolution layer part which consists of a convolution layer with a convolution kernel size of 1 multiplied by 20, a convolution layer with a convolution kernel size of 3 multiplied by 1, a normalization layer and a pooling layer; the second part is a recycled layer part, which consists of two LSTM layers; the last part is a fully connected layer part which consists of two fully connected layers, wherein the output of each layer is used as the input of the next layer, and the sigmoid activation function is used to enable the output value of the network to be in the range of (0, 1);
11 Training the cyclic neural network constructed in step 10) using the training sample set constructed in step 9), the training stage adjusting parameters in the network using a cross entropy loss function, the cross entropy loss function being noted as:
wherein u represents the true tag of the residue to be detected in the protein sequence,representing an output value corresponding to the type of the predicted residue of the network model, wherein Y represents the difference between the predicted output and the real label;
12 Inputting the feature vector of the residue in the sequence S into the model trained in the step 11), outputting the probability value of each residue secondary structure category according to the model, wherein the category corresponding to the maximum probability value is the prediction category of the residue secondary structure.
The embodiment takes the prediction of the secondary structure of the protein 1F2C as an embodiment, and the prediction method of the secondary structure of the protein based on the circulating neural network comprises the following steps:
1) Inputting a protein sequence 1F2C with the number of residues 215 to be subjected to protein secondary structure prediction, and marking the protein sequence as S;
2) The protein sequence S is encoded by a single thermal encoding method to obtain a matrix with the size of 215 multiplied by 20, which is denoted as M 1 ,M 1 The ith row and jth column elements of (b) are represented by the following equations:
wherein A is 1 ,A 2 ,...,A 20 Representing 20 common amino acid types, T i Represents the type of the i-th residue in the protein sequence 1F 2C;
3) Protein sequence S was scored using the PSI-BLAST (https:// www.ebi.ac.uk/Tools/sss/psiblast /) program to generate a protein position-specific scoring matrix, designated PSSM;
4) For M 1 Performing matrix addition operation with PSSM to obtain a matrix with size of Lx20, and recording as M 2 ;
5) Calculating matrix M 2 The average value of each row in the row (C) is obtained, and a vector with the size of L multiplied by 1 is obtained and is recorded as H;
6) Calculating matrix M 2 Obtaining a vector with the size of L multiplied by 1, and recording the vector as V;
7) For the protein sequence S, a feature matrix F with the size of L multiplied by 20 is obtained through the following formula, and the elements in the ith row and the jth column in F are expressed as:
wherein,represents M 2 The ith row and the jth column elements of H i Represents the i-th element in H, V i Represents the i-th element in V;
8) Any row F in matrix F i I=1, 2, L, a feature vector representing the i-th residue in the protein sequence S;
9) Obtaining a protein sequence of a known residue secondary structure class from a PDB (https:// www.rcsb.org /) database as a training set, wherein the residue secondary structure class comprises burial, intermediate state and exposure, and generating feature vectors of all residues in the training set by using the steps 2) -8); combining the residue secondary structure class labels to construct a training sample set;
10 A second-level structure class of residues in the predicted protein sequence S of the cyclic neural network is built, the network consists of three parts, wherein the first part is a convolution layer part which consists of a convolution layer with a convolution kernel size of 1 multiplied by 20, a convolution layer with a convolution kernel size of 3 multiplied by 1, a normalization layer and a pooling layer; the second part is a recycled layer part, which consists of two LSTM layers; the last part is a fully connected layer part which consists of two fully connected layers, wherein the output of each layer is used as the input of the next layer, and the sigmoid activation function is used to enable the output value of the network to be in the range of (0, 1);
11 Training the cyclic neural network constructed in step 10) using the training sample set constructed in step 9), the training stage adjusting parameters in the network using a cross entropy loss function, the cross entropy loss function being noted as:
wherein u represents the true tag of the residue to be detected in the protein sequence,representing an output value corresponding to the type of the predicted residue of the network model, wherein Y represents the difference between the predicted output and the real label;
12 Inputting the feature vector of the residue in the protein sequence S into the model trained in the step 11), outputting the probability value of each residue secondary structure category according to the model, wherein the category corresponding to the maximum probability value is the prediction category of the residue secondary structure.
The prediction of the protein sequence 1F2C using the above method for partitioning was performed using the secondary structure prediction of the protein sequence 1F2C as shown in FIG. 2.
The above description is that the predicted result obtained by taking the secondary structure of the protein sequence 1F2C as an example of the present invention is not limited to the implementation scope of the present invention, and various modifications and improvements can be made thereto without departing from the scope of the present invention, which is not to be excluded from the protection scope of the present invention.
Claims (1)
1. A method for predicting a protein secondary structure based on a recurrent neural network, the method comprising the steps of:
1) Inputting a protein sequence with L residues and to be subjected to protein secondary structure prediction, and marking the protein sequence as S;
2) The protein sequence S is encoded by a single thermal encoding mode to obtain a characteristic matrix with the size of L multiplied by 20, which is marked as M 1 ,M 1 The ith row and jth column elements of (a) are expressed as:
wherein A is 1 ,A 2 ,...,A 20 Residues representing 20 common amino acid types, T i Indicating the type of the i-th residue in sequence S;
3) Protein sequence S was scored using the PSI-BLAST program to generate a protein position-specific scoring matrix, designated PSSM;
4) For M 1 And PSSM performs matrix addition to obtain a matrix with L×20, denoted as M 2 ;
5) Calculating matrix M 2 The average value of each row in the row (C) is obtained, and a vector with the size of L multiplied by 1 is obtained and is recorded as H;
6) Calculating matrix M 2 Obtaining a vector with the size of L multiplied by 1, and recording the vector as V;
7) For the protein sequence S, a feature matrix F with the size of L multiplied by 20 is obtained through the following formula, and the elements in the ith row and the jth column in F are expressed as:
wherein,represents M 2 The ith row and the jth column elements of H i Represents the i-th element in H, V i Represents the i-th element in V;
8) Any row F in matrix F i I=1, 2, L, a feature vector representing the i-th residue in the protein sequence S;
9) Obtaining a protein sequence of a secondary structure class of known residues from a PDB database as a training set, wherein the secondary structure class of the residues comprises burial, intermediate state and exposure, and generating feature vectors of all residues in the training set by using the steps 2) -8); combining the residue secondary structure class labels to construct a training sample set;
10 A second-level structure class of residues in the predicted protein sequence S of the cyclic neural network is built, the network consists of three parts, wherein the first part is a convolution layer part which consists of a convolution layer with a convolution kernel size of 1 multiplied by 20, a convolution layer with a convolution kernel size of 3 multiplied by 1, a normalization layer and a pooling layer; the second part is a recycled layer part, which consists of two LSTM layers; the last part is a fully connected layer part which consists of two fully connected layers, wherein the output of each layer is used as the input of the next layer, and the sigmoid activation function is used to enable the output value of the network to be in the range of (0, 1);
11 Training the network constructed in step 10) using the training sample set constructed in step 9), the training stage adjusting parameters in the network using a cross entropy loss function, the cross entropy loss function being noted as:
wherein u represents the true tag of the residue to be detected in the protein sequence,representing an output value corresponding to the type of the predicted residue of the network model, wherein Y represents the difference between the predicted output and the real label;
12 Inputting the feature vector of the residue in the protein sequence S into the model trained in the step 11), outputting the probability value of each residue secondary structure category according to the model, wherein the category corresponding to the maximum probability value is the prediction category of the residue secondary structure.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110097155.1A CN112837741B (en) | 2021-01-25 | 2021-01-25 | Protein secondary structure prediction method based on cyclic neural network |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110097155.1A CN112837741B (en) | 2021-01-25 | 2021-01-25 | Protein secondary structure prediction method based on cyclic neural network |
Publications (2)
Publication Number | Publication Date |
---|---|
CN112837741A CN112837741A (en) | 2021-05-25 |
CN112837741B true CN112837741B (en) | 2024-04-16 |
Family
ID=75931348
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110097155.1A Active CN112837741B (en) | 2021-01-25 | 2021-01-25 | Protein secondary structure prediction method based on cyclic neural network |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112837741B (en) |
Families Citing this family (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113851192B (en) * | 2021-09-15 | 2023-06-30 | 安庆师范大学 | Training method and device for amino acid one-dimensional attribute prediction model and attribute prediction method |
CN114496068A (en) * | 2022-01-27 | 2022-05-13 | 中国农业银行股份有限公司 | Protein secondary structure prediction method, device, equipment and storage medium |
CN115312119B (en) * | 2022-10-09 | 2023-04-07 | 之江实验室 | Method and system for identifying protein structural domain based on protein three-dimensional structure image |
CN116312754B (en) * | 2023-03-16 | 2023-10-03 | 安庆师范大学 | Protein structure prediction method based on mixed deep learning model |
CN116705150A (en) * | 2023-06-05 | 2023-09-05 | 国家超级计算天津中心 | Method, device, equipment and medium for determining gene expression efficiency |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110689920A (en) * | 2019-09-18 | 2020-01-14 | 上海交通大学 | Protein-ligand binding site prediction algorithm based on deep learning |
CN111667880A (en) * | 2020-05-27 | 2020-09-15 | 浙江工业大学 | Protein residue contact map prediction method based on depth residual error neural network |
CN112085245A (en) * | 2020-07-21 | 2020-12-15 | 浙江工业大学 | Protein residue contact prediction method based on deep residual error neural network |
CN112149885A (en) * | 2020-09-07 | 2020-12-29 | 浙江工业大学 | Ligand binding residue prediction method based on sequence template |
-
2021
- 2021-01-25 CN CN202110097155.1A patent/CN112837741B/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110689920A (en) * | 2019-09-18 | 2020-01-14 | 上海交通大学 | Protein-ligand binding site prediction algorithm based on deep learning |
CN111667880A (en) * | 2020-05-27 | 2020-09-15 | 浙江工业大学 | Protein residue contact map prediction method based on depth residual error neural network |
CN112085245A (en) * | 2020-07-21 | 2020-12-15 | 浙江工业大学 | Protein residue contact prediction method based on deep residual error neural network |
CN112149885A (en) * | 2020-09-07 | 2020-12-29 | 浙江工业大学 | Ligand binding residue prediction method based on sequence template |
Also Published As
Publication number | Publication date |
---|---|
CN112837741A (en) | 2021-05-25 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN112837741B (en) | Protein secondary structure prediction method based on cyclic neural network | |
CN110532900B (en) | Facial expression recognition method based on U-Net and LS-CNN | |
CN110135459B (en) | Zero sample classification method based on double-triple depth measurement learning network | |
WO2016192612A1 (en) | Method for analysing medical treatment data based on deep learning, and intelligent analyser thereof | |
CN116417093A (en) | Drug target interaction prediction method combining transducer and graph neural network | |
CN109192316B (en) | Disease subtype prediction system based on gene network analysis | |
CN117122303B (en) | Brain network prediction method, system, equipment and storage medium | |
CN113096722A (en) | Protein phosphorylation site prediction method based on inner product self-attention neural network | |
CN112364708A (en) | Multi-mode human body action recognition method based on knowledge distillation and antagonistic learning | |
CN106951509B (en) | Multi-tag coring canonical correlation analysis search method | |
CN112434718A (en) | New coronary pneumonia multi-modal feature extraction fusion method and system based on depth map | |
CN114511759A (en) | Method and system for identifying categories and determining characteristics of skin state images | |
CN110781828A (en) | Fatigue state detection method based on micro-expression | |
CN112085245B (en) | Protein residue contact prediction method based on depth residual neural network | |
CN106778252B (en) | Intrusion detection method based on rough set theory and WAODE algorithm | |
CN111695570A (en) | Variational prototype reasoning-based semantic segmentation method under small sample | |
Vauterin et al. | Integrated databasing and analysis | |
CN114299342B (en) | Unknown mark classification method in multi-mark picture classification based on deep learning | |
CN115757855A (en) | Image retrieval method based on graph structure matching | |
CN113571126B (en) | Protein residue contact prediction method based on multi-loss training | |
CN111259176B (en) | Cross-modal Hash retrieval method based on matrix decomposition and integrated with supervision information | |
Sen et al. | DeepGlut: A deep learning framework for prediction of glutarylation sites in proteins | |
US20150302171A1 (en) | Bioimaging grid | |
Narra et al. | Use of extended phylogenetic profiles with E-values and support vector machines for protein family classification | |
Howe et al. | Protein residue contact prediction using support vector machine |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |