CN112530514A - Novel depth model, computer device, storage medium for predicting compound protein interaction based on deep learning method - Google Patents

Novel depth model, computer device, storage medium for predicting compound protein interaction based on deep learning method Download PDF

Info

Publication number
CN112530514A
CN112530514A CN202011499748.2A CN202011499748A CN112530514A CN 112530514 A CN112530514 A CN 112530514A CN 202011499748 A CN202011499748 A CN 202011499748A CN 112530514 A CN112530514 A CN 112530514A
Authority
CN
China
Prior art keywords
model
input
protein
compound
sequence
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202011499748.2A
Other languages
Chinese (zh)
Inventor
王淑栋
刘嘉丽
宋弢
田庆雨
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
China University of Petroleum East China
Original Assignee
China University of Petroleum East China
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by China University of Petroleum East China filed Critical China University of Petroleum East China
Priority to CN202011499748.2A priority Critical patent/CN112530514A/en
Publication of CN112530514A publication Critical patent/CN112530514A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16BBIOINFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR GENETIC OR PROTEIN-RELATED DATA PROCESSING IN COMPUTATIONAL MOLECULAR BIOLOGY
    • G16B5/00ICT specially adapted for modelling or simulations in systems biology, e.g. gene-regulatory networks, protein interaction networks or metabolic networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16BBIOINFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR GENETIC OR PROTEIN-RELATED DATA PROCESSING IN COMPUTATIONAL MOLECULAR BIOLOGY
    • G16B40/00ICT specially adapted for biostatistics; ICT specially adapted for bioinformatics-related machine learning or data mining, e.g. knowledge discovery or pattern finding
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16CCOMPUTATIONAL CHEMISTRY; CHEMOINFORMATICS; COMPUTATIONAL MATERIALS SCIENCE
    • G16C20/00Chemoinformatics, i.e. ICT specially adapted for the handling of physicochemical or structural data of chemical particles, elements, compounds or mixtures
    • G16C20/30Prediction of properties of chemical compounds, compositions or mixtures

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Theoretical Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Biophysics (AREA)
  • General Health & Medical Sciences (AREA)
  • Data Mining & Analysis (AREA)
  • Software Systems (AREA)
  • Evolutionary Computation (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Artificial Intelligence (AREA)
  • Medical Informatics (AREA)
  • Evolutionary Biology (AREA)
  • General Physics & Mathematics (AREA)
  • Spectroscopy & Molecular Physics (AREA)
  • Computational Linguistics (AREA)
  • Biotechnology (AREA)
  • Biomedical Technology (AREA)
  • Mathematical Physics (AREA)
  • General Engineering & Computer Science (AREA)
  • Epidemiology (AREA)
  • Public Health (AREA)
  • Databases & Information Systems (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioethics (AREA)
  • Chemical & Material Sciences (AREA)
  • Crystallography & Structural Chemistry (AREA)
  • Physiology (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The invention discloses a novel depth model for predicting compound protein interaction based on a deep learning method. The novel depth model comprises three bidirectional gating cyclic unit (BiGRU) models, a graph convolution neural network model (GCN) and a Convolution Neural Network (CNN) model, and the whole network architecture is tripleBiGRU/GCN-CNN. The bidirectional gated cyclic unit model comprises a sequence processing model consisting of two gated cyclic units (GRUs), wherein one input is a forward input, the other input is a reverse input, and the two-way recursive neural network is a bidirectional recursive neural network with only an input gate and a forgetting gate. The input of the model is a compound one-dimensional SMILES sequence, a compound two-dimensional molecular graph, a protein structure attribute sequence and a protein amino acid sequence, and the four variables are respectively input into the tripleBiGRU/GCN model. The TripleBiGRU/GCN model output is a feature vector representing the compound and a feature vector representing the protein. The CNN model consists of a convolution layer, a pooling layer and a full-connection layer, and the input of the CNN model is a characteristic vector of a compound and a characteristic vector of a protein; the final output of the TripleBiGRU/GCN-CNN model is the root mean square error value of the predicted compound protein affinity value.

Description

Novel depth model, computer device, storage medium for predicting compound protein interaction based on deep learning method
Technical Field
The invention relates to the field of compound protein molecular feature extraction, in particular to a novel depth model, computer equipment and a storage medium for predicting compound protein interaction based on a deep learning method.
Background
Prediction of drug target interactions is of great significance to the design and development of drugs. While experimental analysis remains the most reliable method of determining drug target interactions, experimental characterization of each possible drug target is daunting due to the enormous cost and labor involved in the experiment. Therefore, the computational prediction of drug targets has grown enormously in recent years, and with the increase of protein structure data sets and ligand data sets in recent years, machine learning-based methods have been rapidly developed. I.e., protein information, ligand information and interactions are integrated into a unified framework.
The introduction of deep learning has proven to be one of the best models to predict drug target binding affinity. The main advantage of deep learning is that by performing nonlinear transformations in each layer, they can better represent the raw data and thus facilitate learning patterns hidden in the data. However, many models of compound representation are only molecular fingerprints, a single SMILES string. This can cause the encoded compound signature to lose important information inherent in many compounds, resulting in inaccuracies in the final prediction of the protein affinity value of the compound.
Disclosure of Invention
The invention aims to solve the problems of losing important information of compound molecules and protein molecules, improving the prediction accuracy and the like, and provides a novel depth model, a computer device and a storage medium for predicting the interaction of the compound molecules and the protein based on a depth learning method.
According to a first aspect of embodiments of the present invention, there is provided a novel depth model for predicting compound protein interactions based on a deep learning approach.
In some optional embodiments, the depth model includes three bidirectional gated cyclic unit (BiGRU) models, a graph convolution neural network model (GCN) and a Convolution Neural Network (CNN) model, and the entire network architecture is TripleBiGRU/GCN-CNN. The bidirectional gated cyclic unit model comprises a sequence processing model consisting of two gated cyclic units (GRUs), wherein one input is a forward input, the other input is a reverse input, and the two-way recursive neural network is a bidirectional recursive neural network with only an input gate and a forgetting gate. The input of the model is a compound one-dimensional SMILES sequence, a compound two-dimensional molecular graph, a protein structure attribute sequence and a protein amino acid sequence, and the four variables are respectively input into the tripleBiGRU/GCN model. The TripleBiGRU/GCN model output is a feature vector representing the compound and a feature vector representing the protein. The CNN model consists of a convolution layer, a pooling layer and a full-connection layer, and the input of the CNN model is a characteristic vector of a compound and a characteristic vector of a protein; the final output of the TripleBiGRU/GCN-CNN model is the root mean square error value of the predicted compound protein affinity value.
Optionally, the bidirectional gating and circulating unit (BiGRU) model enables data to be input from both positive and negative directions, so that information at each moment includes sequence information of previous and subsequent moments, that is, sequence information of a network at a certain specific moment is increased, and information of historical data is fully utilized, thereby enabling prediction to be more accurate. The basic idea of BiGRU is to present each training sequence forward and backward to two separate hidden layers, both connected to the same output layer. The output layer has the complete past and future information for each point in the input sequence. Wherein, the gating cycle unit (GRU) carries out sufficient feature extraction to the multivariate time sequence, continuously learns the long-term dependence of the multivariate time sequence, and the gating cycle unit specifically comprises: firstly, two gating states are obtained through the last transmitted state and the input of the current node, namely a reset gate (reset gate) for controlling reset and a update gate (update gate), after a gating signal is obtained, data after reset is obtained through the reset gate, the data and the input of the current node are spliced, the data are shrunk to the range of-1 to 1 through a hyperbolic tangent function, finally, the states are updated to the range of 0 to 1 through the update gate, and the more gating signals are close to 1, the more data are represented to be memorized.
Optionally, the main algorithm flow of the graph convolution neural network (GCN) model is to calculate neighbor information of each atomic node, and finally form an atomic vector containing the neighbor information.
Optionally, the Convolutional Neural Network (CNN) model is composed of convolution (convolution), activation (activation), and pooling (displacement) structures. The CNN output result is a specific feature space corresponding to a compound protein, and then the feature space output by the CNN is used as an input of a fully connected layer or a fully connected neural network (FCN), and the fully connected layer is used to complete mapping from the affinity value of the input compound feature vector and the protein feature vector.
Optionally, the input to the model is selected 4 variables, the input variables comprising the protein structure attribute sequence and protein amino acid sequence from the UniRef database, the compound SMILES from the stimch database, and the compound two-dimensional molecular map. Wherein the protein structure attribute sequence is encoded by the secondary structure of the protein, the length of the protein amino acid sequence, the physicochemical properties (polarity/non-polarity, acidity/alkalinity) of the protein and the solvent accessibility of the protein.
Optionally, the model is trained using existing affinity values for a large number of protein compounds, and refined model parameters are obtained.
According to a second aspect of an implementation of the present invention, a computer device is provided.
In some optional embodiments, the computer device includes a memory, a graphics card, a central processing unit, and an executable program stored in the memory and capable of being processed by the central processing unit and the graphics card in parallel, wherein the memory is characterized in that the central processing unit executes the program to implement the following steps: constructing a target detection and target prediction model, wherein the target detection and target prediction model comprises the following steps: feature extraction networks and prediction networks. Firstly, a feature extraction network is utilized to extract features of an input compound one-dimensional SMILES sequence, a compound two-dimensional molecular graph, a protein structure attribute sequence and a protein amino acid sequence; and (3) utilizing the extracted characteristic vector matrix to a target prediction model, wherein the target prediction model is to utilize convolution, pooling and full connection to operate the characteristic vector matrix and output root mean square error values of a predicted value and an actual value which are combined with affinity.
Optionally, the bidirectional gating and circulating unit (BiGRU) model enables data to be input from both positive and negative directions, so that information at each moment includes sequence information of previous and subsequent moments, that is, sequence information of a network at a certain specific moment is increased, and information of historical data is fully utilized, thereby enabling prediction to be more accurate. The basic idea of BiGRU is to present each training sequence forward and backward to two separate hidden layers, both connected to the same output layer. The output layer has the complete past and future information for each point in the input sequence. Wherein, the gating cycle unit (GRU) carries out sufficient feature extraction to the multivariate time sequence, continuously learns the long-term dependence of the multivariate time sequence, and the gating cycle unit specifically comprises: firstly, two gating states are obtained through the last transmitted state and the input of the current node, namely a reset gate (reset gate) for controlling reset and a update gate (update gate), after a gating signal is obtained, data after reset is obtained through the reset gate, the data and the input of the current node are spliced, the data are shrunk to the range of-1 to 1 through a hyperbolic tangent function, finally, the states are updated to the range of 0 to 1 through the update gate, and the more gating signals are close to 1, the more data are represented to be memorized.
Optionally, the main algorithm flow of the graph convolution neural network (GCN) model is to calculate neighbor information of each atomic node, and finally form an atomic vector containing the neighbor information.
Optionally, the Convolutional Neural Network (CNN) model is composed of convolution (convolution), activation (activation), and pooling (displacement) structures. The CNN output result is a specific feature space corresponding to a compound protein, and then the feature space output by the CNN is used as an input of a fully connected layer or a fully connected neural network (FCN), and the fully connected layer is used to complete mapping from the affinity value of the input compound feature vector and the protein feature vector.
Optionally, the input to the model is selected 4 variables, the input variables comprising the protein structure attribute sequence and protein amino acid sequence from the UniRef database, the compound SMILES from the stimch database, and the compound two-dimensional molecular map. Wherein the protein structure attribute sequence is encoded by the secondary structure of the protein, the length of the protein amino acid sequence, the physicochemical properties (polarity/non-polarity, acidity/alkalinity) of the protein and the solvent accessibility of the protein.
Optionally, the model is trained using existing affinity values for a large number of protein compounds, and refined model parameters are obtained.
The time-space sequence in the field of medicine is intelligently processed by utilizing an artificial intelligence technology, so that the problems of high development cost, long time consumption, safety and the like of new medicines can be solved. The trend to be able to screen new drugs and therapeutic targets among old drugs and compounds abandoned in use, which have been determined to be safe, is changing the situation of drug development and creating a drug relocation model for new drug development.
It is to be understood that both the foregoing general description and the following detailed description are exemplary and explanatory only and are not restrictive of the invention, as claimed.
Drawings
FIG. 1 is a specific flow diagram of a bidirectional GRU of the present invention
FIG. 2 is a system theme scheme diagram of the present invention
Detailed Description
The following description and the drawings sufficiently illustrate specific embodiments of the invention to enable those skilled in the art to practice them. Other embodiments may incorporate structural, logical, electrical, process, and other changes. The examples merely typify possible variations. Individual components and functions are optional unless explicitly required, and the sequence of operations may vary. Portions and features of some embodiments may be included in or substituted for those of others. The scope of embodiments of the invention encompasses the full ambit of the claims, as well as all available equivalents of the claims. Embodiments may be referred to herein, individually or collectively, by the term "invention" merely for convenience and without intending to voluntarily limit the scope of this application to any single invention or inventive concept if more than one is in fact disclosed. Herein, relational terms such as first and second, and the like may be used solely to distinguish one entity or action from another entity or action without requiring or implying any actual such relationship or order between such entities or actions. Also, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, or apparatus. The use of the phrase "including a" does not exclude the presence of other, identical elements in a process, method or device that includes the recited elements, unless expressly stated otherwise. The embodiments are described in a progressive manner, each embodiment focuses on differences from other embodiments, and the same and similar parts among the embodiments are referred to each other. As for the methods, products and the like disclosed by the embodiments, the description is simple because the methods correspond to the method parts disclosed by the embodiments, and the related parts can be referred to the method parts for description.
Fig. 1 shows an alternative implementation architecture of a novel coding approach for predicting compound protein affinity based on deep learning.
In this optional example, the depth model includes three bidirectional gated cyclic unit (BiGRU) models, a graph convolution neural network model (GCN), and a Convolution Neural Network (CNN) model, and the entire network architecture is TripleBiGRU/GCN-CNN. The bidirectional gated cyclic unit model comprises a sequence processing model consisting of two gated cyclic units (GRUs), wherein one input is a forward input, the other input is a reverse input, and the two-way recursive neural network is a bidirectional recursive neural network with only an input gate and a forgetting gate. The input of the model is a compound one-dimensional SMILES sequence, a compound two-dimensional molecular graph, a protein structure attribute sequence and a protein amino acid sequence, and the four variables are respectively input into the tripleBiGRU/GCN model. The TripleBiGRU/GCN model output is a feature vector representing the compound and a feature vector representing the protein. The CNN model consists of a convolution layer, a pooling layer and a full-connection layer, and the input of the CNN model is a characteristic vector of a compound and a characteristic vector of a protein; the final output of the TripleBiGRU/GCN-CNN model is the root mean square error value of the predicted compound protein affinity value.
Optionally, the bidirectional gating and circulating unit (BiGRU) model enables data to be input from both positive and negative directions, so that information at each moment includes sequence information of previous and subsequent moments, that is, sequence information of a network at a certain specific moment is increased, and information of historical data is fully utilized, thereby enabling prediction to be more accurate. The basic idea of BiGRU is to present each training sequence forward and backward to two separate hidden layers, both connected to the same output layer. The output layer has the complete past and future information for each point in the input sequence. Wherein, the gating cycle unit (GRU) carries out sufficient feature extraction to the multivariate time sequence, continuously learns the long-term dependence of the multivariate time sequence, and the gating cycle unit specifically comprises: firstly, two gating states are obtained through the last transmitted state and the input of the current node, namely a reset gate (reset gate) for controlling reset and a update gate (update gate), after a gating signal is obtained, data after reset is obtained through the reset gate, the data and the input of the current node are spliced, the data are shrunk to the range of-1 to 1 through a hyperbolic tangent function, finally, the states are updated to the range of 0 to 1 through the update gate, and the more gating signals are close to 1, the more data are represented to be memorized.
Optionally, the main algorithm flow of the graph convolution neural network (GCN) model is to calculate neighbor information of each atomic node, and finally form an atomic vector containing the neighbor information.
Optionally, the Convolutional Neural Network (CNN) model is composed of convolution (convolution), activation (activation), and pooling (displacement) structures. The CNN output result is a specific feature space corresponding to a compound protein, and then the feature space output by the CNN is used as an input of a fully connected layer or a fully connected neural network (FCN), and the fully connected layer is used to complete mapping from the affinity value of the input compound feature vector and the protein feature vector.
Optionally, the input to the model is selected 4 variables, the input variables comprising the protein structure attribute sequence and protein amino acid sequence from the UniRef database, the compound SMILES from the stimch database, and the compound two-dimensional molecular map. Wherein the protein structure attribute sequence is encoded by the secondary structure of the protein, the length of the protein amino acid sequence, the physicochemical properties (polarity/non-polarity, acidity/alkalinity) of the protein and the solvent accessibility of the protein.
Optionally, the model is trained using existing affinity values for a large number of protein compounds, and refined model parameters are obtained.
Optionally, the model further includes a training process of the bidirectional gated loop unit model, and a specific embodiment of the training process of the bidirectional gated loop unit model is provided below.
In the embodiment, in the training process of the target detection and target prediction model, firstly, a compound one-dimensional SMILES molecular sequence is input into a BiGRU model, a compound two-dimensional molecular graph is input into a GCN model, a protein structure attribute sequence and a protein amino acid sequence are respectively input into the other two BiGRU models, then two representation methods of a compound are fused into one feature vector, two representation methods of a protein are fused into one feature vector, the two vectors are respectively input into a CNN model, thus, training data was constructed in which the number of cells of the compound BiGRU model and the protein BiGRU model during training was set to 128(cell) and 256(cell), respectively, then the TripleBiGRU/GCN model is trained together with the CNN model, in order to reduce the complexity of the model, the TripleBiGRU/GCN model is trained in advance to fix parameters, and then the model and the CNN model are trained together to determine the parameters of the CNN model. The initial learning rate of the whole model training is 0.0001, a loss function (loss function) is set to be an average absolute error loss (MAE loss), network parameters are adjusted by using an Adam optimizer in the training process through calculating the error between a predicted value and a true value, the weight of the model parameters is adjusted, and then the loss function value is continuously reduced through continuous iteration, so that the network is finally converged.
In an exemplary embodiment, there is also provided a non-transitory computer readable storage medium, such as a memory, comprising instructions executable by a processor to perform the steps of: three bidirectional gating cyclic unit (BiGRU) models, a graph convolution neural network model (GCN) and a Convolution Neural Network (CNN) model are constructed, and the whole network architecture is tripleBiGRU/GCN-CNN. The bidirectional gated cyclic unit model comprises a sequence processing model consisting of two gated cyclic units (GRUs), wherein one input is a forward input, the other input is a reverse input, and the two-way recursive neural network is a bidirectional recursive neural network with only an input gate and a forgetting gate. The input of the model is a compound one-dimensional SMILES sequence, a compound two-dimensional molecular graph, a protein structure attribute sequence and a protein amino acid sequence, and the four variables are respectively input into the tripleBiGRU/GCN model. The TripleBiGRU/GCN model output is a feature vector representing the compound and a feature vector representing the protein. The CNN model consists of a convolution layer, a pooling layer and a full-connection layer, and the input of the CNN model is a characteristic vector of a compound and a characteristic vector of a protein; the final output of the TripleBiGRU/GCN-CNN model is the root mean square error value of the predicted compound protein affinity value.
Optionally, the bidirectional gating and circulating unit (BiGRU) model enables data to be input from both positive and negative directions, so that information at each moment includes sequence information of previous and subsequent moments, that is, sequence information of a network at a certain specific moment is increased, and information of historical data is fully utilized, thereby enabling prediction to be more accurate. The basic idea of BiGRU is to present each training sequence forward and backward to two separate hidden layers, both connected to the same output layer. The output layer has the complete past and future information for each point in the input sequence. Wherein, the gating cycle unit (GRU) carries out sufficient feature extraction to the multivariate time sequence, continuously learns the long-term dependence of the multivariate time sequence, and the gating cycle unit specifically comprises: firstly, two gating states are obtained through the last transmitted state and the input of the current node, namely a reset gate (reset gate) for controlling reset and a update gate (update gate), after a gating signal is obtained, data after reset is obtained through the reset gate, the data and the input of the current node are spliced, the data are shrunk to the range of-1 to 1 through a hyperbolic tangent function, finally, the states are updated to the range of 0 to 1 through the update gate, and the more gating signals are close to 1, the more data are represented to be memorized.
Optionally, the main algorithm flow of the graph convolution neural network (GCN) model is to calculate neighbor information of each atomic node, and finally form an atomic vector containing the neighbor information.
Optionally, the Convolutional Neural Network (CNN) model is composed of convolution (convolution), activation (activation), and pooling (displacement) structures. The CNN output result is a specific feature space corresponding to a compound protein, and then the feature space output by the CNN is used as an input of a fully connected layer or a fully connected neural network (FCN), and the fully connected layer is used to complete mapping from the affinity value of the input compound feature vector and the protein feature vector.
Optionally, the input to the model is selected 4 variables, the input variables comprising the protein structure attribute sequence and protein amino acid sequence from the UniRef database, the compound SMILES from the stimch database, and the compound two-dimensional molecular map. Wherein the protein structure attribute sequence is encoded by the secondary structure of the protein, the length of the protein amino acid sequence, the physicochemical properties (polarity/non-polarity, acidity/alkalinity) of the protein and the solvent accessibility of the protein.
Optionally, the model is trained using existing affinity values for a large number of protein compounds, and refined model parameters are obtained.
The non-transitory computer readable storage medium may be a Read Only Memory (ROM), a Random Access Memory (RAMD), a magnetic tape, an optical storage device, and the like.
The time-space sequence in the field of medicine is intelligently processed by utilizing an artificial intelligence technology, so that the problems of high development cost, long time consumption, safety and the like of new medicines can be solved. The trend to be able to screen new drugs and therapeutic targets among old drugs and compounds abandoned in use, which have been determined to be safe, is changing the situation of drug development and creating a drug relocation model for new drug development.
Those of skill in the art would appreciate that the various illustrative elements and algorithm steps described in connection with the embodiments disclosed herein may be implemented as electronic hardware or combinations of computer software and electronic hardware. Whether such functionality is implemented as hardware or software depends upon the particular application and design constraints imposed on the implementation. Skilled artisans may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the present invention. It is clear to those skilled in the art that, for convenience and brevity of description, the specific working processes of the above-described systems, apparatuses and units may refer to the corresponding processes in the foregoing method embodiments, and are not described herein again.
In the embodiments disclosed herein, it should be understood that the disclosed methods, articles of manufacture (including but not limited to devices, apparatuses, etc.) may be implemented in other ways. For example, the above-described embodiments of the apparatus are merely illustrative, and for example, the division of the units is only one logical division, and other divisions may be realized in practice, for example, a plurality of units or components may be combined or integrated into another system, or some features may be omitted, or not executed. In addition, the shown or discussed mutual coupling or direct coupling or communication connection may be an indirect coupling or communication connection through some interfaces, devices or units, and may be in an electrical, mechanical or other form. The units described as separate parts may or may not be physically separate, and parts displayed as units may or may not be physical units, may be located in one place, or may be distributed on a plurality of network units. Some or all of the units can be selected according to actual needs to achieve the purpose of the solution of the embodiment. In addition, functional units in the embodiments of the present invention may be integrated into one processing unit, or each unit may exist alone physically, or two or more units are integrated into one unit.
It should be understood that the flowchart and block diagrams in the figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods and computer program products according to various embodiments of the present invention. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems which perform the specified functions or acts, or combinations of special purpose hardware and computer instructions. The present invention is not limited to the procedures and structures that have been described above and shown in the drawings, and various modifications and changes may be made without departing from the scope thereof. The scope of the invention is limited only by the appended claims.

Claims (10)

1. A novel depth model for predicting compound protein interaction based on a deep learning method is characterized by comprising three bidirectional gating cycle unit (BiGRU) models, a graph convolution neural network model (GCN) and a Convolution Neural Network (CNN) model, wherein the whole network architecture is TripleBiGRU/GCN-CNN. The bidirectional gated cyclic unit model comprises a sequence processing model consisting of two gated cyclic units (GRUs), wherein one input is a forward input, the other input is a reverse input, and the two-way recursive neural network is a bidirectional recursive neural network with only an input gate and a forgetting gate. The input of the model is a compound one-dimensional SMILES sequence, a compound two-dimensional molecular graph, a protein structure attribute sequence and a protein amino acid sequence, and the four variables are respectively input into the tripleBiGRU/GCN model. The TripleBiGRU/GCN model output is a feature vector representing the compound and a feature vector representing the protein. The CNN model consists of a convolution layer, a pooling layer and a full-connection layer, and the input of the CNN model is a characteristic vector of a compound and a characteristic vector of a protein; the final output of the TripleBiGRU/GCN-CNN model is the root mean square error value of the predicted compound protein affinity value.
2. The novel depth model of claim 1, wherein the bidirectional gated cyclic unit (BiGRU) model enables data to be input simultaneously from both the forward and reverse directions, so that the information at each time includes sequence information of the forward and backward times, which is equivalent to the increase of sequence information of the network at a specific time, and the information of historical data is fully utilized, thereby making prediction more accurate. The basic idea of BiGRU is to present each training sequence forward and backward to two separate hidden layers, both connected to the same output layer. The output layer has the complete past and future information for each point in the input sequence. Wherein, the gating cycle unit (GRU) carries out sufficient feature extraction to the multivariate time sequence, continuously learns the long-term dependence of the multivariate time sequence, and the gating cycle unit specifically comprises: firstly, two gating states, namely a reset gate (reset gate) for controlling reset and a gate (update gate) for controlling update are acquired through the last transmitted state and the input of the current node, after a gating signal is acquired, data after reset is acquired through the reset gate, the data and the input of the current node are spliced, the data are shrunk to the range of-1 to 1 through a hyperbolic tangent function, finally, the states are updated to the range of 0 to 1 through the update gate, and the more gating signals are close to 1, the more data are stored.
3. The feature extraction model of claim 2, wherein the main algorithm flow of the graph convolution neural network (GCN) model is to calculate neighbor information of each atomic node, and finally form an atomic vector containing the neighbor information.
4. The compound feature extraction model of claim 3, wherein the Convolutional Neural Network (CNN) model comprises convolution, activation, and pooling structures. The CNN output result is a specific feature space corresponding to a compound protein, and then the feature space output by the CNN is used as an input of a fully connected layer or a fully connected neural network (FCN), and the fully connected layer is used to complete mapping from the affinity value of the input compound feature vector and the protein feature vector.
5. The entire model of claim 4, wherein the inputs to the model are selected 4 variables, the input variables comprising the protein structure attribute sequence and the protein amino acid sequence from the UniRef database, the compound SMILES from the STITCH database, and the compound two-dimensional molecular map. Wherein the protein structure attribute sequence is encoded by the secondary structure of the protein, the length of the protein amino acid sequence, the physicochemical properties (polarity/non-polarity, acidity/alkalinity) of the protein and the solvent accessibility of the protein.
6. The novel depth model of claim 5, wherein the model is trained using existing affinity values for a plurality of protein compounds and refined model parameters are obtained.
7. A computer device comprising a memory, a graphics card, a central processing unit, and an executable program stored in the memory and capable of being processed in parallel by the central processing unit and the graphics card, wherein the central processing unit implements the following steps when executing the program: constructing a target detection and target prediction model, wherein the target detection and target prediction model comprises the following steps: feature extraction networks and prediction networks. Firstly, a feature extraction network is utilized to extract features of an input compound one-dimensional SMILES sequence, a two-dimensional molecular graph sequence, a protein structure attribute sequence and a protein amino acid sequence; and (3) utilizing the extracted characteristic vector matrix to a target prediction model, wherein the target prediction model is to utilize convolution, pooling and full connection to operate the characteristic vector matrix and output root mean square error values of a predicted value and an actual value which are combined with affinity.
8. The computer device of claim 7, wherein the bidirectional gated cyclic unit model enables data to be input simultaneously from both the front and back directions, such that the information at each time includes sequence information of preceding and following times, which is equivalent to the increase of sequence information of the network at a specific time, and the information of historical data is fully utilized, thereby making prediction more accurate. The gating cycle unit carries out sufficient feature extraction on the multivariate time sequence and continuously learns the long-term dependence of the multivariate time sequence, and the gating cycle unit specifically comprises the following steps: firstly, two gating states are obtained through the last transmitted state and the input of the current node, namely gating for controlling reset and gating for controlling update respectively, after a gating signal is obtained, data after reset is obtained through resetting gating, the data and the input of the current node are spliced, the data are shrunk to the range of-1 to 1 through a hyperbolic tangent function, finally, the states are updated to the range of 0 to 1 through the functions of forgetting and memorizing through the updating gating, and the more the gating signal is close to 1, the more data representing the memory are.
9. The computer device of claim 7, wherein the inputs to the model are selected 4 variables, the input variables comprising a protein structure attribute sequence and a protein amino acid sequence from the UniRef database, a compound SMILES from the STITCH database, and a compound two-dimensional molecular map. Wherein the protein structure attribute sequence is encoded by the secondary structure of the protein, the length of the protein amino acid sequence, the physicochemical properties (polarity/non-polarity, acidity/alkalinity) of the protein and the solvent accessibility of the protein.
10. A storage medium storing a computer program, the memory, when executed by a central processing unit, performing the steps comprising: the system comprises three bidirectional gating cyclic unit (BiGRU) models, a graph convolution neural network model (GCN) and a Convolution Neural Network (CNN) model, and the whole network architecture is tripleBiGRU/GCN-CNN. The bidirectional gated cyclic unit model comprises a sequence processing model consisting of two gated cyclic units (GRUs), wherein one input is a forward input, the other input is a reverse input, and the two-way recursive neural network is a bidirectional recursive neural network with only an input gate and a forgetting gate. The input of the model is a compound one-dimensional SMILES sequence, a compound two-dimensional molecular graph, a protein structure attribute sequence and a protein amino acid sequence, and the four variables are respectively input into the tripleBiGRU/GCN model. The TripleBiGRU/GCN model output is a feature vector representing the compound and a feature vector representing the protein. The CNN model consists of a convolution layer, a pooling layer and a full-connection layer, and the input of the CNN model is a characteristic vector of a compound and a characteristic vector of a protein; the final output of the TripleBiGRU/GCN-CNN model is the root mean square error value of the predicted compound protein affinity value.
CN202011499748.2A 2020-12-18 2020-12-18 Novel depth model, computer device, storage medium for predicting compound protein interaction based on deep learning method Pending CN112530514A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011499748.2A CN112530514A (en) 2020-12-18 2020-12-18 Novel depth model, computer device, storage medium for predicting compound protein interaction based on deep learning method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011499748.2A CN112530514A (en) 2020-12-18 2020-12-18 Novel depth model, computer device, storage medium for predicting compound protein interaction based on deep learning method

Publications (1)

Publication Number Publication Date
CN112530514A true CN112530514A (en) 2021-03-19

Family

ID=75001359

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011499748.2A Pending CN112530514A (en) 2020-12-18 2020-12-18 Novel depth model, computer device, storage medium for predicting compound protein interaction based on deep learning method

Country Status (1)

Country Link
CN (1) CN112530514A (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112735535A (en) * 2021-04-01 2021-04-30 腾讯科技(深圳)有限公司 Prediction model training method, prediction model training device, data prediction method, data prediction device and storage medium
CN113192559A (en) * 2021-05-08 2021-07-30 中山大学 Protein-protein interaction site prediction method based on deep map convolution network
CN113241130A (en) * 2021-06-08 2021-08-10 西南交通大学 Molecular structure prediction method based on graph convolution network
CN117912591A (en) * 2024-03-19 2024-04-19 鲁东大学 Kinase-drug interaction prediction method based on deep contrast learning

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112735535A (en) * 2021-04-01 2021-04-30 腾讯科技(深圳)有限公司 Prediction model training method, prediction model training device, data prediction method, data prediction device and storage medium
CN112735535B (en) * 2021-04-01 2021-06-25 腾讯科技(深圳)有限公司 Prediction model training method, prediction model training device, data prediction method, data prediction device and storage medium
CN113192559A (en) * 2021-05-08 2021-07-30 中山大学 Protein-protein interaction site prediction method based on deep map convolution network
CN113192559B (en) * 2021-05-08 2023-09-26 中山大学 Protein-protein interaction site prediction method based on deep graph convolution network
CN113241130A (en) * 2021-06-08 2021-08-10 西南交通大学 Molecular structure prediction method based on graph convolution network
CN117912591A (en) * 2024-03-19 2024-04-19 鲁东大学 Kinase-drug interaction prediction method based on deep contrast learning
CN117912591B (en) * 2024-03-19 2024-05-31 鲁东大学 Kinase-drug interaction prediction method based on deep contrast learning

Similar Documents

Publication Publication Date Title
CN112530515A (en) Novel deep learning model for predicting protein affinity of compound, computer equipment and storage medium
CN112530514A (en) Novel depth model, computer device, storage medium for predicting compound protein interaction based on deep learning method
US12014257B2 (en) Domain specific language for generation of recurrent neural network architectures
CN112542211A (en) Method for predicting protein affinity of compound based on single attention mechanism, computer device and storage medium
CN112582020A (en) Method for predicting compound protein affinity based on edge attention mechanism, computer device and storage medium
CN112562781A (en) Novel coding scheme, computer device and storage medium for predicting compound protein affinity based on deep learning
CN107862173B (en) Virtual screening method and device for lead compound
CN111259738B (en) Face recognition model construction method, face recognition method and related device
CN106021990B (en) A method of biological gene is subjected to classification and Urine scent with specific character
CN110197099A (en) The method and apparatus of across age recognition of face and its model training
WO2019022854A1 (en) Data2data: deep learning for time series representation and retrieval
CN112070277A (en) Hypergraph neural network-based drug-target interaction prediction method
CN112652358A (en) Drug recommendation system, computer equipment and storage medium for regulating and controlling disease target based on three-channel deep learning
CN109840595B (en) Knowledge tracking method based on group learning behavior characteristics
CN113033603A (en) Weak supervision image multi-label classification method based on meta-learning
CN115346372B (en) Multi-component fusion traffic flow prediction method based on graph neural network
CN112562791A (en) Drug target action depth learning prediction system based on knowledge graph, computer equipment and storage medium
CN112086144A (en) Molecule generation method, molecule generation device, electronic device, and storage medium
CN116206775A (en) Multi-dimensional characteristic fusion medicine-target interaction prediction method
CN114463596A (en) Small sample image identification method, device and equipment of hypergraph neural network
CN114707641A (en) Training method, device, equipment and medium for neural network model of double-view diagram
CN114743590A (en) Drug-target affinity prediction system based on graph convolution neural network, computer device and storage medium
CN114399901B (en) Method and equipment for controlling traffic system
CN114707655B (en) Quantum line conversion method, quantum line conversion system, storage medium and electronic equipment
CN116108195A (en) Dynamic knowledge graph prediction method and device based on time sequence element learning

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
WD01 Invention patent application deemed withdrawn after publication
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20210319