CN115691669A - Protein structure classification system based on quantum convolution neural network - Google Patents

Protein structure classification system based on quantum convolution neural network Download PDF

Info

Publication number
CN115691669A
CN115691669A CN202310000900.5A CN202310000900A CN115691669A CN 115691669 A CN115691669 A CN 115691669A CN 202310000900 A CN202310000900 A CN 202310000900A CN 115691669 A CN115691669 A CN 115691669A
Authority
CN
China
Prior art keywords
quantum
protein
module
loss function
neural network
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202310000900.5A
Other languages
Chinese (zh)
Other versions
CN115691669B (en
Inventor
胡咏梅
刘海建
耿咏忠
李宁
杨昱升
赵立祥
崔国龙
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Zhongke Arc Quantum Software Technology Co ltd
Sinopharm Bio Pharmaceutical Co Ltd
Original Assignee
Beijing Zhongke Arc Quantum Software Technology Co ltd
Sinopharm Bio Pharmaceutical Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Zhongke Arc Quantum Software Technology Co ltd, Sinopharm Bio Pharmaceutical Co Ltd filed Critical Beijing Zhongke Arc Quantum Software Technology Co ltd
Priority to CN202310000900.5A priority Critical patent/CN115691669B/en
Publication of CN115691669A publication Critical patent/CN115691669A/en
Application granted granted Critical
Publication of CN115691669B publication Critical patent/CN115691669B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Landscapes

  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The invention discloses a protein structure classification system based on a quantum convolutional neural network, wherein the system comprises: the coding module of protein sequence amino acid characteristic data is used for extracting and reading protein sequence information and corresponding structural information from the protein structure classification data set; a quantum convolution and pooling module for effecting classification of the protein results by parameterized quantum gates; a build loss function module for obtaining a loss function for characterizing the system performance; and the update quantum wire parameter module is used for updating the quantum wire parameters. Compared with the prior art, the invention realizes an efficient quantum computer convolution neural network system, can efficiently classify protein structures, and greatly accelerates the prediction of the protein structures and the development of drugs by using the model used by the system.

Description

Protein structure classification system based on quantum convolution neural network
Technical Field
The invention belongs to the technical field of quantum computers, and particularly relates to a protein structure classification system based on a quantum convolutional neural network.
Background
Proteins are the main players of life activities, and their functions and structures are closely related. The accuracy of protein structure prediction can be greatly improved by effectively classifying the protein structures. At present, the existing classical machine learning algorithm (neural network, support vector machine, random forest and the like) does much work in the direction of protein structure classification. These efforts first lead to the preprocessing of data based on a protein data set, typically having sequence information, secondary structure information, mutation information, etc. The traditional computer can store the protein data information to the classical bit by means of one-hot coding and the like. The data set is then divided into a training data set and a test data set. On the training data set, extracting features through a machine learning algorithm, training to obtain a prediction model, and then testing the accuracy of the test model on the data set.
The prior art uses machine learning models on classical computers to classify protein structures. The classical computer uses classical bits for calculation, which is different from the quantum ratio used by the quantum computer. The coding of proteins in classical computers does not reflect well the intrinsic information of the protein. The amino acid sequence in the protein has time sequence property, and if the protein is coded on a quantum bit by using a quantum computer, the time sequence property of the amino acid sequence can be well embodied by the entanglement property of the quantum bit. In addition, protein databases are large in number and variety, and classical computers have limited capabilities of storing data, calculating data and the like, so that excessively large data sets cannot be processed.
Disclosure of Invention
In view of the above-mentioned drawbacks in the prior art, the present invention provides a protein structure classification system based on quantum convolutional neural network, which includes: a coding module of protein sequence amino acid characteristic data, a quantum convolution and pooling module, a construction loss function module and an update quantum circuit parameter module,
the coding module of the protein sequence amino acid characteristic data is used for extracting and reading protein sequence information and corresponding structural information from the protein structure classification data set;
a quantum convolution and pooling module for effecting classification of the protein results by parameterized quantum gates;
a build loss function module for obtaining a loss function for characterizing the system performance;
an update quantum wire parameter module for updating the quantum wire parameters.
Wherein the protein structure classification dataset is classified according to 99: the scale of 1 is divided into a training data set and a test data set.
Wherein the quantum convolution and pooling module comprises:
the quantum convolution layer basic unit is used for evolving the quantum state loaded with the protein sequence characteristic information;
a quantum-pooling layer basic unit for mapping information of two qubits onto one qubit.
Wherein the quantum convolution and pooling module is further configured to measure the Polly Z expectation of the last qubit as the final predictor of the protein structure classification by alternating the quantum convolution layer and the quantum pooling layer until only one qubit remains.
Wherein the loss function module is used for constructing protein amino acid sequence characteristic data in each batch b
Figure 416393DEST_PATH_IMAGE002
The input is based on the quantum convolution and pooling module, and each protein amino acid sequence obtains a predicted value through the quantum convolution and pooling module
Figure 67955DEST_PATH_IMAGE004
And then obtaining a loss function for characterizing the system performance by solving the mean square error of the predicted value of all protein amino acid sequences of each batch relative to the real label of the predicted value.
Wherein the loss function is expressed by the following equation:
Figure 441167DEST_PATH_IMAGE005
wherein
Figure 882382DEST_PATH_IMAGE007
And k is the number of protein amino acid sequences contained in the batch b.
The quantum line parameter updating module is specifically configured to solve an analytic gradient of the loss function with respect to the quantum line parameter based on a parameterized circuit movement rule, and then update the quantum line parameter.
Wherein, the solving of the analytic gradient of the loss function with respect to the quantum wire parameter based on the parameterized circuit movement rule specifically includes:
hypothetical measurement operator
Figure 431175DEST_PATH_IMAGE009
In parametric quantum wires
Figure 847113DEST_PATH_IMAGE010
The expected value of (A) can be expressed as
Figure 661616DEST_PATH_IMAGE011
Wherein the content of the first and second substances,
Figure 594937DEST_PATH_IMAGE010
representing parameterized quantum wires composed of quantum convolutional layers and pooling layers,
Figure 857291DEST_PATH_IMAGE013
representing parameters in the quantum convolutional layer and the pooling layer;
then the expected value function
Figure 896660DEST_PATH_IMAGE014
With respect to parameterized quantum line parameters
Figure 385410DEST_PATH_IMAGE016
Can be expressed as
Figure 715897DEST_PATH_IMAGE017
Wherein the system trains a plurality of epochs using the training data set until a desired accuracy is reached.
Compared with the prior art, the invention realizes an efficient quantum computer convolution neural network system, can efficiently classify protein structures, and greatly accelerates the prediction of the protein structures and the development of drugs by using the model used by the system.
Drawings
The above and other objects, features and advantages of exemplary embodiments of the present application will become readily apparent from the following detailed description, which proceeds with reference to the accompanying drawings. Several embodiments of the present application are illustrated by way of example, and not by way of limitation, in the figures of the accompanying drawings and in which like reference numerals refer to similar or corresponding parts and in which:
FIG. 1 is PSSM matrix data showing amino acid signature properties of proteins according to embodiments of the present invention;
fig. 2 (a) is a diagram showing a quantum wire encoding 20-dimensional data of a single amino acid of a protein sequence onto 10 qubits according to an embodiment of the present invention;
FIG. 2 (b) is a block diagram showing an encoded implementation of amino acid sequence characterization data for an entire protein according to an embodiment of the present invention;
FIG. 3 (a) is a block diagram showing a basic cell implementation of a quantum convolutional layer according to an embodiment of the present invention;
FIG. 3 (b) is a block diagram showing a basic cell implementation of a quantum pooling layer according to an embodiment of the present invention;
FIG. 4 is a block diagram illustrating a quantum convolutional neural network-based protein structure classification system according to an embodiment of the present invention;
FIG. 5 is a block diagram illustrating a protein structure classification system based on a quantum convolutional neural network according to an embodiment of the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, the present invention will be described in further detail with reference to the accompanying drawings. All other embodiments, which can be obtained by a person skilled in the art without making any creative effort based on the embodiments in the present invention, belong to the protection scope of the present invention.
The terminology used in the embodiments of the invention is for the purpose of describing particular embodiments only and is not intended to be limiting of the invention. As used in the examples of the present invention and the appended claims, the singular forms "a", "an", and "the" are intended to include the plural forms as well, unless the context clearly indicates otherwise, and "a plurality" typically includes at least two.
It should be understood that, although the terms first, second, third, etc. may be used to describe \8230;, these \8230;, should not be limited to these terms in embodiments of the present invention. These terms are used only to distinguish between \8230; \8230. For example, without departing from the scope of embodiments of the present invention, a first of the methods may be used as a first of the methods for manufacturing a semiconductor device, and the method may be used as a second of the methods for manufacturing a semiconductor device, wherein the first of the methods may be used as a second of the methods for manufacturing a semiconductor device, and the second of the methods may be used as a second of the methods for manufacturing a semiconductor device.
It should be understood that the term "and/or" as used herein is merely one type of association that describes an associated object, meaning that three relationships may exist, e.g., a and/or B may mean: a exists alone, A and B exist simultaneously, and B exists alone. In addition, the character "/" herein generally indicates that the former and latter related objects are in an "or" relationship.
The words "if", as used herein may be interpreted as "at \8230; \8230whenor" when 8230; \8230when or "in response to a determination" or "in response to a detection", depending on the context. Similarly, the phrases "if determined" or "if detected (a stated condition or event)" may be interpreted as "when determined" or "in response to a determination" or "when detected (a stated condition or event)" or "in response to a detection (a stated condition or event)", depending on the context.
It should also be noted that the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a good or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such good or apparatus. Without further limitation, an element defined by the phrases "comprising one of \8230;" does not exclude the presence of additional like elements in an article or device comprising the element.
The related terms of the present application:
PDB (Protein Data Bank) Protein database
NISQ (noise intermediate-scale Quantum) noise-containing mesoscale quantum computer
SCOP (Structural Classification of Proteins) protein structure Classification database
PSSM (Position-specific score matrix) site-specific scoring matrix
The quantum convolution neural network loads the characteristic vector representing the protein sequence amino acid into a quantum state based on an amplitude coding mode, and then processes the quantum state containing the protein sequence amino acid characteristic information through a quantum convolution layer and a quantum pooling layer which respectively correspond to the classical convolution and the pooling. In the process, the dimensionality of the qubits is continuously reduced, finally, one qubit is measured, the information obtained by measurement and the real label of the structural classification of the protein are combined into a loss function, and the parameters are continuously updated according to the loss function until a satisfactory threshold value is reached.
Alternative embodiments of the present invention are described in detail below with reference to the accompanying drawings.
The first embodiment,
The invention provides a protein structure classification system based on a quantum convolutional neural network, which comprises the following components: the system comprises a protein sequence amino acid characteristic data coding module, a quantum convolution and pooling module, a loss function building module and a quantum line parameter updating module.
Wherein, the coding module of the protein sequence amino acid characteristic data is used for extracting protein sequence information and corresponding structure information from a protein structure classification data set (SCOP and the like). The data set was normalized to 99: the scale of 1 is divided into a training data set and a test data set. For amino acid sequence information in the dataset, twenty-dimensional vectors were encoded using the PSSM method. By using
Figure 708124DEST_PATH_IMAGE018
The gyrotron gate acts on qubits (here we use 10 qubits) to load 20-dimensional protein sequence amino acid signature data onto the amplitude of the quantum states.
A quantum convolution and pooling module for building a quantum convolution layer elementary unit and a quantum pooling layer elementary unit, respectively, that can act on two quantum bits, by means of parameterized quantum gates. Then, the quantum convolution layer is formed by the action of the quantum convolution layer basic unit on every two quantum bit pairs of the quantum system, so that the evolution of the quantum state of the previous module loaded with the protein sequence amino acid characteristic information is carried out. Subsequently, a quantum pooling basic unit acts on each quantum bit pair of the quantum system to form a quantum pooling layer, and information of two quantum bits is mapped to one quantum bit, wherein the number of the quantum bits containing high-level information is 5. And then alternately acting the quantum convolution layer and the quantum pooling layer on the remaining 5 qubits, wherein the qubits containing high-level information are 3 qubits, repeating the steps until only one qubit is left by alternately acting the quantum convolution layer and the quantum pooling layer, and finally measuring the Pally Z expected value of the last qubit to be used as a final predicted value of the protein structure classification.
Constructing a loss function module for characterizing the amino acid sequence of the protein in each batch b
Figure 100002_DEST_PATH_IMAGE020
When the protein is input into a quantum convolution neural network built based on the previous module, each protein amino acid sequence can obtain a predicted value through the previous module
Figure 544493DEST_PATH_IMAGE004
And then obtaining a loss function for characterizing the performance of the model by solving the mean square error of the predicted value of all protein amino acid sequences of each batch relative to the real label of the predicted value.
And the updating quantum line parameter module is used for solving the analytic gradient of the loss function of the last module relative to the quantum line parameters based on the existing parameterized circuit movement rule, then updating the quantum line parameters by using a classical computer, finally training a plurality of epochs by using a protein amino acid sequence training data set, and stopping updating after the expected accuracy is reached.
Example II,
In order to further illustrate the method for predicting the protein structure based on the quantum convolution neural network, the following embodiments are provided:
an encoding module for protein sequence amino acid feature data based on PSSM matrix data characterizing the amino acid feature attributes of each protein, as shown in fig. 1, where each letter represents one of 20 amino acids and each amino acid has a feature vector dimension of 20. By passing
Figure 254960DEST_PATH_IMAGE018
The rotaron gate loads protein amino acid sequence data onto the amplitude of the quantum state (here we use 10 quantum bits) based on the characteristic attribute data of each protein amino acid, with the specific quantum wires as shown in fig. 2. Fig. 2 (a) shows quantum wires that encode 20-dimensional data of a single amino acid of a protein sequence onto 10 qubits. FIG. 2 (b) shows the coding of the characteristic data of the entire protein amino acid sequence, here by way of example the protein amino acid sequence shown in the figure, from which it can be seen that the 20-dimensional characteristic data characterizing methionine (M) are first coded in the manner of FIG. 2 (a) onto the amplitudes of the 10 qubit quantum states. Then, the characteristic data of threonine (T) is encoded to the quantum state, and so on until the whole protein sequence is encoded.
The quantum convolution and pooling module comprises a quantum convolution layer basic unit and a quantum pooling layer basic unit which act on two quantum bits, wherein corresponding quantum circuits are shown in figure 3, figure 3 (a) is implemented by the quantum convolution layer basic unit, and figure 3 (b) is implemented by the quantum pooling layer basic unit. Based on the quantum convolution layer basic unit and the quantum convolution pooling layer basic unit, the final predicted value can be obtained by alternately acting the quantum convolution layer and the pooling layer. Specifically, as shown in fig. 4, a block C in the figure represents the quantum convolution layer basic unit in fig. 3 (a), a block P represents the quantum pooling layer basic unit, a block C portion in a dotted line frame represents the first quantum convolution layer, and a block P portion in a dotted line frame represents the first quantum pooling layer. As shown in FIG. 4, we obtain the final predicted value as whether the protein amino acid sequence is an alpha helix structure or not by alternately acting quantum convolution and quantum pooling layers so that the information containing the amino acid characteristics of the protein sequence is finally loaded on one qubit and measuring the Pally Z expected value of the last qubit, wherein the truncated qubit indicates the action of the basic unit without the quantum convolution layer and the pooling layer.
Constructing a loss function module that will eachAmino acid sequence characterization data of proteins in batch b
Figure 576089DEST_PATH_IMAGE020
Inputting the protein into a quantum convolution neural network built by the previous module, and obtaining a corresponding predicted value for each protein amino acid sequence
Figure 813035DEST_PATH_IMAGE004
. Finally, the predicted values obtained by all protein amino acid sequences in the batches are calculated
Figure 882622DEST_PATH_IMAGE004
And (3) calculating the mean square error between the predicted value and the true value of the protein amino acid sequence by combining the true tag of whether the predicted value is in an alpha helical structure or not, so as to obtain a loss function for representing the performance of the quantum convolution neural network model, wherein the expression of the mean square error loss function is shown as the following formula.
Figure 955752DEST_PATH_IMAGE005
Wherein
Figure 831304DEST_PATH_IMAGE007
K is the number of amino acid sequences of the protein contained in the batch b.
Update quantum line parameter module: firstly, a measuring operator \ hat M \]In parametric quantum wires
Figure 798123DEST_PATH_IMAGE010
The expected value of (A) can be expressed as
Figure 406653DEST_PATH_IMAGE011
,(2)
Then the expected value function
Figure 91713DEST_PATH_IMAGE014
About a parameterChemometric line parameters
Figure 770956DEST_PATH_IMAGE016
Can be expressed as
Figure 467647DEST_PATH_IMAGE017
,(3)
In the above formula (3)
Figure 879037DEST_PATH_IMAGE010
Representing parameterized quantum wires composed of quantum convolutional layers and pooling layers,
Figure 176026DEST_PATH_IMAGE016
representing parameters in the quantum convolution layer and the pooling layer.
The above method is called a parameter-shifting rule for solving the gradient of the parameterized quantum wires with respect to the desired value of the operator.
Through the parameter moving rule, the analytical gradient of the mean square error loss function of the previous module, namely the formula (1), on the quantum convolution layer and pooling layer quantum circuit parameters can be obtained. The parameters are then updated by a gradient descent method using a classical computer. And finally, training a plurality of epochs according to a training data set of protein amino acid sequence characteristic data until the protein structure classification predicted based on the quantum convolution neural network provided by the patent is accurate to a desired degree.
EXAMPLE III
As shown in fig. 5, the present invention provides a protein structure classification system based on quantum convolutional neural network, which includes: a coding module of protein sequence amino acid characteristic data, a quantum convolution and pooling module, a construction loss function module and an update quantum line parameter module,
the coding module of the protein sequence amino acid characteristic data is used for extracting and reading protein sequence information and corresponding structural information from the protein structure classification data set;
a quantum convolution and pooling module for effecting classification of the protein results by parameterized quantum gates;
a build loss function module for obtaining a loss function for characterizing the system performance;
an update quantum wire parameter module for updating the quantum wire parameters.
Wherein the protein structure classification dataset is classified according to 99: the scale of 1 is divided into a training data set and a test data set.
Wherein the quantum convolution and pooling module comprises:
the quantum convolution layer basic unit is used for evolving the quantum state loaded with the protein sequence characteristic information;
a quantum-pooling layer basic unit for mapping the information of two qubits onto one qubit.
Wherein the quantum convolution and pooling module is further configured to measure the Polly Z expectation of the last qubit as the final predictor of the protein structure classification by alternating the quantum convolution layer and the quantum pooling layer until only one qubit remains.
Wherein the loss function module is used for constructing protein amino acid sequence characteristic data in each batch b
Figure 65485DEST_PATH_IMAGE020
The input is based on the quantum convolution and pooling module, and each protein amino acid sequence obtains a predicted value through the quantum convolution and pooling module
Figure 380797DEST_PATH_IMAGE004
And then obtaining a loss function for characterizing the system performance by solving the mean square error of the predicted value of all protein amino acid sequences of each batch relative to the real label of the predicted value.
Wherein the loss function is expressed as follows:
Figure 290985DEST_PATH_IMAGE005
wherein k is the number of amino acid sequences of the protein contained in batch b.
The quantum line parameter updating module is specifically used for solving the analytic gradient of the loss function relative to the quantum line parameters based on the parameterized circuit movement rule, and then updating the quantum line parameters.
Wherein, the calculating the analytic gradient of the loss function with respect to the quantum line parameter based on the parameterized circuit movement rule specifically includes:
hypothetical measurement operator
Figure 950636DEST_PATH_IMAGE009
In parametric quantum wires
Figure 519152DEST_PATH_IMAGE010
The expected value of (A) can be expressed as
Figure 49490DEST_PATH_IMAGE011
Wherein the content of the first and second substances,
Figure 130579DEST_PATH_IMAGE010
representing parameterized quantum wires composed of quantum convolutional layers and pooling layers,
Figure 651428DEST_PATH_IMAGE016
representing parameters in the quantum convolutional layer and the pooling layer;
then the expected value function
Figure 148268DEST_PATH_IMAGE014
With respect to parameterized quantum line parameters
Figure 126588DEST_PATH_IMAGE016
Can be expressed as
Figure 660469DEST_PATH_IMAGE017
Wherein the system trains a plurality of epochs using the training data set until a desired accuracy is reached.
Example four,
Embodiments of the present invention provide a non-volatile computer storage medium, where computer-executable instructions are stored, and the computer-executable instructions may perform the method steps described in the above embodiments.
It should be noted that the computer readable medium mentioned above in the present application may be a computer readable signal medium or a computer readable storage medium or any combination of the two. A computer readable storage medium may be, but is not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any combination of the foregoing. More specific examples of the computer readable storage medium may include, but are not limited to: an electrical connection having one or more wires, a portable computer diskette, a hard disk, a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. In the context of this application, a computer readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device. In this application, however, a computer readable signal medium may include a propagated data signal with computer readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated data signal may take many forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof. A computer readable signal medium may also be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device. Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to: electrical wires, optical cables, RF (radio frequency), etc., or any suitable combination of the foregoing.
The computer readable medium may be embodied in the electronic device; or may exist separately without being assembled into the electronic device.
Computer program code for carrying out operations for aspects of the present application may be written in any combination of one or more programming languages, including an object oriented programming language such as Java, smalltalk, C + + or the like and conventional procedural programming languages, such as the "C" programming language or similar programming languages. The program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In the latter scenario, the remote computer may be connected to the user's computer through any type of network, including a local Area Network (AN) or a Wide Area Network (WAN), or the connection may be made to AN external computer (for example, through the internet using AN internet service provider).
The flowchart and block diagrams in the figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods and computer program products according to various embodiments of the present application. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems which perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.
The units described in the embodiments of the present invention may be implemented by software or hardware. Wherein the name of an element does not in some cases constitute a limitation on the element itself.
The foregoing describes preferred embodiments of the present invention, and is intended to make the spirit and scope of the present invention clear and understandable, but not to limit the present invention, and modifications, substitutions and improvements made within the spirit and principle of the present invention are included in the scope of the present invention as outlined by the appended claims.

Claims (9)

1. A system for protein structure classification based on a quantum convolutional neural network, comprising: a coding module of protein sequence amino acid characteristic data, a quantum convolution and pooling module, a construction loss function module and an update quantum line parameter module,
the coding module of the protein sequence amino acid characteristic data is used for extracting and reading protein sequence information and corresponding structural information from the protein structure classification data set;
a quantum convolution and pooling module for effecting classification of the protein results by parameterized quantum gates;
a build loss function module for obtaining a loss function for characterizing the system performance;
an update quantum wire parameter module for updating the quantum wire parameters.
2. The quantum convolutional neural network-based protein structure classification system of claim 1, wherein the protein structure classification dataset is classified according to a 99: a scale of 1 into a training data set and a test data set.
3. The quantum convolutional neural network-based protein structure classification system of claim 1, wherein the quantum convolutional and pooling module comprises:
the quantum convolution layer basic unit is used for evolving the quantum state loaded with the protein sequence characteristic information;
a quantum-pooling layer basic unit for mapping information of two qubits onto one qubit.
4. The system according to claim 3, wherein the QW neural network-based protein structure classification system further comprises a QW and pooling module for alternating the QW layer and the QW pooling layer until only one qubit remains, and then measuring the Pally Z expectation of the last qubit as a final predictor of the protein structure classification.
5. The system for classifying protein structures based on quantum convolutional neural network as claimed in claim 1, wherein said module for constructing loss function is specifically used for characterizing the amino acid sequence of protein in each batch b
Figure DEST_PATH_IMAGE002A
The input is based on the quantum convolution and pooling module, and each protein amino acid sequence obtains a predicted value through the quantum convolution and pooling module
Figure DEST_PATH_IMAGE004A
Then, by solving the mean square error of the predicted value of all protein amino acid sequences of each batch relative to the real label thereof, a loss function for characterizing the system performance is obtained.
6. The system for classifying protein structures based on quantum convolutional neural network as claimed in claim 5, wherein the loss function is expressed as:
Figure DEST_PATH_IMAGE006A
wherein
Figure DEST_PATH_IMAGE008A
And k is the number of protein amino acid sequences contained in the batch b.
7. The system of claim 1, wherein the update quantum wire parameters module is further configured to apply an analytical gradient of the loss function with respect to the quantum wire parameters based on a parameterized circuit motion law, and then update the quantum wire parameters.
8. The system according to claim 1, wherein the step of solving the analytical gradient of the loss function with respect to the quantum wire parameters based on the parameterized circuit motion algorithm comprises:
hypothetical measurement operator
Figure DEST_PATH_IMAGE010A
In parametric quantum wires
Figure DEST_PATH_IMAGE012A
The expected value of (A) can be expressed as
Figure DEST_PATH_IMAGE014A
Wherein a parameterized quantum wire composed of a quantum convolution layer and a pooling layer is represented,
Figure DEST_PATH_IMAGE016
representing parameters in the quantum convolutional layer and the pooling layer;
then the expected value function
Figure DEST_PATH_IMAGE018
In connection with ginsengThe gradient of the parameter of the quantised quantum line may be expressed as
Figure DEST_PATH_IMAGE020
9. The system for classifying protein structures based on quantum convolutional neural network as claimed in claim 2, wherein said system trains a plurality of epochs using said training data set until a desired accuracy is reached.
CN202310000900.5A 2023-01-03 2023-01-03 Protein structure classification system based on quantum convolution neural network Active CN115691669B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202310000900.5A CN115691669B (en) 2023-01-03 2023-01-03 Protein structure classification system based on quantum convolution neural network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202310000900.5A CN115691669B (en) 2023-01-03 2023-01-03 Protein structure classification system based on quantum convolution neural network

Publications (2)

Publication Number Publication Date
CN115691669A true CN115691669A (en) 2023-02-03
CN115691669B CN115691669B (en) 2023-03-17

Family

ID=85057096

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202310000900.5A Active CN115691669B (en) 2023-01-03 2023-01-03 Protein structure classification system based on quantum convolution neural network

Country Status (1)

Country Link
CN (1) CN115691669B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116304883A (en) * 2023-04-25 2023-06-23 电子科技大学 Quantum convolution-based molecular diagram classification method, system, storage medium and terminal

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112289370A (en) * 2020-12-28 2021-01-29 武汉金开瑞生物工程有限公司 Protein structure prediction method and device based on multitask time domain convolutional neural network
US20210183473A1 (en) * 2017-11-08 2021-06-17 Stc.Unm. System and methods for graphic encoding of macromolecules for efficient high-throughput analysis
CN113159239A (en) * 2021-06-28 2021-07-23 北京航空航天大学 Method for processing graph data by quantum graph convolutional neural network
CN114444701A (en) * 2022-02-01 2022-05-06 上海图灵智算量子科技有限公司 Training quantum circuit and data embedding method

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20210183473A1 (en) * 2017-11-08 2021-06-17 Stc.Unm. System and methods for graphic encoding of macromolecules for efficient high-throughput analysis
CN112289370A (en) * 2020-12-28 2021-01-29 武汉金开瑞生物工程有限公司 Protein structure prediction method and device based on multitask time domain convolutional neural network
CN113159239A (en) * 2021-06-28 2021-07-23 北京航空航天大学 Method for processing graph data by quantum graph convolutional neural network
CN114444701A (en) * 2022-02-01 2022-05-06 上海图灵智算量子科技有限公司 Training quantum circuit and data embedding method

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
范兴奎;刘广哲;王浩文;马鸿洋;李伟;王淑梅: "《基于量子卷积神经网络的图像识别新模型》" *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116304883A (en) * 2023-04-25 2023-06-23 电子科技大学 Quantum convolution-based molecular diagram classification method, system, storage medium and terminal

Also Published As

Publication number Publication date
CN115691669B (en) 2023-03-17

Similar Documents

Publication Publication Date Title
EP3514734B1 (en) Method and apparatus for generating a chemical structure using a neural network
CN112561069B (en) Model processing method, device, equipment and storage medium
CN110995459B (en) Abnormal object identification method, device, medium and electronic equipment
CN112994701B (en) Data compression method, device, electronic equipment and computer readable medium
EP4273754A1 (en) Neural network training method and related device
CN115618269B (en) Big data analysis method and system based on industrial sensor production
CN114048468A (en) Intrusion detection method, intrusion detection model training method, device and medium
CN115691669B (en) Protein structure classification system based on quantum convolution neural network
CN112420125A (en) Molecular attribute prediction method and device, intelligent equipment and terminal
CN114912578A (en) Training method and device of structure response prediction model and computer equipment
CN114358216A (en) Quantum clustering method based on machine learning framework and related device
CN117875481A (en) Carbon emission prediction method, electronic device, and computer-readable medium
CN116431807B (en) Text classification method and device, storage medium and electronic device
CN115691654B (en) Method for predicting antibacterial peptide of quantum gate-controlled circulating neural network based on fewer parameters
US11036980B2 (en) Information processing method and information processing system
CN111582313A (en) Sample data generation method and device and electronic equipment
CN115858388A (en) Test case priority ordering method and device based on variation model mapping chart
CN114819140A (en) Model pruning method and device and computer equipment
KR102454742B1 (en) Method for analyzing thickness of cortical region
US11100321B2 (en) Information processing method and information processing system
CN114140246A (en) Model training method, fraud transaction identification method, device and computer equipment
CN113643141A (en) Method, device and equipment for generating explanatory conclusion report and storage medium
US20210295136A1 (en) Improvement of Prediction Performance Using Asymmetric Tanh Activation Function
Zhang et al. MLLBC: A machine learning toolbox for modeling the loss rate of the lining bearing capacity
JP6701467B2 (en) Learning device and learning method

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
CB03 Change of inventor or designer information

Inventor after: Hu Yongmei

Inventor after: Liu Haijian

Inventor after: Geng Yongzhong

Inventor after: Li Ning

Inventor after: Yang Yusheng

Inventor after: Zhao Lixiang

Inventor after: Cui Guolong

Inventor before: Hu Yongmei

Inventor before: Liu Haijian

Inventor before: Geng Yongzhong

Inventor before: Li Ning

Inventor before: Yang Yusheng

Inventor before: Zhao Lixiang

Inventor before: Cui Guolong

CB03 Change of inventor or designer information