CN117172325A - Data coding method - Google Patents

Data coding method Download PDF

Info

Publication number
CN117172325A
CN117172325A CN202311268565.3A CN202311268565A CN117172325A CN 117172325 A CN117172325 A CN 117172325A CN 202311268565 A CN202311268565 A CN 202311268565A CN 117172325 A CN117172325 A CN 117172325A
Authority
CN
China
Prior art keywords
data
quantum
qubit
training
ith
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202311268565.3A
Other languages
Chinese (zh)
Inventor
史舜阳
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shanghai Turing Intelligent Computing Quantum Technology Co Ltd
Original Assignee
Shanghai Turing Intelligent Computing Quantum Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shanghai Turing Intelligent Computing Quantum Technology Co Ltd filed Critical Shanghai Turing Intelligent Computing Quantum Technology Co Ltd
Priority to CN202311268565.3A priority Critical patent/CN117172325A/en
Publication of CN117172325A publication Critical patent/CN117172325A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N10/00Quantum computing, i.e. information processing based on quantum-mechanical phenomena
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Software Systems (AREA)
  • General Physics & Mathematics (AREA)
  • Computing Systems (AREA)
  • Artificial Intelligence (AREA)
  • Mathematical Physics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • General Engineering & Computer Science (AREA)
  • Biophysics (AREA)
  • Molecular Biology (AREA)
  • General Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • Biomedical Technology (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Computational Mathematics (AREA)
  • Condensed Matter Physics & Semiconductors (AREA)
  • Mathematical Analysis (AREA)
  • Mathematical Optimization (AREA)
  • Pure & Applied Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Medical Informatics (AREA)
  • Complex Calculations (AREA)
  • Semiconductor Memories (AREA)

Abstract

The embodiment of the application relates to a data coding method. A method of data encoding comprising: describing the data as a general expansion under the computing substrateWherein the data contains C elements corresponding to the substrate |x respectively 1 >……|x C >The method comprises the steps of carrying out a first treatment on the surface of the Performing one or more quantum gate operations on each qubit, wherein each qubit comprises an ith qubitAnd i=1, 2, … …, n; traversing each qubit in turn to do the following: for the ith qubit, calculate 2 i A parameter expressed as p i,d Where d=1, 2,3, … …,2 i The saidWherein the data is at least one of text, protein sequence, molecular structure, and gene. The data coding method provided by the embodiment of the application can effectively solve the problems encountered in the traditional technology.

Description

Data coding method
Information of divisional application
The application is a divisional application of an application patent application with the application date of 2022, 2 and 1, the application number of 202210115012.3 and the application name of 'training quantum circuit and data embedding method'.
Technical Field
Embodiments of the present application relate generally to the field of quantum computing, and more particularly, to a method of data encoding.
Background
In recent years, machine learning has been applied in a large number of fields. As the core of artificial intelligence, with the continuous increase of data analysis requirements of various industries in the big data age, how to perform deep analysis on complex and diverse data based on machine learning, and more efficient utilization of information becomes the main direction of machine learning research in the current big data environment. To further optimize traditional machine learning, research has begun to be conducted to assist machine learning with quantum circuits. However, current methods of training quantum wires are less.
The present application thus proposes a method of data encoding.
Disclosure of Invention
It is an aim of embodiments of the present application to provide a method of data encoding that reduces the amount of parameters to be trained compared to conventional models, and also significantly reduces the number of storage media, namely qubits, used.
An embodiment of the present application provides a method of data encoding, including: describing the data as a general expansion under the computing substrateWherein the data contains C elements corresponding to the substrate |x respectively 1 >……|x C >The method comprises the steps of carrying out a first treatment on the surface of the Performing one or more quantum gate operations on each qubit, wherein each qubit comprises an i-th qubit, and i = 1,2, … …, n; traversing each qubit in turn to do the following: for the ith qubit, calculate 2 i Parameters ofThe parameter is denoted as p i,d Where d=1, 2,3, … …,2 i Said->Wherein the data is at least one of text, protein sequence, molecular structure, and gene.
Another embodiment of the present application also provides a quantum computer configured to implement the method of data encoding described above.
In yet another embodiment of the present application, a computer readable storage medium is provided, on which a computer program is stored, wherein the processor executes the computer program to implement the method for encoding data as described above.
Compared with the prior art, the data coding method provided by the embodiment of the application can effectively improve the data processing speed and accuracy.
Drawings
The drawings that are necessary to describe embodiments of the present application or the prior art will be briefly described below in order to facilitate the description of the embodiments of the present application. It is apparent that the drawings in the following description are only a few embodiments of the application. It will be apparent to those skilled in the art that other embodiments of the drawings may be made in accordance with the structures illustrated in these drawings without the need for inventive faculty.
Fig. 1 is a schematic diagram of a method of training a quantum wire according to some embodiments of the application.
Fig. 2-4 are schematic structural views of entanglement layers according to some embodiments of the application.
Fig. 5 is a schematic diagram of a method of data embedding according to some embodiments of the present application.
Fig. 6 is a schematic representation of molecular structures according to some embodiments of the applications.
Detailed Description
For a better understanding of the spirit of embodiments of the present application, a further description of some preferred embodiments of the application is provided below.
Embodiments of the present application will be described in detail below. Throughout the present specification, the same or similar components and components having the same or similar functions are denoted by similar reference numerals. The embodiments described herein with respect to the drawings are of illustrative nature, of diagrammatic nature and are provided for the basic understanding of the present application. The embodiments of the present application should not be construed as limiting the application.
In addition, for ease of description, "first," "second," "third," etc. may be used herein to distinguish between different components of a figure or series of figures. The terms "first," "second," "third," and the like are not intended to describe corresponding components.
Before describing the technical scheme of the application, some key terms related in the application are explained first:
1. quantum computing: based on the computation mode of quantum logic, the basic unit for storing data is a quantum bit.
2. Quantum bit: basic unit of quantum computation. Conventional computers use 0 and 1 as basic units of binary. Except that quantum computation can handle both 0 and 1, so that the system can be in a linear superposition of 0 and 1: i psi>=α|0>+β|1>This side α, β represents the complex probability amplitude of the system at 0 and 1. Their modulo square |alpha| 2 ,|β| 2 Representing probabilities at 0 and 1, respectively.
3. Quantum circuit: a representation of a quantum general-purpose computer represents a hardware implementation of the corresponding quantum algorithm/program under a quantum gate model. If the quantum circuit contains adjustable parameters for controlling the quantum gate, the quantum circuit is called parameterized quantum circuit.
4. Quantum gate: often using matrix representation, gates operating on n qubits can be represented by 2 n x 2 n Is a unitary matrix representation of (c). The number of qubits in one gate input and output must be equal. The operation of a quantum gate may be represented by multiplying a matrix representing the quantum gate with a vector representing the state of a qubit.
5. A revolving door: the turnstile is one of quantum gates, and is a group of three 2×2 unitary hermitian matrices (also called unitary matrices). Wherein the rotary x-gate isThe rotary y door is +.> The rotary z door is +.>
6. Quantum classical hybrid computation: the inner layer calculates corresponding physical quantity or loss function by using the quantum circuit, and the outer layer adjusts the calculation range of the variation parameters of the quantum circuit by using a traditional classical optimizer, so that the advantage of quantum calculation can be exerted to the maximum extent, and the method is believed to be one of important directions for potential demonstration of quantum advantage.
The application develops a method for training the quantum circuit for data embedding by using the quantum calculation method, and effectively improves the model learning capacity and the running efficiency by using the quantum calculation method.
Fig. 1 is a schematic diagram of a method of training a quantum wire according to some embodiments of the application. The method includes constructing a quantum wire 101 to be trained; and training with data in a target, which may include a training set or a test set, to update the quantum wire 101. Wherein training quantum wire 101 includes: encoding the data into quantum states (100 in fig. 1) using encoding line u (x); and applying the current quantum wire to the quantum state.
In other implementations of the application, the method of training a quantum wire may further include performing measurements on a computing substrate; and inputting the measured result to the classical neural network.
As shown in fig. 1, the quantum circuit 101 may be a variational circuit, and its effect may be represented by an operator V (θ), which may be trained into a suitable unitary transformation by using the training method of the present application.
The data in the target may be composed of one or more first data in the first database, for example, the data in the target may be data missing one or more first data. The first database is composed of all of the first data.
The method for training the quantum circuit can better reflect the relevance between different first data or between the data and the first data. In particular to the calculation of complex problems, the quantum machine learning is used, so that the calculation speed can be increased, and the calculation precision can be improved.
In some embodiments, the method of training a quantum wire further comprises: the first database is preprocessed. The preprocessing may include mapping each first data in the first database to a different number and mapping each first data to a computing base of the hilbert space by number.
For example, each first data is assigned a digital number (optionally, according to an existing assignment rule, or the co-occurring first data may be encoded into similar numbers, according to different methods) in a certain way, and once agreed, the numbers cannot be changed, and the numbers form the input x of the encoding line U (x) and also determine the quantum state of the encoded output.
In some embodiments, each first data in the first database (set the size of the first database as V) may be mapped to numbers of 0,1,2,3 … … V-1 one by one in some manner, optionally, co-occurring first data may be encoded into similar numbers, and then mapped to 2 of n qubits according to the numbers n The base of the individual hilbert spaces (requirement 2 n Not less than V, i.e. n not less than log 2 V), e.g., 0 first data corresponds to |0>State vector, first data number 634 corresponds to |634>First data of 999 corresponds to |999>And so on.
In some embodiments, after the first database preprocessing is completed, for one input, e.g., any data in the target, the input is provided with C words corresponding to the base |x respectively 1 >、|x 2 >、......、|x C >U (x) will handle qubits(qubit) from the initial state (use |0 for simplicity>Instead of) Transforming into an equal weight superposition state: u (x) in fig. 1 is constructed in turn to encode data in the target used as quantum states.
Unlike the traditional model in which the first data is mapped into a vector of a vector space, the method of mapping each first data into a vector of a Hilbert space can be obtained by training a parameterized quantum circuit, and the data is mapped into the vector of the Hilbert space by training a quantum circuit with a set of suitable unitary transformation parameters, so that the projection of the vector on each computing substrate can reflect the relevance of the data and corresponding data (such as missing one or more first data).
In some embodiments, encoding the data used into quantum states includes: encoding the data used in an equi-weighted superimposed state, e.g.And performing one or more quantum gate operations on each qubit, wherein each qubit comprises an i-th qubit, and i = 1,2, …, n.
In some embodiments, the data may be described as computing an under-substrate general expansionTraversing each qubit in turn to do the following: for the ith qubit, calculate 2 i A parameter, wherein said parameter may be +.>Wherein d=1, 2,3, …,2 i A kind of electronic deviceSequentially performing 2 on the ith qubit i-1 A controlled revolving door. Furthermore, the indices a=0, 1,2, …,2 can be used i-1 -1 denotes these controlled rotation gates, the control qubits being the first i-1 qubits, controlled rotation gate a performing: if the first i-1 qubits are at |a>Then the ith qubit is rotated, e.g., 0>Rotate to +.>Otherwise nothing is done. U (x) was constructed therefrom.
Although the above procedure involves a controlled gate count proportional to 2 n But due to most of C k =0 so most p i,d =0, the number of actually required controlled gates is much smaller than 2 n
The method for encoding the data into the quantum state can encode more information by using fewer quantum bits, and prepares the input data into the quantum state by converting the input data into a quantum superposition state so as to facilitate the quantum computation of the next step. Of course, the encoding of the data used into the quantum state is not limited to the above-described encoding method, and the encoding line U (x) may be constructed according to specific input data.
For example, according to the specific situation of the input data, when a line with 4 qubits is used, the corresponding label is 2, and since the binary representation of 2 is 0010, only the parameters of the controlled rotation gate in U (x) need to be set to 0, pi and 0 respectively, so that the input data is encoded.
In some embodiments, quantum wire 101 may be composed of two sub-modules, one being an entanglement module composed of N entanglement layers [ e.g. L in FIG. 1 ] i (θ)]And secondly, a single-bit rotating module consisting of a single-bit arbitrary rotating gate.
Entanglement module: the training device consists of N entanglement layers, the value of the number of layers N can be selected according to the requirements, and generally, the larger the number of layers N is, the better the training effect is. The effect of the entanglement layer is to apply a two-bit gate, such as a Controlled Not (CNOT) gate, between different quantum bits, thereby generating entanglement. Each entanglement layer can be divided into two parts: parameterized single bit rotation modules and fixed-structure two bit manipulation modules (fixed-structure refers to no adjustable parameters).
Fig. 2-4 are schematic structural views of entanglement layers according to some embodiments of the application. As shown in fig. 2-4, wherein the parameterized single bit manipulation module is composed of 3 rotational manipulations about the Y-axis, Z-axis, and Y-axis, respectively, the adjustable parameter is the angle of rotation. This is an euler rotation decomposition, meaning that such parameterized rotations can be combined into any single bit rotation operation. While two-bit operation modules have a variety of possible designs. As shown in fig. 2-4, the two-bit operation module is three typical entanglement layer structures, respectively, using nearest neighbor CNOT ring structure (105 in fig. 2), next-nearest neighbor CNOT ring structure (106 in fig. 3), and a combination of both (107 in fig. 4). The number of two-bit gates required for these examples is all in O (Nlog 2 V), V is the size of the first database and N is the number of layers of the entanglement. Of course, the entanglement modules of the application are not limited to the examples given.
Single bit rotation module: the module consists of n single bit arbitrary rotation gates for n qubits. The optional revolving door is combined by 3 rotation operations around the Y axis, the Z axis and the Y axis respectively, and the adjustable parameter is the rotation angle. This is an euler rotation angle decomposition, and such parameterized rotations can be combined into arbitrary single bit rotation operations.
The parameters in V (θ) all appear as the rotation angle of a single bit rotation gate when it is desired to measure a desired value of a certain mechanical quantity<M>For a certain parameter theta j The parameter shift rule may be applied and calculated as mentioned above for measuring the expected value and the partial derivative of the expected value.
In some embodiments, the method of training a quantum wire further comprises: a number of measurements (102 in fig. 1) are made on the computing substrate to obtain a model predicted value, e.g., an expected value, and the difference of the model predicted value and a reference value is compared, where the reference value corresponds to the data used.
In some embodiments, the method of training a quantum wire further comprises: the quantum wire 101 is updated according to the loss function using a gradient descent method, for example, the parameter θ in V (θ) is updated.
The data used can be trained repeatedly according to the required precision, the collapse probability distribution on the calculation substrate is obtained through multiple measurements and is used as a model predicted value, such as the probability distribution of what the missing one or more first data are likely to be, the true probability distribution in the first database is used as a reference value, a loss function (such as a cross entropy function) is freely selected to measure the difference between the model predicted value and the reference value, and the loss function and the gradient thereof are calculated according to the expected value and the partial derivative thereof.
For example, as shown in FIG. 1, M is any measurable physical quantity, e.g. using<M>To represent the expected values of a plurality of M to be measuredThe parameter shift rule can be applied, the remaining parameters are kept unchanged, use +.>Instead of theta j Repeating the operation of the line a number of times (the number of times is determined by the accuracy requirement) to measure a desired value +.>The other parameters are kept unchanged, use +.>Instead of theta j Repeating the operation of the line a number of times (the number of times is determined by the accuracy requirement) to measure a desired value +.>Will->Andthe partial derivative to be measured is obtained by subtracting and dividing by 2>
In some embodiments, due to the few parameters used in the quantum circuit, a classical neural network may be further used to perform fitting processing to further improve training efficiency. For example, the measured result, such as a desired value, may be input into a classical neural network, such as one or more perceptrons, e.g. as input to a multi-layer perceptron made up of one or more fully connected layers, from which the probability distribution of data corresponding to the data used in the object, such as "what the missing one or more first data is", is predicted.
In some embodiments, the target may be a training set, and multiple training is performed with different data in the training set to update the quantum wire until the entire training set is traversed, and the entire training set is iteratively trained R times (R is freely selectable) to improve training quality.
In some embodiments, the target may also be a test set, and the method of training the quantum wire further includes determining whether training is complete: and calculating a loss function according to the difference between the predicted value and the reference value and judging whether the loss function is smaller than a selected threshold value or not. If the calculated loss functions for selected data in the test set are all less than some selected threshold, the training is ended, otherwise the training of the quantum wire 101 is continued until the test of the test set can be passed.
The embodiment of the application also provides a data embedding method. Fig. 5 is a schematic diagram of a method of data embedding according to some embodiments of the present application. As shown in fig. 5, the method may include: encoding input data into quantum states (200); constructing a quantum wire 201; and applying the quantum wire 201 to the quantum state, wherein the quantum wire 201 comprises a trained variational wire.
In some embodiments of the present application, the method of training quantum wire 101 described above may be used to obtain a constructed quantum wire 201.
In some embodiments of the present application, encoding input data into quantum states is similar to the method described above for encoding input data into quantum states.
In some embodiments, after measuring the quantum state after the quantum circuit 201 acts, the result may be input into one or more perceptrons, for example, as an input of a multi-layer perceptrons composed of one or more fully connected layers, to complete the embedding of data.
The corresponding data most correlated to the input data is known by taking multiple measurements 202. Because the data embedding method uses a quantum computing method, the computing speed and accuracy are greatly improved.
The method for training the quantum circuit and embedding the data greatly reduces the parameter quantity required to be trained, namely the number of the used storage media, namely the quantum bits, and meanwhile, as the quantum state is normalized naturally, the normalization exponential function layer (softmax) probability normalization operation which consumes very much calculation force in the traditional method can be omitted, so that the method for embedding the data is more convenient and accurate.
The training quantum circuit and the data embedding method can be applied to a plurality of different fields. For example, in the field of Natural Language Processing (NLP), a quantum circuit with proper unitary transformation parameters can be obtained according to the method for training the quantum circuit, so that words and texts are mapped into state vectors of Hilbert space, and the projection of the state vectors on various computing substrates can reflect the relevance of words and texts.
Similarly, in the learning of chemical molecular characteristics, the vector of the substructures of the molecules in the chemical space can be learned by using the training quantum circuit and the data embedding method, and the vector coded by the whole compound can be finally obtained by summing the substructures. By mapping the compound substructures into the state vectors of the Hilbert space, the projection of the state vectors onto the respective computing substrates may reflect the relevance of the respective compound substructures. The method overcomes the defects of the traditional molecular characteristic representation, such as sparsity and bit conflict, and can be used for downstream tasks such as activity prediction, similarity retrieval and the like.
Fig. 6 is a schematic diagram of a molecular structure 10 according to some embodiments of the present application.
According to some embodiments of the application, the preprocessing of the molecular substructure library further comprises the steps of: the collection of compounds can be downloaded from the ZINK and chumbL database, de-duplicated for the known collection of compounds, the chemical informatics tool rdkit is used to obtain each atomic signature 11 for the compounds that remain in the collection of compounds with a molecular weight of 12-600, a heavy atom number of 3-50, a lipid water distribution coefficient clogP of-5 to 7, and compounds that contain only the elements H, B, C, N, O oxygen, F, P, S, cl and Br, the remainder being deleted, then the smiles for each compound are normalized, using the Morgan algorithm, with parameter radii set to 0 and 1, and then all atomic signatures are combined in sequence into the molecular structure 10, as shown in FIG. 6. Wherein an atomic signature for fewer than three occurrences may be replaced with "UNSEEN".
Similarly, in the protein sequence feature learning method, amino acids are mapped to the state vectors of the hilbert space, so that the projection of the state vectors on the respective calculation substrates can reflect the correlation of the respective amino acids. The protein sequence generated by model pre-training effectively extracts a series of physical and chemical properties, and can be used for downstream tasks such as protein family classification, protein structure prediction, protein interaction and the like.
It should be understood that, although the above embodiments apply the methods of training quantum circuits and data embedding in language processing, molecular feature learning and protein sequence feature learning, respectively, this is merely an exemplary embodiment for illustrating some applications provided by the present application and should not be construed as limiting the scope of the present application. According to other embodiments of the present application, the method of training quantum circuits and data embedding can be applied to other fields, such as biology, medicine, finance, etc., and the method of gene learning can also be used.
In some embodiments, the method of gene learning further comprises: and obtaining a model predicted value according to the output of the neural network, and comparing the difference between the model predicted value and a reference value, wherein the reference value is related to the used gene, for example, the reference value is the true probability distribution of the other gene in the coexpression gene pair corresponding to the used gene.
For genes in the target used, for example, one gene in a set of co-expressed gene pairs, the output of the neural network may be a probability distribution of the true other gene in the set of co-expressed gene pairs.
After two genes are input into the gene embedding model, respective corresponding vectors are obtained and combined, for example, the two vectors are overlapped, and interaction between the two genes can be predicted by inputting the two vectors into a gene interaction neural network model (such as a two-layer neural network)
The embodiment of the application also provides a computer readable storage medium, on which a computer program is stored, which when executed by a processor, implements the method for training the quantum circuit and embedding the data.
The embodiment of the application also provides a quantum computer capable of realizing the training quantum circuit and the data embedding method, which is configured to realize the training quantum circuit and the data embedding method. For example, the quantum computer may include a processing unit and a memory device, wherein the memory device includes software capable of performing the above-described methods of training quantum wires and data embedding.
While the technical content and features of the present application have been disclosed above, those skilled in the art may make various substitutions and modifications based on the teachings and disclosure of the present application without departing from the spirit of the present application. Accordingly, the scope of the present application should not be limited to the embodiments disclosed, but should include various alternatives and modifications without departing from the application and be covered by the claims of the present application.

Claims (6)

1. A method of data encoding, comprising:
describing the data as a computing-under-the-substrate general expansionWherein the data contains C elements corresponding to base |x respectively 1 >……|x C >;
Performing one or more quantum gate operations on each qubit, wherein each qubit comprises an i-th qubit, and i = 1,2, … …, n; and
traversing each qubit in turn to do the following: for the ith qubit, calculate 2 i A parameter expressed as p i,d Where d=1, 2,3, … …,2 i The said
Wherein the data is at least one of text, protein sequence, molecular structure, gene.
2. The method of claim 1, further comprising: taking the ith quantum bit as a target bit and the first i-1 quantum bits as control bits, sequentially executing 2 i-1 A controlled revolving door.
3. The method of claim 2, wherein the indices a = 0,1,2, … …,2 are used i-1 -1 represents the controlled rotation gate, the control qubit being the first i-1 qubits, controlled rotation gate a performing: if the first i-1 qubits are at |a>The ith qubit is rotated.
4. The method of claim 3, wherein the rotating comprises rotating |0>Rotated as
5. A quantum computer configured to implement the method of any of the preceding claims 1-4.
6. A computer readable storage medium having stored thereon a computer program, wherein a processor executes the computer program to implement the method of any of the preceding claims 1-4.
CN202311268565.3A 2022-02-01 2022-02-01 Data coding method Pending CN117172325A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202311268565.3A CN117172325A (en) 2022-02-01 2022-02-01 Data coding method

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202210115012.3A CN114444701B (en) 2022-02-01 2022-02-01 Training quantum circuit and data embedding method
CN202311268565.3A CN117172325A (en) 2022-02-01 2022-02-01 Data coding method

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
CN202210115012.3A Division CN114444701B (en) 2022-02-01 2022-02-01 Training quantum circuit and data embedding method

Publications (1)

Publication Number Publication Date
CN117172325A true CN117172325A (en) 2023-12-05

Family

ID=81372415

Family Applications (2)

Application Number Title Priority Date Filing Date
CN202210115012.3A Active CN114444701B (en) 2022-02-01 2022-02-01 Training quantum circuit and data embedding method
CN202311268565.3A Pending CN117172325A (en) 2022-02-01 2022-02-01 Data coding method

Family Applications Before (1)

Application Number Title Priority Date Filing Date
CN202210115012.3A Active CN114444701B (en) 2022-02-01 2022-02-01 Training quantum circuit and data embedding method

Country Status (1)

Country Link
CN (2) CN114444701B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115691669B (en) * 2023-01-03 2023-03-17 北京中科弧光量子软件技术有限公司 Protein structure classification system based on quantum convolution neural network
CN115910220B (en) * 2023-01-03 2023-05-09 北京中科弧光量子软件技术有限公司 Quantum computer-based protein amino acid property encoding method and system

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP6675565B1 (en) * 2019-02-21 2020-04-01 Mdr株式会社 Quantum computing device
CN110674921A (en) * 2019-07-11 2020-01-10 中国科学技术大学 Method for constructing quantum feedforward neural network based on classical training
US11797872B2 (en) * 2019-09-20 2023-10-24 Microsoft Technology Licensing, Llc Quantum bit prediction
CN113012066B (en) * 2021-03-15 2022-06-17 中南大学 Quantum line learning method and image blurring method based on parametric vitreous color sampling
CN113592095B (en) * 2021-08-04 2024-01-16 北京量子信息科学研究院 Model training method and device based on quantum computation

Also Published As

Publication number Publication date
CN114444701A (en) 2022-05-06
CN114444701B (en) 2023-10-27

Similar Documents

Publication Publication Date Title
CN114444701B (en) Training quantum circuit and data embedding method
US11797872B2 (en) Quantum bit prediction
KR20210107584A (en) Method for managing training data
Wang et al. Repetitive reprediction deep decipher for semi-supervised learning
US11460982B1 (en) Number embedding application system
Zhao et al. A review of quantum neural networks: methods, models, dilemma
Wang et al. Development of variational quantum deep neural networks for image recognition
CN114444664A (en) Attention model and neural network model based on quantum computation
US11636374B2 (en) Exponential spin embedding for quantum computers
CN114283903A (en) Method for predicting protein drug binding force based on quantum gated recurrent neural network
CN114444723A (en) Model for data reconstruction
Altares-López et al. AutoQML: Automatic generation and training of robust quantum-inspired classifiers by using evolutionary algorithms on grayscale images
Basterrech et al. Evolutionary Echo State Network: A neuroevolutionary framework for time series prediction
CN114511092B (en) Graph attention machine mechanism implementation method based on quantum circuit
Kashyap et al. Quantum convolutional neural network architecture for multi-class classification
Lin et al. Continuation path learning for homotopy optimization
CN114511097A (en) Mutual learning method and system based on quantum circuit
Vasuki et al. Overview of Quantum Computing in Quantum Neural Network and Artificial Intelligence
Bai et al. Superposition-enhanced quantum neural network for multi-class image classification
Padha et al. Quantum deep neural networks for time series analysis
Zankov et al. Conjugated quantitative structure‐property relationship models: Prediction of kinetic characteristics linked by the Arrhenius equation
CN117994397B (en) Digital human text action generation method, device, computer equipment and storage medium
US20230368036A1 (en) Physics-informed multimodal autoencoder
Pande A Comprehensive Review of Data Encoding Techniques for Quantum Machine Learning Problems
CN114496114A (en) Model and method for molecular structure reasoning

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication