WO2024256595A1 - B-cell epitope prediction - Google Patents
B-cell epitope prediction Download PDFInfo
- Publication number
- WO2024256595A1 WO2024256595A1 PCT/EP2024/066470 EP2024066470W WO2024256595A1 WO 2024256595 A1 WO2024256595 A1 WO 2024256595A1 EP 2024066470 W EP2024066470 W EP 2024066470W WO 2024256595 A1 WO2024256595 A1 WO 2024256595A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- protein
- cell epitope
- amino acid
- true
- machine learning
- Prior art date
Links
- 210000003719 b-lymphocyte Anatomy 0.000 title claims abstract description 220
- 108090000623 proteins and genes Proteins 0.000 claims abstract description 338
- 102000004169 proteins and genes Human genes 0.000 claims abstract description 336
- 238000000034 method Methods 0.000 claims abstract description 152
- 238000010801 machine learning Methods 0.000 claims abstract description 128
- 238000012549 training Methods 0.000 claims abstract description 63
- 150000001413 amino acids Chemical class 0.000 claims description 152
- 125000003275 alpha amino acid group Chemical group 0.000 claims description 82
- 241000894007 species Species 0.000 claims description 41
- 239000000427 antigen Substances 0.000 claims description 36
- 102000036639 antigens Human genes 0.000 claims description 33
- 108091007433 antigens Proteins 0.000 claims description 33
- 238000003556 assay Methods 0.000 claims description 31
- 230000001717 pathogenic effect Effects 0.000 claims description 29
- 244000052769 pathogen Species 0.000 claims description 27
- 229960005486 vaccine Drugs 0.000 claims description 22
- 241000700605 Viruses Species 0.000 claims description 21
- 230000001681 protective effect Effects 0.000 claims description 19
- 108090000765 processed proteins & peptides Proteins 0.000 claims description 16
- 210000004027 cell Anatomy 0.000 claims description 15
- 206010028980 Neoplasm Diseases 0.000 claims description 13
- 108091028043 Nucleic acid sequence Proteins 0.000 claims description 12
- 102000007474 Multiprotein Complexes Human genes 0.000 claims description 9
- 108010085220 Multiprotein Complexes Proteins 0.000 claims description 9
- 239000012472 biological sample Substances 0.000 claims description 9
- 201000011510 cancer Diseases 0.000 claims description 9
- 230000009851 immunogenic response Effects 0.000 claims description 9
- 230000006403 short-term memory Effects 0.000 claims description 8
- 239000002904 solvent Substances 0.000 claims description 8
- 238000004891 communication Methods 0.000 claims description 7
- 208000015181 infectious disease Diseases 0.000 claims description 7
- 238000001914 filtration Methods 0.000 claims description 5
- 238000004590 computer program Methods 0.000 claims description 2
- 230000001419 dependent effect Effects 0.000 claims 1
- 235000018102 proteins Nutrition 0.000 description 270
- 235000001014 amino acid Nutrition 0.000 description 149
- 229940024606 amino acid Drugs 0.000 description 148
- 238000004458 analytical method Methods 0.000 description 58
- 238000004422 calculation algorithm Methods 0.000 description 51
- 239000013598 vector Substances 0.000 description 48
- 230000000875 corresponding effect Effects 0.000 description 36
- 238000002790 cross-validation Methods 0.000 description 34
- 238000012360 testing method Methods 0.000 description 20
- 230000002596 correlated effect Effects 0.000 description 16
- 230000035772 mutation Effects 0.000 description 16
- 238000013459 approach Methods 0.000 description 15
- 230000004044 response Effects 0.000 description 14
- 238000010200 validation analysis Methods 0.000 description 11
- 238000013528 artificial neural network Methods 0.000 description 10
- 230000003612 virological effect Effects 0.000 description 10
- 230000002163 immunogen Effects 0.000 description 9
- 238000012545 processing Methods 0.000 description 9
- 108010032595 Antibody Binding Sites Proteins 0.000 description 8
- 108020004414 DNA Proteins 0.000 description 8
- 241000008904 Betacoronavirus Species 0.000 description 7
- 238000013461 design Methods 0.000 description 7
- 101710198474 Spike protein Proteins 0.000 description 6
- 230000008901 benefit Effects 0.000 description 6
- 238000011161 development Methods 0.000 description 6
- 230000018109 developmental process Effects 0.000 description 6
- 238000002405 diagnostic procedure Methods 0.000 description 6
- 238000010586 diagram Methods 0.000 description 6
- 238000000126 in silico method Methods 0.000 description 6
- 238000013507 mapping Methods 0.000 description 6
- 108010047041 Complementarity Determining Regions Proteins 0.000 description 5
- 238000011156 evaluation Methods 0.000 description 5
- 238000011502 immune monitoring Methods 0.000 description 5
- 230000028993 immune response Effects 0.000 description 5
- 210000000987 immune system Anatomy 0.000 description 5
- 238000011835 investigation Methods 0.000 description 5
- 108020003175 receptors Proteins 0.000 description 5
- 102000005962 receptors Human genes 0.000 description 5
- 239000000523 sample Substances 0.000 description 5
- 230000001225 therapeutic effect Effects 0.000 description 5
- 108091032973 (ribonucleotides)n+m Proteins 0.000 description 4
- 241000894006 Bacteria Species 0.000 description 4
- 241001678559 COVID-19 virus Species 0.000 description 4
- 229940096437 Protein S Drugs 0.000 description 4
- 230000001580 bacterial effect Effects 0.000 description 4
- 230000006870 function Effects 0.000 description 4
- 238000010647 peptide synthesis reaction Methods 0.000 description 4
- 238000002360 preparation method Methods 0.000 description 4
- 230000008569 process Effects 0.000 description 4
- 238000012163 sequencing technique Methods 0.000 description 4
- 229940125575 vaccine candidate Drugs 0.000 description 4
- 108060003951 Immunoglobulin Proteins 0.000 description 3
- 208000025370 Middle East respiratory syndrome Diseases 0.000 description 3
- 210000000988 bone and bone Anatomy 0.000 description 3
- 238000013527 convolutional neural network Methods 0.000 description 3
- 238000010219 correlation analysis Methods 0.000 description 3
- 102000018358 immunoglobulin Human genes 0.000 description 3
- 230000003993 interaction Effects 0.000 description 3
- 230000003211 malignant effect Effects 0.000 description 3
- 230000000873 masking effect Effects 0.000 description 3
- 108020004707 nucleic acids Proteins 0.000 description 3
- 102000039446 nucleic acids Human genes 0.000 description 3
- 150000007523 nucleic acids Chemical class 0.000 description 3
- 238000007781 pre-processing Methods 0.000 description 3
- 102000004196 processed proteins & peptides Human genes 0.000 description 3
- 230000000306 recurrent effect Effects 0.000 description 3
- 230000002787 reinforcement Effects 0.000 description 3
- 239000000243 solution Substances 0.000 description 3
- 238000006467 substitution reaction Methods 0.000 description 3
- 238000002255 vaccination Methods 0.000 description 3
- 238000002424 x-ray crystallography Methods 0.000 description 3
- NFGXHKASABOEEW-UHFFFAOYSA-N 1-methylethyl 11-methoxy-3,7,11-trimethyl-2,4-dodecadienoate Chemical compound COC(C)(C)CCCC(C)CC=CC(C)=CC(=O)OC(C)C NFGXHKASABOEEW-UHFFFAOYSA-N 0.000 description 2
- 108091008875 B cell receptors Proteins 0.000 description 2
- 208000035473 Communicable disease Diseases 0.000 description 2
- 238000002965 ELISA Methods 0.000 description 2
- 241000282414 Homo sapiens Species 0.000 description 2
- 102000002067 Protein Subunits Human genes 0.000 description 2
- 108010001267 Protein Subunits Proteins 0.000 description 2
- 101100194362 Schizosaccharomyces pombe (strain 972 / ATCC 24843) res1 gene Proteins 0.000 description 2
- VYPSYNLAJGMNEJ-UHFFFAOYSA-N Silicium dioxide Chemical compound O=[Si]=O VYPSYNLAJGMNEJ-UHFFFAOYSA-N 0.000 description 2
- 210000001744 T-lymphocyte Anatomy 0.000 description 2
- 230000004913 activation Effects 0.000 description 2
- 239000002671 adjuvant Substances 0.000 description 2
- 230000015572 biosynthetic process Effects 0.000 description 2
- 210000004369 blood Anatomy 0.000 description 2
- 239000008280 blood Substances 0.000 description 2
- 238000007796 conventional method Methods 0.000 description 2
- 238000001514 detection method Methods 0.000 description 2
- 238000003745 diagnosis Methods 0.000 description 2
- 238000013154 diagnostic monitoring Methods 0.000 description 2
- 201000010099 disease Diseases 0.000 description 2
- 208000037265 diseases, disorders, signs and symptoms Diseases 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 230000002209 hydrophobic effect Effects 0.000 description 2
- 230000005847 immunogenicity Effects 0.000 description 2
- 230000002458 infectious effect Effects 0.000 description 2
- 239000012528 membrane Substances 0.000 description 2
- 230000015654 memory Effects 0.000 description 2
- 229940023041 peptide vaccine Drugs 0.000 description 2
- 230000003389 potentiating effect Effects 0.000 description 2
- 230000000069 prophylactic effect Effects 0.000 description 2
- 238000011160 research Methods 0.000 description 2
- 230000011218 segmentation Effects 0.000 description 2
- 238000000926 separation method Methods 0.000 description 2
- 210000002966 serum Anatomy 0.000 description 2
- 238000003860 storage Methods 0.000 description 2
- 238000003786 synthesis reaction Methods 0.000 description 2
- 229940021747 therapeutic vaccine Drugs 0.000 description 2
- 238000009827 uniform distribution Methods 0.000 description 2
- 102000019260 B-Cell Antigen Receptors Human genes 0.000 description 1
- 108010012919 B-Cell Antigen Receptors Proteins 0.000 description 1
- 208000035143 Bacterial infection Diseases 0.000 description 1
- 108020004705 Codon Proteins 0.000 description 1
- 102000004127 Cytokines Human genes 0.000 description 1
- 108090000695 Cytokines Proteins 0.000 description 1
- 238000001712 DNA sequencing Methods 0.000 description 1
- 241000709661 Enterovirus Species 0.000 description 1
- 108090000790 Enzymes Proteins 0.000 description 1
- 102000004190 Enzymes Human genes 0.000 description 1
- 241000282412 Homo Species 0.000 description 1
- 241000282560 Macaca mulatta Species 0.000 description 1
- 241000699666 Mus <mouse, genus> Species 0.000 description 1
- 108091034117 Oligonucleotide Proteins 0.000 description 1
- 241000283973 Oryctolagus cuniculus Species 0.000 description 1
- 238000003559 RNA-seq method Methods 0.000 description 1
- -1 RSA Proteins 0.000 description 1
- 241000700159 Rattus Species 0.000 description 1
- 108091005634 SARS-CoV-2 receptor-binding domains Proteins 0.000 description 1
- 238000012300 Sequence Analysis Methods 0.000 description 1
- 241000251539 Vertebrata <Metazoa> Species 0.000 description 1
- 208000036142 Viral infection Diseases 0.000 description 1
- 239000002250 absorbent Substances 0.000 description 1
- 230000002745 absorbent Effects 0.000 description 1
- 239000002253 acid Substances 0.000 description 1
- 150000007513 acids Chemical class 0.000 description 1
- 230000003213 activating effect Effects 0.000 description 1
- 230000033289 adaptive immune response Effects 0.000 description 1
- 125000000539 amino acid group Chemical group 0.000 description 1
- 230000003321 amplification Effects 0.000 description 1
- 230000005875 antibody response Effects 0.000 description 1
- 239000007864 aqueous solution Substances 0.000 description 1
- 230000001363 autoimmune Effects 0.000 description 1
- 230000005784 autoimmunity Effects 0.000 description 1
- 210000002769 b effector cell Anatomy 0.000 description 1
- 208000022362 bacterial infectious disease Diseases 0.000 description 1
- 230000006399 behavior Effects 0.000 description 1
- 229920001222 biopolymer Polymers 0.000 description 1
- 238000001574 biopsy Methods 0.000 description 1
- 239000012830 cancer therapeutic Substances 0.000 description 1
- 125000003178 carboxy group Chemical group [H]OC(*)=O 0.000 description 1
- 230000010261 cell growth Effects 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 239000003153 chemical reaction reagent Substances 0.000 description 1
- 238000004040 coloring Methods 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 238000012790 confirmation Methods 0.000 description 1
- 229940028617 conventional vaccine Drugs 0.000 description 1
- 238000012937 correction Methods 0.000 description 1
- 238000007821 culture assay Methods 0.000 description 1
- 238000007405 data analysis Methods 0.000 description 1
- 238000010790 dilution Methods 0.000 description 1
- 239000012895 dilution Substances 0.000 description 1
- 238000009826 distribution Methods 0.000 description 1
- 238000009510 drug design Methods 0.000 description 1
- 239000012636 effector Substances 0.000 description 1
- 230000008030 elimination Effects 0.000 description 1
- 238000003379 elimination reaction Methods 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 238000000684 flow cytometry Methods 0.000 description 1
- 239000012634 fragment Substances 0.000 description 1
- 230000002068 genetic effect Effects 0.000 description 1
- 108091005608 glycosylated proteins Proteins 0.000 description 1
- 102000035122 glycosylated proteins Human genes 0.000 description 1
- 238000004128 high performance liquid chromatography Methods 0.000 description 1
- 230000028996 humoral immune response Effects 0.000 description 1
- 238000009396 hybridization Methods 0.000 description 1
- 230000036039 immunity Effects 0.000 description 1
- 238000002649 immunization Methods 0.000 description 1
- 230000002998 immunogenetic effect Effects 0.000 description 1
- 229940072221 immunoglobulins Drugs 0.000 description 1
- 238000009169 immunotherapy Methods 0.000 description 1
- 238000000338 in vitro Methods 0.000 description 1
- 238000001727 in vivo Methods 0.000 description 1
- 238000010212 intracellular staining Methods 0.000 description 1
- 238000002955 isolation Methods 0.000 description 1
- 238000009533 lab test Methods 0.000 description 1
- 238000002032 lab-on-a-chip Methods 0.000 description 1
- 238000011005 laboratory method Methods 0.000 description 1
- 230000005923 long-lasting effect Effects 0.000 description 1
- 210000004698 lymphocyte Anatomy 0.000 description 1
- 230000014759 maintenance of location Effects 0.000 description 1
- 238000004949 mass spectrometry Methods 0.000 description 1
- 238000002493 microarray Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 108091005573 modified proteins Proteins 0.000 description 1
- 102000035118 modified proteins Human genes 0.000 description 1
- 238000003199 nucleic acid amplification method Methods 0.000 description 1
- 238000013450 outlier detection Methods 0.000 description 1
- 244000045947 parasite Species 0.000 description 1
- 238000003909 pattern recognition Methods 0.000 description 1
- 230000002093 peripheral effect Effects 0.000 description 1
- 239000012071 phase Substances 0.000 description 1
- 230000000704 physical effect Effects 0.000 description 1
- 239000013612 plasmid Substances 0.000 description 1
- 238000003752 polymerase chain reaction Methods 0.000 description 1
- 230000002265 prevention Effects 0.000 description 1
- 229940021993 prophylactic vaccine Drugs 0.000 description 1
- 108020001580 protein domains Proteins 0.000 description 1
- 230000012846 protein folding Effects 0.000 description 1
- 230000006916 protein interaction Effects 0.000 description 1
- 238000000455 protein structure prediction Methods 0.000 description 1
- 238000004007 reversed phase HPLC Methods 0.000 description 1
- 239000007790 solid phase Substances 0.000 description 1
- 238000001228 spectrum Methods 0.000 description 1
- 238000010186 staining Methods 0.000 description 1
- 238000012916 structural analysis Methods 0.000 description 1
- 238000000547 structure data Methods 0.000 description 1
- 239000000126 substance Substances 0.000 description 1
- 238000002560 therapeutic procedure Methods 0.000 description 1
- 210000001519 tissue Anatomy 0.000 description 1
- 238000011282 treatment Methods 0.000 description 1
- 210000004881 tumor cell Anatomy 0.000 description 1
- 210000003462 vein Anatomy 0.000 description 1
Classifications
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16B—BIOINFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR GENETIC OR PROTEIN-RELATED DATA PROCESSING IN COMPUTATIONAL MOLECULAR BIOLOGY
- G16B20/00—ICT specially adapted for functional genomics or proteomics, e.g. genotype-phenotype associations
- G16B20/30—Detection of binding sites or motifs
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16B—BIOINFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR GENETIC OR PROTEIN-RELATED DATA PROCESSING IN COMPUTATIONAL MOLECULAR BIOLOGY
- G16B40/00—ICT specially adapted for biostatistics; ICT specially adapted for bioinformatics-related machine learning or data mining, e.g. knowledge discovery or pattern finding
- G16B40/20—Supervised data analysis
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16B—BIOINFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR GENETIC OR PROTEIN-RELATED DATA PROCESSING IN COMPUTATIONAL MOLECULAR BIOLOGY
- G16B15/00—ICT specially adapted for analysing two-dimensional or three-dimensional molecular structures, e.g. structural or functional relations or structure alignment
- G16B15/30—Drug targeting using structural data; Docking or binding prediction
Definitions
- Lymphocytes are a type of white blood and are central components of the immune system of most vertebrates. They include B cells, T cells and natural killer (NK) cells. Their function is to identify and protect against pathogens, such as bacteria, viruses, or malignant cancer cells that harbor mutated protein antigens. Both B- and T-Cells express surface receptors which identify specific molecular components on pathogens called antigens (Ags). The successful recognition of an Ag triggers a wider adaptive immune response which ultimately may eliminate the potentially pathogenic threat.
- antigens Ags
- Antibodies also known as immunoglobulins (Ig) are glycosylated protein molecules present on the surface of B cells (surface immunoglobulins) or B cell eluted Ig molecules, that act as antigen specific B cell receptors (BCR).
- BCR antigen specific B cell receptors
- BCE B-cell epitope
- the secreted soluble Igs and effector BCRs also called antibodies (Abs) will have the same binding specificity as the original Ig-Ag binding interaction event. Abs will then circulate into the serum and bind to the same BCEs to signal the elimination of the pathogenic or cell-stressed cells that correspond to the identified Ags.
- Ab based vaccines have the potential to prepare or educate the immune system of each individual against potentially harmful Ags sourced from infectious pathogenic threats or stressed malignant cancer cells.
- Vaccine design aims for the modification of the Ag in such a way that it bypasses harmful malignant, pathogenic or infectious properties while maintaining potential protective immunes responses driven by immunogenic epitopes.
- the accurate identification or prediction of B-Cell epitopes on antigens is therefore of great importance not only for vaccine design, but also in the domains of molecular diagnostics, immune- monitoring and immunotherapy.
- BCEs can be divided into two main categories: linear B cell epitopes (LBCEs) and conformational B cell epitopes (CBCEs). Both LBCEs and CBCEs can be bound by Abs and, thus, trigger an immune response. It is estimated that > 90% of the BCEs are conformational, while the rest are linear.
- LBCEs linear B cell epitopes
- CBCEs conformational B cell epitopes
- LBCEs are made up of continuous (sequential) sequences of amino-acids. These BCEs can be discovered by identifying positions on the unfolded Ag sequence which can be bound by Abs. LBCEs are often catalogued at the immune epitope database (IEDB) sourced from a large variety of experimental assays. Each assay provides information about the protein sequence tested (that is, the Ag), the LBCE sequence as well as its coordinates on the Ag, and the Ab which the current LBCE was tested on. Moreover, they provide an indicator as to whether the LBCE was bound by the given Ab. The binding information is sometimes described in the IEDB in the form of human readable labels, that is; positive, positive-high, positiveintermediate, positive-low and negative.
- CBCEs are made up of short segments of amino-acids with potential diverse cis, trans, short or distal coordinate distances between each of the amino acid contact points that constitute the BCE. These segments come into proximity during the folding of the Ag in the act of binding to the Ab. CBCEs can be experimentally characterized using X-ray crystallography on Ag-Ab structure complexes. This data is usually catalogued in the Protein Data Base (PDB) in the form of 3D structures, and typically consist of one or more Ag structures bound by at least one Ab structure. PDB structure data provides information about the 3D coordinates of each molecule in the structure. However, the complexity of this method makes the CBCE data somewhat more limited compared to that of LBCEs.
- PDB Protein Data Base
- the 3D structures may not accurately reflect true CBCEs.
- Abs on the complexes may be missing their heavy or light chains, or the 3D structure may only provide Ag-Ab contacts outside the complementaritydetermining regions (CDRs).
- CDRs complementaritydetermining regions
- the expected input is a 3D protein structure file of the Ag.
- the 3D protein structure of the Ag will almost certainly not be available at prediction time in most use cases; which limits the use of the 3D structure input algorithms.
- DiscoTope 3.0 is trained on both predicted and solved antigen structures (i.e. , antibody-antigen structures) merged together, whereas previous versions of that method were trained on experimentally solved antibody-antigen structures.
- B-cell epitope prediction techniques that require the sequence of the protein as input are more user-friendly than those that require the full 3D structure, it can be difficult to capture relationships between spatially distance amino acids, and to distinguish between potentially different CBCEs on the same antigen. There is therefore a desire to improve the prediction of B-Cell epitopes, in particular CBCEs.
- a computer- implemented method of predicting whether a protein comprises a B-cell epitope that is likely to instigate a binding event with an antibody comprising:
- each first reference protein comprising at least one B-cell epitope classified as a true B-cell epitope
- the three-dimensional structure of the BCE may be altered.
- the 3D structure of the BCE within a bound antibody-antigen complex may differ from the 3D structure of the BCE before the binding event was instigated.
- the present invention provides a method of predicting whether a (e.g. query) protein comprises a B-cell epitope using a trained first machine learning model that is trained using a first reference dataset comprising structure and/or surface characteristics of a plurality of first reference proteins in an unbound state.
- Each first reference protein comprises at least one B-cell epitope classified as a true IB- cell epitope.
- the first machine learning model is advantageously trained to predict true B-cell epitopes on a query protein before a binding event has taken place.
- unbound state (or “native state”) is used to mean the state (e.g. the three-dimensional structure) of a protein or BCE without influence from other proteins, structures or interacting molecules (e.g. before a binding event has taken place). Importantly, this unbound state is what the Ab “sees” and initiates the binding event, and not what an Ab has already “seen” and bound on.
- each first reference protein in an unbound state provides a key advantage over state-of-the- art models that are trained on Ag-Ab complexes (i.e. , proteins in a bound state), and which may lead to errors as the predictions are made based on structural data obtained after the binding event has occurred.
- the output of the trained first machine learning model is a probability that is indicative of whether the protein comprises a true B-cell epitope.
- the probability may typically be in the form of a probability that a candidate epitope on the query protein is a true epitope.
- the output may be in the form of a probability that individual amino acids (or groups of amino acids) within the amino acid sequence of the query protein form part of a true B-cell epitope.
- the output may be at a whole protein level (e.g. a probability that the query protein comprises a true BCE).
- the term “true” B-cell epitope is used to mean a B-cell epitope that instigates a binding event with an antibody (e.g. instigating an immunogenic response and/or instigated by an immunogenic response).
- the B-cell epitope may be a conformational B-cell epitope.
- the B-cell epitope may be a linear B-cell epitope.
- the trained first machine learning model of the present invention is typically able to predict the presence of both linear B-cell epitopes and conformational B-cell epitopes, although in some embodiments different models may be employed for predicting linear and conformational B-cell epitopes respectively, based on the reference data used for training.
- the protein may be a protein, protein-domain or a protein sub-unit.
- the term “protein” may include any protein subsequence that may have a viable 3D or functional structure.
- the protein comprises an antigen.
- the protein may comprise a neoantigen.
- one or more (e.g. three- dimensional) structure and/or surface (e.g. surface-exposed) characteristics of the protein are accessed.
- the structural and/or surface characteristics of the protein being investigated are important for predicting B-cell epitopes (in particular CBCEs), as the Abs are very specific to their binding BCEs.
- the one or more structure and/or surface characteristics of the protein may be characteristics at the whole protein level (e.g. secondary structure) or at the amino acid level (e.g. features associated with individual amino acids of the protein).
- the one or more structure and/or surface characteristics may be in the form of continuous features (such as relative solvent accessibility, RSA, or half-sphere exposure, HSA), and/or categorical features such as secondary structure.
- the one or more structure and/or surface characteristics include one or more of: the secondary structure (SS) of the protein; the relative solvent accessibility, RSA; the half-sphere exposure, HSA.
- the HSA values may be divided into values for upper half-sphere exposure and lower half-sphere exposure.
- the SS may be regarded as a structure characteristic, and the RSA and HSE regarded as surface characteristics.
- a particular advantage of the present invention is that the three-dimensional full folding structure of the input protein (which may be difficult to measure, or may contain a large number of errors if predicted) is not required.
- the method accesses one or more structure and/or surface characteristics of the protein which are advantageously more straightforward to obtain than a full folding structure, and typically contain fewer errors.
- the inventors have found that the characteristics of RSA, HSE and secondary structure in particular provide good results for the predictions, as they provide reliable indicators of the three-dimensional structure of the protein. It is envisaged that further structure and/or surface characteristics may also be used as input to the model. In this way, the inventive method can provide good BCE prediction results without requiring the full three-dimensional structure of the protein under investigation.
- the method may further comprise accessing one or more physiochemical characteristics of the protein, and/or the amino acids that constitute the protein, and inputting the one or more physiochemical characteristics into the first machine learning model (e.g. in addition to the one or more structure and/or surface characteristics).
- the first reference dataset typically further comprises one or more physiochemical characteristics of the first reference proteins, and/or the amino acids that constitute them.
- physiochemical characteristics include bulkiness, molar fraction of buried residues, average flexibility index, normalised frequency for alpha helix, known number of codons coding for each amino acid in universal genetic code, polarity, conformational parameter for beta-turn, normalized frequency for beta-sheet, side chain classes, side chain polarity, retention coefficient, hydrophobic constants, normalized frequency of beta-turn, PK1 a-COOH, membrane buried helix parameter, antigenicity.
- Different sets of physiochemical characteristics may be used for predicting CBCEs and LBCEs.
- the one or more physiochemical characteristics may be computed (e.g. from the amino acid sequence of the protein) or measured using techniques known in the art.
- the physiochemical characteristics may be determined for each amino acid of the protein.
- each amino acid (or side chains) that constitutes a protein sequence has physiochemical properties that contribute in some manner to its SS and 3D properties.
- physiochemical characteristics may further enhance the performance of the first machine learning model.
- each of the structure and/or surface characteristic(s) and physiochemical characteristic(s) may contribute separately or in combination to the training and subsequent use of the first machine learning model to predict whether a protein contains a B-cell epitope.
- the one or more structure and/or surface characteristics of each first reference protein are predicted by: accessing the amino acid sequence of the first reference protein; and applying one or more second machine learning model(s) to the amino acid sequence to predict the one or more structure and/or surface characteristics, wherein the one or more second machine learning model(s) are trained on a second reference dataset that comprises a plurality of amino acid sequences of respective second reference proteins and their corresponding structure and/or surface characteristics in an unbound state.
- the one or more second machine learning model(s) are trained on the second reference dataset to learn a relationship between (e.g. one or more features of) the amino acid sequences of the respective second reference proteins and the corresponding structure and/or surface characteristics in an unbound state.
- embodiments of the invention enable the prediction of the structure and/or surface characteristics of the first reference proteins in an unbound state which, as described previously, is a key aspect of the present invention.
- the one or more structure and/or surface characteristics include one or more of: the secondary structure of the protein; the relative solvent accessibility, RSA; the half-sphere exposure, HSE.
- These characteristics of the amino acids in proteins used to train models of the second reference dataset may be accessed using known resources or tools, for example the DSSP and BioPython algorithms applied to the structure files of the second reference proteins (e.g. obtained from the Protein Data Bank, PDB).
- the input to the one or more second machine learning model(s) comprises a plurality of characteristics (e.g. features) of the first reference protein.
- the input characteristics are typically physiochemical characteristics but may in some embodiments include structure and/or surface characteristics. These characteristics are typically computed or measured per amino acid of the first reference protein using techniques known in the art. Examples of such (e.g. physiochemical) characteristics include one or more of hydrophobicity, side chain class, side chain polarity (e.g. for predicting HSE and RSA), and one or more of conformational parameter for coil, conformational parameter for alpha helix (e.g. for predicting secondary structure). Further characteristics may be used.
- the one or more second machine learning model(s) may be trained on the second reference dataset to learn a relationship between one or more (e.g. physiochemical) characteristics of the amino acid sequences of the respective second reference proteins, and the corresponding structure and/or surface characteristics (e.g. one or more of RSA, HSE, Secondary Structure) in an unbound state.
- the second machine learning algorithm(s) utilise a different set of protein characteristics from the first machine learning algorithm.
- the amino acid sequence of the first reference proteins may be obtained using techniques known to the skilled person in the art, for example by one of: oligonucleotide hybridisation methods, nucleic acid amplification based methods (including but not limited to polymerase chain reaction based methods), automated prediction based on DNA or RNA sequencing, de novo peptide sequencing, Edman sequencing or any peptide related sequencing including and not limited to mass spectrometry.
- the amino acid sequence may be downloaded from a bioinformatic depository such as UniProt (www.uniprot.org).
- different second machine learning models may be used to predict different structure and/or surface characteristics.
- one second machine learning model may be used to predict categorical features, and a different second machine learning model used to predict continuous features.
- the input characteristics and reference datasets for the models may differ accordingly.
- a trained second machine learning model may be used to predict the secondary structure of a protein, and a different trained second machine learning model used to predict RSA and HSE characteristics.
- the structure and/or surface characteristics of the protein may be obtained using known techniques.
- known algorithms including computational approaches to predicting full 3D protein structures
- X-ray crystallography techniques may be used to predict or measure the features of the protein.
- RSA, HSE and secondary structure (and other) parameters could be accessed from a predicted or experimental full 3D structure of the protein, without the full structure itself being required as input to the model.
- the one or more structure and/or surface characteristics of the query protein may be predicted by applying the one or more second machine learning model(s) discussed above. This may require accessing the amino acid sequence of the query protein. In this way, structure and/or surface characteristics of the protein under investigation may advantageously be computed directly from the amino acid sequence of the query protein, with no requirement for extra information to be sourced or input by the user.
- Embodiments of the present invention therefore provide a method of obtaining first reference proteins each comprising at least one true B-cell epitope.
- first reference proteins may be obtained by:
- step (iv) for each V H and V chain pair defining a true B-cell epitope on the corresponding antigen; wherein at least some of the first reference proteins of the first reference dataset correspond to an antigen mapped in step (ii) that comprises at least one true IB- cell epitope defined in step (iv).
- the first reference proteins obtained or identified in the manner above are typically mapped onto their corresponding amino acid sequences.
- the structural and/or surface characteristics of each first reference protein in an unbound state may be obtained from the respective amino acid sequence, independent of the protein complex from which it was identified.
- the characteristics may preferably be predicted using the one or more second machine learning model(s) described above.
- other known approaches to predicting the structure and/or surface characteristic(s) may be employed, such as publicly available computational protein structure prediction algorithms.
- the protein complexes accessed in step (i) may be of any organism or entity assigned a taxonomy ID in established authorities such as the NCBI. This advantageously allows the predictions made by the trained first machine learning model to be non-Ab specific.
- each true B-cell epitope is defined as the amino acids of the corresponding antigen in contact with the corresponding V H and V chain pair, preferably wherein the amino acids are defined as being in contact with the corresponding V H and V chain pair if they comprise any atoms within a predefined distance of the complementarity-determining region(s), CDR(s) of the V H and V chain pair, preferably wherein the predetermined distance is equal to or less than 4 Angstroms.
- the method may further comprise: accessing the amino acid sequence of the (e.g. query) protein; and generating one or more candidate encodings of respective one or more candidate B-cell epitopes on the protein, wherein each candidate encoding represents the respective candidate B-cell epitope as a plurality of data elements corresponding to amino acids of the amino acid sequence; wherein the inputs to the trained first machine learning model comprise the amino acid sequence of the protein, the one or more structure and/or surface characteristics of the protein, and the one or more candidate encodings.
- the inventors have realised that a common problem with many conventional sequence-based techniques used to predict the presence of a B-Cell epitope on a protein is that the output prediction is typically in the form of a binary value per amino acid indicating whether each particular amino acid in the sequence of the protein is part of a B-Cell epitope.
- This conventional approach does not allow the user to distinguish between potentially different and distinct BCEs present on the protein that harbor some overlapping amino acid residues, and moreover does not necessarily consider the relationships between spaced-apart amino acids of the entire protein sequence being analysed in 3D conformational BCEs.
- embodiments of the present invention may apply the trained first machine learning model to a combination of the input amino acid sequence of the protein under investigation, the one or more structure and/or surface characteristics, and one or more candidate encodings of respective one or more candidate B-cell epitopes, where each candidate encoding represents the respective candidate B-cell epitope.
- the predictions provided by the trained first machine learning model are made on an “epitope by epitope” basis on the query protein (by considering each candidate encoding corresponding to a candidate epitope), rather than on an “amino acid by amino acid” approach.
- such embodiments consider whether a candidate B-cell epitope, encoded on the amino acid sequence under test, is a true BCE (or not) on the protein.
- This approach is particularly advantageous as it provides improved confidence that an epitope as a whole is indeed present (or not) on the protein, rather than having to infer the presence of epitopes (or multiple overlapping epitopes) based on individual amino acid scores.
- the use of the candidate encodings representative of the candidate epitopes as a whole allows for potential relationships between spaced apart amino acids on the sequence to be considered.
- Providing predictions on such an “epitope by epitope” basis also provides increased accuracy and confidence with regard to predicting multiple (possibly overlapping) epitopes on the same protein.
- candidate encodings to provide predictions on an “epitope by epitope” basis is particularly advantageous in combination with the first machine learning model being trained on unbound data.
- the candidate encodings enable relationships between spaced apart amino acids on the sequence to be considered where, as discussed herein, those relationships may differ slightly between the protein being in a protein complex (bound state) or unbound (native) state.
- the output of the trained first machine learning model is typically a probability that each of the candidate encoding(s) represents a true B-cell epitope on the protein.
- the accessed (e.g. predicted) structure and/or surface characteristics of the protein are in the form of a value and/or a class assigned to each amino acid of the protein amino acid sequence that represents a structure and/or surface characteristic of the protein.
- the input to the first machine learning model may be a set of characteristics per amino acid of the protein.
- the candidate encodings of the B-cell epitopes represent the respective candidate B-cell epitope as a plurality of data elements corresponding to the amino acids of the amino acid sequence.
- each data element of a candidate encoding corresponds to an amino acid of the amino acid sequence of the protein.
- each of the one or more encodings is in the form of a binary vector.
- the binary vector will typically be in the form of a sequence of “1”s and “0”s, with the “1”s each corresponding to an amino acid of the protein being a part of the candidate epitope (e.g. a probable Ag-Ab contact point) , and the “0”s each corresponding to an amino acid of the protein not forming part of the candidate epitope (e.g. not a probable Ag-Ab contact point).
- different encoding protocols may be used (e.g. one-hot encoding) to represent the candidate epitopes.
- the one or more candidate encodings represent all possible combinations of the amino acids within the amino acid sequence. This may include both linear B-cell epitopes (where each of the encoded amino acids will be sequentially positioned in the amino acid sequence), and conformational B-cell epitopes (where at least some of the encoded amino acids are spaced apart on the amino acid sequence).
- prior knowledge of the protein and/or B cell epitope makeup may be used to reduce the search space of the encodings. For example, if the protein contains a neoantigen, the encodings may be fixed in the position of the mutation point (e.g. each encoding will have the same data element at the position of the mutation point on the amino acid sequence).
- the one or more candidate encodings representing combinations of amino acids within the amino acid sequence may be queried through deep reinforcement learning models.
- the one or more candidate encodings representing combinations of the amino acids within the amino acid sequence may be sourced from generative models.
- deep reinforcement learning (DRL) techniques may be used to reduce or query efficiently the large search space.
- the first reference dataset may further comprise: amino acid sequences of each first reference protein; and a plurality of reference encodings of each true B-cell epitope, wherein each reference encoding represents the respective true B-cell epitope as a plurality of data elements corresponding to the amino acids of the amino acid sequence.
- the reference encodings of each true B-cell epitope used for the first reference dataset are typically generated in the same manner as described above.
- the first machine learning model may be (or comprise) any neural network such as convolutional neural networks and feedforward neural networks.
- the first machine learning model is (or comprises) a recurrent neural network.
- the first machine learning model is (or comprises) a long short-term memory network, LSTM, preferably a bi-directional long short-term memory network, BLSTM.
- LSTM long short-term memory network
- BLSTM bi-directional long short-term memory network
- the present invention advantageously enables the discovery of relationships between spatially distant amino acids of the protein sequence which could, for example, be part of the same (e.g. conformational) BCE, and which relationships may not be discovered through the “k-mer” approach.
- the use of a BLSTM is particularly advantageous as this allows the protein amino acid sequence to be processed from both directions.
- the second machine learning model(s) may be (or comprise) any neural network such as convolutional neural networks and feedforward neural networks.
- the one or more trained second machine learning model(s) is (or comprises) a recurrent neural network, preferably a long short-term memory network, LSTM, more preferably a bi-directional long short-term memory network, BLSTM. This provides the same advantages as outline above.
- the present invention may advantageously generate encodings of one or more candidate B-cell epitopes to be tested in order to model and “test” each candidate epitope as a whole.
- the first reference dataset used to train the first machine learning model may only contain observed true positive BCEs (i.e. only “positive” training data). Therefore, in some preferred embodiments, the method may further comprise generating pseudo-random data based on the plurality of reference encodings of each true B-cell epitope of the first reference dataset, and wherein the first reference dataset comprises said pseudo-random data as negative data. In this way, the performance of the first machine learning model may be improved as the model can be trained even in the absence of verified negative data.
- the pseudo-random data is based on the encoded true B-cell epitope(s).
- the pseudo-random data comprise a plurality of permutations of the plurality of data elements representing the true B-cell epitopes.
- the pseudo-random data may be random permutations of the data elements.
- the pseudo-random data may share at least one characteristic with the encodings of the true B-cell epitope(s).
- the characteristic may be physical property of the true B-cell epitope, for example a position on the amino acid sequence of one or more amino acids (e.g. the first and last amino acids of an epitope). In another example, the characteristic may be a linear distance between particular amino acids.
- the method may further comprise: identifying a first B-cell epitope that is predicted to be likely to be a true B-cell epitope on the protein; substituting one or more amino acids forming the identified first B-cell epitope with a different amino acid to thereby generate a modified amino acid sequence of the protein; and using the trained first machine learning model to predict whether the candidate encoding representing the identified first B-cell epitope is likely to correspond to a true B-cell epitope on the modified amino acid sequence of the protein.
- the first machine learning model may be used to predict B-cell epitopes that do not naturally occur on the protein under analysis. This may be advantageous in identifying vaccine candidates that provide protection across possible future viral variants and across a broad range of virus or pathogen species, as described in further detail herein.
- one or more amino acids of the first B-cell epitope (which is typically a wildtype B-cell epitope) is substituted for any other amino acid that exists. This may be referred to as a “mutation” of the amino acid.
- the trained first machine learning model may be used to predict the likelihood of each possible amino acid substitution of the identified first B-cell epitope forming a true B-cell epitope on the protein. Further analysis may be performed on the B- cell epitopes that are predicted to be positive on the modified amino acid sequence. For examples, structural stability analysis may be performed using (e.g. computational) techniques known in the art.
- the candidate encodings are in the form of binary vector
- the amino acids corresponding to the “1”s of the binary vector e.g. representing direct amino acid contact points on the 3D protein sequence with the antibody
- the input to the model comprises the modified amino acid sequence, the one or more structure and/or surface characteristics of the modified protein, and the candidate encoding representing the identified first B-cell epitope.
- the structure and/or surface characteristics may be obtained using the techniques described above, for example by applying the second machine learning model(s) to the modified amino acid sequence.
- method of training a machine learning model comprising generating a reference dataset, the reference dataset comprising: a plurality of first reference proteins, each first reference protein comprising at least one B-cell epitope classified as a true B-cell epitope; and one or more structure and/or surface characteristics of each first reference protein in an unbound state; and training the machine learning model using the reference dataset to learn a relationship between the structure and/or surface characteristic(s) of the first reference proteins in an unbound state, and the corresponding B-cell epitopes classified as true B-cell epitopes.
- the method of training a machine learning model provides the advantages described above, and may be used in combination with any of the described examples of the invention.
- the machine learning model once trained, may be used to predict whether a query protein comprises a B-cell epitope that is likely to instigate a binding event with an antibody.
- the reference dataset further comprises: amino acid sequences of each reference protein; and a plurality of reference encodings of each true B-cell epitope, wherein each reference encoding represents the respective true B-cell epitope as a plurality of data elements corresponding to the amino acids of the amino acid sequence.
- the reference dataset may further comprise one or more physiochemical characteristics of each protein, and/or the amino acids that constitute the reference protein, as described above.
- the machine learning model may be (or comprise) any neural network such as convolutional neural networks and feedforward neural networks.
- the machine learning model is (or comprises) a recurrent neural network.
- the first machine learning model is (or comprises) a long short-term memory network, LSTM, preferably a bi-directional long short-term memory network, BLSTM.
- the invention also provides use of a machine learning model trained using the method described above (e.g. to predict the presence of BCEs on a query protein).
- the invention also provides a machine learning model trained using the method described above.
- the invention may further comprise synthesising one or more proteins that are predicted to contain a B-cell epitope, and/or a B-cell epitope predicted using the techniques described herein.
- Protein(s) that are predicted by the present invention to comprise a B-cell epitope may be used in the treatment or prevention of infectious diseases or other diseases such as autoimmune or immune-related diseases and cancer.
- the present invention finds use in one of more of the following use cases:
- the method may further comprise encoding one or more proteins that are predicted to contain a B-cell epitope, and/or a predicted B-cell epitope, and/or B-cell epitope variant, and/or B-cell epitope predicted or simulated variant into a corresponding protein, peptide, DNA or RNA sequence.
- DNA or RNA sequences may be incorporated into a delivery system for use in a vaccine (e.g. using naked or encapsulated DNA, or encapsulated RNA).
- the method may comprise incorporating the protein, peptide, DNA or RNA sequence into a genome of a bacterial or viral delivery system to create a vaccine.
- a method of creating a vaccine comprising: predicting that a protein comprises a B-cell epitope by a method according to any of the examples discussed above; and synthesising the protein and/or the predicted B-cell epitope, or encoding at least one of the protein, the predicted B-cell epitope, B-cell epitope variant, B- cell epitope predicted or simulated variant into a corresponding protein, peptide, DNA or RNA sequence.
- Such a protein, peptide, DNA or RNA sequence may be delivered in a naked or encapsulated form or incorporated into a genome or cell of a bacterial or viral delivery system to create a vaccine.
- bacterial vectors can be used to deliver the DNA in to vaccinated host cells.
- the identified protein(s) may typically be synthesised as an amino acid sequence or “string”.
- Such a vaccine may be a prophylactic or therapeutic vaccine.
- the method may be used to create a personalised vaccine for an individual, for example a cancer therapeutic vaccine if the protein comprises a neoantigen.
- a method of creating e.g. and/or designing) a diagnostic assay to determine whether a patient has or has had a cancer or prior infection with a pathogen, wherein the diagnostic assay is carried out on a biological sample obtained from a subject, comprising identifying at least one protein of the pathogen or tumor that is predicted to comprise a B-cell epitope, using a method according to any of the examples discussed above; wherein the diagnostic assay comprises the utilisation or identification within the biological sample of the identified at least one protein and/or B-cell epitope.
- the present invention may advantageously be used to create a diagnostic test or assay, by means of a rapid and/or automated molecular target discovery. Query proteins identified as being likely to comprise a true B-cell epitope may be further analysed in laboratory testing in order to create such a diagnostic test or assay, thereby significantly reducing the time taken to develop the test compared to traditional laboratory methods.
- utilisation as used herein is intended to mean that the at least one protein and/or B-cell epitope thereon are used in an assay to identify an (e.g. protective) immune response in a patient.
- the identified protein(s) and/or epitope(s) within are not the target of the assay, but a component of said assay.
- the in vitro diagnostic assay may comprise identification of an immune system component within the biological sample that recognises at least one epitope within the identified protein.
- the diagnostic assay may utilise the at least one identified protein and/or at least one predicted epitope.
- the diagnostic assay will contain the (e.g. synthesised) at least one identified protein, protein subunit and/or predicted epitope.
- the immune system component may be a B-cell.
- the assay may comprise identification of antibodies or B-cells that recognise predicted B-cell epitopes within the identified protein.
- a sample preferably a blood sample, isolated from a patient may be analysed for the presence of B-cells or antibodies within the biological sample that recognise and bind to epitope(s) within the identified protein(s), identified as part of the present invention and that are contained within the assay.
- Suitable diagnostic assays would be appreciated by the skilled person, but may include enzyme-linked immune absorbent spot (ELISPOT) assays, enzyme-linked immunosorbent assays (ELISA), cytokine capture assays, intracellular staining assays, tetramer staining assays, microfluidic devices, lab-on-a-chip, microarrays, flow cytometry, CyTOF, proteomics, molecular sequencing, or limiting dilution culture assays.
- ELISPOT enzyme-linked immune absorbent spot
- ELISA enzyme-linked immunosorbent assays
- cytokine capture assays intracellular staining assays
- tetramer staining assays microfluidic devices
- microfluidic devices lab-on-a-chip
- microarrays microarrays
- flow cytometry CyTOF
- proteomics molecular sequencing
- molecular sequencing or limiting dilution culture assays.
- the amino acid sequence of the one of more proteins to be tested for the presence of BCEs may be chosen based on the desired response to be tested.
- the one or more source proteins may be one or more source proteins of any pathogen or virus (or fragments thereof), such as the SARS-CoV-2 virus.
- the present invention may be used to create a diagnostic test for determining whether a patient has or has had prior infection with the SARS-CoV-2 virus and/or its variants and/or related viral species.
- the one or more source proteins may be from any pathogen (e.g. any virus, parasite, bacterium or cancer indication).
- a diagnostic assay to determine whether a patient has or has had prior infection with a pathogen, wherein the diagnostic assay is carried out on a biological sample obtained from a subject, and wherein the diagnostic assay comprises the utilisation or identification within the biological sample of at least one protein of the pathogen or tumor that is predicted to comprise a B-cell epitope, identified using any of the methods discussed above.
- the diagnostic assay may comprise identification of an immune system component (e.g. a B-cell) within the biological sample that recognises said at least one identified protein and/or at least one predicted epitope.
- a computer program product comprising instructions which, when executed by a computer, cause the computer to perform the method of any of the examples descried above.
- the invention also provides a non-transitory computer readable medium comprising executable instructions that, when executed by a computer, cause the computer to perform the method of any of the examples described above.
- a system for predicting whether a protein comprises a B-cell epitope that is likely to instigate a binding event with an antibody comprising at least one processor in communication with at least one memory device, the at least one memory device having stored thereon instructions for causing the at least one processor to perform a method in accordance with any of the examples of the invention discussed above.
- a method of synthesising a protein comprising: predicting that a protein comprises a B-cell epitope that is likely to instigate a binding event with an antibody according to any of the examples discussed above; and synthesising the predicted protein.
- the synthesis of the predicted protein can be performed using techniques known in the art.
- a protein synthesised using the said method is provided.
- Embodiments of the present invention may advantageously be used to identify one or more B-cell epitopes that are predicted to instigate a protective immunogenic response across a plurality of species of a pathogen or virus.
- Such “broadly protective” B-cell epitopes may be used as vaccine elements or to provide a diagnostic test or assay, for example.
- a (e.g. computer-implemented) method of identifying one or more B-cell epitopes that are predicted to instigate a protective immunogenic response across a plurality of species of a pathogen or virus comprising:
- each identified first B-cell epitope determines the number of different species of the plurality of species of the pathogen or virus on which the first B-cell epitope is predicted to be a true B-cell epitope; and (iii) classifying one or more of the first B-cell epitopes as a broadly protective B-cell epitope, based on the number of different species, or a number of variants of any one given species, on which the first B-cell epitope is predicted to be a true B-cell epitope.
- the present invention may be utilised to classify one or more B-cell epitopes as being predicted to instigate a broad immunogenic response across a plurality of different species of a pathogen or virus, or variants thereof (e.g. cross- reactive).
- the classification is based on a determination of the numbers of different species of the pathogen or virus on which the B-cell epitopes are predicted to be true B-cell epitopes (e.g. viable B-cell epitopes).
- the first B-cell epitope that is predicted to be a viable B-cell epitope on the largest number of different species, or variants thereof, tested may be classified as a broadly protective B-cell epitope (and may, for example, be classified as a vaccine candidate).
- an identified first B-cell epitope may be classified as being broadly protective if the number of species on which the B-cell epitope is predicted to be a true B-cell epitope is greater than a predetermined threshold.
- the one or more candidate encodings may map only onto the receptor binding domain (RBD) of the amino acid sequence of each protein tested.
- RBD receptor binding domain
- the method may further comprise, for at least one of (preferably each of) the first B-cell epitope(s) classified as a broadly protective B- cell epitope: substituting one or more amino acids forming the first B-cell epitope with a different amino acid to thereby generate a modified first B-cell epitope; and, for each of the plurality of proteins of the respective plurality of species of a pathogen of virus; generating a modified amino acid sequence of the protein in accordance with the modified first B-cell epitope; and using the trained first machine learning model to predict whether the candidate encoding representing the first B-cell epitope classified as a broadly protective B-cell epitope is likely to correspond to a true B-cell epitope on the modified amino acid sequence of the protein.
- the method of the present invention may advantageously be used to predict viable B-cell epitopes that may not occur naturally on the proteins under analysis.
- This may advantageously be used to identity broadly protective vaccine candidates that provide a broad protective response across a plurality of different species of a pathogen or virus (and/or variations of these), and which may also be protective against future viral variants or pathogens that have not yet spilled over to be pathogenic in humans.
- Such identified B-cell epitopes may form vaccine candidates or may be used to create a diagnostic test or assay as outlined above.
- modified B-cell epitopes identified in this way are predicted to generate a broader immunogenic response across a larger number of different species of a virus or pathogen, compared to the most cross- reactive wildtype B-cell epitopes.
- any “mutated” B-cell epitopes that are predicted to have low structural stability may be discarded from future analysis.
- Such structural analysis may be performed using (e.g. computational) techniques known in the art.
- Figure 1 is a flow diagram illustrating the steps of a method according to an embodiment of the invention
- Figure 2 is a flow diagram illustrating the steps of a method of training a first machine learning algorithm according to an embodiment of the invention
- Figure 3 is a flow diagram illustrating the steps of a method of training a second machine learning algorithm according to an embodiment of the invention
- Figure 4 is an example of a system suitable for implementing embodiments of the method is shown;
- Figure 5 is an example of a suitable server
- Figures 6(a) to 6(f) show the correlation results of the feature-feature pairs
- Figure 7 shows the Spearman’s correlation coefficients for continuous features against (a) RSA, (b) UHSE and (c) LHSE;
- Figure 8 shows the Kendall’s correlation coefficients for categorical features against (a) RSA, (b) UHSE and (c) LHSE;
- Figure 9 shows the Kendall’s correlation coefficients for continuous features against (a) SS, (b) CBCE and (c) LBCE;
- Figure 10 shows the Mutual Information criterion for categorical features against (a) SS, (b) CBCE and (c) LBCE;
- Figure 11 schematically shows the structure of a Prediction model for RSA and HSE according to an embodiment of the invention
- Figures 12(a) to (f) show cross-validation results for a RSA/HSE model according to an embodiment of the invention
- Figures 13(a) to (f) show test results for a RSA/HSE model according to an embodiment of the invention
- Figures 14(a) to (f) show final training results for a RSA/HSE model according to an embodiment of the invention
- Figures 15(a) to (f) illustrate CV results for a Secondary Structure (SS) prediction model according to the invention
- Figures 16(a) and (b) show the test results for SS model
- Figures 17(a) to (d) show final training results for the SS model according to an embodiment of the invention
- Figures 18(a) and 18(b) show Paratope Analysis for CBCE
- Figure 19(a) schematically illustrates the architecture of a first machine learning algorithm for predicting CBCEs, according to an embodiment of the invention
- Figure 19(b) schematically illustrates the architecture of a first machine learning algorithm for predicting CBCEs, according to a further embodiment of the invention
- Figures 20(a) to (d) illustrate outlier analysis for CBCEs
- Figures 21 (a) to (f) shows results for the precision-recall metric for a first machine learning algorithm for predicting CBCEs, according to a further embodiment of the invention
- Figures 22(a) to (f) show results for the precision-recall metric for a first machine learning algorithm for predicting CBCEs, according to an embodiment of the invention
- Figures 23(a) and (b) show the results for the precision-recall metric for a first machine learning algorithm for predicting CBCEs, according to an embodiment of the invention
- Figures 24(a) and (b) shows the results for the precision-recall metric for a first machine learning algorithm for predicting CBCEs, according to an embodiment of the invention
- Figures 25(a) to (c) shows the results of final training for a first machine learning algorithm for predicting CBCEs, according to an embodiment of the invention
- Figures 26(a) and (b) show the results of final training for a first machine learning algorithm for predicting CBCEs, according to an embodiment of the invention
- Figures 27(a) to (d) illustrates outlier analysis for LBCEs
- Figure 28 schematically illustrates a pipeline according to an embodiment of the invention
- FIG. 29(a) to (c) illustrates the successful mapping of receptor binding domains (RBDs) to betacoronavirus species
- Figure 30 illustrates the prediction of structural epitopes predicted to be broadly immunogenic across a plurality of viral species
- Figures 31(a) and 31(b) illustrate the change in predicted immunogenicity following the implementation of mutation policies.
- Embodiments of the invention provide a B-cell epitope (BCE) prediction algorithm (“first machine learning algorithm” or “first machine learning model”), trained using a reference dataset that is designed such that the trained first machine learning model predicts B-cell epitopes in their unbound state.
- the first machine learning algorithm is trained on unbound 3D Ag protein structures prior to the Ag-Ab binding event, to learn the properties of BCEs on the Ag prior to the Ab binding event, in order to capture the bone fide (“true”) BCEs on a query Ag protein that is input into the model.
- the BCE prediction algorithm predicts BCEs in the form that an antibody “sees” in order to instigate a binding event.
- the first machine learning model accurately predicts BCEs on a given Ag from its primary sequence in combination with one or more 3D structure and/or surface characteristics of the Ag, with no requirement for the full 3D structure of the Ag as an input into the predictor.
- the model is trained using BLSTMs, which allows for the processing of the whole Ag sequence as one data- point, processed from both directions without the need for k-mer segmentation.
- Each amino-acid of the Ag is treated as a time-step in the network, whose characteristics can be affected by both preceding and succeeding amino-acids that flank the amino-acid in question, thus, allowing for the discovery of the contextual relationships between spatially distant amino-acids which could be important parts of conformational BCEs (CBCEs).
- CBCEs conformational BCEs
- the first machine learning model is capable of discovering separate or independent CBCEs that may exist on a single Ag protein sequence.
- the trained model is designed to receive as input an Ag protein sequence and a binary “permutation vector” (“encoding”), representing each single candidate CBCE, as well as the one or more 3D structure and/or surface characteristics of the Ag.
- the output of the model is a single probability indicating the likelihood that the input permutation vector is a bone fide CBCE.
- the user asks the question to the model if the specific query CBCE, encoded as a binary permutation vector on the given Ag sequence, is a true CBCE or not.
- the model may also be used to provide a second output in the form of a probability vector with a probability for each amino-acid in the Ag sequence.
- the second output can be seen as the contribution of each amino acid to the CBCE in question, and may be used to compare the current approach with conventional models that predict on an “amino acid by amino acid” basis. Structural characteristics of the Ag sequence are important for predicting CBCEs.
- embodiments of the present invention negate the need for the complete 3D structure of the protein sequence to be experimentally measured or predicted, and then used as input into the algorithm.
- the first machine learning algorithm does not require the coordinates of each atom in each amino acid in the 3D protein sequence, and instead may predict BCEs using one or more structure and/or surface characteristics of the query protein, such as RSA, HSE and SS.
- top-level most important features such as RSA, HSE and SS
- second machine learning algorithm or “second machine learning model”
- second machine learning model typically receiving as input the amino acid sequence of the protein.
- structure and/or surface characteristics of the protein under investigation may advantageously be computed directly from the amino acid sequence, with no requirement for extra information to be sourced or input by the user.
- Those second algorithms use BLSTM networks also and are shown to have a high performance.
- the first and second machine learning models may together form a “pipeline” which requires as input only the amino acid sequence of a query protein in order to predict whether that query protein comprises a true B-cell epitope.
- a schematic diagram of such a pipeline constituting the first and second machine learning models according to an embodiment of the invention is shown in Figure 28.
- FIG. 1 is a flowchart 100 summarising the steps of a preferred embodiment of the present invention.
- step S101 the amino acid sequence of a protein to be analysed for the presence of B-cell epitopes is accessed.
- the protein under investigation will comprise an antigen.
- Various techniques known to the skilled person may be used to access the amino acid sequence, for example the sequence may be downloaded from a bioinformatic depository such as UniProt.
- one or more (e.g. three-dimensional) structure and/or surface characteristics of the protein in an unbound state are accessed.
- these characteristics of the protein are predicted using one or more trained second machine learning model(s) (schematically illustrated at 300, and described below) that receives the amino acid sequence accessed in step S101 as input.
- the structure and/or surface characteristics typically include the secondary structure of the query protein and relative solvent accessibility.
- the half sphere exposure (both upper and lower half sphere exposure) may also be used. Values or categories for each of these characteristics are typically assigned to each amino acid of the sequence.
- other techniques for obtaining the three- dimensional structure and surface characteristics of the query protein may be employed, such as database entries, X-ray crystallography or computational approaches to predicting full 3D protein folding structures (from which the structure and/or surface characteristics can be obtained).
- candidate encodings of one or more candidate BCEs on the query protein are generated.
- the candidate encodings are each typically in the form of a binary vector, with each data element in the binary vector corresponding to an amino acid of the amino acid sequence forming the respective candidate BCE.
- each amino acid of the sequence forming part of the candidate BCE is assigned a “1” and each amino acid of the sequence not forming part of the candidate BCE is assigned a “0”.
- a candidate conformational B-cell epitope constituted by the amino acids B, D, F and G may be encoded as the binary vector (0101011).
- a candidate encoding for each possible candidate BCE on the protein may be generated.
- the search space may be limited based on physical priors (e.g. knowledge of a mutation point on a neoantigen of the input protein or receptor binding domain on the spike protein of SARS-CoV-2 etc.) or on techniques such as deep reinforcement learning or related generative models.
- step S107 the amino acid sequence accessed in step 101 , the one or more structure and/or surface characteristics accessed in step 103, and the candidate encodings generated in step S105, are input into a trained first machine learning model, which has been trained using BLSTMs.
- One or more physiochemical characteristics assigned to each amino acid of the amino acid sequence may also be used as input to the trained first machine learning model.
- the output of the trained first machine learning model is a probability that each of the candidate encoding(s) represents a true B-cell epitope on the query protein.
- the first machine learning model is trained on unbound data.
- the first machine learning model is trained on a dataset that comprises reference proteins having B-cell epitopes classified as true B-cell epitopes, and structural and/or surface characteristics of the reference proteins in an unbound state.
- the first machine learning model is trained to learn a relationship between the structural and/or surface characteristic(s) of the reference proteins and the classified true B-cell epitopes. Consequently, the trained first machine learning model predicts the presence of B-cell epitopes on the query protein prior to a binding event.
- FIG. 2 is a flow diagram 200 summarising the steps of a preferred embodiment of training the first machine learning model.
- a plurality of protein complexes are accessed. These may be accessed from publicly available sources such as the Protein Databank (PDB) for CBCEs or Immune Epitope Database (IEDB) for LBCEs.
- PDB Protein Databank
- IEDB Immune Epitope Database
- step S203 the protein complexes accessed in step S201 are analysed to define true epitopes and antigens.
- Linear BCEs may be defined in the complexes of the IEDB. Further analysis may be required to define true conformational epitopes from protein complexes accessed from the PDB, and further details on these techniques are described herein.
- each antigen defined in step S203 as having a true B-cell epitope is mapped onto its amino acid sequence.
- each true epitope is encoded onto the amino acid sequence of the corresponding antigen.
- the reference encodings are typically encoded as binary vectors in the same manner as described above in step S105.
- step S209 of the method three-dimensional structure and/or surface characteristics of the antigens in an unbound state are accessed, for example by prediction from the amino acid sequence of the antigen obtained in step S205.
- the structure and/or surface characteristics of the antigens are predicted using the trained second machine learning model(s) (schematically illustrated at 300).
- the characteristics include the secondary structure, half sphere exposure (both UHSE and LHSE), and relative solvent accessibility.
- a first reference dataset is generated.
- the first reference dataset comprises the amino acid sequences of the reference antigens (step S205), the encodings of the true epitopes (step S207) on the corresponding reference antigens, and the accessed (e.g. predicted) structure and/or surface characteristics of the reference antigens in an unbound state.
- the first reference dataset may further comprise one or more physiochemical characteristics assigned to each amino acid of the reference antigens.
- the first machine learning model is trained using the first reference dataset generated in step S211.
- the first machine learning model is or comprises a BLSTM, and may be trained using techniques known in the art.
- the first machine learning model is trained to learn a relationship the structure and/or surface characteristic(s) of the first reference proteins in an unbound state, and the B-cell epitopes classified as true B-cell epitopes.
- the second machine learning model(s) (schematically shown at 300) is used to predict structure and/or surface characteristics both when training the first machine learning model ( Figure 2), and when using the trained first machine learning model to predict the presence of B-cell epitopes on a protein of interest ( Figure 1).
- the second machine learning model is trained to predict structure and/or surface characteristics of a query protein in an unbound state.
- the first machine learning model is configured to predict BCEs in their unbound state, i.e. , having the three dimensional form and characteristics that the Ab “sees” and initiates binding upon.
- Figure 3 is a flow diagram summarising the steps of a preferred embodiment of training a second machine learning model according to an embodiment of the invention.
- a plurality of second reference proteins in an unbound state are accessed. These second reference proteins may be accessed from publicly available databases such as the PDB.
- step S303 the amino acid sequences of the second reference proteins are accessed.
- the three-dimensional structure and/or surface characteristics of the second reference proteins are accessed. Typically these include the secondary structure, the relative solvent accessibility, and the upper and lower half sphere exposures.
- the DSSP algorithm Wilfgang Kabsch and Christian Sander. Dictionary of protein secondary structure: Pattern recognition of hydrogen-bonded and geometrical features. Biopolymers, 22(12):2577-2637, 2018/11/30 December 1983) was used to compute the secondary structure and RSA, and the BioPython package (Peter J A Cock, TiagoAntao, Jeffrey T Chang, Brad A Chapman, Cymon J Cox, Andrew Dalke, Iddo Friedberg, Thomas Hamelryck, Frank Kauff, Bartek Wilczynski, et al.
- Biopython freely available python tools for computational molecular biology and bioinformatics. Bioinformatics, 25(11):1422-1423, 2009) utilised to compute the UHSE and LHSE.
- the DSSP and BioPython algorithms may be applied to the structure files of the second reference proteins (e.g. obtained from the Protein Data Bank, PDB).
- a second reference dataset is generated using the amino acid sequences accessed in step S303, and the structure and/or surface characteristics of the second reference proteins accessed in step S305.
- the second reference dataset may further comprise one or more physiochemical characteristics assigned to each amino acid of the second reference proteins.
- the second machine learning model is trained using the second reference dataset (e.g. to learn a relationship between one or more features of the amino acid sequences of the second reference proteins and the corresponding structure and/or surface characteristics).
- the second machine learning model is or comprises a BLSTM, and may be trained using techniques known in the art. It will be appreciated that different second machine learning models may be trained to predict different respective structure and/or surface characteristics. For example, one second machine learning model may be trained to predict RSA and HSE metrics, and further second machine learning model trained to predict the secondary structure.
- Section 2 explains pre-training techniques used for all the models.
- Section 3 presents some preliminary analyses we performed to choose input features for the models.
- Section 4 presents the structural algorithms for RSA, HSE and SS used in the current BCE algorithms.
- Section 5 presents the structural algorithms for RSA, HSE and SS that are used as features in the BCE model.
- Section 6 presents an example of a first machine learning algorithm according to the invention, used to predict CBCEs, while Section 7 presents an outline of a first model used to predict linear BCEs (LBCEs).
- LBCEs linear BCEs
- AEs auto-encoders
- the intermediate layers may reduce or expand the dimensions of the input.
- Each layer in the main network is converted into an AE, and trained on some compressed amount data.
- the pre-trained weights of the outer layers are used as constants to the subsequent AEs. This is repeated until all the main network layers are pre-trained.
- the pre-trained weights can then be used as starting points for the main training procedure.
- Continuous features provide a single value per amino-acid in the alphabet.
- we averaged the values of each feature by scanning the protein sequence using a sliding window of a peptide length of 9 amino-acids. The average feature value was assigned to the middle amino-acid in the peptide. This was done for every feature on each protein sequence in the entire training and test datasets.
- Correlation coefficient For continuous-categorical features we used Kendall rank correlation coefficient , and for categorical-categorical we used Mutual information criterion. For every pair-wise feature coefficient population we removed outliers using the IQR (box-plot) method. The median of the whole population is then taken as the final coefficient for every pair-wise feature-feature correlation. We defined as correlated continuous-continuous feature pairs with median Spearman’s correlation coefficient > 0.7, continuous-categorical feature pairs with Kendall rank correlation coefficient > 0.7, and categorical-categorical feature pairs with Mutual Information > 2.
- Figure 6 shows the correlation results of the feature-feature pairs for the RSA/HSE model.
- Figures 6(a) to (c) show feature-feature correlations using the data from 6.1 and 7.1.
- Figures 6(d) to (f) show feature-feature correlations using the data from 5.1. Black dots correspond to correlated pairs. The heat-maps are symmetric. Note that not all feature names are written in the x- and y-axis.
- the features for RSA, UHSE, LHSE and SS were predicted using the current models (see 4).
- For the 5.1 dataset those figures were the observed ones, from PDB.
- continuous- continuous feature-pairs resulted in many correlated features.
- For continuous- categorical feature-pairs only a few were correlated, while no correlation was observed between categorical-categorical feature pairs.
- Figure 7 shows the Spearman’s correlation coefficients for continuous features against (a) RSA, (b) UHSE and (c) LHSE.
- the x-axis is the correlation
- the y-axis are the feature names, sorted from highest to lowest absolute correlation of the median of the population. Each plot is from all windows.
- Figure 8 shows the Kendall’s correlation coefficients for categorical features against (a) RSA, (b) UHSE and (c) LHSE.
- the x-axis is the correlation
- the y-axis are the feature names, sorted from highest to lowest absolute correlation of the median of the population. Each plot is from all windows.
- Figure 9 shows the Kendall’s correlation coefficients for continuous features against (a) SS, (b) CBCE and (c) LBCE.
- the x-axis is the correlation
- the y-axis are the feature names, sorted from highest to lowest absolute correlation of the median of the population. Each plot is from all windows.
- Figure 10 shows the Mutual Information criterion for categorical features against (a) SS, (b) CBCE and (c) LBCE.
- the x-axis is the correlation
- the y-axis are the feature names, sorted from highest to lowest absolute correlation of the median of the population. Each plot is from all windows.
- the data-base consisted of 41592 total structures. Those structures contained 70489 protein sequences which resulted in 53524 unique sequences. Subsequences of longer sequences were kept as different data point entries. The reason being that their 3D conformational folding properties may be different because of their shorter length.
- the DSSP algorithm was used to compute the SS and the RSAfor each molecule in each structure file.
- DSSP computes the following secondary structure classes for a protein sequence; a-helix, 310-helix, K-helix, isolated p-bridge, p-strand, turn, bend and coil.
- the BioPython package was used in to compute the UHSE and LHSE.
- each amino-acid in the data base was assigned a value for the RSA, UHSE, LHSE and a class for the SS.
- the mean per amino-acid was taken for RSA, UHSE, LHSE among identical sequences.
- amino-acids of identical sequences but with different SS classes were assigned to the coil class.
- the model predicts RSA, UHSE and LHSE from a primary protein sequence.
- LHSE may not give useful information about the surface position of an amino-acid, it may help predicting UHSE more accurately, since both together are forming a probability metric around each amino-acid.
- the global model loss was the weighted sum of the individual losses with weights: 50, 100 and 125 for RSA, UHSE and LHSE, respectively. The weights were decided by first training the model without them and then see the differences in the magnitude of the three losses. The weights contributed in such a way that the three losses gave the same contribution to the global model loss.
- Figure 11 shows a schematic simplified architecture of the model.
- the first layer is the input layer.
- the input layer may comprise one or more input nodes, for example to receive the input features discussed below.
- Each input node is coupled to a respective masking node in a masking layer.
- each masking node is coupled to a respective BLSTM node in a BLSTM layer.
- Each BLSTM node is coupled to a concatenation node in a concatenation layer.
- the concatenation layer can be coupled to one or more further BLSTM layers, each comprising a corresponding BLSTM node.
- the final of these BLSTM nodes is coupled to an output layer, which in this case comprises a plurality of output nodes corresponding, respectively, to the RSA, UHSE and LHSE output predictions.
- Figure 12 shows the results for RSA/HSE model.
- Figures 12(a)-(c) show the Spearman’s correlation coefficients, and 12(d)-(f) the Mean Square Error (MSE) losses.
- the x-axis is the epochs, the y-axis is the metric values.
- the plots are the average of 5-fold CVs including Gaussian Cl.
- the blue line is for training and the red (dashed) for validation sets.
- Test results We used the remaining ⁇ 20% of the data as test data to assess the performance of the model on completely unseen data. See Figure 13 for the Spearman’s correlation coefficient metric. The models performed well, without overfitting and had high performance for all evaluation metrics.
- Figure 13 shows the results for RSA/HSE model, (a)-(c): Spearman’s correlation coefficients, (d)-(f): MSE losses.
- the x-axis is the epochs, the y-axis is the metric values.
- the solid line is for the training and the dotted for the validation set.
- the final training was done on the entire dataset. Using the optimal number of epochs determined by the cross-validation procedure. See Figure 14 for the Spearman’s correlation coefficient metric. The models seem to perform well for all metrics. The model for trained for 70 epochs, chosen based on the CV and testset analysis.
- Figure 14 shows the results for RSA/HSE model, (a)-(c): Spearman’s correlation coefficients (d)-(f): MSE losses.
- the x-axis is the epochs, the y-axis is the metric values.
- the model uses the categorical cross-entropy loss to assign one of the following classes to each amino-acid in its input sequence; Helices, Strands and Coils.
- the architecture of the model is substantially the same as that illustrated in Figure 11 , with a single output node in the output layer. Note: the input features were computed per amino-acid as they are not averaged over sliding windows.
- Table 2 Features used in the struct 3d ss m3 model.
- Figure 15 illustrates the results for SS model.
- Figures 15(a)-(e) show PR curves.
- the x-axis is the TPR
- the y-axis is the PPV.
- Figure 15(f) shows categorical crossentropy loss.
- the x-axis is the epochs
- the y-axis is the loss values.
- the plot is the average of 5-fold CVs including Gaussian Cl.
- the blue line is for training and the red (dashed) for validation sets.
- the final training was done on the whole data. Using the optimal number of epochs determined by the cross-validation procedure. See Figure 17 for the F1 metric.
- the final trained model performed well, without significant overfitting and gave high performance for all metrics.
- Figure 17 shows results for SS model.
- Figure 17(a)-(c) F1 metrics.
- the x-axis is the epochs, the y-axis is the F1 value.
- Figure 17(d) Categorical cross-entropy loss.
- the x-axis is the epochs, the y-axis is the loss values.
- the final training was done on the whole data. Using the optimal number of epochs determined by the cross-validation procedure.
- Table 4 Features used in the ss3 m6 v1 model.
- the final training was done on the whole data. Using the optimal number of epochs determined by the cross-validation procedure.
- BLSTMs based models. This allowed us to model the whole protein seguences bidirectionally as one observation, without the need of segmenting it. Therefore, distant, and contextual amino-acid relationships are captured by the model. Moreover, the use of BLSTMs allowed us to train different protein lengths simultaneously. The main goal of our models was to create nonAb specific CBCE predictors. This was achieved by the manner we prepared the training data described in section 6.1.
- IMGT The international ImMunoGeneTics information system
- Those genes were downloaded for both V H and V chains, from all the available organisms, that is; human, mouse, rhesus monkey, rabbit and rat.
- the IMGT provides information about the V H and V chains of an Ab, and its paratope regions (the CDRs [Complementary determining region of the antibody] and FRs [framework region of antibody] of each chain).
- V H and V chains in each structure were paired to recreate valid Abs.
- V H and V chains with the minimum mean atom distance were assigned as pairs and assumed to belong to one single Ab.
- Stand-alone V H or V chains were not considered on the downstream analysis as they could not define a complete Ab.
- a paratope analysis provided information about the contacts formed with each Ag. Two amino-acids were assumed to be in contact if any of their atoms were located within a certain probe distance from each other.
- Figure 18(a) shows the normalised mean of those contacts (x-axis shows the amino acids, y-axis shows the paratope parts). Most of the contacts are made with CDR1 and CDR2 of the V H s and CDR1 and CDR3 of the V s. The absence of CDR3 on V H s may be due to the difficulty of identifying it using IgBlast. Moreover, the FR regions do not seem to form as many contacts as the CDR regions, as expected. Additionally, the standard deviation of the total contacts per amino-acid and paratope part is relatively low in Figure 18(b), indicating that similar number of contacts are made between different Ags and Abs. This could be an indication of the possibility of predicting CBCEs without any further information about the actual Abs. Figure 18(b) shows the STD for number of contacts. X-axis shows the amino-acids, y-axis shows the paratope parts.
- V H and V chain pair we defined a single CBCE. This was done by first identifying atoms on any Ag whose distance from the CDRs regions of any V H and V chain pair was ⁇ 4 Angstroms. The amino-acids that those atoms belonged to were defined as contacts between the Ag and the Ab, that is, they defined the CBCE. Multiple CBCEs from different Abs could be defined on the same Ag. Finally, structures that did not define any CBCE within the 4 Angstrom distance were discarded.
- CBCEs were also mapped to similar Ag. Undiscovered CBCEs could increase the false negative rate of the prediction models. To decrease the possibility of assigning undiscovered CBCEs as negative data we copied observed CBCEs to similar Ags. First we identified clusters of similar Ags using BlastPlus with > 90% similarity and at most 2 gaps. CBCEs were copied from an Ag in a cluster to all the other Ags in the same cluster, as long as their corresponding position and distancing was exactly the same based on the mapping from BlastPlus.
- model 13 conf v1 has two main CBCE predictors, namely model 13 conf v1 and model 17 conf v1.
- model 13 conf v1 gives two outputs
- model 17 conf v1 gives one.
- Both models take as input features, among others, a permutation vector. Every sequence in the data is associated with at least one true CBCE, those CBCEs are turned into binary 2D vectors and are given as input to the models.
- the common output of the two models is a probability, so both models are basically asking the question: “Is this permutation vector, on this specific sequence, a true CBCE or not?”.
- the second output of the model 13 conf v1 model is the permutation vector itself. This part of the model works as an AE, where it returns a probability per amino-acid. This output can be seen as a contribution of each amino-acid in the sequence to the specific CBCE in question.
- the goal of those models was to predict CBCEs on the protein Ag sequence, before the Ab binding event took place.
- the dataset used, described in section 6.1 contains the linear protein sequence and the true CBCEs. Therefore, there is no information about any structural or surface characteristic of those protein sequences. However, we need information about how each protein sequence’s secondary structure and surface would look like before the binding event.
- FIG 19(a) schematically shows the architecture of the model 13 conf v1 model.
- the architecture of the model is substantially the same as that shown in Figure 11.
- the output layer comprises two output nodes.
- the BCE amino acid output (BCE_aa) is the actual permutation output, where binary cross-entropy loss was used.
- the BCE perm output is a probability vector indicating if the input permutation is a true CBCE (second position on vector) or not (first position on vector).
- Figure 19(b) shows the architecture of the model 17 conf v1 model.
- the BCE perm represents the same output as the previous model. Note: the input features were computed per amino-acid as they are, not averaged by windows.
- the BCE perm output is computed from amino-acid values.
- the last layer of the output is a Dense layer with sigmoid activation.
- a value in e [0, 1] is returned per amino-acid.
- a 2-class probability vector is computed for that sequence as: where N is the length of the sequence and xi is the dense layer output on aminoacid at position i on the sequence. This 2-class probability vector is then used in the binary cross-entropy loss.
- the model 13 conf v1 model had weighted global loss.
- the global model loss was the weighted sum of the individual ones, that is BCE perm and BCE aa, with weights: 1 , and 1 , respectively. Different weights could be tested as well.
- the positive data would be every amino-acid in a sequence which is a part of any true CBCE, and the negative data would be any other amino-acid on the same sequence.
- the response to this vector is a single probability (that is output BCE perm from section 6.2). Since we only have observed true positive CBCEs from the data in section 6.1 , we do not have any permutation corresponding to negative data.
- Figure 20 shows the total amount of amino-acids from all the observed CBCEs.
- Figures 20(a) and 20(b) show, respectively, density and box-plot of observed amount of total amino-acids in the true CBCEs.
- Figures 20(c) and 20(d) shown log-scales of Figures 20(a) and 20(b) respectively. Where we see that the median amount is at 10 amino-acids.
- Table 5 Features used in the model 13 conf vl and model 17 conf vl models.
- Model model 13 conf yl Figure 21 shows results for the precision-recall metric.
- the model performed well, without significant overfitting and had high performance for all evaluation metrics. Moreover, a small variation is observed between different CV runs for the loss.
- the models were trained for around 466 epochs and apparent bumps in the loss may indicate that the random generator create negative CBCE permutations which were similar to the positives. However, their negative effect became smaller as the epochs proceeded further.
- Figure 21 illustrates results for CBCE model 13 conf v1.
- Figure 21 (a)-(e) PR curves.
- the x-axis is the TPR
- the y-axis is the PPV.
- Figure 21(f) Weighted cross entropy loss.
- the x-axis is the epochs
- the y-axis is the loss values.
- the plot is the average of 5-fold CVs including Gaussian Cl.
- the blue line is for training and the red (dashed) for validation sets.
- Figure 22 shows results for the precision-recall metric.
- the model seems to perform well, without significant overfit before 250 epochs and gives high performance for all metrics. Moreover, small variation is observed between different CV runs for the loss. The models were trained for 455 epochs.
- Figure 22 illustrates results for CBCE model 17 conf v1.
- the x-axis is the TPR
- the y-axis is the PPV.
- the x-axis is the epochs
- the y-axis is the loss values.
- the plot is the average of 5-fold CVs including Gaussian Cl.
- the blue line is for training and the red (dashed) for validation sets.
- Figure 23 shows the results for the precision-recall metric.
- the model performed well, without significant overfitting and gave high performance for all evaluation metrics.
- Figure 23 illustrates the results for CBCE model 13 conf v1.
- Figure 23(a) PR curves.
- the x-axis is the TPR, the y-axis is the PPV.
- the x-axis is the epochs, the y-axis is the loss values.
- the solid line is for the training and the dotted for the validation set.
- Figure 24 shows the results for the precision-recall metric.
- the model seems to perform well, without significant overfit and gives high performance for all metrics.
- Figure 24 illustrates the results for CBCE model 17 conf v1.
- Figure 24(a) PR curve.
- the x-axis is the TPR, the y-axis is the PPV.
- the x-axis is the epochs, the y-axis is the loss values.
- the solid line is for the training and the dotted for the validation set.
- the final training was done on the whole data. Using the optimal number of epochs determined by the cross-validation procedure.
- Figure 25 shows the results for the F1 metric. The model seems to perform well, without significant overfit and gives hight performance for all metrics. The model for trained for 500 epochs, chosen based on the CV analysis. Figure 25 shows the results for CBCE model 13 conf v1.
- Figure 25(a)-(b) F1 metrics. The x-axis is the epochs, the y-axis is the F1 value.
- Figure 25(d) Overall model loss. The x-axis is the epochs, the y-axis is the loss values.
- Figure 26 shows the results for the F1 metric.
- the model seems to perform well, without significant overfit and gives high performance for all metrics.
- Figure 26 shows the results for CBCE model 17 conf v1.
- Figure 26(a) F1 metric.
- the x-axis is the epochs, the y-axis is the F1 value.
- Figure 26(b) Overall model loss.
- the x-axis is the epochs, the y-axis is the loss values.
- BLSTMs models For predicting LBCEs we use BLSTMs models as well. This allows us to model the whole protein sequences as one observation, without the need of segmenting it. Therefore, distant amino-acid relationships should be able to be captured by the model. Moreover, the use of BLSTMs allows us to train different protein lengths simultaneously. Therefore, the model type here is the same as the one for the CBCEs in section 6. The main goal of our models was to create non-Ab specific LBCE predictors. This is achieved by the way we prepare the training data as described in section 7.1.
- LBCEs were also mapped to similar Ag, as we did in section 6.1 for the CBCEs. Undiscovered LBCEs could increase the false negative rate of the prediction models. In an attempt to decrease the possibility of assigning undiscovered LBCEs as negative data we copied observed LBCEs to similar Ags. First we identified clusters of similar Ags using BlastPlus with > 90% similarity and at most 2 gaps. LBCEs were copied from an Ag in a cluster to all the other Ags in the same cluster, as long as their corresponding linear coordinates were exactly the same based on the mapping from BlastPlus.
- Every sequence in the data is associated with at least one true LBCE.
- the model takes as input features, among others, a permutation vector.
- This vector is the LBCEs which are turned into binary 2D vectors and are given as input to the model.
- the output of the model is a probability, so the model is basically asking the question; ”ls this permutation vector, on this specific sequence, a true LBCE or not?”.
- the permutation vectors are the same as described in section 6.2, where the only difference is that they represent LBCEs and not CBCEs.
- the architecture of the model 12 linear v1 model has substantially the same structure as the model depicted in Figure 19(b).
- the bee output is a probability vector indicating if the input permutation is a true LBCE (second position on vector) or not (first position on vector).
- the bee output is computed from aminoacid values as described in section 6.2. Note: the input features were computed per amino-acid as they are, not averaged by windows.
- Figure 27 shows the total amount of amino-acids from all the observed LBCEs. Where we see that the median amount is at 17 amino-acids. After the outlier analysis, we found that valid LBCEs should contain a total amount of amino-acids in range [5, 30], The outliers were finally removed from the data, and the models were trained without them.
- Figure 27 shows outlier analysis for LBCEs.
- Figures 27(a) and (b) show, respectively, density and box-plot of observed amount of total amino-acids in the true LBCEs.
- Figures 27(c) and (d) show, log-scales of Figures 27(a) and 27(b) respectively.
- Table 6 Features used in the model 12 linear v1 model.
- the binding event of an antibody to an antigen is a key event in activating the humoral B Cell immune response against any pathogenic threat, such as viral and bacterial infections, in addition to cancer antigens in malignant tumour cells.
- B Cell epitope mapping is essential in therapeutic antibody development. The prediction of true B-cell epitopes by the present invention not only provides a functional understanding of the critical residues involved antibody-antigen binding; it also aids in the selection of therapeutically relevant antibodies to the prioritized predicted epitopes for further development as therapeutic antibodies.
- B cell epitopes are of direct help to the design of immuno-diagnostic and immune monitoring of therapy strategies and reagents. Integrating prediction approaches for the accurate and comprehensive detection of BCEs can greatly facilitate diagnosis.
- One example of this would be the rapid and accurate detection of BCE epitopes on the spike protein of beta-coronaviruses, allowing for the diagnosis of COVID in a manner that can distinguish infections caused by common cold viruses from the SARS-CoV-2 virus causing COVID.
- the BCE prediction technology can be used to identify immunodominant or humoral reactive epitopes in tumour associated antigens or neoantigens in immuno-diagnostic and immune monitoring clinical applications.
- FIG. 4 schematically illustrates an example of a system suitable for implementing embodiments of the method.
- the system 1100 comprises at least one server 1110 which is in communication with a reference data store 1120.
- the server may also be in communication with an automated peptide synthesis device 1130, for example over a communications network 1140.
- the server may obtain, for example using from the reference data store, an amino acid sequence of one or more query proteins. The server may then predict whether the query protein comprises a BCE that is likely to instigate a binding event with an antibody (e.g. instigate an immunogenic response).
- Query proteins predicted to contain one or more BCEs may be sent to the automated peptide synthesis device 1130 to synthesise the query protein or parts thereof.
- Techniques for automated peptide synthesis are well known in the art and it will be understood that any known technique may be used.
- the query protein or epitope is synthesized using standard solid phase synthetic peptide chemistry and purified using reverse-phase high performance liquid chromatography before being formulated into an aqueous solution. If used for vaccination, prior to administration the peptide solution is usually admixed with an adjuvant before being administered to the patient.
- the server may comprise the functions described above to identify query proteins predicted to contain one or more viable BCEs. It will of course be understood that these functions may be subdivided across different processing entities of a computer network and different processing modules in communication with one another.
- Example vaccine development ecosystems are well known in the art and are described at a high-level for context, but for brevity we do not describe the ecosystem in detail.
- a first, sample, step may be to isolate DNA from a tumor biopsy and matched healthy tissue control.
- the data is sequenced and the variants identified i.e. the mutations.
- the associated mutated peptides may be generated «in silico».
- a candidate protein may be predicted and selected and target epitopes identified for vaccine design. That is, the candidate peptide sequence is chosen based on its predicted binding affinity determined using the technique described herein.
- the target epitopes are then generated synthetically using conventional techniques as described above. Prior to administration the peptide solution is usually admixed with an adjuvant before being administered to the patient (vaccination).
- the target epitopes can be engineered into DNA or RNA, or engineered into the genome of a bacteria or virus, as with any conventional vaccine.
- the proteins predicted by the methods described herein may also be used to create other types of vaccine other than peptide based vaccines.
- the proteins (or predicted epitopes therein) could be encoded into the corresponding DNA or RNA sequence and used to vaccinate the patient.
- the DNA is usually inserted in to a plasmid construct.
- the DNA can be incorporated into the genome of a bacterial or viral delivery system (can be RNA also - depending on the viral delivery system) - which can be used to vaccinate the patient - so the manufactured vaccine in a genetically engineered virus or bacteria which manufactures the targets post immunisation in the patient i.e. in vivo.
- each column can be considered as a single permutation vector shared among the betacoronavirus species counted on the Y- axis.
- the X-axis is sorted in descending order of total unique species. The colouring of the bars is based on Subgenus.
- the permutation vectors represented in the columns in Figure 30 are structural representations of the epitopes predicted by the BCEP predictor. We next generated amino acid presentations of these immunodominant (broadly immunogenic) epitopes and calculated the sequence logos for the top 1 permutation per RBD (ties included). The logos are created by taking the highest column (permutation) from Figure 30 above, and based on the positive BCE prob>0.9 we create the sequence logos of their epitopes.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Medical Informatics (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Theoretical Computer Science (AREA)
- Bioinformatics & Computational Biology (AREA)
- Data Mining & Analysis (AREA)
- Spectroscopy & Molecular Physics (AREA)
- General Health & Medical Sciences (AREA)
- Evolutionary Biology (AREA)
- Biophysics (AREA)
- Biotechnology (AREA)
- Bioethics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Software Systems (AREA)
- Public Health (AREA)
- Evolutionary Computation (AREA)
- Epidemiology (AREA)
- Databases & Information Systems (AREA)
- Artificial Intelligence (AREA)
- Chemical & Material Sciences (AREA)
- Analytical Chemistry (AREA)
- Genetics & Genomics (AREA)
- Molecular Biology (AREA)
- Proteomics, Peptides & Aminoacids (AREA)
- Investigating Or Analysing Biological Materials (AREA)
Abstract
A computer-implemented method of predicting whether a protein comprises a IB-cell epitope that is likely to instigate a binding event with an antibody is disclosed. The method comprises: (a) accessing one or more structure and/or surface characteristics of the protein; and (b) inputting the one or more structure and/or surface characteristics of the protein into a trained first machine learning model to predict whether the protein comprises a true B-cell epitope, wherein the first machine learning model is trained by: generating a first reference dataset that comprises: (i) a plurality of first reference proteins, each first reference protein comprising at least one B-cell epitope classified as a true B-cell epitope; and (ii) one or more structure and/or surface characteristics of each first reference protein in an unbound state; and training the first machine learning model using the first reference dataset to learn a relationship between the structure and/or surface characteristic(s) of the first reference proteins in an unbound state, and the B-cell epitopes classified as true B-cell epitopes.
Description
B-CELL EPITOPE PREDICTION
BACKGROUND
Lymphocytes are a type of white blood and are central components of the immune system of most vertebrates. They include B cells, T cells and natural killer (NK) cells. Their function is to identify and protect against pathogens, such as bacteria, viruses, or malignant cancer cells that harbor mutated protein antigens. Both B- and T-Cells express surface receptors which identify specific molecular components on pathogens called antigens (Ags). The successful recognition of an Ag triggers a wider adaptive immune response which ultimately may eliminate the potentially pathogenic threat.
Antibodies (Abs), also known as immunoglobulins (Ig), are glycosylated protein molecules present on the surface of B cells (surface immunoglobulins) or B cell eluted Ig molecules, that act as antigen specific B cell receptors (BCR). Upon successful Ag recognition, the Ig or membrane-bound BCRs will bind on specific Ag amino acid contact points, collectively named here as a B-cell epitope (BCE). This binding event may trigger the activation of an immune response that includes a clonal B cell expansion which will produce a vast amount of the Ag specific Ig serum eluted molecules or effector B cells harbouring Ag specific BCRs. The secreted soluble Igs and effector BCRs, also called antibodies (Abs), will have the same binding specificity as the original Ig-Ag binding interaction event. Abs will then circulate into the serum and bind to the same BCEs to signal the elimination of the pathogenic or cell-stressed cells that correspond to the identified Ags.
Ab based vaccines have the potential to prepare or educate the immune system of each individual against potentially harmful Ags sourced from infectious pathogenic threats or stressed malignant cancer cells. Vaccine design aims for the modification of the Ag in such a way that it bypasses harmful malignant, pathogenic or infectious properties while maintaining potential protective immunes responses driven by immunogenic epitopes. The accurate identification or prediction of B-Cell epitopes on antigens is therefore of great importance not only
for vaccine design, but also in the domains of molecular diagnostics, immune- monitoring and immunotherapy.
BCEs can be divided into two main categories: linear B cell epitopes (LBCEs) and conformational B cell epitopes (CBCEs). Both LBCEs and CBCEs can be bound by Abs and, thus, trigger an immune response. It is estimated that > 90% of the BCEs are conformational, while the rest are linear.
LBCEs are made up of continuous (sequential) sequences of amino-acids. These BCEs can be discovered by identifying positions on the unfolded Ag sequence which can be bound by Abs. LBCEs are often catalogued at the immune epitope database (IEDB) sourced from a large variety of experimental assays. Each assay provides information about the protein sequence tested (that is, the Ag), the LBCE sequence as well as its coordinates on the Ag, and the Ab which the current LBCE was tested on. Moreover, they provide an indicator as to whether the LBCE was bound by the given Ab. The binding information is sometimes described in the IEDB in the form of human readable labels, that is; positive, positive-high, positiveintermediate, positive-low and negative. The simplicity of LCBE as linear molecular entities, and the well-established experimental methods that are used to characterize them has led to a relatively large amount of available LBCE data compared to CBCE data. Nevertheless, annotation errors are frequent in this type of data, since many amino acids of the identified BCEs are not actually interacting with the Abs under the tertiary conformation of the Ag.
CBCEs are made up of short segments of amino-acids with potential diverse cis, trans, short or distal coordinate distances between each of the amino acid contact points that constitute the BCE. These segments come into proximity during the folding of the Ag in the act of binding to the Ab. CBCEs can be experimentally characterized using X-ray crystallography on Ag-Ab structure complexes. This data is usually catalogued in the Protein Data Base (PDB) in the form of 3D structures, and typically consist of one or more Ag structures bound by at least one Ab structure. PDB structure data provides information about the 3D coordinates of each molecule in the structure. However, the complexity of this
method makes the CBCE data somewhat more limited compared to that of LBCEs. Moreover, the 3D structures may not accurately reflect true CBCEs. For example, Abs on the complexes may be missing their heavy or light chains, or the 3D structure may only provide Ag-Ab contacts outside the complementaritydetermining regions (CDRs). Further, undiscovered CBCEs on any Ag, and the immensely sparse information available on ground truth negative epitopes on an Ag makes the prediction of true CBCEs even more challenging.
There exist various algorithms for predicting either LBCEs and/or CBCEs. However, although many of those algorithms perform well on their test data, they do not produce the same satisfactory results on new independent test data when they are compared against other algorithms of a similar kind.
In many conventional CBCE prediction algorithms, the expected input is a 3D protein structure file of the Ag. However, due to the labour intensive and expensive scale of laboratory experiments required to determine this, the 3D protein structure of the Ag will almost certainly not be available at prediction time in most use cases; which limits the use of the 3D structure input algorithms.
Recently, methods to predict B-cell epitopes have been published that allow for the input of protein sequences without the strict necessity at prediction time for accompanying input 3D structure information. One example is the “DiscoTope 3.0” algorithm (https://doi.org/10.1101/2023.02.05.527174). A machine learning strategy is employed which uses, as training data, epitopes from antibody-antigen complexes, that are mapped onto the antigen. DiscoTope-3.0 is trained on both predicted and solved antigen structures (i.e. , antibody-antigen structures) merged together, whereas previous versions of that method were trained on experimentally solved antibody-antigen structures. Although B-cell epitope prediction techniques that require the sequence of the protein as input are more user-friendly than those that require the full 3D structure, it can be difficult to capture relationships between spatially distance amino acids, and to distinguish between potentially different CBCEs on the same antigen.
There is therefore a desire to improve the prediction of B-Cell epitopes, in particular CBCEs.
SUMMARY OF INVENTION
In accordance with a first aspect of the invention, there is provided a computer- implemented method of predicting whether a protein comprises a B-cell epitope that is likely to instigate a binding event with an antibody, the method comprising:
(a) accessing one or more structure and/or surface characteristics of the protein; and
(b) inputting the one or more structure and/or surface characteristics of the protein into a trained first machine learning model to predict whether the protein comprises a true B-cell epitope, wherein the first machine learning model is trained by: generating a first reference dataset that comprises:
(i) a plurality of first reference proteins, each first reference protein comprising at least one B-cell epitope classified as a true B-cell epitope; and
(ii) one or more structure and/or surface characteristics of each first reference protein in an unbound state; and training the first machine learning model using the first reference dataset to learn a relationship between the structure and/or surface characteristic(s) of the first reference proteins in an unbound state, and the B-cell epitopes classified as true B-cell epitopes.
During binding between a B-cell epitope and an antibody, the three-dimensional structure of the BCE may be altered. In other words, the 3D structure of the BCE within a bound antibody-antigen complex may differ from the 3D structure of the BCE before the binding event was instigated.
The present invention provides a method of predicting whether a (e.g. query) protein comprises a B-cell epitope using a trained first machine learning model that is trained using a first reference dataset comprising structure and/or surface
characteristics of a plurality of first reference proteins in an unbound state. Each first reference protein comprises at least one B-cell epitope classified as a true IB- cell epitope. In this way, the first machine learning model is advantageously trained to predict true B-cell epitopes on a query protein before a binding event has taken place.
Herein, the term “unbound state” (or “native state”) is used to mean the state (e.g. the three-dimensional structure) of a protein or BCE without influence from other proteins, structures or interacting molecules (e.g. before a binding event has taken place). Importantly, this unbound state is what the Ab “sees” and initiates the binding event, and not what an Ab has already “seen” and bound on.
Thus, the use of one or more structure and/or surface characteristics of each first reference protein in an unbound state provides a key advantage over state-of-the- art models that are trained on Ag-Ab complexes (i.e. , proteins in a bound state), and which may lead to errors as the predictions are made based on structural data obtained after the binding event has occurred.
Typically, the output of the trained first machine learning model is a probability that is indicative of whether the protein comprises a true B-cell epitope. As will be described further herein, the probability may typically be in the form of a probability that a candidate epitope on the query protein is a true epitope. In some embodiments the output may be in the form of a probability that individual amino acids (or groups of amino acids) within the amino acid sequence of the query protein form part of a true B-cell epitope. However, it is also envisaged that the output may be at a whole protein level (e.g. a probability that the query protein comprises a true BCE).
Herein, the term “true” B-cell epitope is used to mean a B-cell epitope that instigates a binding event with an antibody (e.g. instigating an immunogenic response and/or instigated by an immunogenic response).
The B-cell epitope may be a conformational B-cell epitope. The B-cell epitope may be a linear B-cell epitope. The trained first machine learning model of the present invention is typically able to predict the presence of both linear B-cell epitopes and conformational B-cell epitopes, although in some embodiments different models may be employed for predicting linear and conformational B-cell epitopes respectively, based on the reference data used for training.
In general, the protein may be a protein, protein-domain or a protein sub-unit. The term “protein” may include any protein subsequence that may have a viable 3D or functional structure. Typically, the protein comprises an antigen. The protein may comprise a neoantigen.
As described above, in step (a) of the method, one or more (e.g. three- dimensional) structure and/or surface (e.g. surface-exposed) characteristics of the protein are accessed. The structural and/or surface characteristics of the protein being investigated are important for predicting B-cell epitopes (in particular CBCEs), as the Abs are very specific to their binding BCEs. The one or more structure and/or surface characteristics of the protein may be characteristics at the whole protein level (e.g. secondary structure) or at the amino acid level (e.g. features associated with individual amino acids of the protein). The one or more structure and/or surface characteristics may be in the form of continuous features (such as relative solvent accessibility, RSA, or half-sphere exposure, HSA), and/or categorical features such as secondary structure.
In embodiments, the one or more structure and/or surface characteristics (e.g. for both the input protein being queried and for the first reference proteins) include one or more of: the secondary structure (SS) of the protein; the relative solvent accessibility, RSA; the half-sphere exposure, HSA. The HSA values may be divided into values for upper half-sphere exposure and lower half-sphere exposure. The SS may be regarded as a structure characteristic, and the RSA and HSE regarded as surface characteristics. A particular advantage of the present invention is that the three-dimensional full folding structure of the input protein (which may be difficult to measure, or may contain a large number of errors
if predicted) is not required. Instead, the method accesses one or more structure and/or surface characteristics of the protein which are advantageously more straightforward to obtain than a full folding structure, and typically contain fewer errors. The inventors have found that the characteristics of RSA, HSE and secondary structure in particular provide good results for the predictions, as they provide reliable indicators of the three-dimensional structure of the protein. It is envisaged that further structure and/or surface characteristics may also be used as input to the model. In this way, the inventive method can provide good BCE prediction results without requiring the full three-dimensional structure of the protein under investigation.
The method may further comprise accessing one or more physiochemical characteristics of the protein, and/or the amino acids that constitute the protein, and inputting the one or more physiochemical characteristics into the first machine learning model (e.g. in addition to the one or more structure and/or surface characteristics). In such embodiments, the first reference dataset typically further comprises one or more physiochemical characteristics of the first reference proteins, and/or the amino acids that constitute them. Examples of such physiochemical characteristics include bulkiness, molar fraction of buried residues, average flexibility index, normalised frequency for alpha helix, known number of codons coding for each amino acid in universal genetic code, polarity, conformational parameter for beta-turn, normalized frequency for beta-sheet, side chain classes, side chain polarity, retention coefficient, hydrophobic constants, normalized frequency of beta-turn, PK1 a-COOH, membrane buried helix parameter, antigenicity. Different sets of physiochemical characteristics may be used for predicting CBCEs and LBCEs. The one or more physiochemical characteristics may be computed (e.g. from the amino acid sequence of the protein) or measured using techniques known in the art. The physiochemical characteristics may be determined for each amino acid of the protein.
Typically, each amino acid (or side chains) that constitutes a protein sequence has physiochemical properties that contribute in some manner to its SS and 3D properties. In this way, embodiments that include the use of physiochemical
characteristics in addition to the one or more structure and/or surface characteristics may further enhance the performance of the first machine learning model.
In general, each of the structure and/or surface characteristic(s) and physiochemical characteristic(s) (where used) may contribute separately or in combination to the training and subsequent use of the first machine learning model to predict whether a protein contains a B-cell epitope.
Preferably, the one or more structure and/or surface characteristics of each first reference protein are predicted by: accessing the amino acid sequence of the first reference protein; and applying one or more second machine learning model(s) to the amino acid sequence to predict the one or more structure and/or surface characteristics, wherein the one or more second machine learning model(s) are trained on a second reference dataset that comprises a plurality of amino acid sequences of respective second reference proteins and their corresponding structure and/or surface characteristics in an unbound state.
Thus, typically, the one or more second machine learning model(s) are trained on the second reference dataset to learn a relationship between (e.g. one or more features of) the amino acid sequences of the respective second reference proteins and the corresponding structure and/or surface characteristics in an unbound state.
In this way, embodiments of the invention enable the prediction of the structure and/or surface characteristics of the first reference proteins in an unbound state which, as described previously, is a key aspect of the present invention. Typically, the one or more structure and/or surface characteristics include one or more of: the secondary structure of the protein; the relative solvent accessibility, RSA; the half-sphere exposure, HSE. These characteristics of the amino acids in proteins used to train models of the second reference dataset may be accessed using
known resources or tools, for example the DSSP and BioPython algorithms applied to the structure files of the second reference proteins (e.g. obtained from the Protein Data Bank, PDB).
Typically, in addition to the amino acid sequence, the input to the one or more second machine learning model(s) comprises a plurality of characteristics (e.g. features) of the first reference protein. The input characteristics are typically physiochemical characteristics but may in some embodiments include structure and/or surface characteristics. These characteristics are typically computed or measured per amino acid of the first reference protein using techniques known in the art. Examples of such (e.g. physiochemical) characteristics include one or more of hydrophobicity, side chain class, side chain polarity (e.g. for predicting HSE and RSA), and one or more of conformational parameter for coil, conformational parameter for alpha helix (e.g. for predicting secondary structure). Further characteristics may be used. In this way, the one or more second machine learning model(s) may be trained on the second reference dataset to learn a relationship between one or more (e.g. physiochemical) characteristics of the amino acid sequences of the respective second reference proteins, and the corresponding structure and/or surface characteristics (e.g. one or more of RSA, HSE, Secondary Structure) in an unbound state. Typically, the second machine learning algorithm(s) utilise a different set of protein characteristics from the first machine learning algorithm.
The amino acid sequence of the first reference proteins may be obtained using techniques known to the skilled person in the art, for example by one of: oligonucleotide hybridisation methods, nucleic acid amplification based methods (including but not limited to polymerase chain reaction based methods), automated prediction based on DNA or RNA sequencing, de novo peptide sequencing, Edman sequencing or any peptide related sequencing including and not limited to mass spectrometry. The amino acid sequence may be downloaded from a bioinformatic depository such as UniProt (www.uniprot.org).
In some embodiments, different second machine learning models may be used to predict different structure and/or surface characteristics. In some embodiments, one second machine learning model may be used to predict categorical features, and a different second machine learning model used to predict continuous features. The input characteristics and reference datasets for the models may differ accordingly. For example, in one embodiment a trained second machine learning model may be used to predict the secondary structure of a protein, and a different trained second machine learning model used to predict RSA and HSE characteristics.
Referring now back to step (a) of the inventive method, the structure and/or surface characteristics of the protein (e.g., in its unbound state) may be obtained using known techniques. For example, known algorithms (including computational approaches to predicting full 3D protein structures) or X-ray crystallography techniques, may be used to predict or measure the features of the protein. For example, RSA, HSE and secondary structure (and other) parameters could be accessed from a predicted or experimental full 3D structure of the protein, without the full structure itself being required as input to the model.
However, in particularly preferred embodiments, the one or more structure and/or surface characteristics of the query protein may be predicted by applying the one or more second machine learning model(s) discussed above. This may require accessing the amino acid sequence of the query protein. In this way, structure and/or surface characteristics of the protein under investigation may advantageously be computed directly from the amino acid sequence of the query protein, with no requirement for extra information to be sourced or input by the user.
One problem with conventional approaches to predicting B-cell epitopes, especially CBCEs, is a limited amount of data defining true CBCEs. Embodiments of the present invention therefore provide a method of obtaining first reference proteins each comprising at least one true B-cell epitope.
Thus, in some embodiments, at least some of the first reference proteins may be obtained by:
(i) accessing a plurality of (preferably non-obligate) protein complexes comprising at least three different protein chains;
(ii) filtering the plurality of protein complexes to leave only those having a protein chain mapped as a valid heavy (VH) chain of an antibody, a protein chain mapped as a valid light (V ) chain of an antibody, and a protein chain mapped as an antigen;
(iii) pairing the VH and V chains to form VH and V chain pairs; and
(iv) for each VH and V chain pair, defining a true B-cell epitope on the corresponding antigen; wherein at least some of the first reference proteins of the first reference dataset correspond to an antigen mapped in step (ii) that comprises at least one true IB- cell epitope defined in step (iv).
The first reference proteins obtained or identified in the manner above are typically mapped onto their corresponding amino acid sequences. In this way, the structural and/or surface characteristics of each first reference protein in an unbound state may be obtained from the respective amino acid sequence, independent of the protein complex from which it was identified. The characteristics may preferably be predicted using the one or more second machine learning model(s) described above. However, other known approaches to predicting the structure and/or surface characteristic(s) (e.g. from the amino acid sequence) may be employed, such as publicly available computational protein structure prediction algorithms.
Typically, the protein complexes accessed in step (i) may be of any organism or entity assigned a taxonomy ID in established authorities such as the NCBI. This advantageously allows the predictions made by the trained first machine learning model to be non-Ab specific.
In embodiments, each true B-cell epitope is defined as the amino acids of the corresponding antigen in contact with the corresponding VH and V chain pair,
preferably wherein the amino acids are defined as being in contact with the corresponding VH and V chain pair if they comprise any atoms within a predefined distance of the complementarity-determining region(s), CDR(s) of the VH and V chain pair, preferably wherein the predetermined distance is equal to or less than 4 Angstroms.
Preferably, the method may further comprise: accessing the amino acid sequence of the (e.g. query) protein; and generating one or more candidate encodings of respective one or more candidate B-cell epitopes on the protein, wherein each candidate encoding represents the respective candidate B-cell epitope as a plurality of data elements corresponding to amino acids of the amino acid sequence; wherein the inputs to the trained first machine learning model comprise the amino acid sequence of the protein, the one or more structure and/or surface characteristics of the protein, and the one or more candidate encodings.
The inventors have realised that a common problem with many conventional sequence-based techniques used to predict the presence of a B-Cell epitope on a protein is that the output prediction is typically in the form of a binary value per amino acid indicating whether each particular amino acid in the sequence of the protein is part of a B-Cell epitope. This conventional approach does not allow the user to distinguish between potentially different and distinct BCEs present on the protein that harbor some overlapping amino acid residues, and moreover does not necessarily consider the relationships between spaced-apart amino acids of the entire protein sequence being analysed in 3D conformational BCEs.
To improve upon these conventional techniques for BCE prediction, embodiments of the present invention may apply the trained first machine learning model to a combination of the input amino acid sequence of the protein under investigation, the one or more structure and/or surface characteristics, and one or more candidate encodings of respective one or more candidate B-cell epitopes, where each candidate encoding represents the respective candidate B-cell epitope.
In this way, the predictions provided by the trained first machine learning model are made on an “epitope by epitope” basis on the query protein (by considering each candidate encoding corresponding to a candidate epitope), rather than on an “amino acid by amino acid” approach. Thus, such embodiments consider whether a candidate B-cell epitope, encoded on the amino acid sequence under test, is a true BCE (or not) on the protein. This approach is particularly advantageous as it provides improved confidence that an epitope as a whole is indeed present (or not) on the protein, rather than having to infer the presence of epitopes (or multiple overlapping epitopes) based on individual amino acid scores. For example, the use of the candidate encodings representative of the candidate epitopes as a whole allows for potential relationships between spaced apart amino acids on the sequence to be considered. Providing predictions on such an “epitope by epitope” basis also provides increased accuracy and confidence with regard to predicting multiple (possibly overlapping) epitopes on the same protein.
The use of candidate encodings to provide predictions on an “epitope by epitope” basis is particularly advantageous in combination with the first machine learning model being trained on unbound data. The candidate encodings enable relationships between spaced apart amino acids on the sequence to be considered where, as discussed herein, those relationships may differ slightly between the protein being in a protein complex (bound state) or unbound (native) state.
In such embodiments in which candidate encodings are used, the output of the trained first machine learning model is typically a probability that each of the candidate encoding(s) represents a true B-cell epitope on the protein.
In such embodiments, typically, the accessed (e.g. predicted) structure and/or surface characteristics of the protein are in the form of a value and/or a class assigned to each amino acid of the protein amino acid sequence that represents a structure and/or surface characteristic of the protein. In this way, the input to the first machine learning model may be a set of characteristics per amino acid of the protein.
As outlined above, the candidate encodings of the B-cell epitopes represent the respective candidate B-cell epitope as a plurality of data elements corresponding to the amino acids of the amino acid sequence. Typically, each data element of a candidate encoding (or at least a subset of the plurality of data elements) corresponds to an amino acid of the amino acid sequence of the protein. Preferably, each of the one or more encodings is in the form of a binary vector. In such embodiments, the binary vector will typically be in the form of a sequence of “1”s and “0”s, with the “1”s each corresponding to an amino acid of the protein being a part of the candidate epitope (e.g. a probable Ag-Ab contact point) , and the “0”s each corresponding to an amino acid of the protein not forming part of the candidate epitope (e.g. not a probable Ag-Ab contact point). In some embodiments, different encoding protocols may be used (e.g. one-hot encoding) to represent the candidate epitopes.
In some embodiments, the one or more candidate encodings represent all possible combinations of the amino acids within the amino acid sequence. This may include both linear B-cell epitopes (where each of the encoded amino acids will be sequentially positioned in the amino acid sequence), and conformational B-cell epitopes (where at least some of the encoded amino acids are spaced apart on the amino acid sequence). However, in some embodiments, prior knowledge of the protein and/or B cell epitope makeup may be used to reduce the search space of the encodings. For example, if the protein contains a neoantigen, the encodings may be fixed in the position of the mutation point (e.g. each encoding will have the same data element at the position of the mutation point on the amino acid sequence).
The one or more candidate encodings representing combinations of amino acids within the amino acid sequence may be queried through deep reinforcement learning models. The one or more candidate encodings representing combinations of the amino acids within the amino acid sequence may be sourced from generative models. In some embodiments, deep reinforcement learning (DRL) techniques may be used to reduce or query efficiently the large search space.
Typically, the first reference dataset may further comprise: amino acid sequences of each first reference protein; and a plurality of reference encodings of each true B-cell epitope, wherein each reference encoding represents the respective true B-cell epitope as a plurality of data elements corresponding to the amino acids of the amino acid sequence.
The reference encodings of each true B-cell epitope used for the first reference dataset are typically generated in the same manner as described above.
The first machine learning model may be (or comprise) any neural network such as convolutional neural networks and feedforward neural networks. Preferably, the first machine learning model is (or comprises) a recurrent neural network. Typically, the first machine learning model is (or comprises) a long short-term memory network, LSTM, preferably a bi-directional long short-term memory network, BLSTM. This is particularly advantageous in embodiments in which the trained first machine learning model takes the amino acid sequence of the query protein as input. Typically, each amino acid of the query protein sequence is treated as a time-step in the network, whose characteristics can be influenced by both previous and following amino acids. Thus, the entire protein amino acid sequence (e.g. paired with the other input features) may be processed as a “single data point” in contrast to conventional approaches which require segmentation of the protein amino acid sequence in to smaller sub-sequences of amino acids, for example into “k-mers”. In this way, the present invention advantageously enables the discovery of relationships between spatially distant amino acids of the protein sequence which could, for example, be part of the same (e.g. conformational) BCE, and which relationships may not be discovered through the “k-mer” approach. The use of a BLSTM is particularly advantageous as this allows the protein amino acid sequence to be processed from both directions.
The second machine learning model(s) may be (or comprise) any neural network such as convolutional neural networks and feedforward neural networks.
Typically, the one or more trained second machine learning model(s) is (or comprises) a recurrent neural network, preferably a long short-term memory network, LSTM, more preferably a bi-directional long short-term memory network, BLSTM. This provides the same advantages as outline above.
As discussed above, in embodiments the present invention may advantageously generate encodings of one or more candidate B-cell epitopes to be tested in order to model and “test” each candidate epitope as a whole. Due to this approach, the first reference dataset used to train the first machine learning model may only contain observed true positive BCEs (i.e. only “positive” training data). Therefore, in some preferred embodiments, the method may further comprise generating pseudo-random data based on the plurality of reference encodings of each true B-cell epitope of the first reference dataset, and wherein the first reference dataset comprises said pseudo-random data as negative data. In this way, the performance of the first machine learning model may be improved as the model can be trained even in the absence of verified negative data.
The pseudo-random data is based on the encoded true B-cell epitope(s). Preferably, the pseudo-random data comprise a plurality of permutations of the plurality of data elements representing the true B-cell epitopes. In some embodiments, the pseudo-random data may be random permutations of the data elements. In some embodiments, the pseudo-random data may share at least one characteristic with the encodings of the true B-cell epitope(s). The characteristic may be physical property of the true B-cell epitope, for example a position on the amino acid sequence of one or more amino acids (e.g. the first and last amino acids of an epitope). In another example, the characteristic may be a linear distance between particular amino acids.
In some embodiments in which the inputs to the trained first machine learning model comprise one or more candidate encodings, the method may further comprise: identifying a first B-cell epitope that is predicted to be likely to be a true B-cell epitope on the protein;
substituting one or more amino acids forming the identified first B-cell epitope with a different amino acid to thereby generate a modified amino acid sequence of the protein; and using the trained first machine learning model to predict whether the candidate encoding representing the identified first B-cell epitope is likely to correspond to a true B-cell epitope on the modified amino acid sequence of the protein.
In this way, by applying one or more amino acid substitutions of an identified B- cell epitope, the first machine learning model may be used to predict B-cell epitopes that do not naturally occur on the protein under analysis. This may be advantageous in identifying vaccine candidates that provide protection across possible future viral variants and across a broad range of virus or pathogen species, as described in further detail herein.
In the step of applying a substitution, one or more amino acids of the first B-cell epitope (which is typically a wildtype B-cell epitope) is substituted for any other amino acid that exists. This may be referred to as a “mutation” of the amino acid. The trained first machine learning model may be used to predict the likelihood of each possible amino acid substitution of the identified first B-cell epitope forming a true B-cell epitope on the protein. Further analysis may be performed on the B- cell epitopes that are predicted to be positive on the modified amino acid sequence. For examples, structural stability analysis may be performed using (e.g. computational) techniques known in the art.
Typically, in the case where the candidate encodings are in the form of binary vector, only the amino acids corresponding to the “1”s of the binary vector (e.g. representing direct amino acid contact points on the 3D protein sequence with the antibody) are substituted (“mutated”).
In the step of using the trained first machine learning model to predict whether the candidate encoding representing the identified first B-cell epitope is likely to correspond to a true B-cell epitope on the modified amino acid sequence of the
protein, the input to the model comprises the modified amino acid sequence, the one or more structure and/or surface characteristics of the modified protein, and the candidate encoding representing the identified first B-cell epitope. The structure and/or surface characteristics may be obtained using the techniques described above, for example by applying the second machine learning model(s) to the modified amino acid sequence.
In a further aspect of the invention there is provided method of training a machine learning model, comprising generating a reference dataset, the reference dataset comprising: a plurality of first reference proteins, each first reference protein comprising at least one B-cell epitope classified as a true B-cell epitope; and one or more structure and/or surface characteristics of each first reference protein in an unbound state; and training the machine learning model using the reference dataset to learn a relationship between the structure and/or surface characteristic(s) of the first reference proteins in an unbound state, and the corresponding B-cell epitopes classified as true B-cell epitopes.
Thus, the method of training a machine learning model provides the advantages described above, and may be used in combination with any of the described examples of the invention. Advantageously, the machine learning model, once trained, may be used to predict whether a query protein comprises a B-cell epitope that is likely to instigate a binding event with an antibody.
Typically, the reference dataset further comprises: amino acid sequences of each reference protein; and a plurality of reference encodings of each true B-cell epitope, wherein each reference encoding represents the respective true B-cell epitope as a plurality of data elements corresponding to the amino acids of the amino acid sequence.
The reference dataset may further comprise one or more physiochemical characteristics of each protein, and/or the amino acids that constitute the reference protein, as described above.
The machine learning model may be (or comprise) any neural network such as convolutional neural networks and feedforward neural networks. Preferably, the machine learning model is (or comprises) a recurrent neural network. Typically, the first machine learning model is (or comprises) a long short-term memory network, LSTM, preferably a bi-directional long short-term memory network, BLSTM.
The invention also provides use of a machine learning model trained using the method described above (e.g. to predict the presence of BCEs on a query protein). The invention also provides a machine learning model trained using the method described above.
The invention may further comprise synthesising one or more proteins that are predicted to contain a B-cell epitope, and/or a B-cell epitope predicted using the techniques described herein.
Protein(s) that are predicted by the present invention to comprise a B-cell epitope may be used in the treatment or prevention of infectious diseases or other diseases such as autoimmune or immune-related diseases and cancer. The present invention finds use in one of more of the following use cases:
• Therapeutic antibody mapping
• Vaccine development and creation (both prophylactic and therapeutic)
• Immuno-diagnostics and immune monitoring.
In some embodiments, the method may further comprise encoding one or more proteins that are predicted to contain a B-cell epitope, and/or a predicted B-cell epitope, and/or B-cell epitope variant, and/or B-cell epitope predicted or simulated variant into a corresponding protein, peptide, DNA or RNA sequence. DNA or RNA sequences may be incorporated into a delivery system for use in a vaccine
(e.g. using naked or encapsulated DNA, or encapsulated RNA). The method may comprise incorporating the protein, peptide, DNA or RNA sequence into a genome of a bacterial or viral delivery system to create a vaccine.
Thus, in accordance with a further aspect of the present invention, there is provided a method of creating a vaccine, comprising: predicting that a protein comprises a B-cell epitope by a method according to any of the examples discussed above; and synthesising the protein and/or the predicted B-cell epitope, or encoding at least one of the protein, the predicted B-cell epitope, B-cell epitope variant, B- cell epitope predicted or simulated variant into a corresponding protein, peptide, DNA or RNA sequence.
Such a protein, peptide, DNA or RNA sequence may be delivered in a naked or encapsulated form or incorporated into a genome or cell of a bacterial or viral delivery system to create a vaccine. In addition, bacterial vectors can be used to deliver the DNA in to vaccinated host cells. For peptide vaccines, the identified protein(s) may typically be synthesised as an amino acid sequence or “string”.
Such a vaccine may be a prophylactic or therapeutic vaccine. For example, the method may be used to create a personalised vaccine for an individual, for example a cancer therapeutic vaccine if the protein comprises a neoantigen.
In accordance with a further aspect of the present invention, there is provided a method of creating (e.g. and/or designing) a diagnostic assay to determine whether a patient has or has had a cancer or prior infection with a pathogen, wherein the diagnostic assay is carried out on a biological sample obtained from a subject, comprising identifying at least one protein of the pathogen or tumor that is predicted to comprise a B-cell epitope, using a method according to any of the examples discussed above; wherein the diagnostic assay comprises the utilisation or identification within the biological sample of the identified at least one protein and/or B-cell epitope.
In this way, the present invention may advantageously be used to create a diagnostic test or assay, by means of a rapid and/or automated molecular target discovery. Query proteins identified as being likely to comprise a true B-cell epitope may be further analysed in laboratory testing in order to create such a diagnostic test or assay, thereby significantly reducing the time taken to develop the test compared to traditional laboratory methods.
The term utilisation as used herein is intended to mean that the at least one protein and/or B-cell epitope thereon are used in an assay to identify an (e.g. protective) immune response in a patient. In this context, the identified protein(s) and/or epitope(s) within are not the target of the assay, but a component of said assay.
The in vitro diagnostic assay may comprise identification of an immune system component within the biological sample that recognises at least one epitope within the identified protein. In this way, the diagnostic assay may utilise the at least one identified protein and/or at least one predicted epitope. Typically, the diagnostic assay will contain the (e.g. synthesised) at least one identified protein, protein subunit and/or predicted epitope. In a preferred embodiment, the immune system component may be a B-cell. For example, the assay may comprise identification of antibodies or B-cells that recognise predicted B-cell epitopes within the identified protein.
As an example of such a diagnostic use, a sample, preferably a blood sample, isolated from a patient may be analysed for the presence of B-cells or antibodies within the biological sample that recognise and bind to epitope(s) within the identified protein(s), identified as part of the present invention and that are contained within the assay.
Suitable diagnostic assays would be appreciated by the skilled person, but may include enzyme-linked immune absorbent spot (ELISPOT) assays, enzyme-linked immunosorbent assays (ELISA), cytokine capture assays, intracellular staining assays, tetramer staining assays, microfluidic devices, lab-on-a-chip, microarrays,
flow cytometry, CyTOF, proteomics, molecular sequencing, or limiting dilution culture assays.
In a method of creating a diagnostic test, the amino acid sequence of the one of more proteins to be tested for the presence of BCEs may be chosen based on the desired response to be tested. For example, the one or more source proteins may be one or more source proteins of any pathogen or virus (or fragments thereof), such as the SARS-CoV-2 virus. In such a case, the present invention may be used to create a diagnostic test for determining whether a patient has or has had prior infection with the SARS-CoV-2 virus and/or its variants and/or related viral species. However, as will be appreciated by the skilled person, the one or more source proteins may be from any pathogen (e.g. any virus, parasite, bacterium or cancer indication).
Further disclosed herein is a diagnostic assay to determine whether a patient has or has had prior infection with a pathogen, wherein the diagnostic assay is carried out on a biological sample obtained from a subject, and wherein the diagnostic assay comprises the utilisation or identification within the biological sample of at least one protein of the pathogen or tumor that is predicted to comprise a B-cell epitope, identified using any of the methods discussed above. The diagnostic assay may comprise identification of an immune system component (e.g. a B-cell) within the biological sample that recognises said at least one identified protein and/or at least one predicted epitope.
In accordance with a further aspect of the invention there is provided a computer program product comprising instructions which, when executed by a computer, cause the computer to perform the method of any of the examples descried above. The invention also provides a non-transitory computer readable medium comprising executable instructions that, when executed by a computer, cause the computer to perform the method of any of the examples described above.
In accordance with a further aspect of the invention, there is provided a system for predicting whether a protein comprises a B-cell epitope that is likely to instigate
a binding event with an antibody, comprising at least one processor in communication with at least one memory device, the at least one memory device having stored thereon instructions for causing the at least one processor to perform a method in accordance with any of the examples of the invention discussed above.
In accordance with a further aspect of the invention, there is provided a method of synthesising a protein, comprising: predicting that a protein comprises a B-cell epitope that is likely to instigate a binding event with an antibody according to any of the examples discussed above; and synthesising the predicted protein. The synthesis of the predicted protein can be performed using techniques known in the art. In accordance with a further aspect of the invention, there is provided a protein synthesised using the said method.
Embodiments of the present invention may advantageously be used to identify one or more B-cell epitopes that are predicted to instigate a protective immunogenic response across a plurality of species of a pathogen or virus. Such “broadly protective” B-cell epitopes may be used as vaccine elements or to provide a diagnostic test or assay, for example.
Thus, further disclosed herein is a (e.g. computer-implemented) method of identifying one or more B-cell epitopes that are predicted to instigate a protective immunogenic response across a plurality of species of a pathogen or virus (e.g. and/or variants thereof), comprising:
(i) performing the method according to any of the examples described above which utilise candidate encodings, for each of a plurality of proteins of a respective plurality of different species of a pathogen or virus, to thereby identify a plurality of first B-cell epitopes that are each predicted to be likely to be a true B-cell epitope on at least one of the plurality of proteins;
(ii) for each identified first B-cell epitope, determining the number of different species of the plurality of species of the pathogen or virus on which the first B-cell epitope is predicted to be a true B-cell epitope; and
(iii) classifying one or more of the first B-cell epitopes as a broadly protective B-cell epitope, based on the number of different species, or a number of variants of any one given species, on which the first B-cell epitope is predicted to be a true B-cell epitope.
In this way, the present invention may be utilised to classify one or more B-cell epitopes as being predicted to instigate a broad immunogenic response across a plurality of different species of a pathogen or virus, or variants thereof (e.g. cross- reactive). The classification is based on a determination of the numbers of different species of the pathogen or virus on which the B-cell epitopes are predicted to be true B-cell epitopes (e.g. viable B-cell epitopes). For example, the first B-cell epitope that is predicted to be a viable B-cell epitope on the largest number of different species, or variants thereof, tested may be classified as a broadly protective B-cell epitope (and may, for example, be classified as a vaccine candidate). In another example, an identified first B-cell epitope may be classified as being broadly protective if the number of species on which the B-cell epitope is predicted to be a true B-cell epitope is greater than a predetermined threshold.
Preferably, in order to reduce computational load, the one or more candidate encodings may map only onto the receptor binding domain (RBD) of the amino acid sequence of each protein tested.
In some embodiments, the method may further comprise, for at least one of (preferably each of) the first B-cell epitope(s) classified as a broadly protective B- cell epitope: substituting one or more amino acids forming the first B-cell epitope with a different amino acid to thereby generate a modified first B-cell epitope; and, for each of the plurality of proteins of the respective plurality of species of a pathogen of virus; generating a modified amino acid sequence of the protein in accordance with the modified first B-cell epitope; and using the trained first machine learning model to predict whether the candidate encoding representing the first B-cell epitope classified as
a broadly protective B-cell epitope is likely to correspond to a true B-cell epitope on the modified amino acid sequence of the protein.
In this way, the method of the present invention may advantageously be used to predict viable B-cell epitopes that may not occur naturally on the proteins under analysis. This may advantageously be used to identity broadly protective vaccine candidates that provide a broad protective response across a plurality of different species of a pathogen or virus (and/or variations of these), and which may also be protective against future viral variants or pathogens that have not yet spilled over to be pathogenic in humans. Such identified B-cell epitopes may form vaccine candidates or may be used to create a diagnostic test or assay as outlined above.
The inventors have found that some modified B-cell epitopes identified in this way are predicted to generate a broader immunogenic response across a larger number of different species of a virus or pathogen, compared to the most cross- reactive wildtype B-cell epitopes.
Further analysis may be performed on the identified modified B-cell epitopes that do not naturally occur in the viruses or pathogens to assess their structural stability. For example, any “mutated” B-cell epitopes that are predicted to have low structural stability may be discarded from future analysis. Such structural analysis may be performed using (e.g. computational) techniques known in the art.
BRIEF DESCRIPTION OF THE DRAWINGS
Embodiments of the invention will now be described with reference to the accompanying figures, in which:-
Figure 1 is a flow diagram illustrating the steps of a method according to an embodiment of the invention;
Figure 2 is a flow diagram illustrating the steps of a method of training a first machine learning algorithm according to an embodiment of the invention;
Figure 3 is a flow diagram illustrating the steps of a method of training a second machine learning algorithm according to an embodiment of the invention;
Figure 4 is an example of a system suitable for implementing embodiments of the method is shown;
Figure 5 is an example of a suitable server;
Figures 6(a) to 6(f) show the correlation results of the feature-feature pairs;
Figure 7 shows the Spearman’s correlation coefficients for continuous features against (a) RSA, (b) UHSE and (c) LHSE;
Figure 8 shows the Kendall’s correlation coefficients for categorical features against (a) RSA, (b) UHSE and (c) LHSE;
Figure 9 shows the Kendall’s correlation coefficients for continuous features against (a) SS, (b) CBCE and (c) LBCE;
Figure 10 shows the Mutual Information criterion for categorical features against (a) SS, (b) CBCE and (c) LBCE;
Figure 11 schematically shows the structure of a Prediction model for RSA and HSE according to an embodiment of the invention;
Figures 12(a) to (f) show cross-validation results for a RSA/HSE model according to an embodiment of the invention;
Figures 13(a) to (f) show test results for a RSA/HSE model according to an embodiment of the invention;
Figures 14(a) to (f) show final training results for a RSA/HSE model according to an embodiment of the invention;
Figures 15(a) to (f) illustrate CV results for a Secondary Structure (SS) prediction model according to the invention;
Figures 16(a) and (b) show the test results for SS model;
Figures 17(a) to (d) show final training results for the SS model according to an embodiment of the invention;
Figures 18(a) and 18(b) show Paratope Analysis for CBCE;
Figure 19(a) schematically illustrates the architecture of a first machine learning algorithm for predicting CBCEs, according to an embodiment of the invention;
Figure 19(b) schematically illustrates the architecture of a first machine learning algorithm for predicting CBCEs, according to a further embodiment of the invention;
Figures 20(a) to (d) illustrate outlier analysis for CBCEs;
Figures 21 (a) to (f) shows results for the precision-recall metric for a first machine learning algorithm for predicting CBCEs, according to a further embodiment of the invention;
Figures 22(a) to (f) show results for the precision-recall metric for a first machine learning algorithm for predicting CBCEs, according to an embodiment of the invention;
Figures 23(a) and (b) show the results for the precision-recall metric for a first machine learning algorithm for predicting CBCEs, according to an embodiment of the invention;
Figures 24(a) and (b) shows the results for the precision-recall metric for a first machine learning algorithm for predicting CBCEs, according to an embodiment of the invention;
Figures 25(a) to (c) shows the results of final training for a first machine learning algorithm for predicting CBCEs, according to an embodiment of the invention;
Figures 26(a) and (b) show the results of final training for a first machine learning algorithm for predicting CBCEs, according to an embodiment of the invention;
Figures 27(a) to (d) illustrates outlier analysis for LBCEs;
Figure 28 schematically illustrates a pipeline according to an embodiment of the invention;
Figures 29(a) to (c) illustrates the successful mapping of receptor binding domains (RBDs) to betacoronavirus species;
Figure 30 illustrates the prediction of structural epitopes predicted to be broadly immunogenic across a plurality of viral species; and
Figures 31(a) and 31(b) illustrate the change in predicted immunogenicity following the implementation of mutation policies.
DETAILED DESCRIPTION
1 . Overview
Embodiments of the invention provide a B-cell epitope (BCE) prediction algorithm (“first machine learning algorithm” or “first machine learning model”), trained using a reference dataset that is designed such that the trained first machine learning model predicts B-cell epitopes in their unbound state. The first machine learning
algorithm is trained on unbound 3D Ag protein structures prior to the Ag-Ab binding event, to learn the properties of BCEs on the Ag prior to the Ab binding event, in order to capture the bone fide (“true”) BCEs on a query Ag protein that is input into the model. In other words, the BCE prediction algorithm predicts BCEs in the form that an antibody “sees” in order to instigate a binding event.
In embodiments, the first machine learning model accurately predicts BCEs on a given Ag from its primary sequence in combination with one or more 3D structure and/or surface characteristics of the Ag, with no requirement for the full 3D structure of the Ag as an input into the predictor. The model is trained using BLSTMs, which allows for the processing of the whole Ag sequence as one data- point, processed from both directions without the need for k-mer segmentation. Each amino-acid of the Ag is treated as a time-step in the network, whose characteristics can be affected by both preceding and succeeding amino-acids that flank the amino-acid in question, thus, allowing for the discovery of the contextual relationships between spatially distant amino-acids which could be important parts of conformational BCEs (CBCEs).
The first machine learning model is capable of discovering separate or independent CBCEs that may exist on a single Ag protein sequence. In particularly advantageous embodiments, the trained model is designed to receive as input an Ag protein sequence and a binary “permutation vector” (“encoding”), representing each single candidate CBCE, as well as the one or more 3D structure and/or surface characteristics of the Ag. The output of the model is a single probability indicating the likelihood that the input permutation vector is a bone fide CBCE. In other words, the user asks the question to the model if the specific query CBCE, encoded as a binary permutation vector on the given Ag sequence, is a true CBCE or not. However, the model may also be used to provide a second output in the form of a probability vector with a probability for each amino-acid in the Ag sequence. The second output can be seen as the contribution of each amino acid to the CBCE in question, and may be used to compare the current approach with conventional models that predict on an “amino acid by amino acid” basis.
Structural characteristics of the Ag sequence are important for predicting CBCEs. However, embodiments of the present invention negate the need for the complete 3D structure of the protein sequence to be experimentally measured or predicted, and then used as input into the algorithm. The first machine learning algorithm does not require the coordinates of each atom in each amino acid in the 3D protein sequence, and instead may predict BCEs using one or more structure and/or surface characteristics of the query protein, such as RSA, HSE and SS.
The “top-level” most important features such as RSA, HSE and SS, may be predicted with one or more trained machine learning algorithms (“second machine learning algorithm” or “second machine learning model”), typically receiving as input the amino acid sequence of the protein. In this way, structure and/or surface characteristics of the protein under investigation may advantageously be computed directly from the amino acid sequence, with no requirement for extra information to be sourced or input by the user. Those second algorithms use BLSTM networks also and are shown to have a high performance.
In this way, the first and second machine learning models may together form a “pipeline” which requires as input only the amino acid sequence of a query protein in order to predict whether that query protein comprises a true B-cell epitope. A schematic diagram of such a pipeline constituting the first and second machine learning models according to an embodiment of the invention is shown in Figure 28.
Figure 1 is a flowchart 100 summarising the steps of a preferred embodiment of the present invention. At step S101 , the amino acid sequence of a protein to be analysed for the presence of B-cell epitopes is accessed. Typically the protein under investigation will comprise an antigen. Various techniques known to the skilled person may be used to access the amino acid sequence, for example the sequence may be downloaded from a bioinformatic depository such as UniProt.
At step S103, one or more (e.g. three-dimensional) structure and/or surface characteristics of the protein in an unbound state are accessed. Typically, these characteristics of the protein are predicted using one or more trained second
machine learning model(s) (schematically illustrated at 300, and described below) that receives the amino acid sequence accessed in step S101 as input. The structure and/or surface characteristics typically include the secondary structure of the query protein and relative solvent accessibility. The half sphere exposure (both upper and lower half sphere exposure) may also be used. Values or categories for each of these characteristics are typically assigned to each amino acid of the sequence. However, other techniques for obtaining the three- dimensional structure and surface characteristics of the query protein may be employed, such as database entries, X-ray crystallography or computational approaches to predicting full 3D protein folding structures (from which the structure and/or surface characteristics can be obtained).
At step S105, candidate encodings of one or more candidate BCEs on the query protein are generated. The candidate encodings are each typically in the form of a binary vector, with each data element in the binary vector corresponding to an amino acid of the amino acid sequence forming the respective candidate BCE. In such a binary “permutation vector” encoding, each amino acid of the sequence forming part of the candidate BCE is assigned a “1” and each amino acid of the sequence not forming part of the candidate BCE is assigned a “0”. For example, for the exemplary amino acid sequence “ABCDEFG”, a candidate conformational B-cell epitope constituted by the amino acids B, D, F and G may be encoded as the binary vector (0101011). In some embodiments, a candidate encoding for each possible candidate BCE on the protein may be generated. However, in order to reduce computational load, the search space may be limited based on physical priors (e.g. knowledge of a mutation point on a neoantigen of the input protein or receptor binding domain on the spike protein of SARS-CoV-2 etc.) or on techniques such as deep reinforcement learning or related generative models.
In step S107, the amino acid sequence accessed in step 101 , the one or more structure and/or surface characteristics accessed in step 103, and the candidate encodings generated in step S105, are input into a trained first machine learning model, which has been trained using BLSTMs. One or more physiochemical characteristics assigned to each amino acid of the amino acid sequence may also
be used as input to the trained first machine learning model. The output of the trained first machine learning model is a probability that each of the candidate encoding(s) represents a true B-cell epitope on the query protein.
Importantly, the first machine learning model is trained on unbound data. In other words, the first machine learning model is trained on a dataset that comprises reference proteins having B-cell epitopes classified as true B-cell epitopes, and structural and/or surface characteristics of the reference proteins in an unbound state. In this way, the first machine learning model is trained to learn a relationship between the structural and/or surface characteristic(s) of the reference proteins and the classified true B-cell epitopes. Consequently, the trained first machine learning model predicts the presence of B-cell epitopes on the query protein prior to a binding event.
Figure 2 is a flow diagram 200 summarising the steps of a preferred embodiment of training the first machine learning model. At step S201 , a plurality of protein complexes are accessed. These may be accessed from publicly available sources such as the Protein Databank (PDB) for CBCEs or Immune Epitope Database (IEDB) for LBCEs.
At step S203, the protein complexes accessed in step S201 are analysed to define true epitopes and antigens. Linear BCEs may be defined in the complexes of the IEDB. Further analysis may be required to define true conformational epitopes from protein complexes accessed from the PDB, and further details on these techniques are described herein.
At step S205, each antigen defined in step S203 as having a true B-cell epitope is mapped onto its amino acid sequence. At step S207, each true epitope is encoded onto the amino acid sequence of the corresponding antigen. In this way, a plurality of reference epitope encodings are generated, corresponding to the true epitopes. The reference encodings are typically encoded as binary vectors in the same manner as described above in step S105.
At step S209 of the method, three-dimensional structure and/or surface characteristics of the antigens in an unbound state are accessed, for example by prediction from the amino acid sequence of the antigen obtained in step S205. Typically, the structure and/or surface characteristics of the antigens are predicted using the trained second machine learning model(s) (schematically illustrated at 300). However, other techniques of accessing the structure and/or surface characteristics of the antigens in an unbound state may be used, for example known computational approaches for predicting 3D protein structures. Typically, the characteristics include the secondary structure, half sphere exposure (both UHSE and LHSE), and relative solvent accessibility.
At step S211 , a first reference dataset is generated. The first reference dataset comprises the amino acid sequences of the reference antigens (step S205), the encodings of the true epitopes (step S207) on the corresponding reference antigens, and the accessed (e.g. predicted) structure and/or surface characteristics of the reference antigens in an unbound state. In some embodiments the first reference dataset may further comprise one or more physiochemical characteristics assigned to each amino acid of the reference antigens.
At step S213, the first machine learning model is trained using the first reference dataset generated in step S211. Here, the first machine learning model is or comprises a BLSTM, and may be trained using techniques known in the art. By training the model using the first reference dataset generated through performing the steps S201 to S211 , the first machine learning model is trained to learn a relationship the structure and/or surface characteristic(s) of the first reference proteins in an unbound state, and the B-cell epitopes classified as true B-cell epitopes.
As discussed above with reference to Figures 1 and 2, in preferred embodiments the second machine learning model(s) (schematically shown at 300) is used to predict structure and/or surface characteristics both when training the first machine learning model (Figure 2), and when using the trained first machine
learning model to predict the presence of B-cell epitopes on a protein of interest (Figure 1). The second machine learning model is trained to predict structure and/or surface characteristics of a query protein in an unbound state. In this way, the first machine learning model is configured to predict BCEs in their unbound state, i.e. , having the three dimensional form and characteristics that the Ab “sees” and initiates binding upon.
Figure 3 is a flow diagram summarising the steps of a preferred embodiment of training a second machine learning model according to an embodiment of the invention.
At step S301 , a plurality of second reference proteins in an unbound state (i.e., single protein structures) are accessed. These second reference proteins may be accessed from publicly available databases such as the PDB.
At step S303, the amino acid sequences of the second reference proteins are accessed.
At step S305, the three-dimensional structure and/or surface characteristics of the second reference proteins are accessed. Typically these include the secondary structure, the relative solvent accessibility, and the upper and lower half sphere exposures. The DSSP algorithm (Wolfgang Kabsch and Christian Sander. Dictionary of protein secondary structure: Pattern recognition of hydrogen-bonded and geometrical features. Biopolymers, 22(12):2577-2637, 2018/11/30 December 1983) was used to compute the secondary structure and RSA, and the BioPython package (Peter J A Cock, TiagoAntao, Jeffrey T Chang, Brad A Chapman, Cymon J Cox, Andrew Dalke, Iddo Friedberg, Thomas Hamelryck, Frank Kauff, Bartek Wilczynski, et al. Biopython: freely available python tools for computational molecular biology and bioinformatics. Bioinformatics, 25(11):1422-1423, 2009) utilised to compute the UHSE and LHSE. The DSSP and BioPython algorithms may be applied to the structure files of the second reference proteins (e.g. obtained from the Protein Data Bank, PDB).
At step S307, a second reference dataset is generated using the amino acid sequences accessed in step S303, and the structure and/or surface characteristics of the second reference proteins accessed in step S305. In some embodiments the second reference dataset may further comprise one or more physiochemical characteristics assigned to each amino acid of the second reference proteins.
At step S309, the second machine learning model is trained using the second reference dataset (e.g. to learn a relationship between one or more features of the amino acid sequences of the second reference proteins and the corresponding structure and/or surface characteristics). Here, the second machine learning model is or comprises a BLSTM, and may be trained using techniques known in the art. It will be appreciated that different second machine learning models may be trained to predict different respective structure and/or surface characteristics. For example, one second machine learning model may be trained to predict RSA and HSE metrics, and further second machine learning model trained to predict the secondary structure.
The various steps outlined in Figures 1-3 will be described in further detail in the following sections. Section 2 explains pre-training techniques used for all the models. Section 3 presents some preliminary analyses we performed to choose input features for the models. Section 4 presents the structural algorithms for RSA, HSE and SS used in the current BCE algorithms. Section 5 presents the structural algorithms for RSA, HSE and SS that are used as features in the BCE model. Section 6 presents an example of a first machine learning algorithm according to the invention, used to predict CBCEs, while Section 7 presents an outline of a first model used to predict linear BCEs (LBCEs). In section 8 we present applications of the present invention.
2 Pre-training Techniques
To ensure reproducible results and avoid the need of random seed in the networks used in our models, we used auto-encoders (AEs). AEs are types of NNs trained on compressed data which mirrors the inputs of the main model. More specifically,
an AE takes an input at layer number / = 0 and processes it through an arbitrary number of layers, say / = 1,
which constitute the encoder part. It then processes it back through a mirrored structure of the encoder part, called the decoder part. Finally, it returns the output which is the same as the input. The intermediate layers may reduce or expand the dimensions of the input.
Each layer in the main network is converted into an AE, and trained on some compressed amount data. The pre-trained weights of the outer layers are used as constants to the subsequent AEs. This is repeated until all the main network layers are pre-trained. The pre-trained weights can then be used as starting points for the main training procedure.
The following procedure were performed once and the pre-trained weights were used on all the cross-validation folds, test of unseen data, and final model.
First, we randomly generated 10000 protein sequences of length between 50 to 1000 amino-acids. The actual amino-acids used was random as well. We then pre-trained each layer of the neural network as an individual AE on the whole generated data. The pre-training was done for maximum 1000 epochs for each AE, with the possibility of early stopping if the loss did not decrease after 10 epochs. The weights of the outer layers were copied to deeper AEs after pretraining, and were also kept constant during the pre-training of those. The last layer of the model (output) was not pre-trained at all.
This procedure was done once and the same pre-trained weights were used on all the validation procedures. This procedure was faster, it used much more data to pre-train and the same initial weights were used for all downstream analyses.
3 Preliminary Analysis and Feature Selection Techniques
We ran various preliminary analyses in order to see how different features correlate with the targets we wanted to predict, as well as with each other. The goal of those analyses was two-fold. First, we tried to find highly feature-targets correlations which would help us decide which features to use in order to predict the targets. Secondly, feature-feature correlations would help us subset the
feature-target correlation space by keeping features which were not correlated with each other, thus, reducing the input space of the models. Note, however, that all the analysed using this selection technique mostly capture linear-relationships between feature-targets and feature-features.
3.1 First Method
For deciding which features will be used as input to predict different targets of the models we performed a correlation analysis using the features in Table 0. The analysis was done using the whole data for those models. Namely, the data-set used for those analyses was the ones described in section 4.1 . This first method of analysis was run for the 3D structure prediction models and not for the CBCE or LBCE models.
3.1.1 Pre-processing continuous features
Continuous features provide a single value per amino-acid in the alphabet. To model the effect of neighbour amino-acids in a protein sequence, we averaged the values of each feature by scanning the protein sequence using a sliding window of a peptide length of 9 amino-acids. The average feature value was assigned to the middle amino-acid in the peptide. This was done for every feature on each protein sequence in the entire training and test datasets.
3.1.2 Correlation between continuous features and continuous targets
These analyses were done for RSA (relative solvent accessibility), UHSE (upper half-sphere exposure) and LHSE (lower half-sphere exposure). First, we pre- processed the continuous features as described in section 3.1.1. The values of the target variables were used as they were, that is not averaged, since their value already represents neighbours. The Spearman’s correlation coefficient was computed for every window-averaged feature in each protein sequence and every target variable. Correlations from all the proteins were then averaged per featuretarget pair. For all the three targets we then chose the 30 most correlated features (positively or negatively), and we chose features that were highly correlated with all the three targets.
3.1.3 Correlation between categorical features and continuous targets
These analyses were done for RSA, UHSE and LHSE. Categorical features provide a single value per amino-acid. Those features were used as they are, per- amino-acid in any given seguence. The point-biserial correlation coefficient between feature and target was computed per class. The targets were also used as they are, that is, no windows, for the same reason as in section 3.1.2. Correlations from all the proteins were then averaged per feature-target pair. The top highest (positively or negatively) correlated features with all the three targets were chosen as inputs in the model.
3.1.4 Correlation between continuous features and categorical targets
This analysis was done for SS. The same procedure was used as in Section 3.1.2 for the windows of the features. The only difference is that the target is categorical. For assessing the correlation with continuous features, we used the point-biserial correlation coefficient. Correlations from all the proteins were then averaged per feature-target pair. Top correlated features (positively or negatively) were used in the models.
3.1.5 Correlation between categorical features and categorical targets
This analysis was done for SS. For assessing the dependency between categorical features-targets we used the Chi-Sguare on each protein. We then did multiple hypothesis correction of the resulting p-values using the Benjamini- Hochberg procedure. Finally, we averaged their corrected p-values from all the seguences, for each feature-response pair, using the harmonic mean. The top most significant feature-target correlations were used in the models.
3.2 Second Method
For deciding which features will be used as input to predict different targets of the models we did a correlation analysis using the features in Table 0. Further information on each feature (e.g. characteristic) is given in the Appendix. The analysis was done using the whole data for those models. Namely, the data-set
used for those analyses was the ones described in section 4.1 (for RSA, UHSE, LHSE, SS) and section 5.1 and 6.1 (for CBCE and LBCE). Which data we used depends on the model tested, accordingly. Moreover, for the CBCE and LBCE analyses, those two data-sets were merged.
Table 0: Table of Global Features
3.2.1 Pre-processing continuous features
The same procedure as in section 3.1.1.
3.2.2 Feature-Feature correlations First, we ran a feature-feature correlation analysis of all the features in Table 0. We did this on two different data-sets, one for the one in section 5.1 and one for the merged one from sections 6.1 and 7.1. The purpose of those analysis was to identify high correlated feature pairs, such that if both features from a correlated pair were correlated with a response in a later analysis, only one of them will be used as input feature for the corresponding model. Thus, reducing the input features space.
We first pre-processed the continuous features as described in section 3.2.1. Categorical features were used as they are, that is, a single value per-amino-acid in every sequence. We then computed correlation metrics for every feature pair on every sequence. For continuous-continuous feature pairs we used Spearman’s
Correlation coefficient . For continuous-categorical features we used Kendall rank
correlation coefficient , and for categorical-categorical we used Mutual information criterion. For every pair-wise feature coefficient population we removed outliers using the IQR (box-plot) method. The median of the whole population is then taken as the final coefficient for every pair-wise feature-feature correlation. We defined as correlated continuous-continuous feature pairs with median Spearman’s correlation coefficient > 0.7, continuous-categorical feature pairs with Kendall rank correlation coefficient > 0.7, and categorical-categorical feature pairs with Mutual Information > 2.
Figure 6 shows the correlation results of the feature-feature pairs for the RSA/HSE model. Figures 6(a) to (c) show feature-feature correlations using the data from 6.1 and 7.1. Figures 6(d) to (f) show feature-feature correlations using the data from 5.1. Black dots correspond to correlated pairs. The heat-maps are symmetric. Note that not all feature names are written in the x- and y-axis. For the 6.1 and 7.1 datasets, the features for RSA, UHSE, LHSE and SS were predicted using the current models (see 4). For the 5.1 dataset, those figures were the observed ones, from PDB. As we can see, for all datasets, continuous- continuous feature-pairs resulted in many correlated features. For continuous- categorical feature-pairs only a few were correlated, while no correlation was observed between categorical-categorical feature pairs.
3.2.3 Feature-response correlations
For each of the following feature-response correlation analyses, we followed the same procedure. For continuous features we did the same pre-processing as described in section 3.1 .1 , but we used windows of 0, 3, 7, 9 and 11 . We computed the correlations based on every window for each feature-response pair in every sequence of the corresponding data. Moreover, we removed outliers using the IQR (box-plot) method, and then computed the median of each correlation population.
We merged all the correlations from every window for each feature into one population, we also removed outliers with the same method, and finally we recomputed the median of that global population. The final “pooled” median was
used to rank the features from highest to lowest correlated with the corresponding response. Categorical responses as well as predicted RSA, UHSE and LHSE were used as they are, that is, no windows.
Each of the following analyses gave us an indication of highly correlated featureresponse pairs. We chose the top correlated features with the responses in such a way that if two features were correlated with the response and also with each other (using the results from section 3.2.3), only the one feature with the highest correlation with the response was used.
3.2.3.1 Correlation between continuous features and continuous targets
These analyses were done for RSA, UHSE and LHSE. We used data from 5.1. The metric used here was Spearman’s Correlation coefficient. See Figure 7 for the results, and section 5 for the models and the features they used.
Figure 7 shows the Spearman’s correlation coefficients for continuous features against (a) RSA, (b) UHSE and (c) LHSE. On each figure, the x-axis is the correlation, the y-axis are the feature names, sorted from highest to lowest absolute correlation of the median of the population. Each plot is from all windows.
3.2.3.2 Correlation between categorical features and continuous targets
These analyses were done for RSA, UHSE and LHSE. We used data from 5.1. The metric used here was Kendall rank correlation coefficient. See Figure 8 for the results, and section 5 for the models and the features they used.
Figure 8 shows the Kendall’s correlation coefficients for categorical features against (a) RSA, (b) UHSE and (c) LHSE. On each figure, the x-axis is the correlation, the y-axis are the feature names, sorted from highest to lowest absolute correlation of the median of the population. Each plot is from all windows.
3.2.3.3 Correlation between continuous features and categorical targets
These analyses were done for SS, CBCE and LBCE. We used data from 5.1 (for SS) and merged 6.1 and 7.1 (for CBCE and LBCE). The metric used here was Kendall rank correlation coefficient . See Figure 9 for the results, and section 5, 6 and 7 for the models and the features they used.
Figure 9 shows the Kendall’s correlation coefficients for continuous features against (a) SS, (b) CBCE and (c) LBCE. On each figure, the x-axis is the correlation, the y-axis are the feature names, sorted from highest to lowest absolute correlation of the median of the population. Each plot is from all windows.
3.2.3.4 Correlation between categorical features and categorical targets
These analyses were done for SS, CBCE and LBCE. We used data from 5.1 (for SS) and merged 6.1 and 7.1 (for CBCE and LBCE). The metric used here was Mutual information criterion. See Figure 10 for the results, and section 5, 6 and 7 for the models and the features they used.
Figure 10 shows the Mutual Information criterion for categorical features against (a) SS, (b) CBCE and (c) LBCE. On each figure, the x-axis is the correlation, the y-axis are the feature names, sorted from highest to lowest absolute correlation of the median of the population. Each plot is from all windows.
4 Predicting 3D features (“Second machine learning algorithm”)
We created two BLSTM models to predict structure and surface characteristics of proteins from their native seguences. Those models are later used in the first machine learning model, to predict structural and surface features of the Ag before the binding event with the Ab takes place. For structural characteristics we chose to create a model for the SS, and for surface characteristics we chose RSA, UHSE and LHSE.
4.1 Data preparation for the 3D features
To accurately predict the SS, RSA, UHSE and LHSE from a native protein seguence, we used all available high guality 3D protein structures from the PDB
from all available organisms. The goal of these models was to predict the most pertinent surface and structural characteristics of single proteins, that are not affected by binding to any other protein interaction, including Ab binding. Therefore, we only kept structures that are not bound by any other structure or molecule. Structures containing more than one copy of the same molecule, but slightly different conformation, are kept in the data. We filtered the structures and kept only those with < 3 Angstrom resolution, ensuring that every atom of each amino-acid is mapped with coordinates, and with protein chains longer than 200 amino-acids.
After filtering, the data-base consisted of 41592 total structures. Those structures contained 70489 protein sequences which resulted in 53524 unique sequences. Subsequences of longer sequences were kept as different data point entries. The reason being that their 3D conformational folding properties may be different because of their shorter length.
The DSSP algorithm was used to compute the SS and the RSAfor each molecule in each structure file. DSSP computes the following secondary structure classes for a protein sequence; a-helix, 310-helix, K-helix, isolated p-bridge, p-strand, turn, bend and coil. We merged those classes into three super-classes; Helices (a-helix, 310-helix and K-helix), Strands (isolated p-bridge and p-strand) and Coils (turn, bend and coil). Finally, the BioPython package was used in to compute the UHSE and LHSE.
At the end of the filtering, each amino-acid in the data base was assigned a value for the RSA, UHSE, LHSE and a class for the SS. To create a unique database, the mean per amino-acid was taken for RSA, UHSE, LHSE among identical sequences. Finally, amino-acids of identical sequences but with different SS classes, were assigned to the coil class.
4.2 Prediction model for RSA and HSE (struct 3d rsa hse m2 model)
We chose to create a single BLSTM model for predicting all surface features. More specifically, the model predicts RSA, UHSE and LHSE from a primary protein
sequence. Although LHSE may not give useful information about the surface position of an amino-acid, it may help predicting UHSE more accurately, since both together are forming a probability metric around each amino-acid.
This is a BLSTM model which takes as inputs a batch of features, each computed per amino-acid from each input sequence and predicts a three-way output. For each protein sequence given as input, a value for each RSA, UHSE and LHSE is predicted per amino-acid.
For all the three outputs the individual losses were calculated on the basis of the mean square error (MSE), but in principle the loss is not limited to this function. The global model loss was the weighted sum of the individual losses with weights: 50, 100 and 125 for RSA, UHSE and LHSE, respectively. The weights were decided by first training the model without them and then see the differences in the magnitude of the three losses. The weights contributed in such a way that the three losses gave the same contribution to the global model loss.
Figure 11 shows a schematic simplified architecture of the model. As shown, the first layer is the input layer. The input layer may comprise one or more input nodes, for example to receive the input features discussed below. Each input node is coupled to a respective masking node in a masking layer. Similarly, each masking node is coupled to a respective BLSTM node in a BLSTM layer. Each BLSTM node is coupled to a concatenation node in a concatenation layer. Thus, a plurality of outputs from the nodes of the preceding BLSTM layer are concentrated together in the concatenation layer. The concatenation layer can be coupled to one or more further BLSTM layers, each comprising a corresponding BLSTM node. The final of these BLSTM nodes is coupled to an output layer, which in this case comprises a plurality of output nodes corresponding, respectively, to the RSA, UHSE and LHSE output predictions.
Note: the input features were computed per amino-acid as they are, not averaged by windows.
4.2.1 Preliminary Analysis and Features Used
See procedure in section 3.1 for the procedure. The 14 features that we chose to use in the model (including the protein-sequence in binary representation) can be seen in Table 1.
Table 1 : Features used in the struct 3d rsa hse m2 model. 4.2.2 Pre-training
See section 2.1.
4.2.3 Cross Validation (CV) results
We used- 80%of the data on 5-fold cross-validation to assess the performance of the model. See Figure 12 for the Spearman’s correlation coefficient metric. The models performed well, without overfitting and had high performance for all the evaluation metrics. Moreover, the model demonstrated high stability, with only a small variation observed between different CV runs.
Figure 12 shows the results for RSA/HSE model. Figures 12(a)-(c) show the Spearman’s correlation coefficients, and 12(d)-(f) the Mean Square Error (MSE) losses. The x-axis is the epochs, the y-axis is the metric values. The plots are the average of 5-fold CVs including Gaussian Cl. The blue line is for training and the red (dashed) for validation sets.
4.2.4 Test results
We used the remaining ~ 20% of the data as test data to assess the performance of the model on completely unseen data. See Figure 13 for the Spearman’s correlation coefficient metric. The models performed well, without overfitting and had high performance for all evaluation metrics.
Figure 13 shows the results for RSA/HSE model, (a)-(c): Spearman’s correlation coefficients, (d)-(f): MSE losses. The x-axis is the epochs, the y-axis is the metric values. The solid line is for the training and the dotted for the validation set.
4.2.5 Final Training
The final training was done on the entire dataset. Using the optimal number of epochs determined by the cross-validation procedure. See Figure 14 for the Spearman’s correlation coefficient metric. The models seem to perform well for all metrics. The model for trained for 70 epochs, chosen based on the CV and testset analysis.
Figure 14 shows the results for RSA/HSE model, (a)-(c): Spearman’s correlation coefficients (d)-(f): MSE losses. The x-axis is the epochs, the y-axis is the metric values.
4.3 Prediction model for SS (struct 3d ss m3 model)
We chose to create a BLSTM model for predicting a three-tier class output for SS. The model uses the categorical cross-entropy loss to assign one of the following classes to each amino-acid in its input sequence; Helices, Strands and Coils. The architecture of the model is substantially the same as that illustrated in Figure 11 , with a single output node in the output layer. Note: the input features were computed per amino-acid as they are not averaged over sliding windows.
4.3.1 Preliminary Analysis and Features Used
See procedure in section 3.1. The 8 features that we chose to use in the model (including the protein-sequence in binary representation) can be seen in Table 2.
Table 2: Features used in the struct 3d ss m3 model.
4.3.2 CV results
We used ~ 80%of the data on 5-fold cross-validation to assess the performance of the model. See Figure 15 for the precision recall metric. The model seems to perform well, without significant overfit and gives hight performance for all metrics.
Moreover, small variation is observed between different CV runs for the loss. Finally, for demonstration purposes, the models were trained for 70 epochs with slight overfit after 60.
Figure 15 illustrates the results for SS model. Figures 15(a)-(e) show PR curves. The x-axis is the TPR, the y-axis is the PPV. Figure 15(f) shows categorical crossentropy loss. The x-axis is the epochs, the y-axis is the loss values. The plot is the average of 5-fold CVs including Gaussian Cl. The blue line is for training and the red (dashed) for validation sets.
4.3.3 Test results We used the rest ~ 20% of the data as test data to assess the performance on completely unseen data. See Figure 16 for the precision-recall metric. The model performed well, without significant overfitting and had high performance for all evaluation metrics. The model was trained for 100 epochs, chosen based on the CV analysis, so slightly more epochs were required than the CV analysis.
Figure 16 shows the results for SS model. Figure 16(a): PR curves. The x-axis is the TPR, the y-axis is the PPV. Figure 16(b): Categorical cross-entropy loss. The x-axis is the epochs, the y-axis is the loss values. The solid line is for the training and the dotted for the validation set.
4.3.4 Final Training
The final training was done on the whole data. Using the optimal number of epochs determined by the cross-validation procedure. See Figure 17 for the F1 metric. The final trained model performed well, without significant overfitting and gave high performance for all metrics. The model for trained for 100 epochs, chosen based on the CV analysis, so slightly more epochs than the CV and test-set analysis.
Figure 17 shows results for SS model. Figure 17(a)-(c): F1 metrics. The x-axis is the epochs, the y-axis is the F1 value. Figure 17(d): Categorical cross-entropy loss. The x-axis is the epochs, the y-axis is the loss values.
5 Predicting 3D features (“Second Machine Learning Algorithm”) (Version 2)
5.1 Data preparation for the 3D features (Version 2)
Same as Section 4.1. The only difference is that for the SS class there was a misplacement of one of the classes. Namely, the isolated p-bridge class was given to the Coil super-class, while here it is moved correctly to the Strand super-class. This small detail only affects some parts of the SS models data.
5.2 Prediction model for RSA and HSE (Version 2) (rsa hse m4 v1 model)
Same as section 4.2, only the preliminary analysis has changed and the input features. The model is deeper as well. The architecture of the model is substantially the same as that illustrated in Figure 11 . Note: the input features were computed per amino-acid as they are, not averaged over sliding windows.
The global model loss was the weighted sum of the individual ones as in the previous model. However, the weights this time were: 1 , 10 and 125 for RSA, UHSE and LHSE, respectively. The weights were decided by first training the model without them, and see the differences in the magnitude of the three losses. 5.2.1 Preliminary Analysis and Features Used
See section 3.2 for the procedure. The 19 features that we chose to use in the model (including the protein-sequence in binary representation) can be seen in Table 3.
Table 3: Features used in the rsa hse m4 v1 model. 5.2.2 Pre-training
See section 2.2.
5.2.3 CV results
We used ~ 80% of the data on 5-fold cross-validation to assess the performance of the model.
5.2.5 Test results
We used the rest ~ 20% of the data as test data to see the performance on completely unseen data.
5.2.6 Final Training
The final training was done on the whole data. Using the optimal number of epochs determined by the cross-validation procedure.
5.3 Prediction model for SS (Version 2) (ss3 m6 v1 model)
Similar to the results described section 4.3. The architecture of the model is substantially the same as that illustrated in Figure 11. Note: the input features were computed per amino-acid as they are, not averaged by windows.
5.3.1 Outliers
Same as section 4.3.
5.3.2 Preliminary Analysis and Features Used
See section 3.2 for the procedure. The 19 features that we chose to use in the model (including the protein-sequence in binary representation) can be seen in table 4.
Table 4: Features used in the ss3 m6 v1 model.
5.3.3 Pre-training
See section 2.2.
5.3.4 CV results We used ~ 80% of the data on 5-fold cross-validation to assess the performance of the model.
5.3.5 Test results
We used the rest ~ 20% of the data as test data to see the performance on completely unseen data. 5.3.6 Final Training
The final training was done on the whole data. Using the optimal number of epochs determined by the cross-validation procedure.
6 First Machine Learning Model (CBCEs)
For predicting CBCEs we used BLSTMs based models. This allowed us to model the whole protein seguences bidirectionally as one observation, without the need
of segmenting it. Therefore, distant, and contextual amino-acid relationships are captured by the model. Moreover, the use of BLSTMs allowed us to train different protein lengths simultaneously. The main goal of our models was to create nonAb specific CBCE predictors. This was achieved by the manner we prepared the training data described in section 6.1.
6.1 Data preparation
For modelling CBCEs we downloaded non-obsolete protein complexes from the PDB. We allowed structures of any resolution and organisms, as long as they had at least three different protein chains. The reason for this is that two of the chains may be the VH and V chains of an Ab, and the third chain may be an Ag. Of course, there may be multiple Abs orAgs in one PDB structure.
We created a local database using all immunoglobulin V-, D- and J-region genes from the IMGT (The international ImMunoGeneTics information system) database. Those genes were downloaded for both VH and V chains, from all the available organisms, that is; human, mouse, rhesus monkey, rabbit and rat. The IMGT provides information about the VH and V chains of an Ab, and its paratope regions (the CDRs [Complementary determining region of the antibody] and FRs [framework region of antibody] of each chain).
To identify the Abs in this local database we used IgBlast (Jian Ye, Ning Ma, Thomas L Madden, and James M Ostell. Igblast: an immunoglobulin variable domain sequence analysis tool. Nucleic Acids Res, 41 (Web Server issue) W34~- 40, Jul 2013) for protein sequences on the IMGT database. For identifying the paratope in a chain we used the Kabat system that the IgBlast provides (G Johnson and T T Wu. Kabat database and its applications: 30 years after the first variability plot. Nucleic acids research, 28(1):214-218, 01 2000). We considered VH and V chains as valid, only if at least their CDR1 and CDR2 were found by IgBlast. Since CDR3 is more difficult to map, we allowed it to be missing in the search. We then aligned all the protein chains of each structure to that database. Structures were filtered out if they did not have at least one chain mapped as a valid VH and one as a valid V . Any other chain that was not mapped as VH or V
was assumed to be an Ag chain. We considered as valid Ag chains, those chains that were longer than 100 amino-acids and were not bound by any other chain other than a VH or V chain. Finally, only structures including at least one VH, V and Ag chains were taken to further analyses.
We then paired VH and V chains in each structure to recreate valid Abs. In case there were multiple VH and V chains in one single structure, we measured the mean distance from every atom on each VH to every atom on each V chain. VH and V chains with the minimum mean atom distance were assigned as pairs and assumed to belong to one single Ab. Stand-alone VH or V chains were not considered on the downstream analysis as they could not define a complete Ab. A paratope analysis provided information about the contacts formed with each Ag. Two amino-acids were assumed to be in contact if any of their atoms were located within a certain probe distance from each other. We computed the total number of Ag amino-acids that form contacts with each paratope’s parts, using probe distances of 4, 6 and 8 Angstroms. Figure 18(a) shows the normalised mean of those contacts (x-axis shows the amino acids, y-axis shows the paratope parts). Most of the contacts are made with CDR1 and CDR2 of the VHs and CDR1 and CDR3 of the V s. The absence of CDR3 on VHs may be due to the difficulty of identifying it using IgBlast. Moreover, the FR regions do not seem to form as many contacts as the CDR regions, as expected. Additionally, the standard deviation of the total contacts per amino-acid and paratope part is relatively low in Figure 18(b), indicating that similar number of contacts are made between different Ags and Abs. This could be an indication of the possibility of predicting CBCEs without any further information about the actual Abs. Figure 18(b) shows the STD for number of contacts. X-axis shows the amino-acids, y-axis shows the paratope parts.
For each VH and V chain pair we defined a single CBCE. This was done by first identifying atoms on any Ag whose distance from the CDRs regions of any VH and V chain pair was < 4 Angstroms. The amino-acids that those atoms belonged to were defined as contacts between the Ag and the Ab, that is, they defined the CBCE. Multiple CBCEs from different Abs could be defined on the same Ag.
Finally, structures that did not define any CBCE within the 4 Angstrom distance were discarded.
Observed CBCEs were also mapped to similar Ag. Undiscovered CBCEs could increase the false negative rate of the prediction models. To decrease the possibility of assigning undiscovered CBCEs as negative data we copied observed CBCEs to similar Ags. First we identified clusters of similar Ags using BlastPlus with > 90% similarity and at most 2 gaps. CBCEs were copied from an Ag in a cluster to all the other Ags in the same cluster, as long as their corresponding position and distancing was exactly the same based on the mapping from BlastPlus.
Lastly, we created a unique Ag local database. Duplicated Ags were completely removed from the data. On the other hand, Ag sequences that were subsequences of longer Ags were kept in the data and treated as different observations. The resulted database consisted of 1003 Ag sequences in FASTA format and 12968 CBCEs, of which 6986 were unique. Each Ag sequence was associated with at least one CBCE.
6.2 CBCE models summary
We have created two main CBCE predictors, namely model 13 conf v1 and model 17 conf v1. There are two differences between the models. Firstly, the negative data generator is different, see section 6.3 for details. Secondly, the model 13 conf v1 model gives two outputs, while model 17 conf v1 gives one. Both models take as input features, among others, a permutation vector. Every sequence in the data is associated with at least one true CBCE, those CBCEs are turned into binary 2D vectors and are given as input to the models. The common output of the two models is a probability, so both models are basically asking the question: “Is this permutation vector, on this specific sequence, a true CBCE or not?”. The second output of the model 13 conf v1 model is the permutation vector itself. This part of the model works as an AE, where it returns a probability per amino-acid. This output can be seen as a contribution of each amino-acid in the sequence to the specific CBCE in question.
The goal of those models was to predict CBCEs on the protein Ag sequence, before the Ab binding event took place. The dataset used, described in section 6.1 , contains the linear protein sequence and the true CBCEs. Therefore, there is no information about any structural or surface characteristic of those protein sequences. However, we need information about how each protein sequence’s secondary structure and surface would look like before the binding event. Therefore, we used our prediction models for RSA, UHSE and LHSE (model struct 3d rsa hse m2 res) and SS (model struct 3d ss m3 res1) described in section 4 to predict those characteristics in every protein in the dataset.
There are three main differences of our models compared to the currently existing publicly available CBCE algorithms. Firstly, we use BLSTMs which allows us to capture relationships between distant and contextual amino-acids which may constitute a CBCE on a native folded 3D protein structure. On the other hand, the other algorithms are segmenting the protein sequences in k-mers and thus, not capturing these important relationships. Secondly, the training dataset is created in such a way to predict true CBCEs before the binding event takes place. This difference brings the model conceptually closer to modelling the reality of Ab binding. As the 3D folded protein structure is exactly what an Ab sees and triggers the binding, and not what an Ab has already seen and bound on.
Thirdly, the currently existing publicly available CBCE algorithms require an experimentally measured, or potentially predicted, 3D protein Ag structure in order to predict epitopes. Our first machine learning model does not require the 3D coordinates of every atom in each amino acid on the Ag or Ab. Finally, our model addresses the epitope question in a conceptually very different manner, but in a manner that is more fined tuned to capture bone fide (“true”) BCEs. We predict probabilities of stand-alone CBCEs on each Ag as a pair or independent class, and therefore multiple CBCEs on the same protein sequence can be discovered and assessed independently. On the other hand, other algorithms provide probabilities per amino-acid, without the ability of separation between different CBCEs.
Figure 19(a) schematically shows the architecture of the model 13 conf v1 model. The architecture of the model is substantially the same as that shown in Figure 11. Here, the output layer comprises two output nodes. The BCE amino acid output (BCE_aa) is the actual permutation output, where binary cross-entropy loss was used. The BCE perm output is a probability vector indicating if the input permutation is a true CBCE (second position on vector) or not (first position on vector). Figure 19(b) shows the architecture of the model 17 conf v1 model. The BCE perm represents the same output as the previous model. Note: the input features were computed per amino-acid as they are, not averaged by windows.
For both models, the BCE perm output is computed from amino-acid values. The last layer of the output is a Dense layer with sigmoid activation. Such that, for each protein sequence, a value in e [0, 1] is returned per amino-acid. Then a 2-class probability vector is computed for that sequence as:
where N is the length of the sequence and xi is the dense layer output on aminoacid at position i on the sequence. This 2-class probability vector is then used in the binary cross-entropy loss.
The model 13 conf v1 model had weighted global loss. The global model loss was the weighted sum of the individual ones, that is BCE perm and BCE aa, with weights: 1 , and 1 , respectively. Different weights could be tested as well.
6.3 Negative Data
If we were to model the CBCEs as per amino-acid, then the positive data would be every amino-acid in a sequence which is a part of any true CBCE, and the negative data would be any other amino-acid on the same sequence. However, since we model the whole CBCE as a single permutation vector of a given protein sequence, the response to this vector is a single probability (that is output BCE perm from section 6.2). Since we only have observed true positive CBCEs from
the data in section 6.1 , we do not have any permutation corresponding to negative data.
We developed a solution to improve the performance of the models, whereby negative permutation data was generated in a manner which resample the ground truth as close as possible. To cover a large spectrum of negative data and make the models more robust, we generated negative permutations for every sequence in the data, on every epoch during training. The two different generation schemes are described in the following sections.
6.3.1 CBCE model (model 13 conf v1)
For this model we generated completely random permutations. For each sequence in the training and validation data we generated 10 completely random permutations and assumed that they are not true CBCEs. Both the total number of amino-acids and the placement of those in the given sequence were random. New permutations were generated on each epoch.
The advantage of this method is that because of the complete randomness, it is quite unlikely that any generated CBCE permutation will be false negative. However, the disadvantage is that the randomly generated CBCE permutations may be extremely different than the true CBCE permutations.
6.3.2 CBCE model (model 17 conf v1)
For this model we applied three different types of negative data generating methods. For each sequence in the training and validation data we generated 30 completely random permutations, 10 from each method, and assumed that they are not true CBCEs. The first method is like the one described in 6.3.1. The second and third methods were applied on every true CBCE of the given protein sequence. The second method kept the first and last amino-acid of a given true CBCE at the correct position, while it randomly shuffled the internal CBCE amino- acids indicators inside the region. The third method kept the total amount of amino-acids of a true CBCE, as well their linear distance, constant. It then
randomly shifted the whole true CBCE on other parts of the protein. New permutations were generated on each epoch from every method.
The advantage of this method is that the randomly generated CBCE permutations cover both extremely similar and extremely different true CBCE permutations. This will probably lead to that the algorithm is more robust to both positive and negative data. On the other side, the disadvantage of the method is that the false negatives may increase . Leading to lower performance on new data.
6.4 Outliers
We removed outliers before any training of either model. We did this by applying Isolation forests (IF) on the total amino-acids of each observed CBCE in our data from section 5.1 . We run the algorithm for every observed amount of total amino- acids as the max samples argument in the IF algorithm from the sklearn Python package (F. Pedregosa, G. Varoquaux, A. Gramfort, V. Michel, B. Thirion, 0. Grisel, M. Blondel, R Prettenhofer, R. Weiss, V. Dubourg, J. Vanderplas, A. Passes, D. Cournapeau, M. Brucher, M. Perrot, and E. Duchesnay. Scikit-learn: Machine learning in Python. Journal of Machine Learning Research, 12:2825- 2830, 2011.) The most common minimum and maximum borders from all the runs were taken as global borders for inliers.
Figure 20 shows the total amount of amino-acids from all the observed CBCEs. Figures 20(a) and 20(b) show, respectively, density and box-plot of observed amount of total amino-acids in the true CBCEs. Figures 20(c) and 20(d) shown log-scales of Figures 20(a) and 20(b) respectively. Where we see that the median amount is at 10 amino-acids. After the outlier analysis, we found that valid CBCEs should contain a total amount of amino-acids in range, which is lower than what discussed in section 1.3. However, those studies did not perform any outlier detection or removal of their observed CBCEs. As we can see from Figure 20, the maximum amount of total amino-acids before outlier removal is around 30 to 35, which corresponds with that reported in the previous studies. The outliers were finally removed from the data, and the models were trained without them.
6.5 Preliminary Analysis and Features Used
See section 3.2 for the procedure. The 14 features that we chose to use in the models (including the protein-sequence in binary representation and the permutation vectors) is shown in in Table 5. Note that the RSA and SS were predicted by models struct 3d rsa hse m2 res and struct 3d ss m3 res1 , respectively, described in section 4. The same features were used for both model model 13 conf v1 and model 17 conf vl .
Table 5: Features used in the model 13 conf vl and model 17 conf vl models.
6.6 Pre-training
See section 2.2.
6.7 CV results
We defined as ’’age” of the protein Ag in our data the date at which the most recent CBCE was registered to the protein sequence. We used ~ 94% of the ’’oldest” proteins in the data on a 5-fold cross-validation in order to assess the performance of the model. We used custom callbacks in Tensorflow to allow for early stopping in case the loss did not decrease anymore.
6.7.1 Model model 13 conf yl
Figure 21 shows results for the precision-recall metric. The model performed well, without significant overfitting and had high performance for all evaluation metrics. Moreover, a small variation is observed between different CV runs for the loss. The models were trained for around 466 epochs and apparent bumps in the loss may indicate that the random generator create negative CBCE permutations which were similar to the positives. However, their negative effect became smaller as the epochs proceeded further.
Figure 21 illustrates results for CBCE model 13 conf v1. Figure 21 (a)-(e): PR curves. The x-axis is the TPR, the y-axis is the PPV. Figure 21(f): Weighted cross entropy loss. The x-axis is the epochs, the y-axis is the loss values. The plot is the average of 5-fold CVs including Gaussian Cl. The blue line is for training and the red (dashed) for validation sets.
6.7.2 Model model 17 conf v1
Figure 22 shows results for the precision-recall metric. The model seems to perform well, without significant overfit before 250 epochs and gives high performance for all metrics. Moreover, small variation is observed between different CV runs for the loss. The models were trained for 455 epochs.
Figure 22 illustrates results for CBCE model 17 conf v1. Figure 22 (a)-(e): PR curves. The x-axis is the TPR, the y-axis is the PPV. Figure 22(f): Weighted cross entropy loss. The x-axis is the epochs, the y-axis is the loss values. The plot is the average of 5-fold CVs including Gaussian Cl. The blue line is for training and the red (dashed) for validation sets.
6.8 Test results
We used ~ 6% of the ’’newest” proteins in the data in order to see the performance on completely unseen data. The reason for keeping the newest data as test set was because the same data was used for comparison with other algorithms. This way we ensure that those algorithms were not trained on that data. The training on the following models was done on the ’’oldest” ~ 94% of the data.
6.8.1 Model model 13 conf v1
Figure 23 shows the results for the precision-recall metric. The model performed well, without significant overfitting and gave high performance for all evaluation metrics. The model for trained for 466 epochs, chosen based on the CV analysis, so slightly more epochs than the CV analysis.
Figure 23 illustrates the results for CBCE model 13 conf v1. Figure 23(a): PR curves. The x-axis is the TPR, the y-axis is the PPV. Figure 23(b): Overall model loss. The x-axis is the epochs, the y-axis is the loss values. The solid line is for the training and the dotted for the validation set.
6.8.2 Model model 17 conf v1
Figure 24 shows the results for the precision-recall metric. The model seems to perform well, without significant overfit and gives high performance for all metrics. The model for trained for 455 epochs, chosen based on the CV analysis, so slightly more epochs than the CV analysis.
Figure 24 illustrates the results for CBCE model 17 conf v1. Figure 24(a): PR curve. The x-axis is the TPR, the y-axis is the PPV. Figure 24(b): Overall model loss. The x-axis is the epochs, the y-axis is the loss values. The solid line is for the training and the dotted for the validation set.
6.9 Final Training
The final training was done on the whole data. Using the optimal number of epochs determined by the cross-validation procedure.
6.9.1 Model model 13 conf v1
Figure 25 shows the results for the F1 metric. The model seems to perform well, without significant overfit and gives hight performance for all metrics. The model for trained for 500 epochs, chosen based on the CV analysis.
Figure 25 shows the results for CBCE model 13 conf v1. Figure 25(a)-(b): F1 metrics. The x-axis is the epochs, the y-axis is the F1 value. Figure 25(d): Overall model loss. The x-axis is the epochs, the y-axis is the loss values.
6.9.2 Model model 17 conf v1
Figure 26 shows the results for the F1 metric. The model seems to perform well, without significant overfit and gives high performance for all metrics. The model for trained for 700 epochs, chosen based on the CV analysis.
Figure 26 shows the results for CBCE model 17 conf v1. Figure 26(a): F1 metric. The x-axis is the epochs, the y-axis is the F1 value. Figure 26(b): Overall model loss. The x-axis is the epochs, the y-axis is the loss values.
7 First Machine Learning Model (LBCEs)
For predicting LBCEs we use BLSTMs models as well. This allows us to model the whole protein sequences as one observation, without the need of segmenting it. Therefore, distant amino-acid relationships should be able to be captured by the model. Moreover, the use of BLSTMs allows us to train different protein lengths simultaneously. Therefore, the model type here is the same as the one for the CBCEs in section 6. The main goal of our models was to create non-Ab specific LBCE predictors. This is achieved by the way we prepare the training data as described in section 7.1.
7.1 Data preparation
For modelling LBCEs we downloaded LBCE data from IEDB We downloaded all non-obsolete assays from the IEDB, from all organisms.
Since the main goal of our models was to create non-Ab specific LBCE predictors, we considered all positive LBCEs for our database. That is, if a LBCE was tested positive against at least one Ab, we included it in the database. Each Ag ID was then mapped to the universal protein resource (UniProt), in order to retrieve the protein sequence in fasta format. Assays were discarded if their IDs were non-
mappable, or if their IDs were mappable but the coordinates of the corresponding LBCE did not give the same sequence on the retrieved Ag sequence.
Observed LBCEs were also mapped to similar Ag, as we did in section 6.1 for the CBCEs. Undiscovered LBCEs could increase the false negative rate of the prediction models. In an attempt to decrease the possibility of assigning undiscovered LBCEs as negative data we copied observed LBCEs to similar Ags. First we identified clusters of similar Ags using BlastPlus with > 90% similarity and at most 2 gaps. LBCEs were copied from an Ag in a cluster to all the other Ags in the same cluster, as long as their corresponding linear coordinates were exactly the same based on the mapping from BlastPlus.
Lastly, we created a unique Ags database. Duplicated Ags were completely removed the data. On the other hand, Ags sequences that were sub-sequences of longer Ags were kept in the data and treated as different observations. The resulted database constituted of 4695 Ag sequences in fasta format and 177782 CBCEs from which 63363 were unique. Each Ag sequence was associated with at least one LBCE.
7.2 LBCE model summary
Every sequence in the data is associated with at least one true LBCE. The model takes as input features, among others, a permutation vector. This vector is the LBCEs which are turned into binary 2D vectors and are given as input to the model. The output of the model is a probability, so the model is basically asking the question; ”ls this permutation vector, on this specific sequence, a true LBCE or not?”. Basically, the permutation vectors are the same as described in section 6.2, where the only difference is that they represent LBCEs and not CBCEs.
The goal of those models was to predict LBCEs before the binding event took place. The dataset used, described in section 7.1 , contains the linear protein sequence and the true LBCEs. Therefore, there is no information about any structural or surface characteristic of those protein sequences. However, we need information about how each protein sequence’s surface would look like before the
binding event. Therefore, we used our prediction model for RSA (model struct 3d rsa hse m2 res), described in section 4, in order to predict those characteristics in every protein in the dataset.
There are three main differences of our models with the publicly available LBCE algorithms. Firstly, we use BLSTMs which allows us to capture relationships between distant amino-acids which may constitute a LBCE. On the other hand, the other algorithms are segmenting the protein sequences and thus, breaking such bonds. Secondly, the training dataset is created in such a way in order to predict true LBCEs before the binding event takes place. Which is exactly what an Ab sees and triggers the binding, and not what an Ab has already seen and bound on. Finally, we predict probabilities of stand-alone LBCEs, and therefore multiple LBCEs on the same protein sequence can be discovered and separated. On the other hand, other algorithms provide probabilities per amino-acid, without the ability of separation between different LBCEs.
The architecture of the model 12 linear v1 model has substantially the same structure as the model depicted in Figure 19(b). The bee output is a probability vector indicating if the input permutation is a true LBCE (second position on vector) or not (first position on vector). The bee output is computed from aminoacid values as described in section 6.2. Note: the input features were computed per amino-acid as they are, not averaged by windows.
7.3 Negative Data
We followed a similar procedure as in section 6.3.1. The only difference here is that we ensured that the randomly generated LBCEs were unbroken. Let N be the length of a given protein sequence. We first generate the total amount X of amino- acids that a random LBCE will have from the uniform distribution in [0,N], Then we generated the starting position i of this LBCE from the uniform distribution in [0,N - X], Finally, we assigned all amino-acids in [i, i + X] as the ones corresponding this random LBCE. A total amount of 10 randomly generated LBCEs was generated pre sequence, for every epoch.
7.4 Outliers
We removed outliers before any training of the model 12 linear v1 model. We followed exactly the same procedure as in section 6.4, and the data we used was the amount of total amino-acids of each observed LBCE in our data from section 7.1.
Figure 27 shows the total amount of amino-acids from all the observed LBCEs. Where we see that the median amount is at 17 amino-acids. After the outlier analysis, we found that valid LBCEs should contain a total amount of amino-acids in range [5, 30], The outliers were finally removed from the data, and the models were trained without them.
Figure 27 shows outlier analysis for LBCEs. Figures 27(a) and (b) show, respectively, density and box-plot of observed amount of total amino-acids in the true LBCEs. Figures 27(c) and (d) show, log-scales of Figures 27(a) and 27(b) respectively.
7.5 Preliminary Analysis and Features Used
See section 3.2 for the procedure. The 13 features that we chose to use in the model (including the protein-sequence in binary representation and the permutation vectors) can be seen in Table 6. Note that the RSA was predicted by the model struct 3d rsa hse m2 res described in Section 4.
The precise prediction of B-cell epitopes (in particular 3D conformational B-cell epitopes) leads to major improvements in diagnostics, drug design and vaccine development in the fields of autoimmunity, infectious-disease, and cancer. Examples are provided below.
The binding event of an antibody to an antigen is a key event in activating the humoral B Cell immune response against any pathogenic threat, such as viral and bacterial infections, in addition to cancer antigens in malignant tumour cells. B Cell epitope mapping is essential in therapeutic antibody development. The prediction of true B-cell epitopes by the present invention not only provides a functional understanding of the critical residues involved antibody-antigen binding; it also aids in the selection of therapeutically relevant antibodies to the prioritized predicted epitopes for further development as therapeutic antibodies.
In antibody driven vaccine design, in silico predictions are performed to predict BCEs that could trigger humoral immune responses. Previous in silico guided vaccine-design initiatives tend to selectively apply BCE prediction with a focus on surface-exposed sites among the antigens. However, by using the approach of the present invention (in particular training on unbound protein structures and the use of candidate BCE encodings (“permutation vectors”) mapped to the antigen protein sequence), we can identify with precision, BCEs to guide Ab based vaccine design with far higher fidelity. BCE vaccines guided by the approach of the present invention can advantageously identify at scale specific antibody epitope interactions capable of avoiding undesirable immune responses, in addition to potentially generating long lasting, potent and universal immunity. Moreover, the method described here bypasses the need for years of target discovery and laborious expensive time-consuming work at a fraction of the cost and time.
8.3 Immuno-diaqnostics and immune monitoring
Prediction of B cell epitopes is of direct help to the design of immuno-diagnostic and immune monitoring of therapy strategies and reagents. Integrating prediction approaches for the accurate and comprehensive detection of BCEs can greatly facilitate diagnosis. One example of this would be the rapid and accurate detection of BCE epitopes on the spike protein of beta-coronaviruses, allowing for the diagnosis of COVID in a manner that can distinguish infections caused by common cold viruses from the SARS-CoV-2 virus causing COVID. Along the same vein, the BCE prediction technology can be used to identify immunodominant or humoral reactive epitopes in tumour associated antigens or neoantigens in immuno-diagnostic and immune monitoring clinical applications.
8.4 Example Systems
Figure 4 schematically illustrates an example of a system suitable for implementing embodiments of the method. The system 1100 comprises at least one server 1110 which is in communication with a reference data store 1120. The server may also be in communication with an automated peptide synthesis device 1130, for example over a communications network 1140.
In certain embodiments the server may obtain, for example using from the reference data store, an amino acid sequence of one or more query proteins. The server may then predict whether the query protein comprises a BCE that is likely to instigate a binding event with an antibody (e.g. instigate an immunogenic response).
Query proteins predicted to contain one or more BCEs may be sent to the automated peptide synthesis device 1130 to synthesise the query protein or parts thereof. Techniques for automated peptide synthesis are well known in the art and it will be understood that any known technique may be used. Typically, the query protein or epitope is synthesized using standard solid phase synthetic
peptide chemistry and purified using reverse-phase high performance liquid chromatography before being formulated into an aqueous solution. If used for vaccination, prior to administration the peptide solution is usually admixed with an adjuvant before being administered to the patient.
Peptide synthesis technology has existed for more than 20 years but has undergone rapid improvements in recent years to the point where synthesis now takes just a few minutes on commercial machines. For brevity we do not describe in detail such machines but their operation would be understood to one skilled in the art and such conventional machines may be adapted to receive a candidate region or epitope from the server.
The server may comprise the functions described above to identify query proteins predicted to contain one or more viable BCEs. It will of course be understood that these functions may be subdivided across different processing entities of a computer network and different processing modules in communication with one another.
The techniques for identifying BCE-containing proteins may integrate into a wider ecosystem for customised vaccine development. Example vaccine development ecosystems are well known in the art and are described at a high-level for context, but for brevity we do not describe the ecosystem in detail.
In an example ecosystem, a first, sample, step may be to isolate DNA from a tumor biopsy and matched healthy tissue control. In a second, sequence, step, the data is sequenced and the variants identified i.e. the mutations. In an immune profiler step the associated mutated peptides may be generated «in silico».
Using the associated mutated peptides, and the techniques described here, a candidate protein may be predicted and selected and target epitopes identified for vaccine design. That is, the candidate peptide sequence is chosen based on its predicted binding affinity determined using the technique described herein.
The target epitopes are then generated synthetically using conventional techniques as described above. Prior to administration the peptide solution is usually admixed with an adjuvant before being administered to the patient (vaccination). In alternatives, the target epitopes can be engineered into DNA or RNA, or engineered into the genome of a bacteria or virus, as with any conventional vaccine.
The proteins predicted by the methods described herein may also be used to create other types of vaccine other than peptide based vaccines. For example the proteins (or predicted epitopes therein) could be encoded into the corresponding DNA or RNA sequence and used to vaccinate the patient. Note that the DNA is usually inserted in to a plasmid construct. Alternatively, the DNA can be incorporated into the genome of a bacterial or viral delivery system (can be RNA also - depending on the viral delivery system) - which can be used to vaccinate the patient - so the manufactured vaccine in a genetically engineered virus or bacteria which manufactures the targets post immunisation in the patient i.e. in vivo.
An example of a suitable server 1110 is shown in Figure 5. In this example, the server includes at least one microprocessor 1200, a memory 1201 , an optional input/output device 1202, such as a keyboard and/or display, and an external interface 1203, interconnected via a bus 1204 as shown. In this example the external interface 1203 can be utilised for connecting the server 1110 to peripheral devices, such as the communications networks 1140, reference data store 1120, other storage devices, or the like. Although a single external interface 1203 is shown, this is for the purpose of example only, and in practice multiple interfaces using various methods (e.g. Ethernet, serial, USB, wireless or the like) may be provided.
In use, the microprocessor 1200 executes instructions in the form of applications software stored in the memory 1201 to allow the required processes to be performed, including communicating with the reference data store 1120 in order to receive and process input data, and/or with a client device to receive sequence
data for one or more query proteins, and to generate immunogenic potential predictions according to the methods described above. The applications software may include one or more software modules, and may be executed in a suitable execution environment, such as an operating system environment, or the like.
Accordingly, it will be appreciated that the server 1200 may be formed from any suitable processing system, such as a suitably programmed client device, PC, web server, network server, or the like. In one particular example, the server 1200 is a standard processing system such as an Intel Architecture based processing system, which executes software applications stored on non- volatile (e.g., hard disk) storage, although this is not essential. However, it will also be understood that the processing system could be any electronic processing device such as a microprocessor, microchip processor, logic gate configuration, firmware optionally associated with implementing logic such as an FPGA (Field Programmable Gate Array), or any other electronic device, system or arrangement. Accordingly, whilst the term server is used, this is for the purpose of example only and is not intended to be limiting.
Whilst the server 1200 is a shown as a single entity, it will be appreciated that the server 1200 can be distributed over a number of geographically separate locations, for example by using processing systems and/or databases 1201 that are provided as part of a cloud based environment. Thus, the above described arrangement is not essential and other suitable configurations could be used.
9. Methodological case study on mutation policy generation
We now describe embodiments in which the present invention may be used to identify one or more B-cell epitopes that are predicted to instigate a protective immunogenic response across a plurality of species of a pathogen or virus. For this analysis we randomly generated unique permutations from a prior distribution for each Sequence and RBD combination. SARS, SARS-COV-2, and MERS receptor binding domains (RBDs) were aligned onto the available spike proteins from 304 different betacoronavirus species.
Figures 29(a) to (c) illustrate the successful mapping of the RBDs to the 304 betacoronavirus species. The species are divided by subgenus groups on the X- axis and the Y-axis depicts the total count of species that the RBD was successfully aligned.
Each permutation vector was then selected from the given RBD region of each spike protein sequence. The random generation was done once per RBD type (SARS, SARS-COV-2, MERS), and the same permutations were used for each sequence. The following outcome emerged from these alignments’ random permutation generation:
1 . 450000 permutation vectors were generated for SARS RBD
2. 450000 permutations vectors generated for SARS-COV-2 RBD
3. 450000 permutations vectors generated for MERS RBD
And those permutations were used in each species having the corresponding RBD aligned onto it.
9.1. Statistics of the results on predictions on all sequences (all RBDs on all species):
Total predictions made per RBD (all species, note the differences in the totals is because not all species have every RBD, so there are fewer permutations created on some RBDs): The permutations are generated from the priors on the RBD regions.
9.2. Immunodominant permutation vectors and eptiopes emerge from the data analysis
Based on the above data of positive (confirmation BCE score > 0.9), we were able to capture robust structural motifs, or B cell immunodominant structural epitopes (broadly immunogenic across most viral species tested). These are depicted in Figure 30. Here we clearly demonstrate that the BCE predictor is capable of capturing potential structural motifs shared across the RBDs of numerous species, i.e., Immunodominant structural motifs. In Figure 30, each panel is an RBD. Filtering for the conformational epitopes used in this plot was BCE>=0.9. The y- axis is the count of total unique species and the x-axis shows one permutation vector (structural motif). Therefore, each column can be considered as a single permutation vector shared among the betacoronavirus species counted on the Y- axis. The X-axis is sorted in descending order of total unique species. The colouring of the bars is based on Subgenus.
The permutation vectors represented in the columns in Figure 30 are structural representations of the epitopes predicted by the BCEP predictor. We next generated amino acid presentations of these immunodominant (broadly immunogenic) epitopes and calculated the sequence logos for the top 1 permutation per RBD (ties included). The logos are created by taking the highest
column (permutation) from Figure 30 above, and based on the positive BCE prob>0.9 we create the sequence logos of their epitopes.
Those are created by taking the highest column (permutation vector) from Figure 30 and based on the positive BCE prob>0.9 we created the sequence logos of their epitopes. It was found that the shared permutation vector also consists of shared reoccurring amino acid sequences, capturing broadly immunogenic B cell epitopes amino acid sequences dominant across most species.
9.3. Implementation of the mutation policies using the B-cell epitope predictor
We next proceeded to integrate the BCEP predictor to in silico mutation policies to 1) identify more potent BCE epitopes and 2) B cell epitopes that are predicted by the model to be even more immunogenic across more of the species analyzed. We demonstrate that mutation policy method here by mutating the most common motif for each RBD, and performed the following steps:
1 . In silico mutate each amino acids on each position of that motif to all the other known acids.
2. For each in silico mutation, we re-predict the permutation vector (candidate epitope) for all the spike proteins from the 304 betacoronavirus species, with the new mutated motif, and count the total species becoming positive (BCE >0.9) for that that mutated motif.
The mutation by its very nature is designed to be broadly immunogenic against the panel of 304 betacoronavirus species. Its original not mutated counterpart was chosen because it was shared across most species and the mutation policies capture an additional 27 species (in terms of predicted immunogenicity) for the most successful mutation, as seen in Figure 31(a) where the highlighted bar 310 representing the mutation is found to be positive on more species compared to the original counterpart 311 . One can speculate that if this mutated spike protein was applied in a vaccination setting, the immunodominant epitopes latent within it would generate antibody responses that could potentially be broadly immunogenic and protective. The mutation policy result illustrated in Figure 31 (a) and (b) is for
demonstration purposes and may be applied to a vast array of shared permutation vectors and motifs across the 304 species, and all 3 of the pathogenic RBDs.
10. Appendix
[1] T P Hopp and K R Woods. Prediction of protein antigenic determinants from amino acid sequences. Proceedings of the National Academy of Sciences of the United States of America, 78(6):3824-3828, 06 1981.
[2] R. Grantham. Amino acid difference formula to help explain protein evolution. Science, 185(4154): 862, 09 1974.
[3] JOEL JANIN. Surface and inside volumes in globular proteins. Nature, 277:491 EP -, 02 1979.
[4] H R Guy. Amino acid side-chain partition energies and distribution of residues in soluble proteins. Biophys J, 47(1):61-70, Jan 1985.
[5] J. K. Mohana Rao and Patrick Argos. A conformational preference parameter to predict helices in integral membrane proteins. Biochimica et Biophysica Acta (BBA) - Protein Structure and Molecular Enzymology, 869(2):197-214, 1986.
[6] Sanzo Miyazawa and Robert L. Jernigan. Estimation of effective interresidue contact energies from protein crystal structures: quasi-chemical approximation. Macromolecules, 18(3): 534-552, 03 1985.
[7] Gang Zhao and Erwin London. An amino acid ’’transmembrane tendency” scale that approaches the theoretical limit to accuracy for prediction of transmembrane helices: relationship to biological hydrophobicity. Protein science : a publication of the Protein Society, 15(8): 1987-2001 , 08 2006.
[8] Cyrus Chothia. The nature of the accessible and buried surfaces in proteins. Journal of Molecular Biology, 105(1):1— 12, 1976.
[9] GD Rose, AR Geselowitz, GJ Lesser, RH Lee, and MH Zehfus. Hydrophobicity of amino acid residues in globular proteins. Science, 229(4716):834, 08 1985.
[10] P Y Chou and G D Fasman. Prediction of the secondary structure of proteins from their amino acid sequence. Adv Enzymol Relat Areas Mol Biol, 47:45-148, 1978.
[11] Shneior Lifson and Christian Sander. Antiparallel and parallel -strands differ in amino acid residue preferences. Nature, 282(5734): 109-111 , 1979.
[12] M. Cooper and Robert E. Hausman. The cell : a molecular approach, volume 4. Washington, DC ASM Press, 2007.
[13] E A Emini, J V Hughes, D S Perlow, and J Boger. Induction of hepatitis a virusneutralizing antibody by a virus-specific synthetic peptide. Journal of virology, 55(3):836-839, 09 1985.
[14] G. Deleage and B. Roux. An algorithm for protein secondary structure prediction based on class prediction. Protein Engineering, Design and Selection, 1 (4):289-294, 08 1987.
[15] P. MANAVALAN and P. K. PONNUSWAMY. Hydrophobic character of amino acid residues in globular proteins. Nature, 275(5681 ):673-674, 1978.
[16] Michael Levitt. Conformational preferences of amino acids in globular proteins. Biochemistry, 17(20):4277-4285, 10 1978.
[17] Serafin Fraga. Theoretical prediction of protein antigenic determinants from amino acid sequences. Canadian Journal of Chemistry, 60(20):2606-2610, 2019/03/01 1982.
[18] Gjalt W. Welling, Wcher J. Weijer, Ruurd van der Zee, and Sytske Welling- Wester. Prediction of sequential antigenic regions in proteins. FEBS Letters, 188(2):215-218, 2019/02/28 1985.
[19] Henry B. Bull and Keith Breese. Surface tension of amino acid solutions: A hydrophobicity scale of the amino acid residues. Archives of Biochemistry and Biophysics, 161 (2):665-670, 1974.
[20] C. A. Browne, H. P. J. Bennett, and S. Solomon. The isolation of peptides by high-performance liquid chromatography using predicted elution positions. Analytical Biochemistry, 124(1 ):201 -208, 1982.
[21] R. Wolfenden, L. Andersson, P. M. Cullis, and C. C. B. Southgate. Affinities of amino acid side chains for solvent water. Biochemistry, 20(4):849-855, 02 1981.
[22] J L Meek. Prediction of peptide retention times in high-pressure liquid chromatography on the basis of amino acid composition. Proceedings of the National Academy of Sciences of the United States of America, 77(3): 1632-1636, 03 1980.
[23] R. BHASKARAN and P. K. PONNUSWAMY. Positional flexibilities of amino acid residues in globular proteins. International Journal of Peptide and Protein Research, 32(4):241-255, 2019/02/28 1988.
[24] Schwartz R.M. Dayhoff, M.O. and B.C. Orcutt. A model of evolutionary change in proteins, volume 5. Atlas of Protein Sequence and Structure. Natl. Biomed. Res. Found., Washington DC, 1978.
[25] David L. (David Lee) Nelson and Michael M. Cox. Lehninger principles of biochemistry. Principles of biochemistry. W.H. Freeman, New York, seventh edition, edition, 2017.
[26] K JWilson, A Honegger, R P Stotzel, and G J Hughes. The behaviour of peptides on reverse-phase supports during high-pressure liquid chromatography. The Biochemical journal, 199(1 ):31- 41 , 10 1981.
[27] A. S. Kolaskar and Prasad C. Tongaonkar. A semi-empirical method for prediction of antigenic determinants on protein antigens. FEBS Letters, 276(1- 2): 172-174, 2018/12/07 December 10, 1990.
[28] Peter McCaldon and Patrick Argos. Oligopeptide biases in protein sequences and their use in predicting protein coding regions in nucleotide sequences. Proteins: Structure, Function, and Bioinformatics, 4(2):99-122, 2019/02/28 1988.
[29] J.M. Zimmerman, Naomi Eliezer, and R. Simha. The characterization of amino acid sequences in proteins by statistical methods. Journal of Theoretical Biology, 21 (2): 170-201 , 1968.
[30] Daniel D. Jones. Amino acid properties and side-chain orientation in proteins: A cross correlation approach, volume 50. 04 1975.
[31] D. Eisenberg, E. Schwarz, M. Komaromy, and R.Wall. Analysis of membrane and surface protein sequences with the hydrophobic moment plot. Journal of Molecular Biology, 179(1 ): 125-142, 1984.
[32] J. M. R. Parker, D. Guo, and R. S. Hodges. New hydrophilicity scale derived from high-performance liquid chromatography peptide retention data: correlation of predicted surface residues with antigenicity and x-ray-derived accessible sites. Biochemistry, 25(19): 5425-5432, 09 1986.
[33] Shaun D. Black and Diane R. Mould. Development of hydrophobicity parameters to analyze proteins which bear post- or cotranslational modifications. Analytical Biochemistry, 193(1):72-82, 1991.
[34] J L. Fauchere and V Pliska. Hydrophobic parameters II of amino acid sidechains from the partitioning of N-acetyl-amino acid amides, volume 18. 01 1983.
[35] Robert M. Sweet and David Eisenberg. Correlation of sequence hydrophobicities measures similarity in three-dimensional protein structure. Journal of Molecular Biology, 171 (4):479-488, 1983.
[36] Jack Kyte and Russell F. Doolittle. A simple method for displaying the hydropathic character of a protein. Journal of Molecular Biology, 157(1): 105-132, 1982.
[37] Charles. Tanford. Contribution of hydrophobic interactions to the stability of the globular conformation of proteins. Journal of the American Chemical Society, 84(22):4240-4247, 11 1962.
[38] Donald J. Abraham and Albert J. Leo. Extension of the fragment method to calculate amino acid zwitterion and side chain partition coefficients. Proteins: Structure, Function, and Bioinformatics, 2(2): 130-152, 2019/02/28 1987.
[39] Richard Cowan and R GWhittaker. Hydrophobicity indices for amino acid residues as determined by HPLC, volume 3. 01 1990.
[40] Mark A. Roseman. Hydrophilicity of polar amino acid side-chains is markedly reduced by flanking peptide bonds. Journal of Molecular Biology, 200(3):513-522, 1988.
[41] A. A. Aboderin. An empirical hydrophobicity scale for alpha-amino-acids and some of its applications. International Journal of Biochemistry, 2:537-544, 1971.
Claims
1. A computer-implemented method of predicting whether a protein comprises a B-cell epitope that is likely to instigate a binding event with an antibody, the method comprising:
(a) accessing one or more structure and/or surface characteristics of the protein; and
(b) inputting the one or more structure and/or surface characteristics of the protein into a trained first machine learning model to predict whether the protein comprises a true B-cell epitope, wherein the first machine learning model is trained by: generating a first reference dataset that comprises:
(i) a plurality of first reference proteins, each first reference protein comprising at least one B-cell epitope classified as a true B-cell epitope; and
(ii) one or more structure and/or surface characteristics of each first reference protein in an unbound state; and training the first machine learning model using the first reference dataset to learn a relationship between the structure and/or surface characteristic(s) of the first reference proteins in an unbound state, and the B-cell epitopes classified as true B-cell epitopes.
2. The method of claim 1 , wherein the output of the trained first machine learning model is a probability that is indicative of whether the protein comprises a true B-cell epitope.
3. The method of claim 1 or claim 2, wherein the one or more structure and/or surface characteristics include one or more of: the secondary structure of the protein; the relative solvent accessibility, RSA; the half-sphere exposure, HSA.
4. The method of any of the preceding claims, wherein the one or more structure and/or surface characteristics of each first reference protein are predicted by:
accessing the amino acid sequence of the first reference protein; and applying one or more second machine learning model(s) to the amino acid sequence to predict the one or more structure and/or surface characteristics, wherein the one or more second machine learning model(s) are trained on a second reference dataset that comprises a plurality of amino acid sequences of respective second reference proteins and their corresponding structure and/or surface characteristics in an unbound state.
5. The method of any of the preceding claims, wherein at least some of the first reference proteins are obtained by:
(i) accessing a plurality of protein complexes comprising at least three different protein chains;
(ii) filtering the plurality of protein complexes to leave only those having a protein chain mapped as a valid heavy (VH) chain of an antibody, a protein chain mapped as a valid light (V ) chain of an antibody, and a protein chain mapped as an antigen;
(iii) pairing the VH and V chains to form VH and V chain pairs; and
(iv) for each VH and V chain pair, defining a true B-cell epitope on the corresponding antigen; wherein at least some of the first reference proteins of the first reference dataset correspond to an antigen mapped in step (ii) that comprises at least one true IB- cell epitope defined in step (iv).
6. The method of any of the preceding claims, further comprising: accessing the amino acid sequence of the protein; and generating one or more candidate encodings of respective one or more candidate B-cell epitopes on the protein, wherein each candidate encoding represents the respective candidate B-cell epitope as a plurality of data elements corresponding to amino acids of the amino acid sequence; wherein the inputs to the trained first machine learning model comprise the amino acid sequence of the protein, the one or more structure and/or surface characteristics of the protein, and the one or more candidate encodings.
7. The method of claim 6, wherein the first reference dataset further comprises: amino acid sequences of each first reference protein; and a plurality of reference encodings of each true B-cell epitope, wherein each reference encoding represents the respective true B-cell epitope as a plurality of data elements corresponding to the amino acids of the amino acid sequence.
8. The method of claim 6 or claim 7, wherein the output of the first machine learning model is a probability that each of the candidate encoding(s) represents a true B-cell epitope on the protein.
9. The method of any of the preceding claims, wherein the first machine leaning model comprises a long short-term memory network, LSTM, preferably a bi-directional long short-term memory network, BLSTM.
10. The method of any of the preceding claims, wherein the B-cell epitope is a conformational B-cell epitope.
11 . The method of any of the preceding claims, wherein the protein comprises an antigen.
12. The method of any of claims 6 to 11 when dependent on claim 6, further comprising: identifying a first B-cell epitope that is predicted to be likely to be a true B-cell epitope on the protein; substituting one or more amino acids forming the identified first B-cell epitope with a different amino acid to thereby generate a modified amino acid sequence of the protein; and using the trained first machine learning model to predict whether the candidate encoding representing the identified first B-cell epitope is likely to
correspond to a true B-cell epitope on the modified amino acid sequence of the protein.
13. A method of creating a vaccine, comprising: predicting that a protein comprises a B-cell epitope by a method according to any of the preceding claims; and synthesising the protein and/or the predicted B-cell epitope, or encoding at least one of the protein, the predicted B-cell epitope, B-cell epitope variant, B- cell epitope predicted or simulated variant into a corresponding protein, peptide, DNA or RNA sequence.
14. A method of creating a diagnostic assay to determine whether a patient has or has had a cancer or prior infection with a pathogen or tumor, wherein the diagnostic assay is carried out on a biological sample obtained from a subject, comprising identifying at least one protein of the pathogen that is predicted to comprise a B-cell epitope, using a method according to any of claims 1 to 12; wherein the diagnostic assay comprises the utilisation or identification within the biological sample of the identified at least one protein and/or B-cell epitope.
15. A method of training a machine learning model, comprising generating a reference dataset, the reference dataset comprising: a plurality of first reference proteins, each first reference protein comprising at least one B-cell epitope classified as a true B-cell epitope; and one or more structure and/or surface characteristics of each first reference protein in an unbound state; and training the machine learning model using the reference dataset to learn a relationship between the structure and/or surface characteristic(s) of the first reference proteins in an unbound state, and the corresponding B-cell epitopes classified as true B-cell epitopes.
16. The method of claim 15, wherein the reference dataset further comprises:
amino acid sequences of each reference protein; and a plurality of reference encodings of each true B-cell epitope, wherein each reference encoding represents the respective true B-cell epitope as a plurality of data elements corresponding to the amino acids of the amino acid sequence.
17. A computer program product comprising instructions which, when executed by a computer, cause the computer to perform the method of any of the preceding claims.
18. Use of a machine learning model trained using the method of claim 15 or claim 16.
19. A machine learning model trained using the method of claim 15 or claim 16.
20. A system for predicting whether a protein comprises a B-cell epitope that is likely to instigate a binding event with an antibody, comprising at least one processor in communication with at least one memory device, the at least one memory device having stored thereon instructions for causing the at least one processor to perform a method according to any of claims 1 to 12.
21 . A method of synthesising a protein, comprising: predicting that a protein comprises a B-cell epitope that is likely to instigate a binding event with an antibody using a method according to any of claims 1 to 12; and synthesising the protein.
22. A protein synthesised using the method of claim 21 .
23. A method of identifying one or more B-cell epitopes that are predicted to instigate a protective immunogenic response across a plurality of species of a pathogen or virus, comprising:
(i) performing the method according to any of claims 6 to 11 for each of a plurality of proteins of a respective plurality of different species of a pathogen or virus, to thereby identify a plurality of first B-cell epitopes that are each predicted to be likely to be a true B-cell epitope on at least one of the plurality of proteins;
(ii) for each identified first B-cell epitope, determining the number of different species of the plurality of species of the pathogen or virus on which the first B-cell epitope is predicted to be a true B-cell epitope; and
(iii) classifying one or more of the first B-cell epitopes as a broadly protective B-cell epitope, based on the number of different species, or the number of variants of any one given species, on which the first B-cell epitope is predicted to be a true B-cell epitope.
24 The method of claim 23, further comprising, for at least one of the first B- cell epitope(s) classified as a broadly protective B-cell epitope: substituting one or more amino acids forming the first B-cell epitope with a different amino acid to thereby generate a modified first B-cell epitope; and, for each of the plurality of proteins of the respective plurality of species of a pathogen of virus: generating a modified amino acid sequence of the protein in accordance with the modified first B-cell epitope; and using the trained first machine learning model to predict whether the candidate encoding representing the first B-cell epitope classified as a broadly protective B-cell epitope is likely to correspond to a true B-cell epitope on the modified amino acid sequence of the protein.
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP23386047.7 | 2023-06-13 | ||
EP23386047.7A EP4478368A1 (en) | 2023-06-13 | 2023-06-13 | B-cell epitope prediction |
EP24165372 | 2024-03-21 | ||
EP24165372.4 | 2024-03-21 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2024256595A1 true WO2024256595A1 (en) | 2024-12-19 |
Family
ID=91433264
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/EP2024/066470 WO2024256595A1 (en) | 2023-06-13 | 2024-06-13 | B-cell epitope prediction |
Country Status (1)
Country | Link |
---|---|
WO (1) | WO2024256595A1 (en) |
Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP0021979A1 (en) | 1979-06-18 | 1981-01-07 | Union Carbide Corporation | Polyarylate blends with copolyesters |
-
2024
- 2024-06-13 WO PCT/EP2024/066470 patent/WO2024256595A1/en unknown
Patent Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP0021979A1 (en) | 1979-06-18 | 1981-01-07 | Union Carbide Corporation | Polyarylate blends with copolyesters |
Non-Patent Citations (55)
Title |
---|
A. A. ABODER: "An empirical hydrophobicity scale for alpha-amino-acids and some of its applications", INTERNATIONAL JOURNAL OF BIOCHEMISTRY, vol. 2, 1971, pages 537 - 544 |
A. S. KOLASKARPRASAD C. TONGAONKAR: "A semi-empirical method for prediction of antigenic determinants on protein antigens", FEBS LETTERS, vol. 276, no. 1-2, 10 December 1990 (1990-12-10), pages 172 - 174 |
BRUNA MOREIRA DA SILVA ET AL: "AI driven B-cell Immunotherapy Design", ARXIV.ORG, CORNELL UNIVERSITY LIBRARY, 201 OLIN LIBRARY CORNELL UNIVERSITY ITHACA, NY 14853, 3 September 2023 (2023-09-03), XP091605103 * |
BUKHARI SYED NISAR HUSSAIN ET AL: "Machine Learning Techniques for the Prediction of B-Cell and T-Cell Epitopes as Potential Vaccine Targets with a Specific Focus on SARS-CoV-2 Pathogen: A Review", vol. 11, no. 2, 24 January 2022 (2022-01-24), pages 146, XP093131196, ISSN: 2076-0817, Retrieved from the Internet <URL:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8879824/pdf/pathogens-11-00146.pdf> DOI: 10.3390/pathogens11020146 * |
C. A. BROWNEH. P. J. BENNETTS. SOLOMON: "The isolation of peptides by high-performance liquid chromatography using predicted elution positions", ANALYTICAL BIOCHEMISTRY, vol. 124, no. 1, 1982, pages 201 - 208, XP024826260, DOI: 10.1016/0003-2697(82)90238-X |
CHARLES. TANFORD: "Contribution of hydrophobic interactions to the stability of the globular conformation of proteins", JOURNAL OF THE AMERICAN CHEMICAL SOCIETY, vol. 84, no. 22, November 1962 (1962-11-01), pages 4240 - 4247 |
CYRUS CHOTHIA: "The nature of the accessible and buried surfaces in proteins", JOURNAL OF MOLECULAR BIOLOGY, vol. 105, no. 1, 1976, pages 1 - 12, XP024011244, DOI: 10.1016/0022-2836(76)90191-1 |
D. EISENBERGE. SCHWARZM. KOMAROMYR.WALL: "Analysis of membrane and surface protein sequences with the hydrophobic moment plot", JOURNAL OF MOLECULAR BIOLOGY, vol. 179, no. 1, 1984, pages 125 - 142 |
DA SILVA BRUNA MOREIRA ET AL: "epitope3D: a machine learning method for conformational B-cell epitope prediction", BRIEFINGS IN BIOINFORMATICS, 17 January 2022 (2022-01-17), England, XP093131857, Retrieved from the Internet <URL:https://oup.silverchair-cdn.com/oup/backfile/Content_public/Journal/bib/23/1/10.1093_bib_bbab423/1/epitope_3d_manuscript_supps_r2_bbab423.pdf?Expires=1710952088&Signature=ECGQoFFLOIhq2u2mx1LUe~0brmiLdi~hI4ut5jm~SRGCti57n2X5pAo46qrNKgO5TJQb4emd5a63lXFRGnAMWgfF8K2L800cGLo0MX3-RUicJn6WHSjXULWy0lgOi-pPq> [retrieved on 20240215], DOI: 10.1093/bib/bbab423 * |
DA SILVA BRUNA MOREIRA ET AL: "epitope3D: a machine learning method for conformational B-cell epitope prediction", vol. 23, no. 1, 17 January 2022 (2022-01-17), GB, XP093131765, ISSN: 1467-5463, Retrieved from the Internet <URL:https://academic.oup.com/bib/article-pdf/23/1/bbab423/42230615/bbab423.pdf> DOI: 10.1093/bib/bbab423 * |
DANIEL D. JONES: "Amino acid properties and side-chain orientation in proteins", A CROSS CORRELATION APPROACH, vol. 50, April 1975 (1975-04-01) |
DAVID L. (DAVID LEE) NELSONMICHAEL M. COX: "Lehninger principles of biochemistry. Principles of biochemistry", 2017, W.H.FREEMAN |
DISCOTOPE 3.0, Retrieved from the Internet <URL:https://doi.org/10.1101/2023.02.05.527174> |
DONALD J. ABRAHAMALBERT J. LEO: "Extension of the fragment method to calculate amino acid zwitterion and side chain partition coefficients", PROTEINS: STRUCTURE, FUNCTION, AND BIOINFORMATICS, vol. 2, no. 2, 28 February 2019 (2019-02-28), pages 130 - 152 |
E A EMINIJ V HUGHESD S PERLOWJ BOGER: "Induction of hepatitis a virus-neutralizing antibody by a virus-specific synthetic peptide", JOURNAL OF VIROLOGY, vol. 55, no. 3, September 1985 (1985-09-01), pages 836 - 839 |
F. PEDREGOSAG. VAROQUAUXA. GRAMFORTV. MICHELB. THIRIONO. GRISELM. BLONDELP. PRETTENHOFERR. WEISSV. OUBOURG: "Scikit-learn: Machine learning in Python", JOURNAL OF MACHINE LEARNING RESEARCH, vol. 12, 2011, pages 2825 - 2830 |
G JOHNSONT T WU: "Kabat database and its applications: 30 years after the first variability plot", NUCLEIC ACIDS RESEARCH, vol. 28, no. 1, January 2000 (2000-01-01), pages 214 - 218, XP009132432 |
G. DELEAGEB. ROUX: "An algorithm for protein secondary structure prediction based on class prediction", PROTEIN ENGINEERING, DESIGN AND SELECTION, vol. 1, no. 4, August 1987 (1987-08-01), pages 289 - 294 |
GANG ZHAOERWIN LONDON: "An amino acid ''transmembrane tendency'' scale that approaches the theoretical limit to accuracy for prediction of transmembrane helices: relationship to biological hydrophobicity", PROTEIN SCIENCE : A PUBLICATION OF THE PROTEIN SOCIETY, vol. 15, no. 8, August 2006 (2006-08-01), pages 1987 - 2001 |
GD ROSEAR GESELOWITZGJ LESSERRH LEEMH ZEHFUS: "Hydrophobicity of amino acid residues in globular proteins", SCIENCE, vol. 229, no. 4716, August 1985 (1985-08-01), pages 834 |
GJALT W. WELLINGWICHER J. WEIJERRUURD VAN DER ZEESYTSKE WELLING-WESTER: "Prediction of sequential antigenic regions in proteins", FEBS LETTERS, vol. 188, no. 2, 28 February 2019 (2019-02-28), pages 215 - 218, XP025607188, DOI: 10.1016/0014-5793(85)80374-4 |
H R GUY: "Amino acid side-chain partition energies and distribution of residues in soluble proteins", BIOPHYS J, vol. 47, no. 1, January 1985 (1985-01-01), pages 61 - 70 |
HENRY B. BULLKEITH BREESE: "Surface tension of amino acid solutions: A hydrophobicity scale of the amino acid residues", ARCHIVES OF BIOCHEMISTRY AND BIOPHYSICS, vol. 161, no. 2, 1974, pages 665 - 670, XP024757834, DOI: 10.1016/0003-9861(74)90352-X |
J L MEEK: "Prediction of peptide retention times in high-pressure liquid chromatography on the basis of amino acid composition", PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, vol. 77, no. 3, March 1980 (1980-03-01), pages 1632 - 1636 |
J L. FAUCHEREV PLISKA, HYDROPHOBIC PARAMETERS II OF AMINO ACID SIDE-CHAINS FROM THE PARTITIONING OF N-ACETYL-AMINO ACID AMIDES, vol. 18, January 1983 (1983-01-01) |
J. K. MOHANA RAOPATRICK ARGOS: "A conformational preference parameter to predict helices in integral membrane proteins", BIOCHIMICA ET BIOPHYSICA ACTA (BBA) - PROTEIN STRUCTURE AND MOLECULAR ENZYMOLOGY, vol. 869, no. 2, 1986, pages 197 - 214, XP025221541, DOI: 10.1016/0167-4838(86)90295-5 |
J. M. R. PARKERD. GUOR. S. HODGES: "New hydrophilicity scale derived from high-performance liquid chromatography peptide retention data: correlation of predicted surface residues with antigenicity and x-ray-derived accessible sites", BIOCHEMISTRY, vol. 25, no. 19, pages 5425 - 5432, XP001062495, DOI: 10.1021/bi00367a013 |
J.M. ZIMMERMANNAOMI ELIEZERR. SIMHA: "The characterization of amino acid sequences in proteins by statistical methods", JOURNAL OF THEORETICAL BIOLOGY, vol. 21, no. 2, 1968, pages 170 - 201 |
JACK KYTERUSSELL F. DOOLITTLE: "A simple method for displaying the hydropathic character of a protein", JOURNAL OF MOLECULAR BIOLOGY, vol. 157, no. 1, 1982, pages 105 - 132 |
JIAN YENING MATHOMAS L MADDENJAMES M OSTELL: "Igblast: an immunoglobulin variable domain sequence analysis tool", NUCLEIC ACIDS RES, vol. 41, July 2013 (2013-07-01), pages W34 - 40 |
JOEL JANIN: "Surface and inside volumes in globular proteins", NATURE, vol. 277, pages 491 |
K JWILSONA HONEGGERR P STOTZELG J HUGHES: "The behaviour of peptides on reverse-phase supports during high-pressure liquid chromatography", THE BIOCHEMICAL JOURNAL, vol. 199, no. 1, October 1981 (1981-10-01), pages 31 - 41 |
LI TONG ET AL: "Artificial intelligence in cancer immunotherapy: Applications in neoantigen recognition, antibody design and immunotherapy response prediction", SEMINARS IN CANCER BIOLOGY, SAUNDERS SCIENTIFIC PUBLICATIONS, PHILADELPHIA, PA, US, vol. 91, 3 March 2023 (2023-03-03), pages 50 - 69, XP087289623, ISSN: 1044-579X, [retrieved on 20230303], DOI: 10.1016/J.SEMCANCER.2023.02.007 * |
M. COOPERROBERT E. HAUSMAN: "The cell : a molecular approach", vol. 4, 2007, DC ASM PRESS |
MARK A. ROSEMAN: "Hydrophilicity of polar amino acid side-chains is markedly reduced by flanking peptide bonds", JOURNAL OF MOLECULAR BIOLOGY, vol. 200, no. 3, 1988, pages 513 - 522, XP024019376, DOI: 10.1016/0022-2836(88)90540-2 |
MICHAEL LEVITT: "Conformational preferences of amino acids in globular proteins", BIOCHEMISTRY, vol. 17, no. 20, October 1978 (1978-10-01), pages 4277 - 4285 |
P Y CHOUG D FASMAN: "Prediction of the secondary structure of proteins from their amino acid sequence", ADV ENZYMOL RELAT AREAS MOL BIOL, vol. 47, 1978, pages 45 - 148 |
P. MANAVALANP. K. PONNUSWAMY: "Hydrophobic character of amino acid residues in globular proteins", NATURE, vol. 275, no. 5681, 1978, pages 673 - 674 |
PETER JA COCKTIAGO ANTAOJEFFREY T CHANGBRAD A CHAPMANCYMON J COXANDREW DALKEIDDO FRIEDBERGTHOMAS HAMELRYCKFRANK KAUFFBARTEK WILCZY: "Biopython: freely available python tools for computational molecular biology and bioinformatics", BIOINFORMATICS, vol. 25, no. 11, 2009, pages 1422 - 1423 |
PETER MCCALDONPATRICK ARGOS: "Oligopeptide biases in protein sequences and their use in predicting protein coding regions in nucleotide sequences", PROTEINS: STRUCTURE, FUNCTION, AND BIOINFORMATICS, vol. 4, no. 2, 28 February 2019 (2019-02-28), pages 99 - 122 |
R. BHASKARANP. K. PONNUSWAMY: "Positional flexibilities of amino acid residues in globular proteins", INTERNATIONAL JOURNAL OF PEPTIDE AND PROTEIN RESEARCH, vol. 32, no. 4, 1988, pages 241 - 255 |
R. GRANTHAM: "Amino acid difference formula to help explain protein evolution", SCIENCE, vol. 862, no. 4154, pages 09 1974 |
R. WOLFENDENL. ANDERSSONP. M. CULLISC. C. B. SOUTHGATE: "ties of amino acid side chains for solvent w", BIOCHEMISTRY, vol. 20, no. 4, February 1981 (1981-02-01), pages 849 - 855 |
REN JING ET AL: "Positive-unlabeled learning for the prediction of conformational B-cell epitopes", 11 September 2015 (2015-09-11), XP093197190, Retrieved from the Internet <URL:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4682424/pdf/1471-2105-16-S18-S12.pdf> * |
RICHARD COWANR GWHITTAKER, HYDROPHOBICITY INDICES FOR AMINO ACID RESIDUES AS DETERMINED BY HPLC, vol. 3, January 1990 (1990-01-01) |
ROBERT M. SWEETDAVID EISENBERG: "Correlation of sequence hydrophobicities measures similarity in three-dimensional protein structure", JOURNAL OF MOLECULAR BIOLOGY, vol. 171, no. 4, 1983, pages 479 - 488, XP024015068, DOI: 10.1016/0022-2836(83)90041-4 |
SANZO MIYAZAWAROBERT L. JERNIGAN: "Estimation of effective interresidue contact energies from protein crystal structures: quasi-chemical approximation", MACROMOLECULES, vol. 18, no. 3, March 1985 (1985-03-01), pages 534 - 552, XP055831084, DOI: 10.1021/ma00145a039 |
SCHWARTZ R.M.DAYHOFF, M.OB.C. ORCUTT: "A model of evolutionary change in proteins", ATLAS OF PROTEIN SEQUENCE AND STRUCTURE. NATL. BIOMED. RES. FOUND., vol. 5, 1978 |
SERAFIN FRAGA: "Theoretical prediction of protein antigenic determinants from amino acid sequences", CANADIAN JOURNAL OF CHEMISTRY, vol. 60, no. 20, 1 March 2019 (2019-03-01), pages 2606 - 2610 |
SHAUN D. BLACKDIANE R. MOULD: "Development of hydrophobicity parameters to analyze proteins which bear post- or cotranslational modifications", ANALYTICAL BIOCHEMISTRY, vol. 193, no. 1, 1991, pages 72 - 82, XP024823173, DOI: 10.1016/0003-2697(91)90045-U |
SHNEIOR LIFSONCHRISTIAN SANDER: "Antiparallel and parallel -strands differ in amino acid residue preferences", NATURE, vol. 282, no. 5734, 1979, pages 109 - 111 |
T P HOPPK R WOODS: "Prediction of protein antigenic determinants from amino acid sequences", PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, vol. 78, no. 6, June 1981 (1981-06-01), pages 3824 - 3828 |
VARDAXIS IOANNIS ET AL: "Deep learning of antibody epitopes using molecular permutation vectors", 21 March 2024 (2024-03-21), XP093197296, Retrieved from the Internet <URL:https://www.biorxiv.org/content/biorxiv/early/2024/03/21/2024.03.20.585661.full.pdf> DOI: 10.1101/2024.03.20.585661 * |
WEN ZHANG ET AL: "Prediction of conformational B-cell epitopes from 3D structures by random forests with a distance-based feature", BMC BIOINFORMATICS, BIOMED CENTRAL , LONDON, GB, vol. 12, no. 1, 17 August 2011 (2011-08-17), pages 341, XP021108658, ISSN: 1471-2105, DOI: 10.1186/1471-2105-12-341 * |
WOLFGANG KABSCHCHRISTIAN SANDER: "Dictionary of protein secondary structure: Pattern recognition of hydrogen-bonded and geometrical features", BIOPOLYMERS, vol. 22, no. 12, 30 November 2018 (2018-11-30), pages 2577 - 2637, XP055076173, DOI: 10.1002/bip.360221211 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Bibi et al. | In silico analysis of epitope-based vaccine candidate against tuberculosis using reverse vaccinology | |
Potocnakova et al. | An introduction to B‐cell epitope mapping and in silico epitope prediction | |
Akbar et al. | Progress and challenges for the machine learning-based design of fit-for-purpose monoclonal antibodies | |
Manavalan et al. | iBCE-EL: a new ensemble learning framework for improved linear B-cell epitope prediction | |
Halperin et al. | Exploring antibody recognition of sequence space through random-sequence peptide microarrays | |
Sela-Culang et al. | Antibody specific epitope prediction—emergence of a new paradigm | |
Kringelum et al. | Reliable B cell epitope predictions: impacts of method development and improved benchmarking | |
Wang et al. | Prediction of B‐cell linear epitopes with a combination of support vector machine classification and amino acid propensity identification | |
Sela-Culang et al. | Using a combined computational-experimental approach to predict antibody-specific B cell epitopes | |
Hegde et al. | The use of databases, data mining and immunoinformatics in vaccinology: where are we? | |
Dalkas et al. | SEPIa, a knowledge-driven algorithm for predicting conformational B-cell epitopes from the amino acid sequence | |
EP4182928B1 (en) | Method, system and computer program product for determining presentation likelihoods of neoantigens | |
Kumar et al. | Multi‐perspectives and challenges in identifying B‐cell epitopes | |
Xu et al. | NetBCE: an interpretable deep neural network for accurate prediction of linear B-cell epitopes | |
Kim et al. | Epitopes of protein binders are related to the structural flexibility of a target protein surface | |
Khan et al. | Analysis of viral diversity for vaccine target discovery | |
Shatnawi | Review of recent protein-protein interaction techniques | |
EP4478368A1 (en) | B-cell epitope prediction | |
CN113762416B (en) | Antigen immunogenicity prediction method and system based on multi-modal depth coding | |
Mubarak et al. | Enhancing tuberculosis vaccine development: a deconvolution neural network approach for multi-epitope prediction | |
Costa et al. | Evaluation and comparison of the ability of online available prediction programs to predict true linear B-cell epitopes | |
Huang et al. | Prediction of linear B-cell epitopes of hepatitis C virus for vaccine development | |
Hooshmand et al. | Prediction of B cell and T‐helper cell epitopes candidates of bovine leukaemia virus (BLV) by in silico approach | |
EP4478367A1 (en) | B-cell epitope prediction | |
Basmenj et al. | Computational epitope-based vaccine design with bioinformatics approach; a review |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 24731619 Country of ref document: EP Kind code of ref document: A1 |