Informational Way to Protein Alphabet: Entropic Classification ... - arXiv

3 downloads 0 Views 285KB Size Report
classifications (Kawashima et al., 2000) and the most popular of them hydrophobic- ity/polarity binary classification revealed many interesting patterns of ...
Informational Way to Protein Alphabet: Entropic Classification of Amino Acids

Alexander Gorban1,2*, Mikhail Kudryashev3, Tatiana Popova2 1

Centre for Mathematical Modeling, University of Leicester, UK, E-mail: [email protected], Institute of Computational Modeling, Akademgorodok, Krasnoyarsk, Russia; 3 Krasnoyarsk State University, Krasnoyarsk, Russia. 2

ABSTRACT Motivation: What are proteins made from, as the working parts of the living cells protein machines? To answer this question, we need a technology to disassemble proteins onto elementary functional details and to prepare lumped description of such details. This lumped description might have a multiple material realization (in amino acids). Our hypothesis is that informational approach to this problem is possible. We propose a way of hierarchical classification that makes the primary structure of protein maximally non-random. The first steps of the suggested research program are realized: the method and the analysis of optimal informational protein binary alphabet. The general method is used to answer several specific questions, for example: • Is there a syntactic difference between Globular and Membrane proteins? • Are proteins random sequences of amino acids (a long discussion)? For these questions, the answers are as follows: • There exists significant syntactic difference between Globular and Membrane proteins, and this difference is described; • Amino acid sequences in proteins are definitely not random. 1 INTRODUCTION Protein is the main engineering material of living cells. The idea of “protein-machine” approach to living cells biophysics was suggested in 1967 (see the review written by the authors of this idea and devoted to its 20th birthday (Chernavskii et al., 1987)). The kinetic approach to protein-machines was reported in (Kurzynski, 1998). The experimentally justified models of conformational transition dynamics are described and applied to the interpretation of a few simple biochemical processes. The engineering approach to protein-machines based on energetic classification of elements was discussed in (Alberts, 1998). Engineers recognize certain fundamental behaviors in nature and then create an idealized element to represent each of those behaviors. The heart of such methods is the simplification and idealization of a real world machine as a composition of discrete elements. Any particular part of a machine might be modeled as consisting of one or more of these basic constituent elements. We are in desperate need of a tool for automatic disassembling of protein machines onto elementary functional details. These functional details should have a lumped description that is maximally independent of the materials: a wheel is the wheel, and the range of material variations can be very wide. The definition of a functional detail is predetermined by its functional properties: flexibility or rigidity, affinity, etc. These properties should be described for a given function, and the set of realizations of the described properties in material (amino acids, for example) should be prepared. This is a long honest way, a physical and chemical description of protein *

To whom correspondence should be addressed.

machines functional details. Our hypothesis is that a “King’s way” is possible for this description. It is the entropic (informational) classification (Bugaenko et al., 1998b). Now, the huge amount of protein primary structures is known. We can assume that nonrandom differences and similarities between these structures are determined by theirs functions. We should just extract these typical differences and similarities, and prepare the collection (a warehouse in a form of a dictionary) of these details and its material realization. In this paper we start from the amino acid alphabet reduction and try to find the rigorous statement and the exact answer to the question: in which alphabet the protein primary structure is the maximally non-random sequence. The following steps will consist in analysis of q-letter (q = 2, 3, 4) fragments of proteins written in reduced alphabet with the same question and further the q-letter fragments alphabet reduction to consider recoded protein again thus increasing words length and extracting functional details. Amino acid classifications and alphabet reduction are widely used in various applications. For example, physical and chemical properties of amino acids provide a number of natural classifications (Kawashima et al., 2000) and the most popular of them hydrophobicity/polarity binary classification revealed many interesting patterns of secondary structure preferences (see recent (Hennetin et al., 2003, Phoenix et al., 2002) and rather old (Eisenberg et al., 1984, West et al., 1995) papers). Recently the idea of amino acid alphabet reduction is developed in two directions. The first one is related to protein design and the main question is: How many amino acids are sufficient to protein folding and function in a correct way (see review (Chan, 1999)). In the second direction we unite various theoretical approaches for amino acid grouping based on substitution matrices, secondary structure, pattern conservation and so on. These methods aim at simplification of amino acid sequences in order to solve some problems of sequence analysis resulted from alphabet size, such as revealing distinct homologs, secondary structure prediction, folds recognition and so on. Unlike protein design approach amino acids are to be classified (not selected) to reduce alphabet size. Theoretical groupings of amino acids mentioned in literature may be attributed to the following main bases: (1) physical, chemical properties (Sirabella et al., 1998) and amino acid environment in proteins (Liu et al., 2003); (2) protein alignments (Irback et al., 1997) and substitution matrices (Liu et al., 2002, Murphy et al., 2002, Li et al, 2003); (3) protein spatial structure (Chen et al., 2002, Solis et al., 2002) and contact potential matrix (Li et al., 1997, Wang et al., 1999, Wang et al., 2000, Wang et al., 2002, Fan et al., 2003, Cieplak et al., 2001); (4) information theory (Bugaenko et al., 1998, Mintseris et al., 2004, Solis et al., 2002). One should understand that pattern conservation, substitution matrices, spatial structures and contact potentials are dependent on physicochemical properties of amino acids, that is why, as it was shown in papers mentioned above, all amino acid classifications based on the famous amino acid related matrices such as BLOSSUM, PAM or MJ-matrix correlate with hydrophobicity/polarity scale. In this paper we consider the method of informational classification of amino acids that have been proposed in (Bugaenko et al., 1998b) for classification of amino acids in protein sequences and recently it was used in (Mintseris et al., 2004) for classification of proteins heavy atoms on the basis of statistical contact potential matrix. The notion of informational classification relates to entropic optimality principle that may have various formulations while the main idea consists in selection such classification that maximizes nonrandomness. Method of informational classification has universal nature: it can reduce any alphabet on the basis of q-letter statistic of some language objects. For example, there was spatial contacts statistic in (Mintseris et al., 2004) or linear contacts statistic in (Bugaenko et al., 1998b).

Informational classifications based on entropic optimality criterion were mentioned in a number of papers while fundamental properties of result such as stability, existence and uniqueness of optimal classifications are still not considered at all. Further we present: (1) complete formulation of optimality principle, (2) application of the method to amino acid classification on the basis of protein sequences, (3) regular analysis of results obtained by applying the method to various groups of proteins, words lengths, and so on, (4) comparison of some reduced amino acid alphabets with informational ones. 2 INFORMATIONAL CLASSIFICATIONS 2.1 Entropic optimality principle The problem of amino acid classification into k groups may be formalized as follows: given a set of 20 amino acids and a number of groups k (1 < k < 20) one should find a mapping f of f amino acids onto groups {A,C,D,E,F,G,H,I,K,L,M,N,P,Q,R,S,T,V,W,Y} ⎯⎯→ {1, …, k} so that every amino acid is assigned to one group and some optimality condition holds. The notion of informational classification relates to the entropic optimality condition. Similar to (Bugaenko et al., 1998b, Mintseris et al., 2004, Gorban et al, 2005) we consider amino acid mapping to be optimal if recorded objects have as informative distribution as possible. Here “informative” means far from random while random distribution can be determined in various ways. Using well-known notion of relative entropy the optimality condition looks as follows. Optimal informational classification provides maximal relative entropy of distribution of recorded objects:

f : D( P | P*) =

P( X f )

∑ P( X f ) ln P * ( X Xf

f

)

→ max ,

(1)

where Xf represents recoded objects, P(Xf) is real distribution, and P*(Xf) is some reference (“random”) distribution of Xf. Relative entropy is non-negative and D( P | P*) = 0 iff P = P* (that means coincidence of real and reference distributions.) D( P | P*) is the convex functional of P, and, therefore, has one zero minimum, but can have many local maxima. It is worth to mention that the relative entropy in (1) measures additional information in the distribution P with respect to the distribution P*. In other words, it is the indicator of the relative order in P against a background of P*. The reference distribution in (1) generalizes criteria introduced in (Bugaenko et al., 1998b, Mintseris et al., 2004, Gorban et al, 2005). X may represent any structural objects of corresponding language. While any classification f determines new alphabet and therefore a set of recoded objects Xf with distribution P(Xf). If one consider protein sequences than X are sequence fragments of some length q; if one consider protein spatial structure than X are the set of amino acid contacts in folded protein (like in (Mintseris et al., 2004)) and so forth. 2.2 Informational classification of amino acids Here we consider amino acid classification on the basis of protein primary structure so we use q-letter words distribution and reconstructed q-letter words distribution as reference one. Reconstructed distribution is calculated according to some model of sequence generation. For example the most popular reconstruction of frequencies of q-letter words (q > 1) is simple multiplication of all q letter frequencies. Markov model can be used as well. However we kept in mind the idea of the maximum entropy principle for frequency reconstruction proposed in (Bugaenko et al., 1998a). According to it reconstructed q-letter words distribution

should have maximum entropy among all q-letter words distributions that could be obtained from the given (q-s)-letter word distribution. Thus, one can get reconstructed theoretical qletter words frequencies on the basis of any less length words distribution. Relative entropy of real distribution with respect to reconstructed one shows how much information is introduced by q-letter words compared to (q-s)-letter words (Gorban et al., 2001). Hence alphabet reduction that maximizes (1) makes the most nonrandom q-letter words distribution given (q-s)-letter words distribution. In more detail, given some protein sequence or a set of sequences one can obtain a vector of frequencies (or frequency dictionary) of protein fragments of the length q (denoted here Wq) According to some classification of amino acids into k groups the Wq can be easily transformed to vector of recoded words frequencies RWq by simple summation. Having chosen the type of reference distribution one can obtain reconstructed frequencies for the same ~ length q RWq . Relative entropy is calculated as follows:

~ D( RW q | RW q ) =



i1 ,i2 ,..iq ∈{1..k }

f i1i2 ...iq f i1i2 ...iq ln ~ . f i1i2 ...iq

(2)

~

Optimal amino acids classification corresponds to maximal value of D ( RWq | RWq ) for a given number of classes k. Reconstructed frequencies according to (Bugaenko et al., 1998a) are calculated as follows. f i1...iq − s ⋅ f i2 ...iq − s +1 ⋅ ... ⋅ f is +1...iq ~ , for q - s > 1 f i1...iq − s ..iq = f i2 ...iq − s ⋅ f i3 ...iq − s +1 ⋅ ... ⋅ f is +1...iq −1

and

(3)

~ f i1...iq = f i1 ⋅ f i2 ⋅ ... ⋅ f iq , for q - s = 1, where iq are symbols of alphabet, i1…iq are q-letter words and f i1...iq are frequencies of corresponding words in the symbol sequence. Such formulation gives a set of reference distributions to use in (2) dependent on the words length q and on the “depth” of reconstruction s. Thus optimal informational classification of alphabet symbols produces a recoded sequence with maximal nonrandomness in frequencies of pairs with respect to symbol frequencies or maximal nonrandomness in frequencies of triplets with respect to pairs and so on in any reasonable combination. In this paper we always use reconstruction on the basis of single letter frequencies (the case q - s = 1 in (3)), its worth to note that this calculation coincides with zero order Markov model (independent letter generation model). 2.3 Binary informational classifications Consider the problem of informational binary classification of amino acids based on 2-letter word distribution calculated for some set of proteins. The problem of binary classification consists in separation of amino acids into two groups: {A,C,D,E,F,G,H,I,K,L,M,N,P,Q,R,S,T,V,W,Y}→{0,1}. Initial vectors of frequencies W1 and W2 are calculated for the set of protein sequences and then they are recalculated to binary ones by simple summation according to current amino acid grouping:

W1 = {fA, fC,…, fY} → RW1 = {f0, f1} W2 = {fAA, fAC,…, fYY} → RW2 = {f00, f01, f10, f11} Reconstructed frequencies for q = 1 are calculated as follows: ~ f ij = f i ⋅ f j , i, j = 0, 1 ,

~ ~ ~ ~ ~ RW2 = { f 00 , f 01 , f 10 , f 11} . Following some scheme of enumeration of amino acid groupings, one calculates the relative entropy

~ D( RW 2 | RW 2 ) =



i1 ,i2 ∈{0,1}

fi i f i1i2 ln ~1 2 f i1i2

(4)

for each grouping and chooses the most informative one. We used the simplest binary problem in order to describe some important things about amino acid classification, such as existence and uniqueness of global maximum of (2), its stability dependent on frequency distribution and so on. In order to achieve global maximum (2) an exhaustive search is needed. We used quasirandom optimization that can be specified as one-step directed descent with a number of random starts. The method consists in following steps: 1) choosing some random classification as a start point and calculating relative entropy (2), 2) calculating relative entropy for all one-letter neighbor classifications, 3) choosing amino acid grouping that produce maximal increment in relative entropy as the next start point, 4) going on to the step 2 until any increment in relative entropy exists. We repeat procedure with 100 random start points and choose the best classification. Method showed more than 99% hit to the global maximum (for 116 protein sequences method produced 115 optimal classifications). 3 DATASETS The method of entropic classification may reduce any alphabet on the basis of some symbol sequence (or sequences) written in this alphabet. As far as amino acids concern, every single protein sequence could have its own optimal informational classification of amino acids that makes it at maximum non-random given word length and alphabet size. Any set of protein sequences could have its own classification of amino acids and so on. There exist some natural types of protein sets which can be considered as a basis of classification: protein families (superfamilies), proteins with similar function, structure or subcellular localization, proteomes, all known protein sequences as a whole and so on. Table 1. The data sets of protein sequences

Dataset 1 (EBI)

Dataset 2 (SwissProt)

Keywords Numb. of proteins Keywords Numb. of proteins Oxidoreductase 452 Transferase 500 Cytochrome 500 Isomerase 578 DNA polyPhytochrome 500 500 merase Oxidase 500 Nitratoreductase 197 ATPase 500 Membrane

10000

Globular

5019

We used several sets of proteins extracted by keyword from databases. Table 1 shows two datasets with the number of protein sequences extracted: Dataset 1 and Dataset 2 were formed according to the keyword in the field Definition from EBI (http://www.ebi.ac.uk) and Swissprot (http://expasy.org) databases correspondingly. As one can see Dataset 1 was completed by sets of proteins with similar function and sets of Dataset 2 can be characterized to have similar structure and subcellular localization. These protein sets were natural choice to elaborate the properties of informational classifications of amino acids. 4 RESULTS AND DISCUSSION 4.1 Informational classification for large groups of proteins Optimal informational binary classifications of amino acids were calculated for sets of proteins as follows. Vectors of frequencies of 2- and 1-letter words were counted through whole sets of proteins and optimal binary classifications were obtained (Tables 2). Table 2. Binary informational classifications for Dataset 1 and 2 Protein dataset ATPase Cytochrome Nitratoreductase Oxidase DNA polymerase Isomerase Transferase Phytochrome Oxidoreductase Globular Membrane

Dmax 0.002 0.0066 0.0027 0.0029 0.0007 0.0006 0.0006 0.0074 0.0024 0.0006 0.0025

A 1 1 1 1 1 1 1 1 1 1 1

C 1 1 0 1 0 0 0 1 1 0 1

D 0 0 0 0 0 1 1 1 0 0 0

E 0 0 0 0 1 1 1 1 0 1 0

F 1 1 1 1 0 0 0 0 0 0 1

G 1 1 1 1 0 0 0 0 1 0 1

H 1 1 1 1 0 0 0 1 0 0 0

I 1 1 1 1 0 1 1 1 0 0 1

K 0 0 0 0 1 1 1 1 1 1 0

L 1 1 1 1 0 1 1 0 1 1 1

M N 1 0 1 0 1 0 1 0 1 0 0 0 1 0 1 1 0 1 1 0 1 0

P 0 0 0 0 0 1 0 0 1 0 1

Q 0 1 0 0 0 1 0 0 0 1 0

R 0 0 0 0 1 1 1 0 1 1 0

S 1 1 1 1 0 0 0 0 1 0 1

T 1 1 1 1 0 0 0 0 0 0 1

V 1 1 1 0 0 1 0 1 1 0 1

WY 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0

As one can see, considered sets of proteins provide different {0, 1} amino acid informational classifications. However more careful consideration shows that there exist similar classifications as well as absolutely different ones. Moreover, the values of maximal relative entropy are close to zero. Zero minimum corresponds to equal real and calculated frequencies of 2-letter words. It means that even in optimal binary classification amino acid sequences are close to random ones if 2-letter words are considered. Our example with Bible (see section 4.4) shows that for English language it is not a case. Here binary classification provides significant value of relative entropy. 4.2 Properties of informational classifications of amino acids 4.2.1 Distribution. Analysis of distribution of binary classifications by relative entropy value shows that there exists one global maximum of relative entropy in the space of all possible binary classifications (Fig.1). The distribution is close to exponential and there are about 10 (as a rule similar) classifications in 90-100% interval of maximal relative entropy for all tested datasets.

number of classifications 1000000 100000 10000 1000 100 10 1 0

50%

100%=0.0066

relative entropy

Fig. 1. Distribution of relative entropy value for all possible binary classifications of amino acids for Cytochrome dataset 4.2.2 Correlation structure. Correlation structure of binary classifications distribution by relative entropy value can become clear by constructing the system of successive “orthogonal” binary classifications. The first optimal classification is chosen to provide maximum of functional (4). The second classification corresponds to constrained maximum of (4) with constrains introduced by the first one and so on. Here we considered two possible types of constraints imposed on new binary classification in hierarchy: 1) it should have Hamming distance equal to 10 to any previous one, and 2) it should have zero Pearson correlation with any previous one (actually, | r | < 0.1). We kept in mind the idea of orthogonality and independency while introducing these constraints. As one can see the system of “orthogonal” classifications can be considered as an analog of principal components method in the problem of depletion of information in symbol sequences: the first classification produce maximal nonrandomness in words distribution, the second one should be independent of the first one and produce maximal nonrandomness and so on. re lative e ntropy 0.008 0.007 0.006 0.005 0.004 0.003 0.002 0.001 0 1

2

Membrane

3

4

5

Globular

6

7 8 9 10 11 orthogonal clas s ifications Phytochrome

Cytochrome

Fig.2. Relative entropy decrease for hierarchical orthogonal classifications for some datasets with Hamming distance constraints. The graph of the relative entropy decrease (Fig.5) shows correlation structure of distributions. For example, one (Cytochrome dataset) and more than two (Phytochrome dataset) independent classifications are present in the top of relative entropy value. 4.2.3 Classification stability. Informational classifications seem to be very sensible to marginal changes in words frequencies. Amino acid frequency profiles seem rather similar for protein sets of Dataset 1 and 2 shown in Fig.3. The same situation is for 2-letter words dis-

tribution. However, optimal classifications vary from similar to absolutely different ones. This kind of instability is rather natural property of optimization in the discrete space. Normalised frequency

0.14 0.12 0.1 0.08 0.06 0.04 0.02 0 A C D E F G H

Globular

I

K

L M N P Q R S

Membrane

Others

T V W Y Amino acids

Fig. 3. Amino acid frequencies in considered sets of proteins From the other side similarity of frequencies should provide high relative entropy of {0,1} recorded protein sequences following any optimal classification. However, it is not always a case. Relative entropy value as well can vary significantly: for some dataset there exists more than 10 times difference in relative entropy (see table 1). And, for example, Phytochrome and Oxidoreductase dataset in any other optimal classification display poor informativity (in average 3.5% and 5% of their maximal value). 4.2.4 q = 2,3,4,5. Binary informational classifications of amino acids on the basis of 3,4,5letter word frequencies are obtained in the same way as for q = 2. The reference distribution is calculated according to (3) for q –s = 1, i.e. word frequencies are reconstructed on the basis of amino acid frequencies. Five protein sets from datasets 1 have equal classifications for all basis word lengths. Other datasets have few equal classifications among the 4 length compared. It means that informational classification that maximizes nonrandomness in two letter words distribution produces maximal or near maximal nonrandomness in 3,4,5-letter words distribution. The {0,1} words frequency profile of recorded datasets appeared similar too. For example, among all 16 4-letter words there is one with prevailed frequency for all datasets: this is 1111 word. It means that maximally non random {0,1} recoded proteins have the structure of repeated homogeneous series. This fact looks promising for revealing common structural elements in protein sequences. 4.2.4 Correlation with some natural classifications. There exists many well known physicochemical properties of amino acids and each of them may serve as a basis of amino acid classification. Here we consider five ones with three various versions of hydrophobicity/polarity (HP), big/small (BS), and charged/uncharged (CU) classifications (Table 3). Table 3. Some natural amino acids binary classifications Property HP I HP II HP III BS CU

A C D E F GH I K LMN P Q R S T 0/ 1 1 0 0 1 1 0 1 0 1 1 0 1 0 0 0 1 1 1 0 0 1 1 0 1 0 1 1 0 0 0 0 0 0 1 1 0 0 1 0 1 1 0 1 1 0 0 0 0 0 0 0 0 0 1 1 0 1 1 1 1 1 0 0 1 1 0 0 0 0 1 1 0 0 1 0 1 0 0 0 0 0 1 0 0

V W Y Source 0/ 1 1 1 1 1 1 1 2 1 1 1 3 0 1 1 1 0 0 0 1

Source: 1. http://en.wikipedia.org/wiki/Amino_acid; 2. (Eisenberg et al.,,1984); 3. (Cid et al., 1992) Direct comparison of optimal informational binary classifications and natural binary classifications shows similarity between Hydrophobic / Polar classification and informational ones for ATPase, Cytochrome, Nitratoreductase, Oxidase and Membrane datasets (r ≈ 0.7). The same classifications have moderate similarity with Charged / Uncharged binary pattern (r ≈ 0.5). The detailed results for Dataset 2 are shown in Table 6. 4.3 Globular vs Membrane comparison Consider informational classification for Globular (G) and Membrane (M) datasets in more details. We got two amino acid groupings: G: {A,E,K,L,M,Q,R} and {C,D,F,G,H,I,N,P,S,T,V,W,Y}, M: {D,E,H,K,N,Q,R,W,Y} and {A,C,F,G,I,L,M,P,S,T,V}. Hamming distance for these classifications is equal to 8. It means that they are practically independent (orthogonal) Amino acid frequency profiles for these datasets are similar (fig. 3). However, classifications are rather different. Membrane binary classification is nicely correlated with Hydrophbic/Polar I and Charged/Uncharged classification (r=0.7, and 0.63, Hamming distance=2 and 4). Globular classification has moderate correlation with Big/Small property (r=0.45, Hamming distance=6). So, approximate coincidence of Hydrophbic/Polar and informational classification for Membrane dataset shows principal importance of that feature for these proteins structural organization. This is not surprising. But it seems more important that for globular proteins the main amino acid classification is very specific and has almost no similarity to usual classifications (only the size of amino acid matters slightly). 4.4 En the netennent ten the Tent, ent the Tent ten teth Tet, ent the Tent ten Tet Method of informational classification is rather universal: having some text one can reduce its alphabet and the text written in reduced alphabet would be at most non-random. For example, we consider the text of King James Bible http://bible.cc written in English language. 2-letter frequencies were calculated over the whole words. Gaps and numbers were ignored. Reference frequencies were calculated as product of symbol frequencies. Optimal classifications of alphabet symbols in 2-6 groups are shown in Table 4. Table 4. Entropic classification of letters for English language in Bible text Relative Entropy 0.767926 0.934107 1.096432 1.171895 1.227138

Groups 1 aeioudgt aeiou aeiou aeiou aeiou

2 Bcfhjklmnpqrsvwxyz Bcdfgklmnpqrstvwxyz Bcfklmnpqrsvxz Bcfklmprsvxyz Bcfklmpqrsvxyz

3 hj hj hj hj

4 dgtwy dgqtw n tw n

5

6

dg

The most frequent letter is written in bold underlined according to standard frequency table of English language letters (Lewand et al., 2002). Surprisingly, letters that correspond to a vowel appeared to constitute the group. It means that recoding text according to vowel and consonant one gets very non-random text. The informational classification “discovered” the syllabic structure of the text (with some approximation error). It is very natural. For example, in the optimal binary classification the

text should be maximally ordered locally. It means either the presence of long one-class clusters (0000… or 11111…), or local order of the form 0101… The first possibility cannot be realized for the natural language. Hence, for the best binary classification the encoded texts should be locally close to 0101… . And it is the syllabic structure (approximate syllabic structure, to be exact). And now we can explain the title of the section: "In the beginning was the Word, and the Word was with God, and the Word was God. The same was in the beginning with God. All things were made by him" (Jn. 1:1-3). Table 5. The phrase coded by the most frequent letters from classes (with an example of {01} coding) Number The recoded phrase of classes 01 110 101011011 101 2 En nne nenennenn nen Et the tetettett tet 3 En the netennent ten 4 En the setennent tes 5 En the setennend tes 6 Initial In the beginning was phrase

110 1011 011 110 1011 101 1011 101 011 110 1011 101 101 nne Nenn, enn nne Nenn nen nenn Nen, enn nne Nenn nen Nen the Teth, eth the Teth tet teth Teh, eth the Teth tet Teh the Tent, ent the Tent ten teth Tet, ent the Tent ten Tet the Test, ent the Test tes teth Set, ent the Test tes Set the Tesd, end the Tesd tes teth Ded, end the Tesd tes Ded the Word, and the Word was with God, and the Word was God

5 CONCLUSION AND OUTLOOK We presented the method of informational classification of amino acids based on protein sequences and considered some interesting properties of informational binary classifications. Informational binary classification makes protein sequence at maximum non random. However, relative entropy of recoded protein sequences is close to zero and it is not a case for Bible. So, even in optimal binary alphabet protein sequences slightly differ from zero order model. We revealed that informational classification that maximizes nonrandomness in two letter words distribution produces maximal or near maximal nonrandomness in 3,4,5-letter words distribution. We found similar 4-letter binary words frequency profile for all datasets, so, maximally non random {0,1} recoded proteins have the structure of repeated homogeneous series. This fact looks promising for revealing common structural elements in protein sequences. We showed the existence and uniqueness of optimal classification. We observed correlation between informational classifications and natural classifications of amino acids such as Hydrophobic/Polar and Charged/Uncharged ones for some datasets. Informational binary classifications are sensitive to amino acid distribution. However, similarity of one and two letter words frequency profiles in a number of considered protein sets provides as a rule high relative entropy for protein sequences recoded in any optimal classification.

Table 6. Hamming distances between various binary classifications

Membrane Globular Membrane+Globular (Cieplak et al., 2001)* (Irback et al., 1997) (Murphy et al., 2002) (Liu et al., 2002) MJ (Liu et al., 2002) BLOSSUM

HP I

HP II

HP III

BSl

CU

2 7 5 3 3 3 3

5 9 7 4 2 3 3

7 9 7 2 2 5 3

6 6 10 7 7 8 8

4 6 4 7 7 2 8

0 8 6 7 7 2 6

8 0 8 7 9 6 8

Membr. +Globr 6 8 0 7 9 4 10

4

3

3

6

8

8

10

10

Membr. Glob.

*(Chen et al.,2002) and (Li et al, 2003) have the same amino acid groupings. Finally, in Table 6 we present a pair comparison of amino acid binary classifications obtained from three sources: 1) informational classifications for Membrane, Globular dataset and Dataset 2 as a whole; 2) some natural classifications; 3) classifications taken from recent publications. Amino acid groupings mentioned in reviewed papers do have similarity with Hydrophobic/Polar classification while some informational classifications show similarity to Hydrophobic/Polar and Charged/Uncharged property. Classification of (Murphy et al., 2002) is the only one to be rather close to informational classifications. The idea of amino acids to be structural units of the sequence which contain all the information to protein fold and function is the base metaphor in studying amino acids informational classifications. The simplest binary informational classification gives the most nonrandom recoded sequences which allows one to study words distribution and to extract functional units. Fundamental property of informational classification method is its universal nature: the method may reduce any alphabet on the basis of frequencies of words of any length. Thus one can consider other structural units which probably determine protein fold and function. For example, the 400-letter alphabet of amino acid pairs can be supposed to be structural units. And in this case an alphabet reduction is the key operation. Presented method can perform alphabet reduction to be formal, automatic and the most informative. Another possible application of the method is amino acid classification based on contacts statistics. To the moment only pair contacts are considered while the method allows using an arbitrary set of contacts. Informational classification makes recoded sets of contacts to be at the most nonrandom. The nearest steps in informational classification method development and application could be as follows: (1) Computation of optimal 2, 3, 4, and 5-letters alphabets for databases of known protein primary structures; (2) The informational classifications analysis of 2 and 3-letters elements of primary structures (after reduction of the initial alphabet to 2-5 symbols) and so on, producing hierarchy of constructive details. How can we answer now the main question: What are proteins made from? Are they made from the following details: • Amino Acids (AAs)? • Short sequences of such classes? • Short sequences of AAs? • Anything else? • Classes of equivalent AAs?

For the set of available globular proteins we find the following main binary classification of AAs: G: {A,E,K,L,M,Q,R}U{C,D,F,G,H,I,N,P,S,T,V,W,Y}. For the set of available membrane proteins produces the following main binary classification which almost coincides with hydrophobic/hydrophilic classification: M: {D,E,H,K,N,Q,R,W,Y}U{A,C,F,G,I,L,M,P,S, T,V}. The primary sequences of proteins for these classes are definitely not random, and this nonrandomness (the correlations between AAs in sequences) is significantly different for these classes. If we combine two classifications into one “GorM” classification, then we get four groups {A,L,M}U{C,F,G,I,P,S,T,V}U{E,K,Q,R}U{D,H,N,W,Y}. In each class, we select two most frequent AAs. The symbol of the most frequent AA is written in bold underlined. The symbol of the second AA is written in italic underlined. Using these AAs we can call the classes as L-A (Leucine-Alanin) group {A,L,M}, G-S (Glycine-Serine) group {C,F,G,I,P,S,T,V}, K-E (Lysine-Glutamic Acid) group {E,K,Q,R}, and D-N (Aspartic Acid-Asparagin) group {D,H,N,W,Y}. So, we have new candidates for a minimal set of amino acids (one AA from a group or two AAs from a group). But, perhaps, it is wiser to classify couples and triples of amino acids. Classes of such couples and triples are, perhaps, the elementary details of proteins. ACKNOWLEDGEMENTS. We are grateful to M. Gromov for the question about difference between informational classifications for Membrane and Globular proteins. REFERENCES Alberts, B. (1998) The Cell as a Collection of Protein Machines: Preparing the Next Generation of Molecular Biologists. Cell, 92, 291–294. Bugaenko, N.N., Gorban, A.N., Sadovsky, M.G. (1998) Maximum entropy method in analysis of genetic text and measurement of its information content. Open Sys. & Information Dyn. 5(3), 265-278. Bugaenko, N.N., Gorban, A.N., Sapozhnikov, A.N. (1998) Minimum entropy principle and classification of symbols for revealing statistical regularities in text. Proc.of the I Int. Conf. on Bioinformatics of Genome Regulation and Structure, 2, 280-282. Online:http://www.bionet.nsc.ru/bgrs/author/gorban.html. Chan, H.S. Folding Alphabets. (1999) Nature Struct. Biol. 6, 994-996. Chen, H., Zhou, X., Ou-Yang, Z.-C. (2002) Classification of amino acids based on statistical results of known structures and cooperativity of protein folding. Phys. Rev. E 65, 061907. Chernavskii, D.S., Khurgin, Iu.I, Shnol', S.E. (1987) The "protein-machine" concept and its consequences. Biofizika, 32,: 775-781. Cid, H., Bunster, M., Canales, M. and Gazitua, F. (1992) Hydrophobicity and structural classes in proteins J Protein Eng 5, 373-375. Cieplak, M., Holter, N.S., Maritan, A., Banavar, J.R. (2001) Amino acids classes and protein folding problem. J. Chem. Phys., 114, 1420-1423. Eisenberg, D., Weiss, R.M., Terwilliger, T.C. (1984) The hydrophobic moment detects periodicity in protein hydrophobicity. Proc. Natl. Acad. Sci. USA, 81(1): 140-144. Fan, K, Wang, W. (2003) What is the minimum number of letters required to fold a protein? J. Mol. Biol. 328(4), 921-926. Gorban, A.N., Kudryashev, M., Popova, T. (2005) On the Way to Protein Alphabet: Informational Classification of Amino Acids in Comparison to Other Classifications, arXiv:q-bio/0501019v3 [q-bio.BM].

Gorban, A.N., Popova, T.G., Sadovsky, M.G., Wunsch, D.C. (2001) Information content of the frequency dictionaries, reconstruction, transformation and classification of dictionaries and genetic texts. Smart Eng. System Design 11, 657-663. Hennetin, J., Le T.K., Canard, L., Colloc'h, N., Mornon, J.P., Callebaut, I. (2003) Nonintertwined binary patterns of hydrophobic/nonhydrophobic amino acids are considerably better markers of regular secondary structures than nonconstrained patterns. Proteins, 51(2): 236-244. Irback, A., Potthast, F. (1997) Binary assignments of amino acids from pattern conservation. Protein Eng. 10(9), 1013-1017. Kawashima, S., Kanehisa, M. (2000) AAindex: Amino Acid index database. Nucleic Acids Research 28(1): 374. Kurzynski, M. (1998) A synthetic picture of intramolecular dynamics of proteins. Towards a contemporary statistical theory of biochemical processes. Prog. Biophys. Mol. Biol. 69(1): 23-82. Lewand, R.E. (2002) Cryptological Mathematics, The Mathematical Association of America Press, Washington DC. Li, H., Tang, C., Wingreen, N. S. (1997) Nature of driving force for protein folding: a result from analysing the statistical potential. Phys. Rev. Lett, 79, 765-768. Li, T., Fan, K., Wang, J., Wang, W. (2003) Reduction of protein sequence complexity by residue grouping. Protein Eng. 16(5), 323-330. Liu, X., Liu, D., Qi, J., Zheng, W.-M. (2002) Simplified amino acids alphabets based on deviation of conditional probability from random background. Phys. Rev. E, 66, 021906. Liu, X., Zhang, L.-M., Guan, S., Zheng, W.-M. (2003) Distances and classification of amino acids for different protein secondary structure. Phys. Rev. E, 67, 051927. Mintseris, J., Weng, Zh. (2004) Optimizing Protein Representations with Information Theory. Genome Informatics. 15(1), 160-169. Murphy, L.R., Wallqvist, A., Levy, R.M. (2000) Simplified amino acid alphabets for protein fold recognition and implications for folding. Protein Eng. 13(3), 149-52. Phoenix, D.A., Harris, F. (2002) The hydrophobic moment and its use in the classification of amphiphilic structures (review). Mol. Membr. Biol., 19(1), 1-10. Sirabella, P., Guiliani, A., Colosimo, A. (1998) Oligopeptides' frequencies in the classification of proteins' primary structure. E-print: arXiv: physics / 9807059. Solis, A.D., Rackovsky, S. (2002) Optimally informative backbone structural propensities in proteins. Proteins 48(3), 463-86. Wang, J., Wang, W. (1999) A computational approach to simplifying the protein folding alphabet. Nature Struct. Biol. 6, 1033-1038. Wang, J., Wang, W. (2000) Modeling study on the validity of a possibly simplified representation of proteins. Phys. Rev. E, 61(6 Pt B), 6981-6986. Wang, J., Wang, W. (2002) Grouping of residues based on their contact interactions. Phys. Rev. E, 65(4 Pt 1), 041911. West, M.W., Hecht, M.H. (1995) Binary patterning of polar and nonpolar amino acids in the sequences and structures of native proteins. Protein Sci. 4(10), 2032-2039.