Statistical analysis of the Indus script using -grams
Abstract
The Indus script is one of the major undeciphered scripts of the ancient world. The small size of the corpus, the absence of bilingual texts, and the lack of definite knowledge of the underlying language has frustrated efforts at decipherment since the discovery of the remains of the Indus civilisation. Recently, some researchers have questioned the premise that the Indus script encodes spoken language. Building on previous statistical approaches, we apply the tools of statistical language processing, specifically -gram Markov chains, to analyse the Indus script for syntax. Our main results are that the script has well-defined signs which begin and end texts, that there is directionality and strong correlations in the sign order, and that there are groups of signs which appear to have identical syntactic function. All these require no a priori suppositions regarding the syntactic or semantic content of the signs, but follow directly from the statistical analysis. Using information theoretic measures, we find the information in the script to be intermediate between that of a completely random and a completely fixed ordering of signs. Our study reveals that the Indus script is a structured sign system showing features of a formal language, but, at present, cannot conclusively establish that it encodes natural language. Our -gram Markov model is useful for predicting signs which are missing or illegible in a corpus of Indus texts. This work forms the basis for the development of a stochastic grammar which can be used to explore the syntax of the Indus script in greater detail.
Index Terms:
Formal languages, Markov processes, Sequences1 Introduction
The civilisation that flourished from 7000 BCE to 1500 BCE in the valley of the river Indus and its surroundings was the first urban civilisation in the Indian subcontinent. At its peak, in the period between 2600 BCE and 1900 BCE, it covered approximately a million square kilometers [1], making it the largest urban civilisation of the ancient world. The remains of the civilisation were first found in Harappa and, following historical convention, is called the Harappan civilisation. The terms Harappan civilisation and Indus civilisation are used interchangeably in this paper.

The Indus people used a system of signs, which has mainly survived on seals (see Fig. 1 for an example), pottery, and other artifacts made of durable materials such as stone, terracotta, copper. The sign system is yet to be deciphered. The use of these signs is usually in short texts, numbering not more than signs in a single line of text. Around distinct signs have been identified [2, 3], though some researchers identify up to distinct signs [4]. The total texts number to about . The lack of decipherment of the sign system is attributed to the paucity of the working material, the absence of any bilingual texts that might aid in decipherment, and a lack of definite knowledge of the underlying language, assuming the script was used to write language. The sign system of the Indus remains controversial, with contested claims of decipherment, and even the suggestion that it does not encode language [5]. Here we shall use the terms sign system and script interchangeably to refer to the Indus script.
The main methodological difficulty in attempting any interpretation of the script is that, due to the paucity of information on the context of the writing, one is perforce required to make an a priori assumption regarding the information the script is meant to convey. Thus, the range of opinion on what the script encodes varies from an Indo-Aryan language [6], a Dravidian language [3], a purely numerical system [7], to non-linguistic symbols [5]. There is no consensus on any of the above interpretations.
A line of research, which does not require such a priori assumptions, is the statistical approach. At its heart lies the notion of recognising patterns through counting. While such an approach cannot shed light on the semantics of the script, it can reveal important features of its syntax. Research using the statistical approach was initiated by a Soviet team in 1965, further developed by a Finnish team in 1969 (for review of various attempts see [8, 9, 10]), continued by Siromoney [11] in the 1980s and followed up more recently by some of us [12, 13, 14].
In this article, we bring to bear the tools of -gram modelling for a thorough statistical analysis of sequences in the Indus script. An -gram is a sequence of units, or tokens. -gram models provide a very useful method of modelling sequences111The examples of English language used in the paper to explain various concepts related to n-gram modelling are given for the ease of comprehension, as a representative of the general class of sequences. These linguistic examples do not imply that any linguistic assumption has been made for the Indus script., whether they be the sequence of letters or words in a natural language, the base pairs in the genetic code, or the notes in a musical score. This generality is possible because -gram models are indifferent to the semantic content of the units or tokens making up the sequence (the words, the letters, the base pairs or the notes) but, nonetheless, reveal the syntax, if any, that the sequences follow. The -gram approach, then, forms the ideal framework in which the Indus script can be analysed without making any a priori assumptions.
Our work in the paper uses a bigram model, that is an -gram model with , to model the sequence of signs in the Indus script. Our analysis reveals that the sign system of the Indus is structured, with specific text beginner and text ender signs, that there is directionality, strong correlations in the order of the signs, and groups of signs which appear to have the same syntactic function. In Section 2, we provide details of the corpus used in this study and present results of an empirical statistical analysis of the corpus. The empirical analysis shows that a small number of signs account for the majority of the usage, a fact we are able to capture by fitting the frequency distribution of signs to the well-known Zipf-Mandelbrot law. We note in passing, that word frequency distributions in natural languages also follow this law. In Section 3 we provide a summary of the theory of -gram models, paying special attention to the problem of unseen -grams. We describe how this problem is resolved using smoothing and backoff algorithms and demonstrate how this modifies the statistics obtained from the empirical analysis. Our main results, summarised above, are presented in Section 4 using a bigram model. The Indus script corpora contains many instances of texts in which signs are damaged beyond recognition or otherwise illegible. The bigram model can be used to suggest a set of signs, ranked by probability, for the restoration of the damaged or illegible signs. In Section 5 we outline the theory behind this method and provide examples on known texts from the corpus. In the final section, we discuss future work based on the -gram approach and end with a summary of the present work.
2 The corpus and empirical analysis
Any work on the Indus script, statistical or otherwise, must begin with a corpus. Three major corpora of the Indus texts, by Mahadevan [2], Parpola [3] and Wells [4] are available. We use the electronic concordance of Mahadevan, henceforth referred to as M77, which records unique signs222The serial number of the signs used in this paper is as given by Mahadevan in his concordance (M77). As a convention followed in the present paper, the texts depicted by pictures are to be read from right to left, whereas the texts represented by just strings of sign numbers are to be read from left to right (see M77 for discussion on direction of texts). in lines of texts. In order to remove ambiguity, an Extended Basic Unique Data Set (EBUDS) is created by removing from the concordance all texts containing lost, damaged, illegible, or doubtfully read parts [12]. Texts spread across multiple lines are also removed. For texts that occur more than once, only a single copy is retained. Variations due to the archaeological context of the sites, stratigraphy, and type of object on which the texts are inscribed are, at present, not taken into account in the interests of retaining a reasonable sample size.
The reasons for discarding texts which are spread over multiple lines are twofold. First, it is not clear if each line of a multi-line text is to be treated as a single text having a continuity of sequence, or if it is to be regarded as separate text. Second, assuming a continuity of sequence, the order in which the texts are to be read across lines remains ambiguous [2].
The pruned EBUDS dataset contains texts. In EBUDS, signs out of present in the sign list of M77 do not make their appearance. Out of these removed signs, one sign () appears times, one sign () appears times, two signs (, ) appear times, three signs (, , ) appear twice and thirty-three signs appear only once in M77. Fig. 2 compares the distribution of texts lengths in various datasets and shows the effect of the removal of texts on the final dataset (EBUDS). We have already shown in our earlier work [12], that the relative frequency distribution of signs in EBUDS is comparable to M77 and hence EBUDS is a true representation of M77, with a reduction in total sign occurrences, but not in the percentage of total sign occurrences.

We first present the results of an empirical statistical analysis of the EBUDS corpus. Fig. 3 shows the frequency distribution of signs in EBUDS. The sign corresponding to in M77 (), is the most frequent sign, followed by signs (
), (
), and (
). The relative frequencies have no significant change in the M77 and EBUDS corpora.
2.1 Zipf-Mandelbrot law
The same data can be plotted in a different way, as a rank-ordered frequency distribution. The most frequent sign is given rank and its frequency is denoted by , the next most frequent sign is given rank and its frequency is denoted as and so on, till all signs are exhausted. The rank-ordered frequency is then plotted against the rank , on double logarithmic axes, as shown in Fig. 4. Remarkably, the data can be fit very well to the Zipf-Mandelbrot law, [15]. The Zipf-Mandelbrot law is commonly used to fit the rank-ordered frequency distribution of words in corpora of natural languages. Qualitatively, a distribution which follows the Zipf-Mandelbrot law has a small number of tokens (words, signs) which contribute to the bulk of the distribution, but also a large number of rare signs which contribute to a long tail. To emphasise this point, it is estimated that English has approximately a million words, though a college graduate might know only between to of these, and yet be a completely competent user of the language [16]. The Indus script seems to follow the same pattern, with a small number of signs accounting for the majority of use, but with a large number of signs which are used infrequently.


2.2 Cumulative distribution
To further follow up this point, we plot the cumulative frequency distribution of the signs in EBUDS in Fig. 5. As can be seen from the graph about signs account for of the total data and the most frequent sign () alone accounts for of the complete EBUDS. This observation is consistent with previous analysis by Mahadevan for M77 corpus [2]. In the same graph, we plot the cumulative distribution of text beginners, that is the sign beginning the text, and text enders, that is the sign ending the text. Here, an interesting asymmetry is evident. Approximately text beginners account for of the text beginner usage, but only text enders are needed to account for the same percentage of text ender usage. Since the possible number of text beginners and text enders can be any of the signs, the numbers above indicate that both text beginners and text enders are well-defined, with text enders being more strictly defined than text beginners. This fact alone indicates that there is directionality in the use of the signs in Indus texts. Note that this conclusion of directionality is independent of any assumed direction of reading the sign sequence. Assuming the texts are read left to right would, instead, imply that text beginners are more strictly defined than text enders, again implying directionality.
The analysis above has only been concerned with frequency distributions of single signs. We may extend the analysis to pairs of signs, triplets of signs and so on, as done by some of us earlier [12, 13]. This allows us to explore the very important feature of order and correlations between the signs, which are the outward manifestations of syntax. Syntax can be recognised even when the semantics is doubtful. For example, Chomsky’s famous sentence “Colourless green ideas sleep furiously” becomes syntactically incorrect when the word order is changed to give, say, “Green furiously ideas colourless sleep”. We recognise the first sentence as syntactically correct English, but of doubtful semantic value, but the second sentence is immediately recognised as syntactically invalid, even before we pause to think of its semantic content. It is this feature of the sign sequence of the Indus script, word order and correlation as an indicator of syntax, that we want to explore in detail, and which forms the main body of the present work. Such features are best explored in the -gram model to be presented in the next section.

2.3 Probability of sequences
Before exploring -gram models, we briefly mention what sort of frequency distribution we should expect in the absence of correlations and also establish the notation we will be using for the remaining part of the paper. We denote a sequence of signs by , where each is one of the possible Indus signs. Each of these is what we have been referring to as a text. The EBUDS corpus contains texts of maximum length . By our analysis above, we have obtained frequency distributions for the signs , by counting the number of times that sign occurs in the corpus, and then dividing it by the total size of the corpus. This is identified with the probability, in the sense of maximum likelihood, of seeing the sign in a text,
(1) |
In the absence of correlations, the joint probability that we see sign after sign is independent of , and is just the product of their individual probabilities
(2) |
Generalising, the probability of the string is simply a product of the individual probabilities
(3) |
In the absence of correlations, then, we have a generalisation of die throwing, where instead of possible outcomes, we have possible outcomes in each throw, and the outcome has a probability . Each throw outputs a single sign, and the outcome of a throw is independent of all previous throws. In throws, we generate a line of text . In probability theory, this process is called a Bernoulli scheme. In the next section, we will show that this is not adequate to model the sequence of Indus signs, indicating that there are significant correlations. We can then expect that an arbitrary permutation of signs in an Indus text will produce a syntactically invalid sequence.

3 -gram model for the Indus script
The main purpose of an -gram model is to capture the correlations that exist between tokens in a sequence . Thus, conditional probabilities form the core of an -gram model. Specifically, for a string the -gram model is a specification of conditional probabilities of the form , quantifying the probability that the previous signs of the sub string is followed by the sign . Given the -gram conditional probability, and the relation between joint and conditional probabilities , the probability of the string can be written as,
Recursively applying to the leftmost terms, we obtain the probability as a product over conditional probabilities
(5) |
In the above, it is understood that is a special token indicating the start of the string. Note that the above expression is an identity that follows from the basic rules of probability and contains no approximations. As an example, the probability of a string of length three, , is given as a product of trigram, bigram and unigram probabilities
(6) |
Clearly, for an -gram model to be tractable, only a finite number of such probabilities can be retained. In the simplest bigram model, all correlations beyond the preceding sign are discarded, so that
(7) |
In a trigram model, all correlations beyond two preceding signs are discarded, so that
(8) |
In a general -gram model, all correlations beyond the preceding signs are discarded. An -gram model can then be thought of as an order Markov chain in a state space consisting of the signs . The introduction of the Markov assumption is the main approximation in -gram models. -grams were first used by Markov to analyse the probability of a consonant following a vowel in Pushkin’s Eugene Onegin. Shannon applied word -grams to model sentences in English text and with obtained sentences which have a remarkable syntactic similarity to English. Since then, -gram models have found wide use in many fields where sequences are to be analysed, including bioinformatics, speech processing and music.
The theory and applications of -grams are described in several excellent textbooks [15, 17] and need not be repeated here. We summarise below our method of obtaining probabilities from counts, the use of smoothing and backoff to account for unseen -grams, the measures we use for evaluating the -gram model, and finally the tests we use to assign a statistical significance to the correlations.
-gram probabilities are obtained from counts. For single signs, the counts are unambiguous. However, for sign pairs, it is possible that a rare combination may not be present in a small sized corpus. The count, and the resulting probability estimate for that sign pair, then is zero. However, in the absence of reasons to the contrary, common sense dictates that no sign pair should have a strictly zero probability. This intuition is quantified by various rules which remove probability weight from seen sign pairs and distributes it to sign pairs which have never been seen, but are nonetheless not impossible. Common amongst such “smoothing” rules are Laplace’s add-one rule, a method developed by Turing and Good in breaking the Enigma code, and a more recent algorithm due to Witten and Bell [18]. Here, we use the Witten-Bell algorithm to smooth our -gram models. In Fig. 6 we show the estimate of the probability of a sign being followed by sign , before smoothing and after Witten-Bell smoothing. In above panel, the only non-zero probabilities are those corresponding to signs and . These probabilities sum to one, indicating that other sign pairs are impossible. The Witten-Bell smoothing algorithm restores a finite, but small probability to the unseen sign pairs, ensuring again that all probabilities sum to one. Apart from being a more reasonable way of estimating probabilities from counts, it also ensures that the resulting Markov chain is ergodic333A Markov chain is ergodic if it is possible to reach every state from every state (not necessarily in one move).. An ergodic Markov chain is essential in such applications, since otherwise, probabilities of all strings containing unseen -grams vanish.
Smoothing is not the only way of correcting for unseen -grams. Another method, called backoff uses probabilities of -grams to estimate -gram probabilities. Thus, the probability of unseen trigrams can be estimated from that of seen bigrams and unigrams. Here, we used the Katz backoff algorithm to estimate bigram, trigram and quadrigram probabilities when appropriate.
The estimation of -gram probabilities from -gram counts is called learning. A learned -gram model can then be tested to see if it produces -grams in the corpus. To avoid circularity, the corpus is usually divided into a training set, from which the probabilities are estimated, and a test set, on which the model is evaluated. There are several standard measures for evaluating the goodness of an -gram model. Here, we use a standard measure, the perplexity, which is related to the information theoretic cross-entropy. We also do a cross-validation test using standard procedures.
Finally, we need tests of association to ascertain the significance of sign pairs which appear more or less frequently than what would be predicted by the Bernoulli scheme model. For this, we use a log-likelihood test, testing the null hypothesis that there is no association between sign pairs.
4 -gram analysis of the corpus
In this section we present the -gram probablities obtained from the EBUDS corpus and draw inferences on the structure of the script from an analysis of the -gram probabilities.
We introduce, following usual practice, two additional tokens <s> and </s> which indicate the beginning and end of a text respectively. By convention, the unigram probability for the start token is unity, , since every text must begin with <s>. The probability of sign being a beginner is then , since . The probability of sign being an ender is . It is also assumed, as usual, that -gram probabilities do not depend on the location of the signs in a text, or more formally, that the Markov chain is stationary.
In any -gram study, a maximum value of has to be chosen in the interest of tractability, beyond which correlations are discarded. This can be done in an empirical fashion, balancing the needs of accuracy and computational complexity, using measures from information theory which discriminate between -grams models with increasing [15, 17], or by more sophisticated methods like the Akaike information criterion which directly provides an optimal value for [21].
In previous work, some of us have shown that bigram and trigram frequencies in the EBUDS corpus differ significantly from frequencies expected from a Bernoulli scheme [12]. The small size of the corpus limits the ability to assess significance of quadrigrams and beyond, when using the method in [12]. In our subsequent work [13] it has been shown that of the texts of length 5 and above can be segmented using frequent unigrams, bigrams, trigrams and quadrigrams and complete texts of length 2, 3 and 4. Moreover, frequent bigrams or texts of length 2 alone account for of the segmented corpus. Thus the bulk of the corpus can be segmented with -grams with not exceeding , and almost half the corpus can be segmented into bigrams alone. This indicates that correlations beyond quadrigrams have relatively less significance.
Here, we use an information theoretic measure, the perplexity, which we explain in detail below, to estimate an optimal value for . Summarising the results of this analysis, which is presented in Table. III, we find that the perplexity monotonically decreases as ranges from to (corresponding to unigram, bigram and trigram correlations), but then saturates beyond (corresponding to quadrigram and higher correlations). This confirms the observation made in [13] that pentagram and higher correlations are of relatively less significance in the EBUDS corpus. From the differential reduction in perplexity with increase in model order, it is clear that the most significant correlations are due to bigrams, with somewhat modest trigram correlations, and almost negligible quadrigram correlations. Our subsequent analysis is, therefore, wholly based on bigrams. The main conclusions that we draw in this paper on the structure of the script are expected to remain broadly unaltered with the inclusion of trigram and quadrigram correlations. We expect a modest improvement in the predictive power of the -gram model when trigram and quadrigram correlations are included. The role of higher-order correlations will be more fully explored in forthcoming work.
Sign Pair | Rank | Frequency | Significant | Rank | LL Value |
---|---|---|---|---|---|
(Naive) | (EBUDS) | Sign Pair | (LL) | ||
4.1 Inferences from bigram matrix
We now present the results of a bigram based analysis of the sequence of signs in the EBUDS corpus. To remind ourselves, in the bigram model, it is assumed that the probability depends only on the immediately preceding sign and is the same as . The bigram model is fully specified by the unigram probabilities and the bigram probabilities .
We compare the bigram probabilities in the absence of correlations with the bigram probabilities for EBUDS corpus after Witten-Bell smoothing in the two plots of Fig. 7. If there is no correlation between and , we expect , that is the conditional probability of is identical to the marginal probability. We show this marginal probability in the first plot of Fig. 7. In the second plot of Fig. 7 we show the matrix of bigram probabilities that sign follows sign in the corpus. It is clear even by a casual observation, that there are significant differences between the two plots. This signals the presence of correlations in the script.








In Fig. 10 we show the text beginner and text ender sign probability distributions. The more careful analysis using smoothed bigram probabilities confirms our earlier conclusion, based on a raw counts, that text enders are more strictly defined than text beginners.
We can further analyse the nature of correlations of a sign with other signs preceding or following it using the results of bigram analysis. As an example, we explore the correlations of the three most frequent text beginners (sign numbers , and ) and the three most frequent text enders (sign numbers , and ) shown in Fig. 10 with other signs. It can be inferred from the plots of conditional probabilities (Fig. 10) i.e. , and for the text beginners and , and for the text enders, that the text beginners , and are more selective in terms of the number of signs which can follow them in comparison to the text enders , and which can be preceded by relatively larger number of signs. Thus, there is greater flexibility for signs preceding the text enders than the signs which tend to follow the text beginners. Moreover, if we further go into the details of the restricted number signs , which follow the text beginners we find that the nature of the most frequent signs following the text beginners is quite similar to the text enders (though in reverse direction), in terms of the number of signs which follow them as shown in the plots of and (Fig. 10). In other words, the correlation between the text beginners and the signs following them is stronger than the correlation of these immediate neighbours of text beginners with the sign following them, a fact which helps us in finding the weaker and stronger junctions in the text [13]. The diagonal elements of the matrix of bigram probabilities are the probabilities of sign pairs with same signs and the most frequent sign pairs with repeating signs in the corpus are (, ) and (, ).
Measure | Random | EBUDS |
---|---|---|
Entropy (H) | ||
Mutual information (I) |
Perplexity () | 68.82 | 26.69 | 26.09 | 25.26 | 25.26 |
---|---|---|---|---|---|
6.10 | 4.74 | 4.71 | 4.66 | 4.66 |
4.2 Test of significance
To quantitatively assess the significance of the correlations, and to obtain the statistically significant sign pairs, we need to test the null hypothesis that signs and are independent. In Table I we give the most frequent sign pairs as well as the ones which are most statistically significant. Here, we enumerate the most significant sign pairs on the basis of log-likelihood measure of association for bigrams. It is interesting to note the most significant sign pairs are not always the most frequent ones given in the first column of Table I for comparison.
![]() |
![]() |
![]() |
![]() |
![]() |
4.3 Entropy, Mutual Information and Perplexity
Text No. | Text | Incomplete Text | Most Probable Restoration | Probable Restored Sign |
4312 |
![]() |
![]() |
![]() |
![]() ![]() ![]() |
4016 |
![]() |
![]() |
![]() |
![]() ![]() ![]() |
5237 |
![]() |
![]() |
![]() |
![]() ![]() |
2653 |
![]() |
![]() |
![]() |
![]() ![]() ![]() |
5073 |
![]() |
![]() |
![]() |
![]() ![]() ![]() |
3360 |
![]() |
![]() |
![]() |
![]() |
9071 |
![]() |
![]() |
![]() |
![]() ![]() ![]() |
4081 |
![]() |
![]() |
![]() |
![]() ![]() |
Text No. | Text | Incomplete Text | Most Probable Restoration | Probable Restored Sign |
8302 |
![]() |
![]() |
![]() |
![]() ![]() |
5317 |
![]() |
![]() |
![]() |
![]() |
1193 |
![]() |
![]() |
![]() |
![]() |
1407 |
![]() |
![]() |
![]() |
![]() ![]() ![]() |
2179 |
![]() |
![]() |
![]() |
![]() |
3396 |
![]() |
![]() |
![]() |
![]() ![]() ![]() |
8101 |
![]() |
![]() |
![]() |
![]() |
2802 |
![]() |
![]() |
![]() |
![]() ![]() |
Information theoretic measures can also be used to make an estimate of the role of correlations. We use the informational theoretic entropy and the mutual information ,
(9) | |||||
(10) |
to quantify the amount of correlation in the sequence.
In Table II we compare the entropy and mutual information of the corpus with that of a completely random sequence with no correlations. For this, the probability of signs is uniform and the joint probability is . This gives an entropy of and a vanishing mutual information. In contrast, the entropy of the EBUDS corpus is and the mutual information is . This indicates the presence of correlations, but the difference between the entropy and mutual information also indicates that there is flexibility in the sign usage, and the probability of a sign is not completely determined by the preceding sign.
A related information-theoretic measure, the cross-entropy, is useful in discriminating between -gram models with different . The main goal of -gram analysis is to construct a good model for the probability of sequences in a corpus. A metric which quantifies how close the estimated probability distribution is to the actual probability distribution is the cross-entropy. For a true distribution and its estimate , the cross entropy is defined as,
(11) |
The cross-entropy is minimum when the true and estimated probability distributions coincide, . The perplexity , which is the measure commonly used in natural language processing literature, is an exponential of the cross-entropy,
(12) |
Of course, the true probability distribution is not known, but it can be shown [15, 17] that for a corpus of size obtained from a stationary and ergodic Markov chain, the cross-entropy is given by the limit,
(13) |
In other words, the cross-entropy of the corpus, as the corpus size approaches infinity becomes the cross-entropy of the probability distribution from which the corpus was generated. The previous formula does not require knowledge of , and can then be used to give an estimate of the cross-entropy for a large, but finite, corpus. The relation has obvious generalisations to joint probability distributions , , of bigrams, trigrams and higher -grams. We denote the -gram cross-entropy by .
Here, we measure the cross-entropy against a held out corpus. We see that the perplexity is reduced considerably when bigram correlations are taken into account. We have also evaluated the perplexity for trigram and quadrigram models, and this shows a monotonic reduction in perplexity as shown in Table III. This implies that correlations beyond the preceding sign are important, though the most important correlations comes from the preceeding sign. The perplexity of the bigram model is 26.69 which is significantly lower than that of unigram model which equals 68.82. As discussed in the beginning of this section, this motivates our choice of retaining only bigram correlations for the present study.
The bigram model can now be used to generate texts according to the Markov chain defined by the unigram and bigram probabilities. In Table IV we show examples of texts which have been generated by the bigram model. The performance of the model is discussed in Section 6.
Our results in this section establish without doubt that there are significant correlations between signs in the Indus script, and reveal the presence of what can reasonably be called syntax. There are well-defined text beginners and text enders, sign order is important, and many statistically significant pairs of signs occur in the corpus.
5 Restoring illegible signs
An important practical use of the bigram model is to restore signs which are not legible in the corpus due to damage or other reasons. The basic idea is to use the bigram model to evaluate the probability of a suggested restoration, and chose the restoration with the highest probability. For example, consider the three sign text in which the middle sign is illegible. We use the bigram model to evaluate the probability of the string for different choices of by
(14) |
The most probable sign with which to restore the text is, then, the maximum of this probability over all possible signs . Since there are possible signs, this can be accomplished by a direct enumeration. When the number of illegible signs is more, the space over which the maximisation needs to be done grows rapidly. With illegible signs, there are values from which to pick a maximum. In such instances, instead of a direct search, a dynamic programming algorithm may be applied. Here, we use the well-known Viterbi algorithm[15] for finding the most probable state sequence for a given observation in a hidden Markov model, suitably modified to make it applicable to a Markov chain, to find most probable sequence of signs. Our results for sign restorations are summarised in Table V. We list the original text, a randomly chosen deletion for that text, the most probable restoration, and the next probable restorations obtained using the bigram model. We see that in all cases, the bigram model is successful in reproducing the deleted sign. This gives us confidence that the bigram model can be used to suggest restorations of illegible signs in various corpora. Table VI gives few examples of how the model can be used for restoration of doubtfully read signs in the texts of M77 corpus.

Test Set | Train Set | Mean Sensitivity | STDEV |
---|---|---|---|
{, , , } | |||
{, , , } | |||
{, , , } | |||
{, , , } | |||
{, , , } | |||
Average | 74 | 2 |
We can also use the model to find most probable texts of various lengths. In Table VIII we reproduce the most probable texts of length and as predicted by the bigram model. It is quite remarkable that there are exact instances of the most probable texts of length and slight variant of most probable text of length in the corpus.
Text length | Text with blank positions | Most probable predicted text | Text No. | Closest matching text from M77 |
4 |
</s>
![]() |
![]() |
1232 |
![]() |
|
2580 |
![]() |
||
5 |
</s>
![]() |
![]() |
2476 |
![]() |
6 |
</s>
![]() |
![]() |
1322 |
![]() |
Sl. No | Test/Measure | Results | Fig. / Table No. | Conclusions |
1. | Zipf-Mandelbrot law | Best fit for a= 15.4, b = 2.6, | Fig. 4 | Small number of signs account for |
c = 44.5 (95% Confidence interval) | bulk of the data while a large number | |||
of signs contribute to a long tail. | ||||
2. | Cumulative frequency | 69 signs: 80 % of EBUDS | Fig. 5 | Indicates asymmetry in usage of 417 |
distribution | 23 signs: 80 % of Text Enders | distinct signs. Suggests directionality | ||
82 signs: 80 % of Text Beginners | and structure in writing. | |||
3. | Bigram probability | Conditional probability matrix for | Fig. 7 | Indicates presence of significant |
EBUDS is strikingly different from | correlations between signs. | |||
the matrix assuming no correlations. | ||||
4. | Conditional probabilities | Restricted number of signs follow | Fig. 10 | Indicates presence of signs having |
of text beginners and | frequent text beginners whereas | similar syntactic functions. | ||
text enders | large number of signs precede | |||
frequent text enders. | ||||
5. | Log-likelihood | Significant sign pairs extracted. | Table I | The most significant sign pairs are not |
significance test | always the most frequent ones. | |||
6. | Entropy | Random: 8.73; EBUDS: 6.68 | Table II | Indicates presence of correlations. |
7. | Mutual information | Random: 0; EBUDS: 2.24 | Table II | Indicates flexibility in sign usage. |
8. | Perplexity | Monotonic reduction as n | Table III | Indicates presence of long range |
increases from 1 to 5. | correlations, see also [12, 13]. | |||
9. | Sign Restoration | Restoration of missing and | Tables V, VI | Model can restore illegible signs, |
illegible signs. | according to probability. | |||
10. | Cross validation | Sensitivity of the model = 74 % | Table VII, Fig. 11 | Model can predict signs with |
74% accuracy. |
6 Model performance evaluation: Cross Validation
We have already discussed perplexity in section 4. Here we discuss another method to evaluate the performance of the model called cross validation. In cross validation, the corpus is divided into a training set, from which the probabilities are estimated, and a test set, on which the model is evaluated. The measure of goodness of the model calculated is sensitivity which is defined as
(15) |
where is the count of true positives and is the count of false negatives. The ratio of training to test set size used is : . We divide EBUDS into equal parts i.e. , , , and each being one-fifth of the corpus. We start with first part, selecting that to be the test set and concatenate the remaining parts to form the first training set. We use the training set to learn the parameters of the model and use the test set to evaluate the goodness of the model. We drop out signs randomly from the test set and ask the model to fill in the dropped signs. We then count the number of true positives, that is, the number of times the predicted signs matches with the signs under 90% area of the probability curve; otherwise, they are considered false negatives. This is repeated about 100 times with the first test set and training set and we plot the histogram of the sensitivity for these runs.
The cross validation test described above is repeated by taking each of the five equal parts of EBUDS as test set and concatenating remaining parts as training sets. The results are shown in Table VII and Fig. 11. As can be seen from the plots the sensitivity of the model considering all signs under 90% area of the cumulative probability curve as true positives is 74%.
7 Summary and conclusions
We conclude that a bigram model, equivalent to a first-order Markov chain, can be used to capture the correlations of signs in the Indus texts and gives important insights into the structure of Indus writing. Our study has revealed that the script carries syntax, with well-defined text beginners and text enders, directionality and sign order, and strong correlations between signs. Information theoretic measures indicate that the correlations are not completely rigid. This indicates that the script can certainly be considered as a formal language, but it remains to be seen if these features imply an underlying natural language. The tests from which we draw this conclusion are summarised in Table IX. The bigram model can be used to suggest probable restorations of illegible signs from a corpus.
To the best of our knowledge, our work represents the first use of the methods of n-gram analysis to an undeciphered scipt. We believe stochastic methods hold considerable promise in elucidating syntax in undeciphered scripts. We will present further results on the syntax of Indus script, based on stochastic methods, in our forthcoming work.
Acknowledgements
We wish to acknowledge the research grant of the Jamsetji Tata Trust that enabled us to do this work. We wish to acknowledge the generous assistance of Indus Research Centre of the Roja Muthiah Research Library. We sincerely thank Prof. P. P. Divakaran for enabling a very fruitful collaboration between TIFR and IMSc. R. Adhikari gratefully acknowledges Pinaki Mitra for first arousing his interest in probability theory and the study of language. Rajesh Rao would like to acknowledge the support of the National Science Foundation (NSF) and the Packard Foundation. We are grateful to Harappa.com for their kind permission to use the picture of an Indus seal in the paper.
References
- [1] G. L. Possehl, The Indus Civilization A contemporary perspective, Vistaar Publications, New Delhi (2002).
- [2] I. Mahadevan, THE INDUS SCRIPT Texts, Concordance and Tables, Memoirs of the Archaeological Survey of India (1977).
- [3] A. Parpola, Deciphering the Indus Script, Cambridge University Press (1994).
- [4] B. K. Wells, Epigraphic Approaches To Indus Writing, Harvard University, Cambridge, Massachusetts (2006).
- [5] S. Farmer, R. Sproat, and M. Witzel, The collapse of the indus-script thesis: The myth of a literate harappan civilization. Electronic Journal of Vedic Studies (EJVS) (2004), Vol. 11.
- [6] S. R. Rao, The Decipherment of the Indus Script, Asia Publishing House, Bombay (1982).
- [7] B. V. Subbarayappa, INDUS SCRIPT: Its Nature and Structure, New Era Publications, Madras (1997).
- [8] G. L. Possehl, Indus Age: The writing System, Oxford IBH Publishing Co. Pvt. Ltd. (1996).
- [9] I. Mahadevan, Aryan or dravidian or neither? a study of recent attempts to decipher the indus script (1995-2000). Electronic Journal of Vedic Studies (EJVS) (2002), Vol. 2.
- [10] A. Parpola, Study of the indus script. 50th ICSE Tokyo (2005).
- [11] G. Siromoney and A. Haq, Indus script: Challenge and response. The Sunday Standard, Madras (1979).
- [12] N. Yadav, M. N. Vahia, I. Mahadevan, and H. Joglekar, A statistical approach for pattern search in indus writing. International Journal of Dravidian Linguistics (2008), Vol. 37, pp. 39–52, http://www.harappa.com/script/tata-writing/tata-indus-script-research.html.
- [13] N. Yadav, M. N. Vahia, I. Mahadevan, and H. Joglekar, Segmentation of indus texts. International Journal of Dravidian Linguistics (2008), Vol. 37, pp. 53–72, http://www.harappa.com/script/tata-writing/tata-indus-script-research.html.
- [14] R. P. N. Rao, N. Yadav, M. N. Vahia, R. Adhikari, H. Joglekar, and I. Mahadevan, A probabilistic model for analyzing undeciphered scripts and its application to the 4500-year-old indus script (2008), in preparation.
- [15] C. Manning and H. Schütze, Foundations of Statistical Natural Language Processing, MIT Press, Cambridge (1999).
- [16] D. Crystal, Cambridge Encyclopedia of the English Language, Cambridge University Press (1995).
- [17] D. Jurafsky and J. H. Martin, Speech and Language Processing: An introduction to speech recognition, computational linguistics and natural language processing, 2nd ed., Pearson Prentice Hall (2008).
- [18] S. F. Chen and J. Goodman, An empirical study of smoothing techniques for language modeling. Technical Report TR-10-98, Computer Science Group, Harvard University, Cambridge University, Massachusetts (1998).
- [19] A. Stolcke, Srilm - an extensible language modeling toolkit (2002).
- [20] S. Banerjee and T. Pedersen, The design, implementation and use of the ngram statistics package. Proceedings of the Fourth International Conference on Intelligent Text Processing and Computational Linguistics, Mexico City (2003).
- [21] H. Tong, Determination of the order of a markov chain by akaike’s information criterion. Journal of Applied Probability (1975), Vol. 37, pp. 488–497.