3
Views
1
CrossRef citations to date
0
Altmetric
Original Articles

Predictive Entropy of Hindi and Tamil

, FIETE
Pages 41-45 | Received 29 Aug 1983, Published online: 02 Jun 2015
 

Abstract

Any sequence of written symbols is called a message for convenience. The information associated with any one symbol depends on the relative frequency of occurrence and the average over the various symbols of a language is termed entropy. A more accurate description should take into account a larger number of previous symbols or blocks of previous symbols and their effect on the succeeding symbol. The increasing dependency of the symbols on those preceding them reduces the entropy per symbol.

Shannon has suggested a fairly reliable method for finding the reduction in entropy with increasing length of the message. In this method, the number of trials a subject makes to get the correct letter, is used as the score for that letter. These scores for the various positions or order N, are used to find the entropy.

Shannon's predictive entropy method is repeated for Indian languages on phoneme basis and the entropies for various orders are obtained for Hindi and Tamil. The entropies on grapheme basis are also obtained. The entropy at the hundredth phoneme or grapheme may be taken to indicate the true value of the entropy in the natural language.

Reprints and Corporate Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

To request a reprint or corporate permissions for this article, please click on the relevant link below:

Academic Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

Obtain permissions instantly via Rightslink by clicking on the button below:

If you are unable to obtain permissions via Rightslink, please complete and submit this Permissions form. For more information, please visit our Permissions help page.