Entropy estimates for natural languages are useful for a number of reasons. For example, they can be used to estimate the length of a translated text or the amount of text required to make a brute-force attack on an encrypted message feasible. This paper briefly reviews the development of techniques for entropy estimation and then applies modern techniques to Irish text. We believe this addresses a gap in the literature giving an entropy estimate for the Irish language. We discuss our results in the context of entropy estimates for equivalent English text
We compared entropy for texts written in natural languages (English, Spanish) and artificial languag...
Beyond the local constraints imposed by grammar, words concatenated in long sequences carrying a com...
<p>For each language, blue bars represent the average entropy of the random ...
Entropy estimates for natural languages are useful for a number of reasons. For example,...
Shannon estimates the entropy of the set of words in printed English as 11.82 bits per word. As this...
The goal of this paper is to show the dependency of the entropy of English text on the subject of th...
This work is a discussion of algorithms for estimating the Shannon entropy h of finite symbol sequen...
The choice associated with words is a fundamental property of natural languages. It lies at the hear...
The choice associated with words is a fundamental property of natural languages. It lies at the hear...
This paper presents a quantitative approach to poetry, based on the use of several statistical measu...
The purpose of this study is to estimate and compare the entropy and redundancy of written English a...
Since Shannon's original experiment in 1951, several methods have been applied to the problem o...
We estimate the n-gram entropies of natural language texts in word-length representation and find th...
Previous work on estimating the entropy of written natural language has focused primarily on English...
Entropy is a fundamental property of a repertoire. Here, we present an efficient algorithm to estima...
We compared entropy for texts written in natural languages (English, Spanish) and artificial languag...
Beyond the local constraints imposed by grammar, words concatenated in long sequences carrying a com...
<p>For each language, blue bars represent the average entropy of the random ...
Entropy estimates for natural languages are useful for a number of reasons. For example,...
Shannon estimates the entropy of the set of words in printed English as 11.82 bits per word. As this...
The goal of this paper is to show the dependency of the entropy of English text on the subject of th...
This work is a discussion of algorithms for estimating the Shannon entropy h of finite symbol sequen...
The choice associated with words is a fundamental property of natural languages. It lies at the hear...
The choice associated with words is a fundamental property of natural languages. It lies at the hear...
This paper presents a quantitative approach to poetry, based on the use of several statistical measu...
The purpose of this study is to estimate and compare the entropy and redundancy of written English a...
Since Shannon's original experiment in 1951, several methods have been applied to the problem o...
We estimate the n-gram entropies of natural language texts in word-length representation and find th...
Previous work on estimating the entropy of written natural language has focused primarily on English...
Entropy is a fundamental property of a repertoire. Here, we present an efficient algorithm to estima...
We compared entropy for texts written in natural languages (English, Spanish) and artificial languag...
Beyond the local constraints imposed by grammar, words concatenated in long sequences carrying a com...
<p>For each language, blue bars represent the average entropy of the random ...