The distribution of word probabilities in the monkey model of Zipf’s law is associated with two universality properties: (1) the exponent in the approximate power law approaches −1 as the alphabet size increases and the letter probabilities are specified as the spacings from a random division of the unit interval for any distribution with a bounded density function on [0,1] ; and (2), on a logarithmic scale the version of the model with a finite word length cutoff and unequal letter probabilities is approximately normally distributed in the part of the distribution away from the tails. The first prop...
Some authors have recently argued that a finite-size scaling law for the text-length dependence of w...
Background: Zipf's law states that the relationship between the frequency of a word in a text and it...
We investigate the origin of Zipf's law for words in written texts by means of a stochastic dynamic ...
H. Simon proposed a simple stochastic process for explaining Zipf’s law for word frequencies. Here w...
The word-frequency distribution provides the fundamental building blocks that generate discourse in ...
We study a model of generating words with independent unequal probabilities of letters. We prove tha...
In recent years, researchers have realized the difficulties of fitting power-law distributions prope...
The word-frequency distribution provides the fundamental building blocks that generate discourse in ...
The frequency of words and letters in bodies of text has been heavily studied for several purposes, ...
We investigate the origin of Zipf's law for words in written texts by means of a stochastic dynamic ...
This paper studies the limits of language models' statistical learning in the context of Zipf's law....
In a language corpus, the probability that a word occurs n times is often proportional to 1/ n 2. As...
The paper studies the probabilities of the occurrence of m- word phrases (m=2,3,...) in relation wit...
In many practical situations, we have probability distributions for which, for large values of the c...
In the last years, researchers have realized the difficulties of fitting power-law distributions pro...
Some authors have recently argued that a finite-size scaling law for the text-length dependence of w...
Background: Zipf's law states that the relationship between the frequency of a word in a text and it...
We investigate the origin of Zipf's law for words in written texts by means of a stochastic dynamic ...
H. Simon proposed a simple stochastic process for explaining Zipf’s law for word frequencies. Here w...
The word-frequency distribution provides the fundamental building blocks that generate discourse in ...
We study a model of generating words with independent unequal probabilities of letters. We prove tha...
In recent years, researchers have realized the difficulties of fitting power-law distributions prope...
The word-frequency distribution provides the fundamental building blocks that generate discourse in ...
The frequency of words and letters in bodies of text has been heavily studied for several purposes, ...
We investigate the origin of Zipf's law for words in written texts by means of a stochastic dynamic ...
This paper studies the limits of language models' statistical learning in the context of Zipf's law....
In a language corpus, the probability that a word occurs n times is often proportional to 1/ n 2. As...
The paper studies the probabilities of the occurrence of m- word phrases (m=2,3,...) in relation wit...
In many practical situations, we have probability distributions for which, for large values of the c...
In the last years, researchers have realized the difficulties of fitting power-law distributions pro...
Some authors have recently argued that a finite-size scaling law for the text-length dependence of w...
Background: Zipf's law states that the relationship between the frequency of a word in a text and it...
We investigate the origin of Zipf's law for words in written texts by means of a stochastic dynamic ...