Lexical density
From Wikipedia, the free encyclopedia
| This article does not cite any references or sources. (November 2007) Please help improve this article by adding citations to reliable sources. Unverifiable material may be challenged and removed. |
In computational linguistics, lexical density constitutes the estimated measure of content per functional (grammatical) and lexical units (lexemes) in total. Specifically, this is a coefficient of the word type to token ratio of a text. The main application of the lexical density coefficient is to be found in corpus linguistics.
Lexical density may be determined thus:
(Tp / Tn)x100 = Ld
Where:
Tp = the number of word types in the analysed text
Tn = the number of all tokens in the analysed text
Ld = the analysed text's lexical density
(Please note that the variable symbols applied herein are by no means conventional, they are simply random chosen designations that serve to illustrate the example in question.)

