Lexical density

From Wikipedia, the free encyclopedia

In computational linguistics, lexical density constitutes the estimated measure of content per functional (grammatical) and lexical units (lexemes) in total. Specifically, this is a coefficient of the word type to token ratio of a text. The main application of the lexical density coefficient is to be found in corpus linguistics.

Lexical density may be determined thus:


(Tp / Tn)x100 = Ld


Where:

Tp = the number of word types in the analysed text

Tn = the number of all tokens in the analysed text

Ld = the analysed text's lexical density


(Please note that the variable symbols applied herein are by no means conventional, they are simply random chosen designations that serve to illustrate the example in question.)