A coding theorem on generalized \(R\)-norm entropy (Q2761621)
From MaRDI portal
| This is the item page for this Wikibase entity, intended for internal use and editing purposes. Please use this page instead for the normal view: A coding theorem on generalized \(R\)-norm entropy |
scientific article; zbMATH DE number 1686037
| Language | Label | Description | Also known as |
|---|---|---|---|
| English | A coding theorem on generalized \(R\)-norm entropy |
scientific article; zbMATH DE number 1686037 |
Statements
28 August 2003
0 references
Shannon entropy
0 references
entropy in coding theory
0 references
mean codeword length
0 references
\(R\)-norm entropy
0 references
A coding theorem on generalized \(R\)-norm entropy (English)
0 references
The generalized \(R\)-norm entropy of a distribution \(P= (p_1,\dots, p_N)\), \(p_i> 0\), \(\sum p_i= 1\), is defined by NEWLINE\[NEWLINEH_{{R\over 2-\beta}}(P)= {R\over R+\beta-2} \Biggl[1- (\sum p^{{R\over 2-\beta}}_i)^{{2-\beta\over R}}\Biggr],\text{ where }0< \beta< 2,\;R+\beta\neq 2;NEWLINE\]NEWLINE if \(\beta= 1\) and \(R\to 1\) this approaches the Shannon entropy. Applications of \(R\)-norm entropy in coding theory are studied; a theorem on bounds of generalized mean codeword length in terms of the \(R\)-norm entropy is proven.
0 references