-
Notifications
You must be signed in to change notification settings - Fork 5
Expand file tree
/
Copy path68P30-EntropyEncoding.tex
More file actions
53 lines (45 loc) · 1.84 KB
/
68P30-EntropyEncoding.tex
File metadata and controls
53 lines (45 loc) · 1.84 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
\documentclass[12pt]{article}
\usepackage{pmmeta}
\pmcanonicalname{EntropyEncoding}
\pmcreated{2013-03-22 12:32:22}
\pmmodified{2013-03-22 12:32:22}
\pmowner{vampyr}{22}
\pmmodifier{vampyr}{22}
\pmtitle{entropy encoding}
\pmrecord{4}{32782}
\pmprivacy{1}
\pmauthor{vampyr}{22}
\pmtype{Definition}
\pmcomment{trigger rebuild}
\pmclassification{msc}{68P30}
\pmclassification{msc}{94A24}
\pmsynonym{entropy encoder}{EntropyEncoding}
\pmsynonym{entropy coding}{EntropyEncoding}
\pmrelated{HuffmanCoding}
\endmetadata
% this is the default PlanetMath preamble. as your knowledge
% of TeX increases, you will probably want to edit this, but
% it should be fine as is for beginners.
% almost certainly you want these
\usepackage{amssymb}
\usepackage{amsmath}
\usepackage{amsfonts}
% used for TeXing text within eps files
%\usepackage{psfrag}
% need this for including graphics (\includegraphics)
%\usepackage{graphicx}
% for neatly defining theorems and propositions
%\usepackage{amsthm}
% making logically defined graphics
%%%\usepackage{xypic}
% there are many more packages, add them here as you need them
% define commands here
\begin{document}
An \emph{entropy encoding} is a coding scheme that involves assigning codes to symbols so as to match code lengths with the probabilities of the symbols. Typically, entropy encoders are used to compress data by replacing symbols represented by equal-length codes with symbols represented by codes proportional to the negative logarithm of the probability. Therefore, the most common symbols use the shortest codes.
According to Shannon's theorem, the optimal code length for a symbol is
$$-\log_b P$$
where $b$ is the number of symbols used to make output codes and $P$ is the probability of the input symbol.
Two of the most common entropy encoding techniques are Huffman encoding and arithmetic encoding.
%%%%%
%%%%%
\end{document}