Introduction To Multimedia Compression: National Chiao Tung University Chun-Jen Tsai 2/21/2012
Introduction To Multimedia Compression: National Chiao Tung University Chun-Jen Tsai 2/21/2012
2/22
3/22
The paper provides many insights into the essence of the communication problem
In particular, Shannon perceived that all communication is essentially digital !
4/22
Source Encoder
Channel Encoder
Channel
Channel Decoder
Source Decoder
Get most compact Representation of information Get most robust Representation of information
The information content of a source and the information capacity of a channel can be identified using the concept of entropy
5/22
6/22
Entropy in Thermodynamics
In thermodynamics, entropy is a measure of thermal energy of a body of gas
Low entropy
High entropy
8/22
Some Information
Check the entropy of the following messages
My dog cannot fly My dog runs faster than a chicken My dog is a lady dog My dog runs slower than a chicken My dog can sing
It seems that, a rare message carries more information than a common message
9/22
Frequency-based Coding
Morse code
Invented in 1838 by Morse for electrical telegraph, and expanded by Vail in 1844 To shorten the transmission of messages, English text was coded based on relative frequencies of occurrence The efficiency of Morse code can only be improved by 15% using modern theory Questions: efficient for all languages?
Context-based Coding
Braille code, by Louis Braille in 1825 Grade 1 Braille
Letters and numbers
A or 1
B or 2
C or 3
D or 4
E or 5
F or 6
G or 7
H or 8
I or 9
J or 0
Symbols
or ?
( or )
Grade 2 Braille
AND
CH
SH
ST
TH
11/22
Model-based Coding
Statistical structure is not the only way of compression. Describing things using models is usually less wordy For example, what is the minimal precise description of ?
Shannons idea the unpredictability of patterns of digits in Kolmogorovs idea the size of a program that computes
12/22
13/22
http://www.nhk.or.jp/digital/en/super_hi/
14/22
compression
xc
decompression
Lossless compression: when y is equal to x Lossy compression: when y is different from x Compression ratio: |X| : |Xc| or (|X| |Xc|/ |X|) *100%
For example, |X| = 65536 bytes, |Xc| = 16384 bytes, the compression ration is 4:1 or 75%.
Data rate: for time-varying data, the number of bits per second (or sample) required to represent the data
15/22
A common data compression scheme is to encode a description of the model, and a description of how the data differ from the model (aka, residual)
By encode, we mean to put the data in binary digits
Data Source Are they similar? Model
17/22
...
Model: xn = n + 8, n = 1, 2, Residual: en = xn x n = 0, 1, 0, 1, 1, 1, 0,
18/22
...
19/22
frequency
time
spectrogram
waveform
20/22
21/22
If fixed length coding (FLC) is used: 3 bits per symbol If variable length coding (VLC) is used: 2.58 bits per symbol 1.16 : 1 compression ratio
22/22