|
582650 Information-Theoretic Modeling (4 cr)
Department of Computer Science, University of Helsinki
An advanced course in the Algorithms and Machine Learning
sub-programme. Students in other programmes, as well as students
in mathematics and statistics are welcome too.
This course provides an introduction to information theory from the
computer science perspective. Much of the course can be viewed as
consequences of Shannon's central result known as the Noiseless Source
Coding Theorem. The theoretical results will be illustrated by various
practical data compression systems from Huffman coding
to Rissanen's arithmetic coding. In order to demonstrate the wide
applicability of information-theoretic concepts in intelligent
systems, we discuss information-theoretic principles in statistical
modeling, and especially the Minimum Description Length (MDL) principle.
| Jorma Rissanen, the
inventor of the MDL principle (left)
receiving a best paper award from Claude Shannon, the father of
information theory.
|
|
Note: Since the contents of this course overlap with the
earlier
Three Concepts: Information (3CI) course, it is not
possible to include both of them in your degree. This also applies to
the project.
- Lecturer
- Teemu Roos, A332,
teemu.roos at cs.helsinki.fi
- Teaching assistant
- Anupam Arohi,
anupam.arohi at cs.helsinki.fi
Language: All course material and lectures will be in
English.
Lectures: Period I: 8.9.–16.10. Tue & Fri 10–12
in C222 (except 25.9. in CK112). All the lectures will be recorded
on video and made available the following day.
Exercises: 18.9.–16.10. Fri 12–14 in C222.
Final exam: Tue 20.10. 9-12AM in B123
The course is followed by a related project in Period II. It is
recommended that you take both the course and the project. (More
information will be posted later.)
Prerequisites:
- Elementary calculus (limits and convergence, convexity)
- Basic probability (random variables, conditional and joint
distributions, expectation)
- Good programming skills (no specific language required)
Contents
The schedule below is tentative and subject to change.
If you don't have a high-speed connection, you may prefer
the downloadable
versions of the videos.
- Lecture 1. (Tue 8.9.)
- Introduction.
participation obligatory
• introduction to the course •
overview of contents
slides
video
exercises #1
- Lecture 2. (Fri 11.9.)
- Noisy Channel Coding.
• error correcting codes • noisy channel coding theorem
• Hamming codes
slides
video
- Lecture 3. (Tue 15.9.)
- Mathematical Preliminaries.
• calculus • probability •
law of large numbers • Jensen's and
Gibbs' inequalities
slides
video
exercises #2 |
files
- Lecture 4. (Fri 18.9.)
- Source Coding: Theory.
• entropy and information • asymptotic equipartition
property (AEP) • noiseless source coding theorem
slides
video
- Lecture 5. (Tue 22.9.)
- Source Coding: Practice.
• symbol codes
• Kraft inequality • entropy coding
slides
video
exercises #3
- Lecture 6. (Fri 25.9.)
- Special Guest Lecture: Daniel Schmidt
• Minimum Message Length (MML) principle
slides
video
- Lecture 7. (Tue 29.9.)
- Source Coding: Practice (contd.).
• Shannon-Fano code • Huffman code
• arithmetic coding
slides
video
exercises #4
- Lecture 8. (Fri 2.10.)
- Universal Source Coding.
• two-part codes • mixture codes • normalized
maximum likelihood
slides
video
- Lecture 9. (Tue 6.10.)
- MDL Principle.
• Occam's razor • old-style & modern MDL
slides |
more slides
video
exercises #5 |
data
- Lecture 10. (Fri 9.10.)
- MDL Principle (contd.).
• example applications
slides
video
- Lecture 11. (Tue 13.10.)
- Further Topics
• Kolmogorov complexity • gambling • lossy
coding • etc.
slides
video
- Lecture 12. (Fri 16.10.)
- Redundant Lecture
• summary of course contents • questions and answers before
final exam • introduction to project
video
Additional Material (by Lecture)
Here you can find recommended (but not required) additional material,
categorized according to (our) lectures.
- Lecture 2. Noisy Channel Coding
-
- Information and Entropy
- an MIT OpenCourseWare elementary level course •
for material by topic, click here
⇒ material on noise and errors
- Lecture 4. Source Coding: Theory
-
- Cover & Thomas, Elements of Information Theory, Chapter 2
- (see course folder, C127)
- Claude E. Shannon, "A mathematical theory of communication",
The Bell Systems Technical Journal, Vol. 27, pp.
379–423, 623–656, July-October, 1948.
- ⇒ a reprint (with corrections) from Bell Labs
- Joaquin Candela, Probability, Information Theory and Bayesian Inference,
on videolectures.net
- • part 1: probability
- • parts 2 & 3: entropy and information
- • part 4: Bayesian inference
- Lectures 5 & 7. Source Coding: Practice
-
- Witten, Neal & Cleary, "Arithmetic coding for data compression", Communications of
the ACM, Vol. 30, No 6, pp. 520–540, 1987.
- Cover & Thomas, Elements of Information Theory, Chapter 5
- (see course folder, C127)
- Lelewer & Hirschberg, "Data compression",
ACM Computing Surveys,
Vol. 19, No. 3, pp. 261–296, 1987.
- Ken Huffman, "Huffman algorithm",
Huffman Coding Blog, Dec. 24, 2008.
-
a mostly informal blog entry written by David Huffman's nephew
&bull relates Huffman code to mobiles (no, not cell phones)
- David Solomon, Data Compression: The Complete Reference,
Springer, 2004, 926 pages
- the title says it all: a comprehensive reference;
more recent editions available
- Lecture 6. Minimum Message Length Principle
-
- Lloyd Allison, Minimum Message Length, web resource
- brief introduction to basic ideas • many useful links
- Lecture 8: Universal Source Coding
-
- Peter Grünwald, The Minimum Description Length
Principle, MIT Press, 2007, 504 pages
- • part 2: universal coding
- Lectures 9 & 10: MDL Principle
-
- Peter Grünwald, The Minimum Description Length
Principle, MIT Press, 2007, 504 pages
- • chapter 1: learning, regularity, and compression
⇒ PDF
available
- Peter Grünwald, MDL Tutorial, on videolectures.net
- Jorma Rissanen, Minimum Description Length Principle,
Scholarpedia, 3(8):6727.
- Kontkanen & Myllymäki, "A linear-time algorithm for computing the multinomial stochastic
complexity", Information Processing Letters, Vol. 103,
No. 6, pp. 227–233, 2007.
- a simple yet very useful result • quite complex proof
- Lecture 11: Further Topics
-
- Li & Vitanyi, Kolmogorov Complexity and Its Applications,
Springer, 1997.
- the book about Kolmogorov complexity
- John L. Kelly, Jr., "A new interpretation of information rate",
Bell Systems Technical Journal, Vol. 35, pp. 917–926,
1956.
- ⇒
PDF
available
- Video Coding
Fundamentals & H.264, blog
- •
image coding
• video coding
|
|
|
September 8 First lecture (C222); participation obligatory.
September 9 Video of first lecture and first set of exercises
available (see Contents below).
September 18 First exercises (C222).
September 25 We have a special guest lecturer, Dr.
Daniel
Schmidt, who talks about the Minimum Message Length (MML)
principle. Note changed room: CK112.
October 20, 9-12AM Final exam, B123.
November 20, 16-20PM Separate exam in case you missed
or failed the first one, A111.
|
|