views:

175

answers:

4

Information theory comes into play where ever encoding & decoding is present. For example: compression(multimedia), cryptography.

In Information Theory we encounter terms like "Entropy", "Self Information", "Mutual Information" and entire subject is based on these terms. Which just sound nothing more than abstract. Frankly, they don't really make any sense.

Is there any book/material/explanation (if you can) which explains these things in a practical way?

EDIT:

An Introduction to Information Theory: symbols, signals & noise by John Robinson Pierce is The Book that explains it the way I want (practically). Its too good. I started reading it.

+7  A: 

Shanon's original paper "A mathematical theory of communication" is one very very important resource for studying this theory. Nobody NOBODY should miss it.

By reading it you will understand how Shanon arrived at the theory which should clear most of the doubts.

Also studying workings of Huffman compression algorithm will be very helpful.

EDIT:

An Introduction to Information Theory

John R. Pierce

seems good according to the amazon reviews (I haven't tried it).

[by Googleing "information theory layman" ]

TheMachineCharmer
oh god!! not Shanon again.
claws
@ claws why? :D
TheMachineCharmer
I'm studying this because, multimedia compression is fascinates me. But the result of Information "Theory" is this fascinating subject. This Shanon dude is the root cause of this information theory which bores me to hell.
claws
@claws haha.. there is no escape from him though :D. Give him a try he won't hurt. Believe me HE IS EINSTEIN! That paper of his contains more "information" than any other text on the topic. (I like the way you address Shanon dude bro!!;)
TheMachineCharmer
You may be in a bit of a bind. A serious study *must* eventually deal with the math and use the language of the field. That said, just jumping into the math is not for everyone.
dmckee
@dmckee +1 Agreed. I think we still need to find a way to make information theory fun for those who don't want to do any math somehow?!
TheMachineCharmer
Actually, because Shannon's paper is at the beginning of the field, it is probably one of the most readable.
GregS
@dmckee +1 Agreed.
claws
+1 for Shannon's 1948 paper, which as GregS said, is not that bad to read. Pierce's book is a gentle introduction to beginners, too. (Also cheap.)
Steve
+1  A: 

I was going to recommend Feynman for pop-sci purposes, but on reflection I think it might be a good choice for easing into a serious study as well. You can't really know this stuff without getting the math, but Feynman is so evocative that he sneaks the math into without scaring the horses.

Feynman Lectures on Computation

Covers rather more ground than just information theory, but good stuff and pleasant to read. (Besides, I am obligated to pull for Team Physics. Rah! Rah! Rhee!)

dmckee
+1 for Feynman. The man was a bloody genius.
duffymo
@duffymo: The Lectures on Computation are *good*, but his little pop-sci book on QED is *mind-boggling*. It approaches perfection. I used to admire the man, but having read the QED book I've lapsed into something approaching hero worship. I *so* want to be that smart.
dmckee
duffymo
@dmckee - "I so want to be that smart." - me too, brother. Reading Feynman gave me a sense of how his contemporaries must have felt: a mix of awe, admiration, and jealousy. Tremendous stuff.
duffymo
+3  A: 

My own view on "Information Theory" is that it's essentially just applied math / statistics but because it's being applied to communications / signals it's been called "Information Theory".

The best way to start understanding the concepts is to set yourself a real task. Say for example take a few pages of your favourite blog save it as a text file and then attempt to reduce the size of the file whilst ensuring you can still reconstruct the file completely (I.e. lossless compression). You'll start for example replacing all the instances of and with a 1 for example....

I'm always of the opinion learning by doing will be the best approach

dangerstat
+1  A: 

I remember articles in, I think, Personal Computer World that presented a version of ID3 for identifying coins, though it used a heuristic alternative to the log formula. I think it minimised sums of squares rather than maximising entropy - but it was a long time ago. There was another article in (I think) Byte that used the log formula for information (not entropy) for similar things. Things like that gave me a handle that made the theory easier to cope with.

EDIT - by "not entropy" I mean I think it used weighted averages of information values, but didn't use the name "entropy".

I think construction of simple decision trees from decision tables is a very good way to understand the relationship between probability and information. It makes the link from probability to information more intuitive, and it provides examples of the weighted average to illustrate the entropy-maximizing effect of balanced probabilities. A very good day-one kind of lesson.

And what's also nice is you can then replace that decision tree with a Huffman decoding tree (which is, after all, a "which token am I decoding?" decision tree) and make that link to coding.

BTW - take a look at this link...

Mackay has a free downloadable textbook (and available in print), and while I haven't read it all, the parts I have read seemed very good. The explanation of "explaining away" in Bayes, starting page 293, in particular, sticks in mind.

CiteSeerX is a very useful resource for information theory papers (among other things).Two interesting papers are...

Though CN2 probably isn't day one material.

Steve314