Logo

Information Theory and Coding

Small book cover: Information Theory and Coding

Information Theory and Coding
by

Publisher: University of Cambridge
Number of pages: 75

Description:
The aims of this course are to introduce the principles and applications of information theory. The course will study how information is measured in terms of probability and entropy, and the relationships among conditional and joint entropies; how these are used to calculate the capacity of a communication channel, with and without noise; coding schemes, including error correcting codes; how discrete channels and measures of information generalize to their continuous forms; etc.

Home page url

Download or read it online for free here:
Download link
(1.4MB, PDF)

Similar books

Book cover: Lecture Notes on Network Information TheoryLecture Notes on Network Information Theory
by - arXiv
Network information theory deals with the fundamental limits on information flow in networks and optimal coding and protocols. These notes provide a broad coverage of key results, techniques, and open problems in network information theory.
(16537 views)
Book cover: Entropy and Information TheoryEntropy and Information Theory
by - Springer
The book covers the theory of probabilistic information measures and application to coding theorems for information sources and noisy channels. This is an up-to-date treatment of traditional information theory emphasizing ergodic theory.
(20933 views)
Book cover: Algorithmic Information TheoryAlgorithmic Information Theory
by - CWI
We introduce algorithmic information theory, also known as the theory of Kolmogorov complexity. We explain this quantitative approach to defining information and discuss the extent to which Kolmogorov's and Shannon's theory have a common purpose.
(12279 views)
Book cover: Around Kolmogorov Complexity: Basic Notions and ResultsAround Kolmogorov Complexity: Basic Notions and Results
by - arXiv.org
Algorithmic information theory studies description complexity and randomness. This text covers the basic notions of algorithmic information theory: Kolmogorov complexity, Solomonoff universal a priori probability, effective Hausdorff dimension, etc.
(8375 views)