**A primer on information theory, with applications to neuroscience**

by Felix Effenberger

**Publisher**: arXiv 2013**Number of pages**: 58

**Description**:

This chapter is supposed to give a short introduction to the fundamentals of information theory; not only, but especially suited for people having a less firm background in mathematics and probability theory. Regarding applications, the focus will be on neuroscientific topics.

Download or read it online for free here:

**Download link**

(1MB, PDF)

## Similar books

**Exploring Randomness**

by

**Gregory J. Chaitin**-

**Springer**

This book presents the core of Chaitin's theory of program-size complexity, also known as algorithmic information theory. LISP is used to present the key algorithms and to enable computer users to interact with the author's proofs.

(

**9742**views)

**Information Theory, Inference, and Learning Algorithms**

by

**David J. C. MacKay**-

**Cambridge University Press**

A textbook on information theory, Bayesian inference and learning algorithms, useful for undergraduates and postgraduates students, and as a reference for researchers. Essential reading for students of electrical engineering and computer science.

(

**12414**views)

**Logic and Information**

by

**Keith Devlin**-

**ESSLLI**

An introductory, comparative account of three mathematical approaches to information: the classical quantitative theory of Claude Shannon, a qualitative theory developed by Fred Dretske, and a qualitative theory introduced by Barwise and Perry.

(

**5101**views)

**Information Theory, Excess Entropy and Statistical Complexity**

by

**David Feldman**-

**College of the Atlantic**

This e-book is a brief tutorial on information theory, excess entropy and statistical complexity. From the table of contents: Background in Information Theory; Entropy Density and Excess Entropy; Computational Mechanics.

(

**6857**views)