Download as pptx, pdf, or txt
Download as pptx, pdf, or txt
You are on page 1of 19

Entropy

and
Information Theory

Tahereh Toosi
IPM
Recap

[Churchland and Abbott, 2012]


Outline
• Some general problems in Neuroscience
• Information Theory
• IT in Neuroscience
• Entropy
• Mutual Information
• How to use Mutual Information?
• Summary
3
Some general problems
• Coding scheme :
• What is being encoded?
• How is it being encoded?
• With what precision?
• With what limitations?
• Multi trial or single trial?
• Goodness of our neural decoding model
4

[Borst and Theunissen, 1999]


Stimulus space
Stimulus parameter S1
Distribution of stimuli
P[S|{ti}]
given spike train

Stimulus parameter S2

Estimate of stimulus
given spike train accessible region for
all stimuli in P[S]

[Spikes : Exploring the neural code]


Statistical significance of how
neural responses vary with different stimuli

[Borst and Theunissen, 1999]


Shannon helps…

[Borst and Theunissen, 1999]


Information Theory
𝑷
  [ 𝒓 ] : 𝑝𝑟𝑜𝑏𝑎𝑏𝑖𝑙𝑖𝑡𝑦 𝑜𝑓 𝑜𝑏𝑠𝑒𝑟𝑣𝑖𝑛𝑔 𝑎 𝑟𝑒𝑠𝑝𝑜𝑛𝑠𝑒
𝒉  ( 𝑷 [ 𝒓 ] ) : 𝑠𝑢𝑟𝑝𝑟𝑖𝑠𝑒 !
h  ( P[ r 1]P[r 2]) = h ( P[ r 1])+ h( P[r 2] )❑

information is reported in units of ‘bits’, with :


𝒉(
  𝑷 [𝒓 ])=−𝒍𝒐𝒈 𝑷 [𝒓 ]
𝟐

 The minus sign makes h a decreasing function of its argument as required.


 Note that information is really a dimensionless number.

[Abbott and Dayan, 2001]


Encoding numbers in a digital
code!
one hand can carry
log2(6) ≈22.58 bits

[Spikes : Exploring the neural code]


Channel capacity of a neuron

10

[Borst and Theunissen, 1999]


Information loss

11
Entropy
𝒉(𝑷
  [𝒓 ])=−𝒍𝒐𝒈 𝟐 𝑷 [𝒓 ] quantifies the surprise or unpredictability associated
with a particular response

Shannon’s entropy is just this measure averaged over all responses:


 
𝐻=− ∑ 𝑃 [ 𝑟 ] 𝑙𝑜𝑔 2 𝑃 [𝑟 ]
𝑟

H =−(1− P[r+]) log2(1− P[r+])− P[r+] log2 P[r+]

12
The entropy of a binary code

[Abbott and Dayan, 2001]


Mutual Information idea
• To convey information about a set of stimuli, neural responses
must be different for different stimuli.
• Entropy is a measure of response variability
• Mutual Information idea : comparing the responses obtained
using a different stimulus on every trial with those measured
in trials involving repeated presentations of the same stimulus
• The mutual information is the difference between the total
response entropy and the average response entropy on trials
that involve repetitive presentation of the same stimulus

13
Mutual Information
The entropy of the responses to a given stimulus :

The noise entropy:


This is the entropy associated with that part of the response variability
that is not due to changes in the stimulus, but arises from other sources.

Mutual Information
14
How to use Information Theory:
1. Show your system stimuli.
2. Measure neural responses.
3. Estimate: P( neural response | stimulus presented )
4. From that, Estimate: P( neural repsones )
5. Compute: H(neural response) and
H(neural response | stimulus presented)
6. Calculate: I(response ; stimulus)

15
Entropy : spike count
Stimuli responses spike count

11001010 4

01000110 3

00110111 5

11000000 2

16
How to screw it up:
• Choose stimuli which are not representative.
• Measure the “wrong” aspect of the response.
• Don’t take enough data to estimate P( ) well.
• Use a crappy method of computing H( ).
• Calculate I( ) and report it without comparing it
to anything...

17

[Math for Neuroscience, Stanford]


Summary
• Strength of IT in Neuroscience:
• Coding efficiency : Comparing the overall information
transfer to maximum spike train entropy
• Calculating the absolute amount of information
transmitted
• Identifying system nonlinearities and validating any
nonlinear system
• Neural code precision : The measure of Mutual
Information
• Goodness of our neuronal decoding model : comparison
of upper/lower estimate to direct estimate 18
THANK YOU! 19

You might also like