Maximizing sensory information with neural populations of arbitrary size

E Doi, L Paninski and E P Simoncelli

Published in Computational and Systems Neuroscience (CoSyNe), (III-36), Feb 2008.

This paper has been superseded by:
Efficient coding of spatial information in the primate retina
E Doi, J Gauthier, G Field, J Shlens, A Sher, M Greschner, T Machado, L Jepson, K Mathieson, D Gunning, A Litke, L Paninski, EJ Chichilnisky and E P Simoncelli.
J. Neuroscience , vol.32(46), pp. 16256--16264, Nov 2012.


One fundamental goal of sensory systems is to transmit sensory information efficiently under the condition that both the input and the neural representation are noisy, and that the number of available neurons is limited. Previous results based on this idea provide a great deal of insight into the optimality of sensory coding, in particular with regard to retinal receptive field structure and contrast sensitivity [1-4]. These results relied on the rather strong assumption that the retinal ganglion cell population is shift-invariant (i.e., receptive fields are centered on each photoreceptor and have identical profiles, regardless of their location). But mammalian retinas are clearly not shift-invariant: the cone mosaic is fairly irregular, the mapping from cones to ganglion cells is often convergent (i.e., the ganglion cells are sampled at lower density), and this convergence is also irregular and varies substantially with eccentricity. As such, ganglion cell receptive fields are highly irregular [5], making it difficult to directly compare the theory with experimental results.

In previous work [6], we relaxed the shift-invariance and the population size assumptions, solving for receptive fields that minimize the mean squared reconstruction error. Here, we relax those same assumptions in the context of maximizing the sensory information, deriving a set of conditions that the optimal receptive field populations should satisfy, and compute examples of receptive field populations that reach the theoretical limit of information rate. The basic problem setting is the same as the previous models [1-4]: obtain a population of linear receptive fields that maximize information transmitted about a signal that is solely characterized by its covariance. The input and neural noise are both assumed to be additive, independent, and Gaussian-distributed. The difference is that the individual receptive fields are allowed to vary, both in terms of their spatial extent (i.e., the subset of cones from which their responses are constructed), as well as their tuning properties (i.e., the weights applied to each cone). In addition, the number of ganglion cells is adjustable, and need not be the same as the number of cones. We are currently working to compare these results directly to a newly obtained experimental data set in which both cone locations and ganglion cell receptive field properties are known (with Chichilnisky lab). The theory may then be compared with the data in terms of information preserved, as well as the detailed properties of the receptive fields.

Acknowledgments:
This work was supported by NIH grant RO1-EY018003-01 (ED, LP, EPS) and HHMI (EPS).We would like to thank EJ Chichilnisky, Jeff Gauthier, and Greg Field for fruitful discussions.

References:
[1] An application of the principle of maximum information preservation to linear systems. R. Linsker, NIPS*1988:186-194, 1989.
[2] Towards a theory of early visual processing. J.J. Atick & A.N. Redlich, Neural Comp. 2:308-320, 1990.
[3] A theory of maximizing sensory information. J.H. van Hateren, Biol. Cybern. 68:23-29, 1992.
[4] Designing receptive fields for highest fidelity. D.L. Ruderman, Network 5:147-155, 1994.
[5] Fidelity of the ensemble code for visual motion in primate retina. E.S. Frechette, A.Sher, M.I. Grivich, D.Petrusca, A.M. Litke, and E.J. Chichilnisky, Journal of Neurophysiology 94:119-135, 2005.
[6] A theory of retinal population coding. E. Doi & M. S. Lewicki, NIPS*2006:353-360, 2007.


  • Listing of all publications