ANALYZING NEURAL CODING: JOINTLY TYPICAL SEQUENCES AND QUANTIZATION

 

 

Alexander G.  Dimitrov and John P. Miller

Center for Computational Biology

Montana State University

 

 The nature and information content of neural signals have   been discussed extensively in the neuroscience community. They are an important ingredient of many theories on neural function, yet there is still no agreement on the details of neural coding. There   have been various suggestions about how information is encoded in   neural spike trains: by the number of spikes, by temporal   correlations, through single spikes, or by spike patterns in one, or across many neurons. The latter scheme is most general, and   encompasses many others. We shall describe our progress in modeling   it through jointly typical sequences. The search for pattern codes   requires exponentially more data than the search for mean rate or   correlation codes. We will also describe a method that enables   optimal use of limited quantities of data, through quantization to a   reproduction set of small finite size. To asses the quality of the quantization we use an information-based distortion measure.  The   quantization is optimized to have minimal distortion for a fixed   size of the reproduction. This method allows us to study coarse models of coding schemes which can be refined as more data becomes available.