The computational resources available to the brain can be summarized in terms of a number of numerical parameters. These include the number of neurons, the average number of neurons each neuron synapses with, and the basic response time of a neuron to its inputs. A fourth parameter is the strength of influence that a neuron has through its synapses on a typical neuron with which it synapses. Given the limited resources that the brain has, as summarized by these four parameters as currently understood, it turns out to be highly challenging to explain how the brain can perform even the most basic information processing tasks. In particular, the sparsity of the interconnection pattern in conjunction with generally weak synapses appears to impose very considerable constraints. The aim of the proposed research is to explore, using abstract computational models, how information processing is possible at all under these constraints. In particular the challenge is to explain how multiple tasks, such as memorization, association and inductive learning, can be supported together in such a system.

Agency
National Science Foundation (NSF)
Institute
Division of Computer and Communication Foundations (CCF)
Application #
0432037
Program Officer
Elliott Francis
Project Start
Project End
Budget Start
2004-08-15
Budget End
2008-07-31
Support Year
Fiscal Year
2004
Total Cost
$200,000
Indirect Cost
Name
Harvard University
Department
Type
DUNS #
City
Cambridge
State
MA
Country
United States
Zip Code
02138