Abstract
Short-term synaptic plasticity is modulated by long-term synaptic changes. There is, however, no general agreement on the computational role of this interaction. Here, we derive a learning rule for the release probability and the maximal synaptic conductance in a circuit model with combined recurrent and feedforward connections that allows learning to discriminate among natural inputs. Short-term synaptic plasticity thereby provides a nonlinear expansion of the input space of a linear classifier, whereas the random recurrent network serves to decorrelate the expanded input space. Computer simulations reveal that the twofold increase in the number of input dimensions through short-term synaptic plasticity improves the performance of a standard perceptron up to 100%. The distributions of release probabilities and maximal synaptic conductances at the capacity limit strongly depend on the balance between excitation and inhibition. The model also suggests a new computational interpretation of spikes evoked by stimuli outside the classical receptive field. These neuronal activitiesmay reflect decorrelation of the expanded stimulus space by intracortical synaptic connections.
Dokumententyp: | Zeitschriftenartikel |
---|---|
Publikationsform: | Publisher's Version |
Fakultät: | Biologie > Department Biologie II > Neurobiologie |
Themengebiete: | 500 Naturwissenschaften und Mathematik > 570 Biowissenschaften; Biologie |
URN: | urn:nbn:de:bvb:19-epub-14751-5 |
ISSN: | 1530-888X |
Bemerkung: | Copyright of Neural Computation is the property of MIT Press and its content may not be copied or emailed to multiple sites or posted to a listserv without the copyright holder's express written permission. However, users may print, download, or email articles for individual use. |
Sprache: | Englisch |
Dokumenten ID: | 14751 |
Datum der Veröffentlichung auf Open Access LMU: | 12. Mrz. 2013, 10:48 |
Letzte Änderungen: | 04. Nov. 2020, 12:55 |