A spiking recurrent neural network for semantic processing
Sentence processing requires the ability to establish
thematic relations between constituents. Here we
investigate the computational basis of this ability in a
neurobiologically motivated comprehension model.
The model has a tripartite architecture where input
representations are supplied by the mental lexicon
to a network that performs incremental thematic role
assignment. Roles are combined into a representation
of sentence-level meaning by a downstream system
(semantic unification). Recurrent, sparsely connected,
spiking networks were used which project a time-varying
input signal (word sequences) into a high-dimensional,
spatio-temporal pattern of activations. Local, adaptive
linear read-out units were then calibrated to map the
internal dynamics to desired output (thematic role
sequences) [1]. Read-outs were adjusted on network
dynamics driven by input sequences drawn from
argument-structure templates with small variation in
function words and larger variation in content words.
Models were trained on sequences of 10K words for
200ms per word at a 1ms resolution, and tested on novel
items generated from the language. We found that a
static, random recurrent spiking network outperformed
models that used only local word information without
context. To improve performance, we explored various
ways of increasing the model’s processing memory
(e.g., network size, time constants, sparseness, input
strength, etc.) and employed spiking neurons with
more dynamic variables (leaky integrate-and-fire versus
Izhikevich-neurons). The largest gain was observed
when the model’s input history was extended to include
previous words and/or roles. Model behavior was also
compared for localist and distributed encodings of word
sequences. The latter were obtained by compressing
lexical co-occurrence statistics into continuous-valued
vectors [2]. We found that performance for localist-input
was superior even though distributed representations
contained extra information about word context and
semantic similarity. Finally, we compared models that
received input enriched with combinations of semantic
features, word-category, and verb sub-categorization
labels. Counter-intuitively, we found that adding this
information to the model’s lexical input did not further
improve performance. Consistent with previous results,
however, performance improved for increased variability
in content words [3]. This indicates that the approach
to comprehension taken here might scale to more
diverse and naturalistic language input. Overall, the
results suggest that active processing memory beyond
pure state-dependent effects is important for sentence
interpretation, and that memory in neurobiological
systems might be actively computing [4]. Future work
therefore needs to address how the structure of word representations interacts with enhanced processing
memory in adaptive spiking networks. [1] Maass
W., Natschläger T., & Markram H. (2002). Real-time
computing without stable states: A new framework for
neural computation based on perturbations. Neural
Computation, 14: 2531-2560. [2] Mikolov, T., Chen, K.,
Corrado, G., & Dean, J. (2013). Efficient estimation of
word represen-tations in vector space. Proceedings of the
International Conference on Learning Represen-tations,
Scottsdale/AZ. [3] Fitz, H. (2011). A liquid-state model
of variability effects in learning nonadjacent dependencies.
Proceedings of the 33rd Annual Conference of the
Cognitive Science Society, Austin/TX. [4] Petersson,
K.M., & Hagoort, P. (2012). The neurobiology of syntax:
Beyond string-sets. Philo-sophical Transactions of the
Royal Society B 367: 1971-1883.
Publication type
PosterPublication date
2014
Share this page