An experience-based model of syntactic ambiguity resolution using dynamic grammars and recursive neural networks
One of the central problems in the study of human language processing is ambiguity resolution: How do people resolve the extremely pervasive ambiguity of the language they encounter? One possible answer to this question is suggested by experience-based models, which claim that people typically resolve ambiguities in a way which has been successful in the past. In order to determine the course of action that has been "successful in the past" when faced with some ambiguity, it is necessary to generalize over past experience. In this talk, I will present a computational experience-based model, which learns to generalize over linguistic experience from exposure to syntactic structures in a corpus. The model is a hybrid system, which uses symbolic grammars to build and represent syntactic structures, and neural networks to rank these structures on the basis of its experience. The key features of the model are that it uses a dynamic grammar (Lombardo and Sturt, in press), which provides a very tight correspondence between grammatical derivations and incremental processing, and recursive neural networks (Frasconi et al 1998), which are able to deal with the complex hierarchical structures produced by the grammar. I will show that the model reproduces a number of the structural preferences found in the experimental psycholinguistics literature, and also performs well on unrestricted text.