Jul 7 – 15, 2016
SISSA main building
Europe/Rome timezone

Connectionist Semantic Systematicity in Language Production

Not scheduled
20m
Meeting room (7th floor) (SISSA main building)

Meeting room (7th floor)

SISSA main building

via Bonomea 265, 34136, Trieste, Italy
Student

Speaker

Mr Jesus Calvillo (Saarland University)

Description

A defining characteristic of human language is systematicity: “the ability to produce/understand some sentences is intrinsically connected to the ability to produce/understand certain others” (Fodor & Pylyshyn, 1988). Further, Fodor and Pylyshyn (1988) argue that connectionist models are not able to display systematicity without implementing a classical symbol system. The connectionist comprehension model developed by Frank, Haselager, and van Rooij (2009), however, challenges this highly debated assertion, by developing a connectionist model of comprehension which is argued to achieve relevant levels of systematicity. Their model constructs a a situation model (see Zwaan and Radvansky (1998)) of the state-ofaffairs described by a sentence that also incorporates world knowledge-driven inferences. When the model processes a sentence like ‘a boy plays soccer’, for instance, it not only recovers the explicit, literal propositional content, but also constructs a more complete situation model in which a boy is likely playing outside on a field, with a ball, with others, and so forth. Crucially, Frank et al. (2009)’s model generalizes to both sentences and situations that it has not seen during training, exhibiting different levels of semantic systematicity and is argued to provide an important step in the direction of psychologically plausible models of language comprehension. In the present paper, we examine whether the approach developed by Frank et al. (2009) is equally well suited to language production, and present a connectionist production model that generates sentences from these situation models. We employ an extended Simple Recurrent Neural Network architecture (SRN) (Elman, 1990). Our architecture is broadly similar to the one used by Frank et al. (2009), with the main difference being that the inputs and outputs are reversed; it maps situation model representations onto sequences of localist word representations. In order to assess the performance of the model, we tested it on 5 different conditions representing different levels of generalization or systematicity. In all cases, the queried sentence type has never been seen by the model. We defined a similarity score to evaluate the results. On the training set, the model achieved an average similarity score of 99.43% (and 98.23% perfect matches). On the testing set, the average similarity score across all conditions is of 97.1%, with 88.57% of perfect matches. Although the performance of the model is very high, the model elicits some mistakes that allow us to get some insight into the internal mechanism of the model. After an analysis of the output, we see that all the sentences produced are syntactically correct and semantically felicitous. The vast majority of the elicited mistakes occur when the model produces a sentence that is semantically highly similar to the one expected. We hypothesize that this happens because the model is able to roughly reconstruct the semantic space, putting together representations that are semantically similar and thus assigning similar linguistic representations to them. We conclude that our model successfully learns to produce sentences from the situation models. Importantly, we demonstrate that this model is able to describe both unseen situations, demonstrating semantic systematicity similar to Frank et al. (2009), as well as produce alternative encodings (e.g. active/passive) for a given situation, that were not seen during training and thus demonstrating syntactic systematicity.

Primary author

Mr Jesus Calvillo (Saarland University)

Co-authors

Dr Harm Brouwer (Saarland University) Prof. Matthew Crocker (Saarland University)

Presentation materials

There are no materials yet.