Assessing the Unitary RNN as an End-to-End Compositional Model of Syntax

Jean-Philippe Bernardy
(University of Gothenburg)
Shalom Lappin
(University of Gothenburg, Queen Mary University of London, and King's College London)

We show that both an LSTM and a unitary-evolution recurrent neural network (URN) can achieve encouraging accuracy on two types of syntactic patterns: context-free long distance agreement, and mildly context-sensitive cross serial dependencies. This work extends recent experiments on deeply nested context-free long distance dependencies, with similar results. URNs differ from LSTMs in that they avoid non-linear activation functions, and they apply matrix multiplication to word embeddings encoded as unitary matrices. This permits them to retain all information in the processing of an input string over arbitrary distances. It also causes them to satisfy strict compositionality. URNs constitute a significant advance in the search for explainable models in deep learning applied to NLP.

In Michael Moortgat and Gijs Wijnholds: Proceedings End-to-End Compositional Models of Vector-Based Semantics (E2ECOMPVEC), NUI Galway, 15-16 August 2022, Electronic Proceedings in Theoretical Computer Science 366, pp. 9–22.
Published: 10th August 2022.

ArXived at: https://dx.doi.org/10.4204/EPTCS.366.4 bibtex PDF
References in reconstructed bibtex, XML and HTML format (approximated).
Comments and questions to: eptcs@eptcs.org
For website issues: webmaster@eptcs.org