Skip to main content
eScholarship
Open Access Publications from the University of California

Distributional Language Models and the Representation of Multiple Kinds of Semantic Relations

Creative Commons 'BY' version 4.0 license
Abstract

Distributional models (such as neural network language mod- els) have been successfully used to model a wide range of lin- guistic semantic behaviors. However, they lack a way to dis- tinctly represent different kinds of semantic relations within a single semantic space. Here, we propose that neural network language models can sensibly be interpreted as representing syntagmatic (co-occurrence) relations using their input-output mappings, and as representing paradigmatic (similarity) rela- tions using the similarity of their internal representations. We tested and found support for this hypothesis on four neural net- work architectures (SRNs, LSTMs, Word2Vec and GPT-2) us- ing a carefully constructed artificial language corpus. Using this corpus, we show that the models display interesting but understandable differences in their ability to represent these two kinds of relationships. This work demonstrates distribu- tional models can simultaneously learn multiple kinds of re- lationships, and that systematic investigation of these models can lead to a deeper understanding of how they work.

Main Content
For improved accessibility of PDF content, download the file to your device.
Current View