Skip to yearly menu bar Skip to main content


Poster
in
Workshop: Symmetry and Geometry in Neural Representations

Geometric Signatures of Compositionality Across a Language Model’s Lifetime

Jin Hwa Lee · Thomas Jiralerspong · Lei Yu · Emily Cheng

Keywords: [ compositionality ] [ language models ] [ geometry ] [ intrinsic dimension ]


Abstract:

Compositionality, the notion that the meaning of an expression is constructed from the meaning of its parts and syntactic rules, permits the infinite productivity of human language. For the first time, artificial language models (LMs) are able to match human performance ina number of compositional generalization tasks. However, much remains to be understood about the computational mechanisms underlying these abilities. We take a geometric approach to this problem by relating the degree of compositionality in data to the intrinsicdimensionality of their representations under an LM, a measure of feature complexity. We show that the degree of dataset compositionality is reflected in representations’ intrinsic dimensionality, and that the relationship between compositionality and geometric complexity arises due to learned linguistic features over training. Overall, our results highlight that linear and nonlinear dimensionality measures capture different and complementary views of data complexity.

Chat is not available.