Home  | Publications | KKS20

Are Pretrained Language Models Symbolic Reasoners Over Knowledge?

MCML Authors

Link to Profile Hinrich Schütze PI Matchmaking

Hinrich Schütze

Prof. Dr.

Principal Investigator

Abstract

How can pretrained language models (PLMs) learn factual knowledge from the training set? We investigate the two most important mechanisms: reasoning and memorization. Prior work has attempted to quantify the number of facts PLMs learn, but we present, using synthetic data, the first study that investigates the causal relation between facts present in training and facts learned by the PLM. For reasoning, we show that PLMs seem to learn to apply some symbolic reasoning rules correctly but struggle with others, including two-hop reasoning. Further analysis suggests that even the application of learned reasoning rules is flawed. For memorization, we identify schema conformity (facts systematically supported by other facts) and frequency as key factors for its success.

inproceedings


CoNLL 2020

24th Conference on Computational Natural Language Learning. Virtual, Nov 19-20, 2020.
Conference logo
A Conference

Authors

N. Kassner • B. Krojer • H. Schütze

Links

DOI

Research Area

 B2 | Natural Language Processing

BibTeXKey: KKS20

Back to Top