Abstract
We investigate the extent to which pre-trained language models acquire analytical and deductive logical reasoning capabilities as a side effect of learning word prediction. We present AnaLog, a natural language inference task designed to probe models for these capabilities, controlling for different invalid heuristics the models may adopt instead of learning the desired generalisations. We test four languagemodels on AnaLog, finding that they have all learned, to a different extent, to encode information that is predictive of entailment beyond shallow heuristics such as lexical overlap and grammaticality. We closely analyse the best performing language model and show that while it performs more consistently than other language models across logical connectives and reasoning domains, it still is sensitive to lexical and syntactic variations in the realisation of logical statements.
Original language | English |
---|---|
Title of host publication | Proceedings of the 11th Joint Conference on Lexical and Computational Semantics |
Publisher | Association for Computational Linguistics |
Pages | 55-68 |
Number of pages | 14 |
ISBN (Electronic) | 978-1-955917-98-8 |
DOIs | |
Publication status | Published - 14 Jul 2022 |
Event | The 11th Joint Conference on Lexical and Computational Semantics - Seattle, United States Duration: 14 Jul 2022 → 15 Jul 2022 https://sites.google.com/view/starsem2022/ |
Conference
Conference | The 11th Joint Conference on Lexical and Computational Semantics |
---|---|
Abbreviated title | *SEM 2022 |
Country/Territory | United States |
City | Seattle |
Period | 14/07/22 → 15/07/22 |
Internet address |
Bibliographical note
AcknowledgementsWe would like to thank the anonymous ARR and *SEM 2022 reviewers for their feedback and suggestions, as well as Ece Takmaz for her comments. Samuel Ryb and Arabella Sinclair worked on this project while affiliated with the University of Amsterdam. The project has received funding from the European Research Council (ERC) under the European Union’s Horizon 2020 research and innovation programme (grant agreement No. 819455).
1The dataset is available at https://github.com/dmg-illc/analog