AnaLog: Testing Analytical and Deductive Logic Learnability in Language Models

Samuel Ryb, Mario Giulianelli, Arabella Sinclair, Raquel Fernández

Research output: Chapter in Book/Report/Conference proceedingPublished conference contribution

1 Citation (Scopus)

Abstract

We investigate the extent to which pre-trained language models acquire analytical and deductive logical reasoning capabilities as a side effect of learning word prediction. We present AnaLog, a natural language inference task designed to probe models for these capabilities, controlling for different invalid heuristics the models may adopt instead of learning the desired generalisations. We test four languagemodels on AnaLog, finding that they have all learned, to a different extent, to encode information that is predictive of entailment beyond shallow heuristics such as lexical overlap and grammaticality. We closely analyse the best performing language model and show that while it performs more consistently than other language models across logical connectives and reasoning domains, it still is sensitive to lexical and syntactic variations in the realisation of logical statements.
Original languageEnglish
Title of host publicationProceedings of the 11th Joint Conference on Lexical and Computational Semantics
PublisherAssociation for Computational Linguistics
Pages55-68
Number of pages14
ISBN (Electronic)978-1-955917-98-8
DOIs
Publication statusPublished - 14 Jul 2022
EventThe 11th Joint Conference on Lexical and Computational
Semantics
- Seattle, United States
Duration: 14 Jul 202215 Jul 2022
https://sites.google.com/view/starsem2022/

Conference

ConferenceThe 11th Joint Conference on Lexical and Computational
Semantics
Abbreviated title*SEM 2022
Country/TerritoryUnited States
CitySeattle
Period14/07/2215/07/22
Internet address

Bibliographical note

Acknowledgements
We would like to thank the anonymous ARR and *SEM 2022 reviewers for their feedback and suggestions, as well as Ece Takmaz for her comments. Samuel Ryb and Arabella Sinclair worked on this project while affiliated with the University of Amsterdam. The project has received funding from the European Research Council (ERC) under the European Union’s Horizon 2020 research and innovation programme (grant agreement No. 819455).

1The dataset is available at https://github.com/dmg-illc/analog

Fingerprint

Dive into the research topics of 'AnaLog: Testing Analytical and Deductive Logic Learnability in Language Models'. Together they form a unique fingerprint.

Cite this