UCL Discovery
UCL home » Library Services » Electronic resources » UCL Discovery

Language Models as Knowledge Bases?

Petroni, F; Rocktäschel, T; Lewis, P; Bakhtin, A; Wu, Y; Miller, AH; Riedel, S; (2019) Language Models as Knowledge Bases? In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing. (pp. pp. 2463-2473). Association for Computational Linguistics: Hong Kong, China. Green open access

[thumbnail of Rocktaschel_D19-1250.pdf]
Preview
Text
Rocktaschel_D19-1250.pdf - Published Version

Download (364kB) | Preview

Abstract

Recent progress in pretraining language models on large textual corpora led to a surge of improvements for downstream NLP tasks. Whilst learning linguistic knowledge, these models may also be storing relational knowledge present in the training data, and may be able to answer queries structured as "fill-in-the-blank" cloze statements. Language models have many advantages over structured knowledge bases: they require no schema engineering, allow practitioners to query about an open class of relations, are easy to extend to more data, and require no human supervision to train. We present an in-depth analysis of the relational knowledge already present (without fine-tuning) in a wide range of state-of-the-art pretrained language models. We find that (i) without fine-tuning, BERT contains relational knowledge competitive with traditional NLP methods that have some access to oracle knowledge, (ii) BERT also does remarkably well on open-domain question answering against a supervised baseline, and (iii) certain types of factual knowledge are learned much more readily than others by standard language model pretraining approaches. The surprisingly strong ability of these models to recall factual knowledge without any fine-tuning demonstrates their potential as unsupervised open-domain QA systems. The code to reproduce our analysis is available at https://github.com/facebookresearch/LAMA.

Type: Proceedings paper
Title: Language Models as Knowledge Bases?
Event: 2019 Conference on Empirical Methods in Natural Language Processing
Open access status: An open access version is available from UCL Discovery
DOI: 10.18653/v1/D19-1250
Publisher version: https://doi.org/10.18653/v1/D19-1250
Language: English
Additional information: © 1963–2019 ACL; other materials are copyrighted by their respective copyright holders. Materials prior to 2016 here are licensed under the Creative Commons Attribution-NonCommercial-ShareAlike 3.0 International License. Permission is granted to make copies for the purposes of teaching and research. Materials published in or after 2016 are licensed on a Creative Commons Attribution 4.0 International License.
UCL classification: UCL
UCL > Provost and Vice Provost Offices > UCL BEAMS
UCL > Provost and Vice Provost Offices > UCL BEAMS > Faculty of Engineering Science
UCL > Provost and Vice Provost Offices > UCL BEAMS > Faculty of Engineering Science > Dept of Computer Science
URI: https://discovery.ucl.ac.uk/id/eprint/10084428
Downloads since deposit
Loading...
143Downloads
Download activity - last month
Loading...
Download activity - last 12 months
Loading...
Downloads by country - last 12 months
1.China
7
2.United States
3
3.Hong Kong
3
4.Germany
2
5.Singapore
2
6.Russian Federation
2
7.France
2
8.Indonesia
1
9.Iran, Islamic Republic of
1
10.United Kingdom
1

Archive Staff Only

View Item View Item