0
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: not found

      How Can We Know What Language Models Know?

      1 , 1 , 2 , 1
      Transactions of the Association for Computational Linguistics
      MIT Press

      Read this article at

      ScienceOpenPublisher
      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          Recent work has presented intriguing results examining the knowledge contained in language models (LMs) by having the LM fill in the blanks of prompts such as “ Obama is a __ by profession”. These prompts are usually manually created, and quite possibly sub-optimal; another prompt such as “ Obama worked as a __ ” may result in more accurately predicting the correct profession. Because of this, given an inappropriate prompt, we might fail to retrieve facts that the LM does know, and thus any given prompt only provides a lower bound estimate of the knowledge contained in an LM. In this paper, we attempt to more accurately estimate the knowledge contained in LMs by automatically discovering better prompts to use in this querying process. Specifically, we propose mining-based and paraphrasing-based methods to automatically generate high-quality and diverse prompts, as well as ensemble methods to combine answers from different prompts. Extensive experiments on the LAMA benchmark for extracting relational knowledge from LMs demonstrate that our methods can improve accuracy from 31.1% to 39.6%, providing a tighter lower bound on what LMs know. We have released the code and the resulting LM Prompt And Query Archive (LPAQA) at https://github.com/jzbjyb/LPAQA .

          Related collections

          Most cited references44

          • Record: found
          • Abstract: not found
          • Article: not found

          Assessing the Ability of LSTMs to Learn Syntax-Sensitive Dependencies

            Bookmark
            • Record: found
            • Abstract: not found
            • Conference Proceedings: not found

            Context dependent recurrent neural network language model

              Bookmark
              • Record: found
              • Abstract: not found
              • Conference Proceedings: not found

              Language Models as Knowledge Bases?

                Bookmark

                Author and article information

                Journal
                Transactions of the Association for Computational Linguistics
                Transactions of the Association for Computational Linguistics
                MIT Press
                2307-387X
                December 2020
                December 2020
                : 8
                : 423-438
                Affiliations
                [1 ]Language Technologies Institute, Carnegie Mellon University.
                [2 ]Bosch Research North America.
                Article
                10.1162/tacl_a_00324
                19dbfb37-c522-4f0e-bccf-6e413d7ac927
                © 2020
                History

                Comments

                Comment on this article