10
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      Accurate, interpretable predictions of materials properties within transformer language models

      research-article

      Read this article at

      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Summary

          Property prediction accuracy has long been a key parameter of machine learning in materials informatics. Accordingly, advanced models showing state-of-the-art performance turn into highly parameterized black boxes missing interpretability. Here, we present an elegant way to make their reasoning transparent. Human-readable text-based descriptions automatically generated within a suite of open-source tools are proposed as materials representation. Transformer language models pretrained on 2 million peer-reviewed articles take as input well-known terms such as chemical composition, crystal symmetry, and site geometry. Our approach outperforms crystal graph networks by classifying four out of five analyzed properties if one considers all available reference data. Moreover, fine-tuned text-based models show high accuracy in the ultra-small data limit. Explanations of their internal machinery are produced using local interpretability techniques and are faithful and consistent with domain expert rationales. This language-centric framework makes accurate property predictions accessible to people without artificial-intelligence expertise.

          Highlights

          • Text descriptions are efficient in representing materials for property prediction

          • Pretrained language models outperform graph neural networks in most cases

          • Explanations provided by language models are consistent with human rationales

          The bigger picture

          Description, prediction, and explanation are traditionally acknowledged as central aims of science. In the field of materials informatics, the second objective receives the most attention, while the understanding of the resulting structure-property relationships is less emphasized. In this study, we reconcile large-scale language models and human-readable descriptions of crystal structure to facilitate materials design insights. The presented approach surpasses the state of the art in property prediction and provides transparency in the machinery of artificial-intelligence algorithms, thereby possibly improving the trust of materials scientists. In addition, the clarity of text-based representation and maturity of associated explainability methods make the approach appealing for educational uses.

          Abstract

          Black-boxed algorithms dominate in materials property prediction. Pretrained large-scale models in conjunction with interpretability techniques counteract the unfavorable tendency by providing clear explanations of suggested outputs.

          Related collections

          Most cited references133

          • Record: found
          • Abstract: not found
          • Article: not found

          Random Forests

            Bookmark
            • Record: found
            • Abstract: not found
            • Article: not found

            Regression Shrinkage and Selection Via the Lasso

              Bookmark
              • Record: found
              • Abstract: found
              • Article: not found

              Attention Is All You Need

              The dominant sequence transduction models are based on complex recurrent or convolutional neural networks in an encoder-decoder configuration. The best performing models also connect the encoder and decoder through an attention mechanism. We propose a new simple network architecture, the Transformer, based solely on attention mechanisms, dispensing with recurrence and convolutions entirely. Experiments on two machine translation tasks show these models to be superior in quality while being more parallelizable and requiring significantly less time to train. Our model achieves 28.4 BLEU on the WMT 2014 English-to-German translation task, improving over the existing best results, including ensembles by over 2 BLEU. On the WMT 2014 English-to-French translation task, our model establishes a new single-model state-of-the-art BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small fraction of the training costs of the best models from the literature. We show that the Transformer generalizes well to other tasks by applying it successfully to English constituency parsing both with large and limited training data. 15 pages, 5 figures
                Bookmark

                Author and article information

                Contributors
                Journal
                Patterns (N Y)
                Patterns (N Y)
                Patterns
                Elsevier
                2666-3899
                02 August 2023
                13 October 2023
                02 August 2023
                : 4
                : 10
                : 100803
                Affiliations
                [1 ]Department of Chemistry, Lomonosov Moscow State University, 119991 Moscow, Russia
                Author notes
                []Corresponding author korolev@ 123456colloid.chem.msu.ru
                [2]

                Lead contact

                Article
                S2666-3899(23)00158-7 100803
                10.1016/j.patter.2023.100803
                10591138
                37876904
                6ab3d66d-223f-44a2-8d17-94db105b5b42
                © 2023 The Authors

                This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/4.0/).

                History
                : 28 March 2023
                : 6 June 2023
                : 4 July 2023
                Categories
                Article

                property prediction,explainable artificial intelligence,language models,transformers,fine-tuning

                Comments

                Comment on this article