9
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      Predicting article quality scores with machine learning: The U.K. Research Excellence Framework

      Read this article at

      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          National research evaluation initiatives and incentive schemes choose between simplistic quantitative indicators and time-consuming peer/expert review, sometimes supported by bibliometrics. Here we assess whether machine learning could provide a third alternative, estimating article quality using more multiple bibliometric and metadata inputs. We investigated this using provisional three-level REF2021 peer review scores for 84,966 articles submitted to the U.K. Research Excellence Framework 2021, matching a Scopus record 2014–18 and with a substantial abstract. We found that accuracy is highest in the medical and physical sciences Units of Assessment (UoAs) and economics, reaching 42% above the baseline (72% overall) in the best case. This is based on 1,000 bibliometric inputs and half of the articles used for training in each UoA. Prediction accuracies above the baseline for the social science, mathematics, engineering, arts, and humanities UoAs were much lower or close to zero. The Random Forest Classifier (standard or ordinal) and Extreme Gradient Boosting Classifier algorithms performed best from the 32 tested. Accuracy was lower if UoAs were merged or replaced by Scopus broad categories. We increased accuracy with an active learning strategy and by selecting articles with higher prediction probabilities, but this substantially reduced the number of scores predicted.

          Related collections

          Most cited references70

          • Record: found
          • Abstract: not found
          • Article: not found

          Bibliometrics: The Leiden Manifesto for research metrics.

            Bookmark
            • Record: found
            • Abstract: not found
            • Article: not found

            Bias in peer review

              Bookmark
              • Record: found
              • Abstract: found
              • Article: not found

              The Matthew effect in science funding

              Why do scientists with similar backgrounds and abilities often end up achieving very different degrees of success? A classic explanation is that academic achievement exhibits a “Matthew effect”: Early successes increase future success chances. We analyze data from a large academic funding program that present a unique opportunity to quantify the Matthew effect and identify generative mechanisms. Our results show that winners just above the funding threshold accumulate more than twice as much funding during the subsequent eight years as nonwinners with near-identical review scores that fall just below the threshold. This effect is partly caused by nonwinners ceasing to compete for other funding opportunities, revealing a “participation” mechanism driving the Matthew effect. A classic thesis is that scientific achievement exhibits a “Matthew effect”: Scientists who have previously been successful are more likely to succeed again, producing increasing distinction. We investigate to what extent the Matthew effect drives the allocation of research funds. To this end, we assembled a dataset containing all review scores and funding decisions of grant proposals submitted by recent PhDs in a €2 billion granting program. Analyses of review scores reveal that early funding success introduces a growing rift, with winners just above the funding threshold accumulating more than twice as much research funding (€180,000) during the following eight years as nonwinners just below it. We find no evidence that winners’ improved funding chances in subsequent competitions are due to achievements enabled by the preceding grant, which suggests that early funding itself is an asset for acquiring later funding. Surprisingly, however, the emergent funding gap is partly created by applicants, who, after failing to win one grant, apply for another grant less often.
                Bookmark

                Author and article information

                Contributors
                Journal
                Quantitative Science Studies
                MIT Press
                2641-3337
                2023
                May 01 2023
                2023
                May 01 2023
                May 01 2023
                : 4
                : 2
                : 547-573
                Article
                10.1162/qss_a_00258
                62d54eec-9b3d-4f3f-9bfe-f1259edbb110
                © 2023

                http://creativecommons.org/licenses/by/4.0/

                History

                Comments

                Comment on this article