0
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      From Revisions to Insights: Converting Radiology Report Revisions into Actionable Educational Feedback Using Generative AI Models

      research-article

      Read this article at

      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          Expert feedback on trainees’ preliminary reports is crucial for radiologic training, but real-time feedback can be challenging due to non-contemporaneous, remote reading and increasing imaging volumes. Trainee report revisions contain valuable educational feedback, but synthesizing data from raw revisions is challenging. Generative AI models can potentially analyze these revisions and provide structured, actionable feedback. This study used the OpenAI GPT-4 Turbo API to analyze paired synthesized and open-source analogs of preliminary and finalized reports, identify discrepancies, categorize their severity and type, and suggest review topics. Expert radiologists reviewed the output by grading discrepancies, evaluating the severity and category accuracy, and suggested review topic relevance. The reproducibility of discrepancy detection and maximal discrepancy severity was also examined. The model exhibited high sensitivity, detecting significantly more discrepancies than radiologists ( W = 19.0, p < 0.001) with a strong positive correlation ( r = 0.778, p < 0.001). Interrater reliability for severity and type were fair (Fleiss’ kappa = 0.346 and 0.340, respectively; weighted kappa = 0.622 for severity). The LLM achieved a weighted F1 score of 0.66 for severity and 0.64 for type. Generated teaching points were considered relevant in ~ 85% of cases, and relevance correlated with the maximal discrepancy severity (Spearman ρ = 0.76, p < 0.001). The reproducibility was moderate to good (ICC (2,1) = 0.690) for the number of discrepancies and substantial for maximal discrepancy severity (Fleiss’ kappa = 0.718; weighted kappa = 0.94). Generative AI models can effectively identify discrepancies in report revisions and generate relevant educational feedback, offering promise for enhancing radiology training.

          Supplementary Information

          The online version contains supplementary material available at 10.1007/s10278-024-01233-4.

          Related collections

          Most cited references25

          • Record: found
          • Abstract: found
          • Article: not found

          Language Models are Few-Shot Learners

          Recent work has demonstrated substantial gains on many NLP tasks and benchmarks by pre-training on a large corpus of text followed by fine-tuning on a specific task. While typically task-agnostic in architecture, this method still requires task-specific fine-tuning datasets of thousands or tens of thousands of examples. By contrast, humans can generally perform a new language task from only a few examples or from simple instructions - something which current NLP systems still largely struggle to do. Here we show that scaling up language models greatly improves task-agnostic, few-shot performance, sometimes even reaching competitiveness with prior state-of-the-art fine-tuning approaches. Specifically, we train GPT-3, an autoregressive language model with 175 billion parameters, 10x more than any previous non-sparse language model, and test its performance in the few-shot setting. For all tasks, GPT-3 is applied without any gradient updates or fine-tuning, with tasks and few-shot demonstrations specified purely via text interaction with the model. GPT-3 achieves strong performance on many NLP datasets, including translation, question-answering, and cloze tasks, as well as several tasks that require on-the-fly reasoning or domain adaptation, such as unscrambling words, using a novel word in a sentence, or performing 3-digit arithmetic. At the same time, we also identify some datasets where GPT-3's few-shot learning still struggles, as well as some datasets where GPT-3 faces methodological issues related to training on large web corpora. Finally, we find that GPT-3 can generate samples of news articles which human evaluators have difficulty distinguishing from articles written by humans. We discuss broader societal impacts of this finding and of GPT-3 in general. 40+32 pages
            Bookmark
            • Record: found
            • Abstract: not found
            • Article: not found

            Leveraging GPT-4 for Post Hoc Transformation of Free-Text Radiology Reports into Structured Reporting: A Multilingual Feasibility Study

              Bookmark
              • Record: found
              • Abstract: found
              • Article: not found

              Interpretive Error in Radiology

              Although imaging technology has advanced significantly since the work of Garland in 1949, interpretive error rates remain unchanged. In addition to patient harm, interpretive errors are a major cause of litigation and distress to radiologists. In this article, we discuss the mechanics involved in searching an image, categorize omission errors, and discuss factors influencing diagnostic accuracy. Potential individual- and system-based solutions to mitigate or eliminate errors are also discussed.
                Bookmark

                Author and article information

                Contributors
                shawn.kt.lyo@gmail.com
                Journal
                J Imaging Inform Med
                J Imaging Inform Med
                Journal of Imaging Informatics in Medicine
                Springer International Publishing (Cham )
                2948-2925
                2948-2933
                19 August 2024
                19 August 2024
                April 2025
                : 38
                : 2
                : 1265-1279
                Affiliations
                Department of Radiology, Hospital of the University of Pennsylvania, ( https://ror.org/02917wp91) Philadelphia, PA USA
                Author information
                http://orcid.org/0000-0002-5999-6194
                http://orcid.org/0000-0002-4025-115X
                http://orcid.org/0000-0003-1292-5265
                http://orcid.org/0000-0003-3969-6351
                http://orcid.org/0000-0003-4765-9358
                http://orcid.org/0000-0002-6007-7267
                Article
                1233
                10.1007/s10278-024-01233-4
                11950553
                39160366
                efbb3d51-0ccf-4230-b6c2-3ec7335dd8a9
                © The Author(s) 2024

                Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.

                History
                : 13 May 2024
                : 7 August 2024
                : 8 August 2024
                Categories
                Original Paper
                Custom metadata
                © Society for Imaging Informatics in Medicine 2025

                radiology training,generative artificial intelligence,large language models,report revisions,education,precision radiology education

                Comments

                Comment on this article

                scite_
                6
                0
                3
                0
                Smart Citations
                6
                0
                3
                0
                Citing PublicationsSupportingMentioningContrasting
                View Citations

                See how this article has been cited at scite.ai

                scite shows how a scientific paper has been cited by providing the context of the citation, a classification describing whether it supports, mentions, or contrasts the cited claim, and a label indicating in which section the citation was made.

                Similar content48

                Cited by2

                Most referenced authors331