7
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      Towards AI-Driven Healthcare: Systematic Optimization, Linguistic Analysis, and Clinicians’ Evaluation of Large Language Models for Smoking Cessation Interventions

      research-article

      Read this article at

      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          Creating intervention messages for smoking cessation is a labor-intensive process. Advances in Large Language Models (LLMs) offer a promising alternative for automated message generation. Two critical questions remain: 1) How to optimize LLMs to mimic human expert writing, and 2) Do LLM-generated messages meet clinical standards? We systematically examined the message generation and evaluation processes through three studies investigating prompt engineering (Study 1), decoding optimization (Study 2), and expert review (Study 3). We employed computational linguistic analysis in LLM assessment and established a comprehensive evaluation framework, incorporating automated metrics, linguistic attributes, and expert evaluations. Certified tobacco treatment specialists assessed the quality, accuracy, credibility, and persuasiveness of LLM-generated messages, using expert-written messages as the benchmark. Results indicate that larger LLMs, including ChatGPT, OPT-13B, and OPT-30B, can effectively emulate expert writing to generate well-written, accurate, and persuasive messages, thereby demonstrating the capability of LLMs in augmenting clinical practices of smoking cessation interventions.

          Related collections

          Most cited references83

          • Record: found
          • Abstract: found
          • Article: not found

          Language Models are Few-Shot Learners

          Recent work has demonstrated substantial gains on many NLP tasks and benchmarks by pre-training on a large corpus of text followed by fine-tuning on a specific task. While typically task-agnostic in architecture, this method still requires task-specific fine-tuning datasets of thousands or tens of thousands of examples. By contrast, humans can generally perform a new language task from only a few examples or from simple instructions - something which current NLP systems still largely struggle to do. Here we show that scaling up language models greatly improves task-agnostic, few-shot performance, sometimes even reaching competitiveness with prior state-of-the-art fine-tuning approaches. Specifically, we train GPT-3, an autoregressive language model with 175 billion parameters, 10x more than any previous non-sparse language model, and test its performance in the few-shot setting. For all tasks, GPT-3 is applied without any gradient updates or fine-tuning, with tasks and few-shot demonstrations specified purely via text interaction with the model. GPT-3 achieves strong performance on many NLP datasets, including translation, question-answering, and cloze tasks, as well as several tasks that require on-the-fly reasoning or domain adaptation, such as unscrambling words, using a novel word in a sentence, or performing 3-digit arithmetic. At the same time, we also identify some datasets where GPT-3's few-shot learning still struggles, as well as some datasets where GPT-3 faces methodological issues related to training on large web corpora. Finally, we find that GPT-3 can generate samples of news articles which human evaluators have difficulty distinguishing from articles written by humans. We discuss broader societal impacts of this finding and of GPT-3 in general. 40+32 pages
            Bookmark
            • Record: found
            • Abstract: found
            • Article: found
            Is Open Access

            Performance of ChatGPT on USMLE: Potential for AI-assisted medical education using large language models

            We evaluated the performance of a large language model called ChatGPT on the United States Medical Licensing Exam (USMLE), which consists of three exams: Step 1, Step 2CK, and Step 3. ChatGPT performed at or near the passing threshold for all three exams without any specialized training or reinforcement. Additionally, ChatGPT demonstrated a high level of concordance and insight in its explanations. These results suggest that large language models may have the potential to assist with medical education, and potentially, clinical decision-making.
              Bookmark
              • Record: found
              • Abstract: not found
              • Article: not found

              The Psychological Meaning of Words: LIWC and Computerized Text Analysis Methods

                Bookmark

                Author and article information

                Contributors
                Journal
                101620299
                41882
                Proc SIGCHI Conf Hum Factor Comput Syst
                Proceedings of the SIGCHI conference on human factors in computing systems. CHI Conference
                3 June 2024
                May 2024
                11 May 2024
                21 June 2024
                : 2024
                : 436
                Affiliations
                School of Computer Science, University of Oklahoma Norman, Oklahoma, USA
                TSET Health Promotion Research Center, Stephenson Cancer Center, University of Oklahoma Health Sciences Center Oklahoma City, Oklahoma, USA
                School of Computer Science, University of Oklahoma Norman, Oklahoma, USA
                School of Public Health, The University of Texas Health Science Center at Houston Austin, TX, USA
                TSET Health Promotion Research Center, Stephenson Cancer Center; Department of Family and Preventive Medicine, University of Oklahoma Health Sciences Center Oklahoma City, Oklahoma, USA
                TSET Health Promotion Research Center, Stephenson Cancer Center; Department of Family and Preventive Medicine, University of Oklahoma Health Sciences Center Oklahoma City, Oklahoma, USA
                TSET Health Promotion Research Center, Stephenson Cancer Center; Department of Family and Preventive Medicine, University of Oklahoma Health Sciences Center Oklahoma City, Oklahoma, USA
                School of Computer Science, University of Oklahoma Norman, Oklahoma, USA
                Author notes
                [*]

                These authors contributed equally to this work.

                Article
                NIHMS1998814
                10.1145/3613904.3641965
                11192205
                38912297
                94db72e7-8796-41b9-8031-d8285ea43e53

                This work is licensed under a Creative Commons Attribution International 4.0 License.

                History
                Categories
                Article

                large language model,message generation,computational linguistic analysis,expert review,smoking cessation intervention

                Comments

                Comment on this article