2
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: not found
      • Article: not found

      When dignity meets evidence

      The Lancet
      Elsevier BV

      Read this article at

      ScienceOpenPublisher
      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Related collections

          Most cited references7

          • Record: found
          • Abstract: found
          • Article: found
          Is Open Access

          Evidence based medicine: a movement in crisis?

          Trisha Greenhalgh and colleagues argue that, although evidence based medicine has had many benefits, it has also had some negative unintended consequences. They offer a preliminary agenda for the movement’s renaissance, refocusing on providing useable evidence that can be combined with context and professional expertise so that individual patients get optimal treatment
            Bookmark
            • Record: found
            • Abstract: found
            • Article: not found
            Is Open Access

            Epistemic injustice in academic global health

            This Viewpoint calls attention to the pervasive wrongs related to knowledge production, use, and circulation in global health, many of which are taken for granted. We argue that common practices in academic global health (eg, authorship practices, research partnerships, academic writing, editorial practices, sensemaking practices, and the choice of audience or research framing, questions, and methods) are peppered with epistemic wrongs that lead to or exacerbate epistemic injustice. We describe two forms of epistemic wrongs, credibility deficit and interpretive marginalisation, which stem from structural exclusion of marginalised producers and recipients of knowledge. We then illustrate these forms of epistemic wrongs using examples of common practices in academic global health, and show how these wrongs are linked to the pose (or positionality) and the gaze (or audience) of producers of knowledge. The epistemic injustice framework shown in this Viewpoint can help to surface, detect, communicate, make sense of, avoid, and potentially undo unfair knowledge practices in global health that are inflicted upon people in their capacity as knowers, and as producers and recipients of knowledge, owing to structural prejudices in the processes involved in knowledge production, use, and circulation in global health.
              Bookmark
              • Record: found
              • Abstract: found
              • Article: found
              Is Open Access

              Will COVID-19 be evidence-based medicine’s nemesis?

              Once defined in rhetorical but ultimately meaningless terms as “the conscientious, judicious and explicit use of current best evidence in making decisions about the care of individual patients” [1], evidence-based medicine rests on certain philosophical assumptions: a singular truth, ascertainable through empirical enquiry; a linear logic of causality in which interventions have particular effect sizes; rigour defined primarily in methodological terms (especially, a hierarchy of preferred study designs and tools for detecting bias); and a deconstructive approach to problem-solving (the evidence base is built by answering focused questions, typically framed as ‘PICO’—population-intervention-comparison-outcome) [2]. The trouble with pandemics is that these assumptions rarely hold. A pandemic-sized problem can be framed and contested in multiple ways. Some research questions around COVID-19, most notably relating to drugs and vaccines, are amenable to randomised controlled trials (and where such trials were possible, they were established with impressive speed and efficiency [3, 4]). But many knowledge gaps are broader and cannot be reduced to PICO-style questions. Were care home deaths avoidable [5]? Why did the global supply chain for personal protective equipment break down [6]? What role does health system resilience play in controlling the pandemic [7]? And so on. Against these—and other—wider questions, the neat simplicity of a controlled, intervention-on versus intervention-off experiment designed to produce a definitive (i.e. statistically significant and widely generalisable) answer to a focused question rings hollow. In particular, upstream preventive public health interventions aimed at supporting widespread and sustained behaviour change across an entire population (as opposed to testing the impact of a short-term behaviour change in a select sample) rarely lend themselves to such a design [8, 9]. When implementing population-wide public health interventions—whether conventional measures such as diet or exercise, or COVID-19 related ones such as handwashing, social distancing and face coverings—we must not only persuade individuals to change their behavior but also adapt the environment to make such changes easier to make and sustain [10–12]. Population-wide public health efforts are typically iterative, locally-grown and path-dependent, and they have an established methodology for rapid evaluation and adaptation [9]. But evidence-based medicine has tended to classify such designs as “low methodological quality” [13]. Whilst this has been recognised as a problem in public health practice for some time [11], the inadequacy of the dominant paradigm has suddenly become mission-critical. Whilst evidence-based medicine recognises that study designs must reflect the nature of question (randomized trials, for example, are preferred only for therapy questions [13]), even senior scientists sometimes over-apply its hierarchy of evidence. An interdisciplinary group of scholars from the UK’s prestigious Royal Society recently reviewed the use of face masks by the general public, drawing on evidence from laboratory science, mathematical modelling and policy studies [14]. The report was criticised by epidemiologists for being “non-systematic” and for recommending policy action in the absence of a quantitative estimate of effect size from robust randomized controlled trials [15]. Such criticisms appear to make two questionable assumptions: first, that the precise quantification of impact from this kind of intervention is both possible and desirable, and second, that unless we have randomized trial evidence, we should do nothing. It is surely time to turn to a more fit-for-purpose scientific paradigm. Complex adaptive systems theory proposes that precise quantification of particular cause-effect relationships is both impossible (because such relationships are not constant and cannot be meaningfully isolated) and unnecessary (because what matters is what emerges in a particular real-world situation). This paradigm proposes that where multiple factors are interacting in dynamic and unpredictable ways, naturalistic methods and rapid-cycle evaluation are the preferred study design. The 20th-century logic of evidence-based medicine, in which scientists pursued the goals of certainty, predictability and linear causality, remains useful in some circumstances (for example, the drug and vaccine trials referred to above). But at a population and system level, we need to embrace 21st-century epistemology and methods to study how best to cope with uncertainty, unpredictability and non-linear causality [16]. In a complex system, the question driving scientific inquiry is not “what is the effect size and is it statistically significant once other variables have been controlled for?” but “does this intervention contribute, along with other factors, to a desirable outcome?”. Multiple interventions might each contribute to an overall beneficial effect through heterogeneous effects on disparate causal pathways, even though none would have a statistically significant impact on any predefined variable [11]. To illuminate such influences, we need to apply research designs that foreground dynamic interactions and emergence. These include in-depth, mixed-method case studies (primary research) and narrative reviews (secondary research) that tease out interconnections and highlight generative causality across the system [16, 17]. Table 1 lists some philosophical contrasts between the evidence-based medicine and complex-systems paradigms. Ogilvie et al have argued that rather than pitting these two paradigms against one another, they should be brought together [9]. As illustrated in (Fig 1), these authors depict randomized trials (what they call the “evidence-based practice pathway”) and natural experiments (the “practice-based evidence pathway”) in a complementary and recursive relationship rather than a hierarchical one. They propose that “…intervention studies [e.g. trials] should focus on reducing critical uncertainties, that non-randomised study designs should be embraced rather than tolerated and that a more nuanced approach to appraising the utility of diverse types of evidence is required.” (page 203) [9]. 10.1371/journal.pmed.1003266.g001 Fig 1 Ogilvie et al’s model of two complementary modes of evidence generation: evidence-based practice and practice-based evidence. Reproduced under CC-BY-4.0 licence from authors’ original [9]. 10.1371/journal.pmed.1003266.t001 Table 1 Evidence-based medicine versus complex systems research paradigms. Adapted under Creative Commons licence from Greenhalgh and Papoutsi [16]. Evidence-based medicine paradigm Complex systems paradigm Perspective on scientific truth Singular, independent of the observer, ascertainable through empirical inquiry Multiple, influenced by mode of inquiry and perspective taken Goal of research Establishing the truth; finding more or less universal and generalisable solutions to well-defined problems Exploring tensions; generating insights and wisdom; exposing multiple perspectives; viewing complex systems as moving targets Assumed model of causality Linear, cause-and-effect causality (perhaps incorporating mediators and moderators) Emergent causality: multiple interacting influences account for a particular outcome but none can be said to have a fixed ‘effect size’ Typical format of research question “What is the effect size of the intervention on the predefined outcome, and is it statistically significant?” “What combination of influences has generated this phenomenon? What does the intervention of interest contribute? What happens to the system and its actors if we intervene in a particular way? What are the unintended consequences elsewhere in the system?” Mode of representation Attempt to represent science in one authoritative voice Attempt to illustrate the plurality of voices inherent in the research and phenomena under study Good research is characterised by Methodological ‘rigour’, i.e. strict application of structured and standardised design, conventional approaches to generalisability and validity Strong theory, flexible methods, pragmatic adaptation to emerging circumstances, contribution to generative learning and theoretical transferability Purpose of theorising Disjunctive: simplification and abstraction; breaking problems down into analysable parts Conjunctive: drawing parts of the problem together to produce a rich, nuanced picture of what is going on and why Approach to data Research should continue until data collection is complete Data will never be complete or perfect; decisions often need to be made despite incomplete or contested data Analytic focus Dualisms: A versus B; influence of X on Y Dualities: inter-relationships and dynamic tensions between A, B, C and other emergent aspects In the current fast-moving pandemic, where the cost of inaction is counted in the grim mortality figures announced daily, implementing new policy interventions in the absence of randomized trial evidence has become both a scientific and moral imperative. Whilst it is hard to predict anything in real time, history will one day tell us whether adherence to “evidence-based practice” helped or hindered the public health response to Covid-19—or whether an apparent slackening of standards to accommodate “practice-based evidence” was ultimately a more effective strategy.
                Bookmark

                Author and article information

                Journal
                The Lancet
                The Lancet
                Elsevier BV
                01406736
                February 2023
                February 2023
                : 401
                : 10374
                : 340-341
                Article
                10.1016/S0140-6736(23)00176-9
                3624f7bd-726b-492e-bb3b-b1574cd10a1c
                © 2023

                https://www.elsevier.com/tdm/userlicense/1.0/

                https://doi.org/10.15223/policy-017

                https://doi.org/10.15223/policy-037

                https://doi.org/10.15223/policy-012

                https://doi.org/10.15223/policy-029

                https://doi.org/10.15223/policy-004

                History

                Comments

                Comment on this article