5
views
0
recommends
+1 Recommend
1 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: not found

      HRL4EC: Hierarchical reinforcement learning for multi-mode epidemic control

      research-article

      Read this article at

      ScienceOpenPublisherPMC
      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          Infectious diseases, such as Black Death, Spanish Flu, and COVID-19, have accompanied human history and threatened public health, resulting in enormous infections and even deaths among citizens. Because of their rapid development and huge impact, laying out interventions becomes one of the most critical paths for policymakers to respond to the epidemic. However, the existing studies mainly focus on epidemic control with a single intervention, which makes the epidemic control effectiveness severely compromised. In view of this, we propose a Hierarchical Reinforcement Learning decision framework for multi-mode Epidemic Control with multiple interventions called HRL4EC. We devise an epidemiological model, referred to as MID-SEIR, to describe multiple interventions' impact on transmission explicitly, and use it as the environment for HRL4EC. Besides, to address the complexity introduced by multiple interventions, this work transforms the multi-mode intervention decision problem into a multi-level control problem, and employs hierarchical reinforcement learning to find the optimal strategies. Finally, extensive experiments are conducted with real and simulated epidemic data to validate the effectiveness of our proposed method. We further analyze the experiment data in-depth, conclude a series of findings on epidemic intervention strategies, and make a visualization accordingly, which can provide heuristic support for policymakers' pandemic response.

          Related collections

          Most cited references35

          • Record: found
          • Abstract: found
          • Article: not found

          Human-level control through deep reinforcement learning.

          The theory of reinforcement learning provides a normative account, deeply rooted in psychological and neuroscientific perspectives on animal behaviour, of how agents may optimize their control of an environment. To use reinforcement learning successfully in situations approaching real-world complexity, however, agents are confronted with a difficult task: they must derive efficient representations of the environment from high-dimensional sensory inputs, and use these to generalize past experience to new situations. Remarkably, humans and other animals seem to solve this problem through a harmonious combination of reinforcement learning and hierarchical sensory processing systems, the former evidenced by a wealth of neural data revealing notable parallels between the phasic signals emitted by dopaminergic neurons and temporal difference reinforcement learning algorithms. While reinforcement learning agents have achieved some successes in a variety of domains, their applicability has previously been limited to domains in which useful features can be handcrafted, or to domains with fully observed, low-dimensional state spaces. Here we use recent advances in training deep neural networks to develop a novel artificial agent, termed a deep Q-network, that can learn successful policies directly from high-dimensional sensory inputs using end-to-end reinforcement learning. We tested this agent on the challenging domain of classic Atari 2600 games. We demonstrate that the deep Q-network agent, receiving only the pixels and the game score as inputs, was able to surpass the performance of all previous algorithms and achieve a level comparable to that of a professional human games tester across a set of 49 games, using the same algorithm, network architecture and hyperparameters. This work bridges the divide between high-dimensional sensory inputs and actions, resulting in the first artificial agent that is capable of learning to excel at a diverse array of challenging tasks.
            Bookmark
            • Record: found
            • Abstract: not found
            • Article: not found

            A Contribution to the Mathematical Theory of Epidemics

              Bookmark
              • Record: found
              • Abstract: not found
              • Article: not found

              The Mathematics of Infectious Diseases

                Bookmark

                Author and article information

                Journal
                Inf Sci (N Y)
                Inf Sci (N Y)
                Information Sciences
                Elsevier Inc.
                0020-0255
                1872-6291
                9 May 2023
                September 2023
                9 May 2023
                : 640
                : 119065
                Affiliations
                [a ]School of Artificial Intelligence, Jilin University, Changchun 130012, China
                [b ]Engineering Research Center of Knowledge-Driven Human-Machine Intelligence, Ministry of Education, China
                [c ]College of Computer Science and Technology, Jilin University, Changchun 130012, China
                [d ]Key Laboratory of Symbolic Computation and Knowledge Engineering of Ministry of Education, China
                [e ]School of Computer Science and Engineering, Nanyang Technological University, Singapore
                Author notes
                [* ]Corresponding author at: School of Artificial Intelligence, Jilin University, Changchun 130012, China.
                Article
                S0020-0255(23)00650-3 119065
                10.1016/j.ins.2023.119065
                10167779
                4687c67d-2210-4d17-a947-0b578fa5c1eb
                © 2023 Elsevier Inc. All rights reserved.

                Since January 2020 Elsevier has created a COVID-19 resource centre with free information in English and Mandarin on the novel coronavirus COVID-19. The COVID-19 resource centre is hosted on Elsevier Connect, the company's public news and information website. Elsevier hereby grants permission to make all its COVID-19-related research that is available on the COVID-19 resource centre - including this research content - immediately available in PubMed Central and other publicly funded repositories, such as the WHO COVID database with rights for unrestricted research re-use and analyses in any form or by any means with acknowledgement of the original source. These permissions are granted for free by Elsevier for as long as the COVID-19 resource centre remains active.

                History
                : 7 December 2022
                : 21 March 2023
                : 27 April 2023
                Categories
                Article

                hierarchical reinforcement learning,epidemic control,multi-mode intervention,deep reinforcement learning

                Comments

                Comment on this article