VacSIM: Learning Effective Strategies for COVID-19 Vaccine Distribution using Reinforcement Learning

A COVID-19 vaccine is our best bet for mitigating the ongoing onslaught of the pandemic. However, vaccine is also expected to be a limited resource. An optimal allocation strategy, especially in countries with access inequities and temporal separation of hot-spots, might be an effective way of halti...

Full description

Saved in:
Bibliographic Details
Published inarXiv.org
Main Authors Awasthi, Raghav, Guliani, Keerat Kaur, Saif Ahmad Khan, Vashishtha, Aniket, Mehrab Singh Gill, Bhatt, Arshita, Nagori, Aditya, Gupta, Aniket, Kumaraguru, Ponnurangam, Sethi, Tavpritesh
Format Paper
LanguageEnglish
Published Ithaca Cornell University Library, arXiv.org 04.12.2021
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:A COVID-19 vaccine is our best bet for mitigating the ongoing onslaught of the pandemic. However, vaccine is also expected to be a limited resource. An optimal allocation strategy, especially in countries with access inequities and temporal separation of hot-spots, might be an effective way of halting the disease spread. We approach this problem by proposing a novel pipeline VacSIM that dovetails Deep Reinforcement Learning models into a Contextual Bandits approach for optimizing the distribution of COVID-19 vaccine. Whereas the Reinforcement Learning models suggest better actions and rewards, Contextual Bandits allow online modifications that may need to be implemented on a day-to-day basis in the real world scenario. We evaluate this framework against a naive allocation approach of distributing vaccine proportional to the incidence of COVID-19 cases in five different States across India (Assam, Delhi, Jharkhand, Maharashtra and Nagaland) and demonstrate up to 9039 potential infections prevented and a significant increase in the efficacy of limiting the spread over a period of 45 days through the VacSIM approach. Our models and the platform are extensible to all states of India and potentially across the globe. We also propose novel evaluation strategies including standard compartmental model-based projections and a causality-preserving evaluation of our model. Since all models carry assumptions that may need to be tested in various contexts, we open source our model VacSIM and contribute a new reinforcement learning environment compatible with OpenAI gym to make it extensible for real-world applications across the globe. (http://vacsim.tavlab.iiitd.edu.in:8000/).
ISSN:2331-8422