Controlling Action Space of Reinforcement Learning-based Energy Management in Batteryless Applications

Duty cycle management is critical for energy-neutral operation of batteryless devices. Many efforts have been made to develop an effective duty cycling method, including machine learning-based approaches, but existing methods can barely handle the dynamic harvesting environments of batteryless devic...

Full description

Saved in:
Bibliographic Details
Published inIEEE internet of things journal Vol. 10; no. 11; p. 1
Main Authors Ahn, JunIck, Kim, Daeyong, Ha, Rhan, Cha, Hojung
Format Journal Article
LanguageEnglish
Published Piscataway IEEE 01.06.2023
The Institute of Electrical and Electronics Engineers, Inc. (IEEE)
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:Duty cycle management is critical for energy-neutral operation of batteryless devices. Many efforts have been made to develop an effective duty cycling method, including machine learning-based approaches, but existing methods can barely handle the dynamic harvesting environments of batteryless devices. Specifically, most machine learning-based methods require the harvesting patterns to be collected in advance, as well as manual configuration of the duty-cycle boundaries. In this paper, we propose a configuration-free duty cycling scheme for batteryless devices, called CTRL, with which energy harvesting nodes tune the duty cycle themselves adapting to the surrounding environment without user intervention. This approach combines reinforcement learning (RL) with a control system to allow the learning algorithm to explore all possible search space automatically. The learning algorithm sets the target state of charge (SoC) of the energy storage, instead of explicitly setting the target task frequency at a given time. The control system then satisfies the target SoC by controlling the duty cycle. An evaluation based on real implementation of the system using publicly available trace data shows that CTRL outperforms state-of-the-art approaches, resulting in 40% less frequent power failures in energy-scarce environments, while achieving more than ten times the task frequency in energy-rich environments.
ISSN:2327-4662
2327-4662
DOI:10.1109/JIOT.2023.3234905