Model Compression for Deep Reinforcement Learning Through Mutual Information

One of the most important limitation of deep learning and deep reinforcement learning, is the number of parameters in their models (dozens to hundreds of millions). Different model compression techniques, such as policy distillation, have been proposed to alleviate this limitation. However, they nee...

Full description

Saved in:
Bibliographic Details
Published inAdvances in Artificial Intelligence – IBERAMIA 2022 pp. 196 - 207
Main Authors García-Ramírez, Jesús, Morales, Eduardo F., Escalante, Hugo Jair
Format Book Chapter
LanguageEnglish
Published Cham Springer International Publishing
SeriesLecture Notes in Computer Science
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:One of the most important limitation of deep learning and deep reinforcement learning, is the number of parameters in their models (dozens to hundreds of millions). Different model compression techniques, such as policy distillation, have been proposed to alleviate this limitation. However, they need a high number of instances to obtain acceptable performance and the use of the source model. In this work, we propose a model compression method based on the comparison of mutual information between the distribution layers of the network. This method automatically determines how much the model should be reduced, and the number of instances required to obtain acceptable performance is considerably lower than the state-of-the-art solutions (19M). It also requires lower resources because only the last two layers of the network are fine-tuned.
ISBN:3031224183
9783031224188
ISSN:0302-9743
1611-3349
DOI:10.1007/978-3-031-22419-5_17