Foundations and Trends® in Machine Learning > Vol 16 > Issue 6

Reinforcement Learning, Bit by Bit

By Xiuyuan Lu, DeepMind, USA, lxlu@deepmind.com | Benjamin Van Roy, DeepMind, USA, benvanroy@deepmind.com | Vikranth Dwaracherla, DeepMind, USA, vikranthd@deepmind.com | Morteza Ibrahimi, DeepMind, USA, mibrahimi@deepmind.com | Ian Osband, DeepMind, USA, iosband@deepmind.com | Zheng Wen, DeepMind, USA, zhengwen@deepmind.com

 
Suggested Citation
Xiuyuan Lu, Benjamin Van Roy, Vikranth Dwaracherla, Morteza Ibrahimi, Ian Osband and Zheng Wen (2023), "Reinforcement Learning, Bit by Bit", Foundations and TrendsĀ® in Machine Learning: Vol. 16: No. 6, pp 733-865. http://dx.doi.org/10.1561/2200000097

Publication Date: 11 Jul 2023
© 2023 X. Lu et al.
 
Subjects
Reinforcement Learning
 

Free Preview:

Download extract

Share

Download article
In this article:
1. Introduction
2. Environments and Agents
3. Elements of Agent Design
4. Cost-Benefit Analysis
5. Retaining Information
6. Seeking Information
7. Computational Examples
8. Closing Remarks
Acknowledgements
Appendices
References

Abstract

Reinforcement learning agents have demonstrated remarkable achievements in simulated environments. Data efficiency poses an impediment to carrying this success over to real environments. The design of data-efficient agents calls for a deeper understanding of information acquisition and representation. We discuss concepts and regret analysis that together offer principled guidance. This line of thinking sheds light on questions of what information to seek, how to seek that information, and what information to retain. To illustrate concepts, we design simple agents that build on them and present computational results that highlight data efficiency.

DOI:10.1561/2200000097
ISBN: 978-1-63828-254-9
150 pp. $99.00
Buy book (pb)
 
ISBN: 978-1-63828-255-6
150 pp. $150.00
Buy E-book (.pdf)
Table of contents:
1. Introduction
2. Environments and Agents
3. Elements of Agent Design
4. Cost-Benefit Analysis
5. Retaining Information
6. Seeking Information
7. Computational Examples
8. Closing Remarks
Acknowledgements
Appendices
References

Reinforcement Learning, Bit by Bit

Reinforcement learning agents have demonstrated remarkable achievements in simulated environments. Data efficiency, however, significantly impedes carrying this success over to real environments. The design of data-efficient agents that address this problem calls for a deeper understanding of information acquisition and representation. This tutorial offers a framework that can guide associated agent design decisions. This framework is inspired in part by concepts from information theory that has grappled with data efficiency for many years in the design of communication systems.

In this tutorial, the authors shed light on questions of what information to seek, how to seek that information, and what information to retain. To illustrate the concepts, they design simple agents that build on them and present computational results that highlight data efficiency.

This book will be of interest to students and researchers working in reinforcement learning and information theorists wishing to apply their knowledge in a practical way to reinforcement learning problems.

 
MAL-097