
Efficient reinforcement learning through uncertainties
- Event Type
- Seminar/Symposium
- Sponsor
- Industrial and Enterprise Systems Engineering, Dept. Head office
- Location
- Room 303, Transportation Building
- Date
- Mar 3, 2023 10:00 - 11:00 am
- Speaker
- Dongruo Zhou
- Contact
- BuuLinh Quach
- bquach@illinois.edu
- Phone
- 217-265-5220
- Views
- 83
- Originating Calendar
- ISE Faculty Candidates
*Presentation will be recorded.
Abstract:
Reinforcement learning (RL) has achieved great empirical success in many real-world problems in the last few years. However, many RL algorithms are inefficient due to their data-hungry nature. Whether there exists a universal way to improve the efficiency of existing RL algorithms remains an open question.
In this talk, I will give a selective overview of my research, which suggests that efficient (and optimal) RL can be built through the lens of uncertainties. I will show that uncertainties can not only guide RL to make decisions efficiently, but also have the ability to accelerate the learning of the optimal policy over a finite number of data samples collected from the unknown environment. By utilizing the proposed uncertainty-based framework, I design computationally efficient and statistically optimal RL algorithms under various settings, which improve existing baseline algorithms from both theoretical and empirical aspects. At the end of this talk, I will briefly discuss several additional works, and my future research plan for designing next-generation decision making algorithms.
Bio:
Dongruo Zhou is a final-year PhD student in the Department of Computer Science at UCLA, advised by Prof. Quanquan Gu. His research is broadly on the foundation of machine learning, with a particular focus on reinforcement learning and stochastic optimization. He aims to provide a theoretical understanding of machine learning methods, as well as to develop new machine learning algorithms with better performance. He is a recipient of the UCLA dissertation year fellowship.