skip to main content
Caltech

Mathematics & Machine Learning Seminar

Tuesday, October 14, 2025
3:00pm to 4:00pm
Add to Cal
East Bridge 114
Action Gaps and Advantages in Continuous-Time Distributional Reinforcement Learning
Yash Jhaveri, Postdoctoral Associate, Department of Mathematics and Computer Science, Rutgers University-Newark,

CANCELED

When decisions are made at high frequency, traditional reinforcement learning (RL) agents struggle to accurately estimate the values of their actions (their action-values). In turn, their performance is inconsistent and often poor. To what extent the performance of distributional RL (DRL) agents suffers similarly, however, is unknown. For instance, does estimating the full distribution of action-conditioned returns lessen this struggle?

In this talk, we will show that DRL agents are just as sensitive as their traditional counterparts and how to make them more robust, to decision frequency. We will introduce distributional perspectives on action gaps and advantages. In particular, we will introduce the superiority as a probabilistic generalization of the advantage function —the core object in approaches to mitigating performance issues in high-frequency value-based RL. In addition, we will build a collection of superiority-based DRL algorithms. Through simulations in an option-trading domain, we will show that proper modeling of the superiority distribution produces improved controllers at high decision frequencies.

For more information, please contact Math Department by phone at 626-395-4335 or by email at [email protected].