RL Debates 6: Thomas “no reward for you” Ringstrom
In our 6th and final RL Debates presentation, Tom introduced a rigorous framework for compositional planning that replaces the standard scalar reward with pr...
Next session: Thursday, January 08, 2026
Topic: The Final Debate/Synthesis (concluding the first round of RL Debate Series)
Presenters: Eli Sennesh, Fritz Sommer, Niels Leadholm, Adam Lowet, Anne Collins, Thomas J. Ringstrom (moderated by Hadi Vafaii)
In our 6th and final RL Debates presentation, Tom introduced a rigorous framework for compositional planning that replaces the standard scalar reward with pr...
We explored DD-DC, a provocative normative theory proposing that neurons are not just information processors, but active feedback controllers that steer thei...
In our 5th RL Debates presentation, Anne argued that reward-based learning is not always driven by RL computations. Sometimes it’s working memory combined wi...
In our 4th RL Debates presentation, Adam presented broad topics on RL and the brain, including his distributional RL paper.
In our 3rd RL Debates presentation, Niels from the Thousand Brains Project shared how their team is reverse-engineering the cortex to build sensorimotor syst...
In our 2nd RL Debates presentation, Fritz introduced a first-principles, information-theoretic approach to modeling exploration through the maximization of ‘...
In our 3rd meeting (the 1st in the RL Debate Series), we kicked off a contentious discussion on the role of the value function in RL, and why Eli believes it...
In our sceond meeting, we covered BONE, a unifying framework for online Bayesian learning when the environment is changing.
In our inaugural meeting, we covered BONG, which shows us how to do online variational inference in a principled manner.