Program

Wednesday 17.09

8:00h
Welcome coffee
9:15-9:30h
Opening remarks
9:30-10:15h
Keynote: Amy Zhang "Successor Measures and Self-supervised Reinforcement Learning "
10:15-10:30h
Contributed talk: Ahmet Onur Akman "Routing Autonomous Vehicles Using Reinforcement Learning: Progress and Future Directions"
10:30-10:45h
Contributed talk: Hao Ma "Provably Efficient Online Learning in Real-World Cyber-Physical and Robotic Systems"
10:45-11:15h
Coffee break
11:15-12:30h
Poster session A
12:30-13:30h
Lunch
13:30-14:15h
Keynote: Gerhard Neumann "From Extended Action Representations to Versatile Policy Learning in Reinforcement Learning "
Abstract

Reinforcement learning (RL) with primitive actions often leads to inefficient exploration and brittle behaviors. Extended action representations, such as motion primitives (MPs), offer a more structured approach: they encode trajectories with a concise set of parameters, naturally yielding smooth behaviors and enabling exploration in parameter space rather than in raw action space. This parametrization allows black-box RL algorithms to adapt MP parameters to diverse contexts and initial states, providing a pathway toward versatile skill acquisition. However, standard MP-based approaches result in open-loop policies; to address this, we extend them with online replanning of MP trajectories and off-policy learning strategies that exploit single-time step information. Building on this foundation, we introduce a novel algorithm for skill discovery with MPs that leverages maximum entropy RL and mixture-of-expert models to autonomously acquire diverse, reusable skills. Finally, we present diffusion policies as a more expressive policy class for maximum entropy RL, and highlight their advantageous properties for stability, flexibility, and scalability in complex domains. Together, these contributions demonstrate how extended action representations and advanced policy models can advance the efficiency and versatility of RL.

14:15-14:30h
Contributed talk: Pierre Clavier "Bringing back Bellman to LLMs"
14:30-14:45h
Contributed talk: Théo Vincent "Optimizing the Learning Trajectory of Reinforcement Learning Agents"
14:45-15:15h
Coffee break
15:15-17:00h
Poster session B
17:45h
Social: Punting on the Neckar river (meeting point) or Visit of the Museum of Ancient Cultures
19:00h
Opening Reception at Neckawa


Thursday 18.09

8:00h
Welcome coffee
9:30-10:15h
Keynote: Peter Dayan "Liking, Shaping and Biological Alignment"
Abstract

As reinforcement learners, humans and other animals are excellent at improving their otherwise miserable lot in life. This is often described in terms of optimizing utility. However, understanding utility in a non-circular manner is surprisingly difficult. I will talk about an example of the complexity that has important psychological and neural resonance - namely the distinct concepts of 'liking' and 'wanting'. The former characterizes an immediate hedonic experience; and the latter the motivational force associated with that experience. How could it be that we, or an agent, could `want' something that it does not `like', or `like' something that it would not be willing to exert any effort to acquire? I will suggest a framework for answering these questions through the medium of potential-based shaping - in which 'liking' provides immediate, but preliminary and ultimately cancellable, information about the true, long-run worth of outcomes.

10:15-10:30h
Contributed talk: Miguel Suau "Policy Confounding: Causes, Consequences, and Corrections"
10:30-10:45h
Contributed talk: Alberto Maria Metelli "A Modern Perspective on Inverse Reinforcement Learning"
10:45-11:15h
Coffee break
11:15-12:30h
Poster session C
12:30-13:30h
Lunch
13:30-14:00h
Townhall meeting
14:00-14:45h
Keynote: Leif Döring "ADDQ: Continuing the struggle of overestimation in Q-learning"
14:45-15:00h
Contributed talk: Matteo Papini "Accelerating Policy Gradient Algorithms with Data Reuse"
15:00-15:15h
Contributed talk: Liam Schramm "Online Convex Optimization as a Theoretical Framework for Long-Horizon Exploration"
15:15-15:45h
Coffee break
15:45-17:30h
Poster session D
19:00h
Conference banquet at Castle Hohentübingen


Friday 19.09

8:00h
Welcome coffee
9:30-10:15h
Keynote: Katja Hoffman "World and Human Action Models for gameplay ideation"
Abstract

Modeling complex environments and realistic human behaviors within them is a key goal of artificial intelligence research. Progress towards this goal has exciting potential for applications in video games, from new tools that empower game developers to realize new creative visions, to enabling new kinds of immersive player experiences. This talk focuses on recent research advances from the Game Intelligence team at Microsoft Research, towards scalable machine learning architectures that effectively model human gameplay, and our vision of how these innovations could empower creatives in the future.

10:15-10:30h
Contributed talk: Theresa Eimer "Reinforcement Learning Algorithms Are Misbehaved Black Boxes"
10:30-10:45h
Contributed talk: Christian Gumbsch "Brave New World Models: Enhancing Model-Based Reinforcement Learning through Structured and Semantic Representations"
10:45-11:15h
Coffee break
11:15-12:30h
Poster session E
12:30-13:30h
Lunch
13:30-14:15h
Keynote: Marcus Hutter "Universal Reinforcement Learning"
14:15-14:30h
Contributed talk: Antoine Moulin "Optimistic Planning in Infinite-Horizon Markov Decision Processes" and "Confidence Sequences for Generalized Linear Models via Regret Analysis"
14:30-14:45h
Contributed talk: Hamish Flynn "Confidence Sequences for Generalized Linear Models via Regret Analysis"
14:45h
End of conference



Poster Sessions

Poster Session A (Wednesday Morning)

Poster Session B (Wednesday Afternoon)

Poster Session C (Thursday Morning)

Poster Session D (Thursday Afternoon)

Poster Session E (Friday Morning)