Divergence_Minimization

Paper accepted at ICLR!

Our Paper, LS-IQ: Implicit Reward Regularization for Inverse Reinforcement Learning, was accepted at the International Conference on Learning Representations (ICLR)! We achieve fast and stable inverse reinforcement learning by using a squared reward regularizer on a mixture distribution between the expert and the policy distribution. We show that this specific choice of regularizer results in a bounded Divergence, a bounded optimal reward function, and a bounded Q-function. This starkly contrasts the previously used regularizer that mainly resulted in an unbounded reward function causing instability. Also, we show that this regularizer gives a unique reinforcement learning perspective on the original perspective.

We evaluate our approach on complex locomotion tasks such as on the Atlas robot.

Interested? Here is our paper and here you go to our GitHub Repo.

Robot Learning Workshop

I am excited to announce that I will be co-organizing the Next-Gen Robot Learning Symposium at the Technical University of Darmstadt on 4th November 2024!

Read More »

Paper accepted at HUMANOIDs!

Our paper, Exciting Action: Investigating Efficient Exploration for Learning Musculoskeletal Humanoid Locomotion, was accepted at the International Conference on Humanoid Robots. In this work, we

Read More »

Paper accepted at ICLR!

Our Paper, Time-Efficient Reinforcement Learning with Stochastic Stateful Policies, was accepted at the International Conference on Learning Representations (ICLR) 2024! We introduce a novel training

Read More »