Repository logo

Infoscience

  • English
  • French
Log In
Logo EPFL, École polytechnique fédérale de Lausanne

Infoscience

  • English
  • French
Log In
  1. Home
  2. Academic and Research Output
  3. Conferences, Workshops, Symposiums, and Seminars
  4. Robust Inverse Reinforcement Learning under Transition Dynamics Mismatch
 
conference paper not in proceedings

Robust Inverse Reinforcement Learning under Transition Dynamics Mismatch

Viano, Luca
•
Huang, Yu-Ting
•
Parameswaran, Kamalaruban  
Show more
2021
35th Conference on Neural Information Processing Systems (NeurIPS 2021)

We study the inverse reinforcement learning (IRL) problem under a transition dynamics mismatch between the expert and the learner. Specifically, we consider the Maximum Causal Entropy (MCE) IRL learner model and provide a tight upper bound on the learner’s performance degradation based on the `1-distance between the transition dynamics of the expert and the learner. Leveraging insights from the Robust RL literature, we propose a robust MCE IRL algorithm, which is a principled approach to help with this mismatch. Finally, we empirically demonstrate the stable performance of our algorithm compared to the standard MCE IRL algorithm under transition dynamics mismatches in both finite and continuous MDP problems.

  • Files
  • Details
  • Metrics
Loading...
Thumbnail Image
Name

robust_inverse_reinforcement_l-Supplementary Material1.pdf

Type

Postprint

Version

Accepted version

Access type

openaccess

License Condition

Copyright

Size

1.61 MB

Format

Adobe PDF

Checksum (MD5)

b948502fbb23b56dba3dfc3b7b4b5b36

Logo EPFL, École polytechnique fédérale de Lausanne
  • Contact
  • infoscience@epfl.ch

  • Follow us on Facebook
  • Follow us on Instagram
  • Follow us on LinkedIn
  • Follow us on X
  • Follow us on Youtube
AccessibilityLegal noticePrivacy policyCookie settingsEnd User AgreementGet helpFeedback

Infoscience is a service managed and provided by the Library and IT Services of EPFL. © EPFL, tous droits réservés