Inverse reinforcement learning

tags
Reinforcement learning, Reinforcement learning with human feedback, Reward shaping

Recovering an unknown reward function from expert demonstrations such that the demonstrated behavior is optimal under it.

It is a classical setting extended in modern post-training via RLHF, adversarial IRL, and demonstration-conditioned implicit rewards.

Last changed | authored by

Comments

Loading comments...

Leave a comment

Back to Notes