Learning Cost Function Through Inverse Reinforcement Learning.
Project information
- Navigation through unstructured environments remains a task in which humans are substantially more competent than robots.
- We target the fundamental issue to robotics, perception-control relationship. We are Developing end-to-end policies for robotic applications that prioritize objective-driven representations, enabling robust decision-making under uncertain and incomplete environmental conditions.
- Approach:
- Address fundamentally poor perception, dynamics, and unobservability in unstructured environments.
- Algorithmically tune MPC parameters to compensate for perception through end-to-end learning
- Learn components of the objective that are difficult to manually define. Ex: how to behave when there is no sensor information from the ground
- Integrate pereption into POMDP-based planning, leveraging imitation learning without assuming convergence to an underlying MDP.