Teaching Robots to Learn from Human Preferences

Teaching Robots to Learn from Human Preferences

Combining Vision and Language to Improve Embodied Manipulation

VLP (Vision-Language Preference learning) framework enables robots to learn complex manipulation tasks by leveraging human feedback converted into preference signals.

  • Eliminates costly and time-consuming manual preference labeling
  • Combines vision-language models with reinforcement learning
  • Enhances robot learning in embodied manipulation tasks
  • Demonstrates significant performance improvements over baseline approaches

This research advances engineering applications by making robot training more efficient and scalable, potentially accelerating deployment of robotic systems in manufacturing, healthcare, and domestic environments.

VLP: Vision-Language Preference Learning for Embodied Manipulation

93 | 168