2024 DrEurekaLanguageModelGuidedSimT

From GM-RKB
Jump to navigation Jump to search

Subject Headings:

Notes

Cited By

Quotes

Abstract

Transferring policies learned in simulation to the real world is a promising strategy for acquiring robot skills at scale. However, sim-to-real approaches typically rely on manual design and tuning of the task reward function as well as the simulation physics parameters, rendering the process slow and human-labor intensive. In this paper, we investigate using Large Language Models (LLMs) to automate and accelerate sim-to-real design. Our LLM-guided sim-to-real approach requires only the physics simulation for the target task and automatically constructs suitable reward functions and domain randomization distributions to support real-world transfer. We first demonstrate our approach can discover sim-to-real configurations that are competitive with existing human-designed ones on quadruped locomotion and dexterous manipulation tasks. Then, we showcase that our approach is capable of solving novel robot tasks, such as quadruped balancing and walking atop a yoga ball, without iterative manual design.

Body

  • QUOTE: “Specifically, to generate the highest quality of reward functions, we build on Eureka, a state-of-the-art LLM-based reward design algorithm that can generate free-form, effective reward functions in code.”
  • QUOTE: “DrEureka decomposes the optimization into three stages: an LLM first synthesizes reward functions, then an initial policy is rolled out in perturbed simulations to create a suitable sampling range for physics parameters, which is finally used by the LLM to generate valid domain randomization configurations.”
  • QUOTE: “Our experiments primarily focus on quadruped locomotion and dexterous manipulation because reward design, domain randomization, and sim-to-real reinforcement learning at large have already been established as critical components of effective policy learning strategies within these domains.”

References

;

 AuthorvolumeDate ValuetitletypejournaltitleUrldoinoteyear
2024 DrEurekaLanguageModelGuidedSimTYuke Zhu
Jason Ma
William Liang
Hungju Wang
Sam Wang
Jim Fan
Osbert Bastani
Dinesh Jayaraman
DrEureka: Language Model Guided Sim-To-Real Transfer2024