arXiv Analytics

Sign in

arXiv:2209.00465 [cs.AI]AbstractReferencesReviewsResources

On Grounded Planning for Embodied Tasks with Language Models

Bill Yuchen Lin, Chengsong Huang, Qian Liu, Wenda Gu, Sam Sommerer, Xiang Ren

Published 2022-08-29Version 1

Language models (LMs) are shown to have commonsense knowledge of the physical world, which is fundamental for completing tasks in everyday situations. However, it is still an open question whether LMs have the ability to generate grounded, executable plans for embodied tasks. It is very challenging because LMs do not have an "eye" or "hand" to perceive the realistic environment. In this work, we show the first study on this important research question. We first present a novel problem formulation named G-PlanET, which takes as input a high-level goal and a table of objects in a specific environment. The expected output is a plan consisting of step-by-step instructions for agents to execute. To enable the study of this problem, we establish an evaluation protocol and devise a dedicated metric for assessing the quality of plans. In our extensive experiments, we show that adding flattened tables for encoding environments and using an iterative decoding strategy can both improve the LMs' ability for grounded planning. Our analysis of the results also leads to interesting non-trivial findings.

Related articles: Most relevant | Search more
arXiv:2305.03742 [cs.AI] (Published 2023-05-05)
Improved Logical Reasoning of Language Models via Differentiable Symbolic Programming
arXiv:2403.14443 [cs.AI] (Published 2024-03-21)
Language Models Can Reduce Asymmetry in Information Markets
arXiv:2311.10098 [cs.AI] (Published 2023-10-31)
Automated Parliaments: A Solution to Decision Uncertainty and Misalignment in Language Models