CLARA: Classifying and Disambiguating User Commands for Reliable Interactive Robotic Agents. (arXiv:2306.10376v3 [cs.RO] UPDATED)
By: <a href="http://arxiv.org/find/cs/1/au:+Park_J/0/1/0/all/0/1">Jeongeun Park</a>, <a href="http://arxiv.org/find/cs/1/au:+Lim_S/0/1/0/all/0/1">Seungwon Lim</a>, <a href="http://arxiv.org/find/cs/1/au:+Lee_J/0/1/0/all/0/1">Joonhyung Lee</a>, <a href="http://arxiv.org/find/cs/1/au:+Park_S/0/1/0/all/0/1">Sangbeom Park</a>, <a href="http://arxiv.org/find/cs/1/au:+Chang_M/0/1/0/all/0/1">Minsuk Chang</a>, <a href="http://arxiv.org/find/cs/1/au:+Yu_Y/0/1/0/all/0/1">Youngjae Yu</a>, <a href="http://arxiv.org/find/cs/1/au:+Choi_S/0/1/0/all/0/1">Sungjoon Choi</a> Posted: June 23, 2023
In this paper, we focus on inferring whether the given user command is clear,
ambiguous, or infeasible in the context of interactive robotic agents utilizing
large language models (LLMs). To tackle this problem, we first present an
uncertainty estimation method for LLMs to classify whether the command is
certain (i.e., clear) or not (i.e., ambiguous or infeasible). Once the command
is classified as uncertain, we further distinguish it between ambiguous or
infeasible commands leveraging LLMs with situational aware context in a
zero-shot manner. For ambiguous commands, we disambiguate the command by
interacting with users via question generation with LLMs. We believe that
proper recognition of the given commands could lead to a decrease in
malfunction and undesired actions of the robot, enhancing the reliability of
interactive robot agents. We present a dataset for robotic situational
awareness, consisting pair of high-level commands, scene descriptions, and
labels of command type (i.e., clear, ambiguous, or infeasible). We validate the
proposed method on the collected dataset, pick-and-place tabletop simulation.
Finally, we demonstrate the proposed approach in real-world human-robot
interaction experiments, i.e., handover scenarios.
Provided by:
http://arxiv.org/icons/sfx.gif