Joint Inference of States, Robot Knowledge, and Human (False-)Beliefs
release_qjguj52anrad7l6h4y6ptftopa
by
Tao Yuan, Hangxin Liu, Lifeng Fan, Zilong Zheng, Tao Gao, Yixin Zhu, Song-Chun Zhu
2020
Abstract
Aiming to understand how human (false-)belief--a core socio-cognitive
ability--would affect human interactions with robots, this paper proposes to
adopt a graphical model to unify the representation of object states, robot
knowledge, and human (false-)beliefs. Specifically, a parse graph (pg) is
learned from a single-view spatiotemporal parsing by aggregating various object
states along the time; such a learned representation is accumulated as the
robot's knowledge. An inference algorithm is derived to fuse individual pg from
all robots across multi-views into a joint pg, which affords more effective
reasoning and inference capability to overcome the errors originated from a
single view. In the experiments, through the joint inference over pg-s, the
system correctly recognizes human (false-)belief in various settings and
achieves better cross-view accuracy on a challenging small object tracking
dataset.
In text/plain
format
Archived Files and Locations
application/pdf 7.6 MB
file_is3qvy3g3jgddhif532kj2yava
|
arxiv.org (repository) web.archive.org (webarchive) |
2004.12248v1
access all versions, variants, and formats of this works (eg, pre-prints)