Take an Emotion Walk: Perceiving Emotions from Gaits Using Hierarchical
Attention Pooling and Affective Mapping
release_ii5zakhk4bd2hczenkxxj5hjfu
by
Uttaran Bhattacharya, Christian Roncal, Trisha Mittal, Rohan Chandra,
Aniket Bera, Dinesh Manocha
2019
Abstract
We present an autoencoder-based semi-supervised approach to classify
perceived human emotions from walking styles obtained from videos or from
motion-captured data and represented as sequences of 3D poses. Given the motion
on each joint in the pose at each time step extracted from 3D pose sequences,
we hierarchically pool these joint motions in a bottom-up manner in the
encoder, following the kinematic chains in the human body. We also constrain
the latent embeddings of the encoder to contain the space of
psychologically-motivated affective features underlying the gaits. We train the
decoder to reconstruct the motions per joint per time step in a top-down manner
from the latent embeddings. For the annotated data, we also train a classifier
to map the latent embeddings to emotion labels. Our semi-supervised approach
achieves a mean average precision of 0.84 on the Emotion-Gait benchmark
dataset, which contains gaits collected from multiple sources. We outperform
current state-of-art algorithms for both emotion recognition and action
recognition from 3D gaits by 7% -- 23% on the absolute.
In text/plain
format
Archived Files and Locations
application/pdf 5.9 MB
file_jwjdgybljrgfnbohx3otu3pvgy
|
arxiv.org (repository) web.archive.org (webarchive) |
1911.08708v1
access all versions, variants, and formats of this works (eg, pre-prints)