TY - GEN
T1 - Robot-Centric Activity Prediction from First-Person Videos
T2 - 10th Annual ACM/IEEE International Conference on Human-Robot Interaction, HRI 2015
AU - Ryoo, M. S.
AU - Fuchs, Thomas J.
AU - Xia, Lu
AU - Aggarwal, J. K.
AU - Matthies, Larry
N1 - Publisher Copyright:
© 2015 ACM.
PY - 2015/3/2
Y1 - 2015/3/2
N2 - In this paper, we present a core technology to enable robot recognition of human activities during human-robot interactions. In particular, we propose a methodology for early recognition of activities from robot-centric videos (i.e., first-person videos) obtained from a robot's viewpoint during its interaction with humans. Early recognition, which is also known as activity prediction, is an ability to infer an ongoing activity at its early stage. We present an algorithm to recognize human activities targeting the camera from streaming videos, enabling the robot to predict intended activities of the interacting person as early as possible and take fast reactions to such activities (e.g., avoiding harmful events targeting itself before they actually occur). We introduce the novel concept of 'onset' that efficiently summarizes pre-activity observations, and design a recognition approach to consider event history in addition to visual features from first-person videos. We propose to represent an onset using a cascade histogram of time series gradients, and we describe a novel algorithmic setup to take advantage of such onset for early recognition of activities. The experimental results clearly illustrate that the proposed concept of onset enables better/earlier recognition of human activities from first-person videos collected with a robot.
AB - In this paper, we present a core technology to enable robot recognition of human activities during human-robot interactions. In particular, we propose a methodology for early recognition of activities from robot-centric videos (i.e., first-person videos) obtained from a robot's viewpoint during its interaction with humans. Early recognition, which is also known as activity prediction, is an ability to infer an ongoing activity at its early stage. We present an algorithm to recognize human activities targeting the camera from streaming videos, enabling the robot to predict intended activities of the interacting person as early as possible and take fast reactions to such activities (e.g., avoiding harmful events targeting itself before they actually occur). We introduce the novel concept of 'onset' that efficiently summarizes pre-activity observations, and design a recognition approach to consider event history in addition to visual features from first-person videos. We propose to represent an onset using a cascade histogram of time series gradients, and we describe a novel algorithmic setup to take advantage of such onset for early recognition of activities. The experimental results clearly illustrate that the proposed concept of onset enables better/earlier recognition of human activities from first-person videos collected with a robot.
KW - activity recognition
KW - first-person videos
KW - human-robot interaction
UR - http://www.scopus.com/inward/record.url?scp=84943521902&partnerID=8YFLogxK
U2 - 10.1145/2696454.2696462
DO - 10.1145/2696454.2696462
M3 - Conference contribution
AN - SCOPUS:84943521902
T3 - ACM/IEEE International Conference on Human-Robot Interaction
SP - 295
EP - 302
BT - HRI 2015 - Proceedings of the 2015 ACM/IEEE International Conference on Human-Robot Interaction
PB - IEEE Computer Society
Y2 - 2 March 2015 through 5 March 2015
ER -