Authors
Alireza Fathi, Ali Farhadi, James M Rehg
Publication date
2011/11/6
Conference
2011 international conference on computer vision
Pages
407-414
Publisher
IEEE
Description
We present a method to analyze daily activities, such as meal preparation, using video from an egocentric camera. Our method performs inference about activities, actions, hands, and objects. Daily activities are a challenging domain for activity recognition which are well-suited to an egocentric approach. In contrast to previous activity recognition methods, our approach does not require pre-trained detectors for objects and hands. Instead we demonstrate the ability to learn a hierarchical model of an activity by exploiting the consistent appearance of objects, hands, and actions that results from the egocentric context. We show that joint modeling of activities, actions, and objects leads to superior performance in comparison to the case where they are considered independently. We introduce a novel representation of actions based on object-hand interactions and experimentally demonstrate the superior performance …
Total citations
2010201120122013201420152016201720182019202020212022202320243115153056645248383947253010
Scholar articles
A Fathi, A Farhadi, JM Rehg - 2011 international conference on computer vision, 2011