ETRI-Knowledge Sharing Plaform

KOREAN
논문 검색
Type SCI
Year ~ Keyword

Detail

Conference Paper Spatio-Temporal Relationship Match: Video Structure Comparison for Recognition of Complex Human Activities
Cited 495 time in scopus Download 6 time Share share facebook twitter linkedin kakaostory
Authors
M. S. Ryoo, J. K. Aggarwal
Issue Date
2009-09
Citation
International Conference on Computer Vision (ICCV) 2009, pp.1593-1600
Language
English
Type
Conference Paper
DOI
https://dx.doi.org/10.1109/ICCV.2009.5459361
Project Code
09MC3200, Hybrid u-Robot Service System Technology Development for u-City, Wonpil Yu
Abstract
Human activity recognition is a challenging task, especially when its background is unknown or changing, and when scale or illumination differs in each video. Approaches utilizing spatio-temporal local features have proved that they are able to cope with such difficulties, but they mainly focused on classifying short videos of simple periodic actions. In this paper, we present a new activity recognition methodology that overcomes the limitations of the previous approaches using local features. We introduce a novel matching, spatio-temporal relationship match, which is designed to measure structural similarity between sets of features extracted from two videos. Our match hierarchically considers spatio-temporal relationships among feature points, thereby enabling detection and localization of complex non-periodic activities. In contrast to previous approaches to 'classify' videos, our approach is designed to 'detect and localize' all occurring activities from continuous videos where multiple actors and pedestrians are present. We implement and test our methodology on a newly-introduced dataset containing videos of multiple interacting persons and individual pedestrians. The results confirm that our system is able to recognize complex non-periodic activities (e.g. 'push' and 'hug') from sets of spatio-temporal features even when multiple activities are present in the scene. ©2009 IEEE.