ETRI-Knowledge Sharing Plaform

KOREAN
논문 검색
Type SCI
Year ~ Keyword

Detail

Conference Paper 4D Effect Classification by Encoding CNN Features
Cited 2 time in scopus Share share facebook twitter linkedin kakaostory
Authors
Thomhert S. Siadari, Mikyong Han, Hyunjin Yoon
Issue Date
2017-09
Citation
International Conference on Image Processing (ICIP) 2017, pp.1812-1816
Publisher
IEEE
Language
English
Type
Conference Paper
DOI
https://dx.doi.org/10.1109/ICIP.2017.8296594
Abstract
4D effects are physical effects simulated in sync with videos, movies, and games to augment the events occurring in a story or a virtual world. Types of 4D effects commonly used for the immersive media may include seat motion, vibration, flash, wind, water, scent, thunderstorm, snow, and fog. Currently, the recognition of physical effects from a video is mainly conducted by human experts. Although 4D effects are promising in giving immersive experience and entertainment, this manual production has been the main obstacle to faster and wider application of 4D effects. In this paper, we utilize pretrained models of Convolutional Neural Networks (CNNs) to extract local visual features and propose a new representation method that combines extracted features into video level features. Classification tasks are conducted by employing Support Vector Machine (SVM). Comprehensive experiments are performed to investigate different architecture of CNNs and different type of features for 4D effect classification task and compare baseline average pooling method with our proposed video level representation. Our framework outperforms the baseline up to 2-3% in terms of mean average precision (mAP).
KSP Keywords
CNN features, Classification task, Convolution neural network(CNN), Immersive experience, Immersive media, Manual production, Pooling method, Representation method, Support VectorMachine(SVM), Virtual world, local visual features