ETRI-Knowledge Sharing Plaform

KOREAN
논문 검색
Type SCI
Year ~ Keyword

Detail

Journal Article Character Detection in Animated Movies Using Multi-Style Adaptation and Visual Attention
Cited 6 time in scopus Share share facebook twitter linkedin kakaostory
Authors
Hayeon Kim, Eun-Cheol Lee, Yongseok Seo, Dong-Hyuck Im, In-Kwon Lee
Issue Date
2021-06
Citation
IEEE Transactions on Multimedia, v.23, pp.1990-2004
ISSN
1520-9210
Publisher
IEEE
Language
English
Type
Journal Article
DOI
https://dx.doi.org/10.1109/TMM.2020.3006372
Abstract
Automatic identification of fictional characters is one of the primary analysis techniques for video content. A common approach to detect characters in live-action movies involves detecting human faces; however, this approach cannot be used in non-realistic domains, such as animated movies. Detection of characters in animated movies presents two major challenges: the same subject of character can be expressed in various unique styles, and there are no stylistic or other restrictions on the nature and design of character objects. To address these challenges, we introduce the 'animation adaptive region-based convolutional neural network' model to detect characters in animated movies and determine whether the detected characters are human or non-human types. Our model extends the Faster R-CNN model, which is a two-stage object detector, in the following manner: 1) we add a hierarchical animation adaptation module to learn the variety of unique styles from animated movies using a single model; 2) we incorporate a double-detector architecture to focus on the regions that are visually important in determining the character class. We build a new dataset for the animated character detection task. Experiments on this dataset show that our model outperforms other existing representative object detector models in terms of character detection. Furthermore, our model achieves significant performance improvements compared with previous state-of-the-art methods used for the character dictionary generation task. Our model is robust for a variety of animation styles and can find common visual representations of all types of characters, providing an effective way to detect animated characters.
KSP Keywords
Animated characters, CNN model, Character detection, Convolution neural network(CNN), Detection task, Faster R-CNN, Region-based, Two-Stage, Video content, Visual Representation, Visual attention