ETRI-Knowledge Sharing Plaform

KOREAN
논문 검색
Type SCI
Year ~ Keyword

Detail

Conference Paper A Study of Evaluation Metrics and Datasets for Video Captioning
Cited 13 time in scopus Download 2 time Share share facebook twitter linkedin kakaostory
Authors
Jaehui Park, Chibon Song, Ji-hyeong Han
Issue Date
2017-11
Citation
International Conference on Intelligent Informatics and Biomedical Sciences (ICIIBMS) 2017, pp.172-175
Language
English
Type
Conference Paper
DOI
https://dx.doi.org/10.1109/ICIIBMS.2017.8279760
Project Code
17HS2400, Basic Technology for Extracting High-level Information from Multiple Sources Data base on Intelligent Analysis, Yoo Jang-Hee
Abstract
With the fast growing interest in deep learning, various applications and machine learning tasks are emerged in recent years. Video captioning is especially gaining a lot of attention from both computer vision and natural language processing fields. Generating captions is usually performed by jointly learning of different types of data modalities that share common themes in the video. Learning with the joining representations of different modalities is very challenging due to the inherent heterogeneity resided in the mixed information of visual scenes, speech dialogs, music and sounds, and etc. Consequently, it is hard to evaluate the quality of video captioning results. In this paper, we introduce well-known metrics and datasets for evaluation of video captioning. We compare the the existing metrics and datasets to derive a new research proposal for the evaluation of video descriptions.
KSP Keywords
Computer Vision(CV), Inherent heterogeneity, Natural Language Processing, Research proposal, Video Captioning, Visual scenes, deep learning(DL), evaluation metrics, machine Learning