ETRI-Knowledge Sharing Plaform

KOREAN
논문 검색
Type SCI
Year ~ Keyword

Detail

Conference Paper Learning a Video-Text Joint Embedding using Korean Tagged Movie Clips
Cited 0 time in scopus Share share facebook twitter linkedin kakaostory
Authors
Gyeong-June Hahm, Chang-Uk Kwak, Sun-Joong Kim
Issue Date
2020-10
Citation
International Conference on Information and Communication Technology Convergence (ICTC) 2020, pp.1158-1160
Publisher
IEEE
Language
English
Type
Conference Paper
DOI
https://dx.doi.org/10.1109/ICTC49870.2020.9289342
Abstract
For intelligent multimedia services, video contents understanding is a major challenge. In the existing video retrieval approaches, manual descriptive sentence data is necessary for retrieving desired videos against user's search intent. To overcome these limitations, modeling visual concepts included in video and sentence is necessary to learn a mapping of video and text into a common vector space, where relevant videos and texts are close to each other. In this study, we construct a new dataset containing 250 Korean movies with manual text description in Korean. Also, video-text joint embedding model and its quantitative and qualitative search results are introduced. With our proposed model, video manual tagging is no longer necessary for video retrieval services.
KSP Keywords
Embedding model, Intelligent Multimedia, Multimedia Service, Proposed model, Search intent, Search results, Video and text, Video contents, Video retrieval, vector space, visual concepts