ETRI-Knowledge Sharing Plaform

ENGLISH

성과물

논문 검색
구분 SCI
연도 ~ 키워드

상세정보

학술대회 Video Prediction Recalling Long-term Motion Context via Memory Alignment Learning
Cited 61 time in scopus Download 24 time Share share facebook twitter linkedin kakaostory
저자
이상민, 김학구, 최대휘, 김형일, 노용만
발행일
202106
출처
Conference on Computer Vision and Pattern Recognition (CVPR) 2021, pp.3054-3063
DOI
https://dx.doi.org/10.1109/CVPR46437.2021.00307
협약과제
20HS5300, 장기 시각 메모리 네트워크 기반의 예지형 시각지능 핵심기술 개발, 문진영
초록
Our work addresses long-term motion context issues for predicting future frames. To predict the future precisely, it is required to capture which long-term motion context (e.g., walking or running) the input motion (e.g., leg movement) belongs to. The bottlenecks arising when dealing with the long-term motion context are: (i) how to predict the long-term motion context naturally matching input sequences with limited dynamics, (ii) how to predict the long-term motion context with high-dimensionality (e.g., complex motion). To address the issues, we propose novel motion context-aware video prediction. To solve the bottleneck (i), we introduce a long-term motion context memory (LMC-Memory) with memory alignment learning. The proposed memory alignment learning enables to store long-term motion contexts into the memory and to match them with sequences including limited dynamics. As a result, the long-term context can be recalled from the limited input sequence. In addition, to resolve the bottleneck (ii), we propose memory query decomposition to store local motion context (i.e., low-dimensional dynamics) and recall the suitable local context for each local part of the input individually. It enables to boost the alignment effects of the memory. Experimental results show that the proposed method outperforms other sophisticated RNN-based methods, especially in long-term condition. Further, we validate the effectiveness of the proposed network designs by conducting ablation studies and memory feature analysis. The source code of this work is available.
KSP 제안 키워드
Context aware, Feature Analysis, High dimensionality, Leg movement, Local context, Low-dimensional, Source Code, Video prediction, motion context, network design