ETRI-Knowledge Sharing Plaform

KOREAN
논문 검색
Type SCI
Year ~ Keyword

Detail

Conference Paper Segmentation-Based Masked Sampling for text-to-animated image synthesis in disaster scenarios
Cited 1 time in scopus Share share facebook twitter linkedin kakaostory
Authors
Ru-Bin Won, Minji Choi, Ji Hoon Choi, Byungjun Bae
Issue Date
2023-10
Citation
International Conference on Information and Communication Technology Convergence (ICTC) 2023, pp.1524-1527
Publisher
IEEE
Language
English
Type
Conference Paper
DOI
https://dx.doi.org/10.1109/ICTC58733.2023.10392655
Abstract
Generative AI has demonstrated significant capabilities in text-to-video synthesis, using advanced models characterized by extensive parameters. Given that current disaster alert services are mostly text-based and can be less accessible to many, this paper proposes a new approach. We aim to provide animated disaster images that precisely mirrors text prompts, thereby enhancing the efficiency and accessibility of disaster alerts. Our methodology combines the strengths of segmentation models and pre-trained Vision Transformer (ViT) mechanisms. By using a unique image selection based on CLIPScore and processing it with the CLIPSeg segmentation model, we generate an animated representation of disaster scenarios. This offers a simple, fast, and effective solution to the challenges of the current disaster alert systems.
KSP Keywords
Advanced models, Disaster Scenarios, Disaster alert, Image selection, New approach, animated image, image synthesis, segmentation-based, text-based, video synthesis