ETRI-Knowledge Sharing Plaform

KOREAN
논문 검색
Type SCI
Year ~ Keyword

Detail

Journal Article Deep Learning-Based Optimization of Visual–Auditory Sensory Substitution
Cited 3 time in scopus Download 205 time Share share facebook twitter linkedin kakaostory
Authors
Mooseop Kim, Yunkyung Park, Kyeongdeok Moon, Chi Yoon Jeong
Issue Date
2023-02
Citation
IEEE Access, v.11, pp.14169-14180
ISSN
2169-3536
Publisher
IEEE
Language
English
Type
Journal Article
DOI
https://dx.doi.org/10.1109/ACCESS.2023.3243641
Abstract
Visual-auditory sensory substitution systems can aid visually impaired people in traveling to various places and recognizing their own environments without help from others. Although several such systems have been developed, they are either not widely used or are limited to laboratory-scale research. Among various factors that hinder the widespread use of these systems, one of the most important issues to consider is the optimization of the algorithms for sensory substitution. This study is the first attempt at exploring the possibility of using deep learning for the objective quantification of sensory substitution. To this end, we used generative adversarial networks to investigate the possibility of optimizing the vOICe algorithm, a representative visual-auditory sensory substitution method, by controlling the parameters of the method for converting an image to sound. Furthermore, we explored the effect of the parameters on the conversion scheme for the vOICe system and performed frequency-range and frequency-mapping-function experiments. The process of sensory substitution in humans was modeled to use generative models to assess the extent of visual perception from the substituted sensory signals. We verified the human-based experimental results against the modeling results. The results suggested that deep learning could be used for evaluating the efficiency of algorithms for visual-auditory sensory substitutions without labor-intensive human behavioral experiments. The introduction of deep learning for optimizing the visual-auditory conversion method is expected to facilitate studies on various aspects of sensory substitution, such as generalization and estimation of algorithm efficiency.
KSP Keywords
Behavioral experiments, Conversion method, Laboratory scale, Sensory Substitution, Substitution method, Substitution systems, Visual Perception, Visually Impaired People, algorithm efficiency, deep learning(DL), generative adversarial network
This work is distributed under the term of Creative Commons License (CCL)
(CC BY)
CC BY