Cargando…

Computer Vision System for Expressing Texture Using Sound-Symbolic Words

The major goals of texture research in computer vision are to understand, model, and process texture and ultimately simulate human visual information processing using computer technologies. The field of computer vision has witnessed remarkable advancements in material recognition using deep convolut...

Descripción completa

Detalles Bibliográficos
Autores principales: Yamagata, Koichi, Kwon, Jinhwan, Kawashima, Takuya, Shimoda, Wataru, Sakamoto, Maki
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8529034/
https://www.ncbi.nlm.nih.gov/pubmed/34690855
http://dx.doi.org/10.3389/fpsyg.2021.654779
_version_ 1784586379563368448
author Yamagata, Koichi
Kwon, Jinhwan
Kawashima, Takuya
Shimoda, Wataru
Sakamoto, Maki
author_facet Yamagata, Koichi
Kwon, Jinhwan
Kawashima, Takuya
Shimoda, Wataru
Sakamoto, Maki
author_sort Yamagata, Koichi
collection PubMed
description The major goals of texture research in computer vision are to understand, model, and process texture and ultimately simulate human visual information processing using computer technologies. The field of computer vision has witnessed remarkable advancements in material recognition using deep convolutional neural networks (DCNNs), which have enabled various computer vision applications, such as self-driving cars, facial and gesture recognition, and automatic number plate recognition. However, for computer vision to “express” texture like human beings is still difficult because texture description has no correct or incorrect answer and is ambiguous. In this paper, we develop a computer vision method using DCNN that expresses texture of materials. To achieve this goal, we focus on Japanese “sound-symbolic” words, which can describe differences in texture sensation at a fine resolution and are known to have strong and systematic sensory-sound associations. Because the phonemes of Japanese sound-symbolic words characterize categories of texture sensations, we develop a computer vision method to generate the phonemes and structure comprising sound-symbolic words that probabilistically correspond to the input images. It was confirmed that the sound-symbolic words output by our system had about 80% accuracy rate in our evaluation.
format Online
Article
Text
id pubmed-8529034
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-85290342021-10-22 Computer Vision System for Expressing Texture Using Sound-Symbolic Words Yamagata, Koichi Kwon, Jinhwan Kawashima, Takuya Shimoda, Wataru Sakamoto, Maki Front Psychol Psychology The major goals of texture research in computer vision are to understand, model, and process texture and ultimately simulate human visual information processing using computer technologies. The field of computer vision has witnessed remarkable advancements in material recognition using deep convolutional neural networks (DCNNs), which have enabled various computer vision applications, such as self-driving cars, facial and gesture recognition, and automatic number plate recognition. However, for computer vision to “express” texture like human beings is still difficult because texture description has no correct or incorrect answer and is ambiguous. In this paper, we develop a computer vision method using DCNN that expresses texture of materials. To achieve this goal, we focus on Japanese “sound-symbolic” words, which can describe differences in texture sensation at a fine resolution and are known to have strong and systematic sensory-sound associations. Because the phonemes of Japanese sound-symbolic words characterize categories of texture sensations, we develop a computer vision method to generate the phonemes and structure comprising sound-symbolic words that probabilistically correspond to the input images. It was confirmed that the sound-symbolic words output by our system had about 80% accuracy rate in our evaluation. Frontiers Media S.A. 2021-10-07 /pmc/articles/PMC8529034/ /pubmed/34690855 http://dx.doi.org/10.3389/fpsyg.2021.654779 Text en Copyright © 2021 Yamagata, Kwon, Kawashima, Shimoda and Sakamoto. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Psychology
Yamagata, Koichi
Kwon, Jinhwan
Kawashima, Takuya
Shimoda, Wataru
Sakamoto, Maki
Computer Vision System for Expressing Texture Using Sound-Symbolic Words
title Computer Vision System for Expressing Texture Using Sound-Symbolic Words
title_full Computer Vision System for Expressing Texture Using Sound-Symbolic Words
title_fullStr Computer Vision System for Expressing Texture Using Sound-Symbolic Words
title_full_unstemmed Computer Vision System for Expressing Texture Using Sound-Symbolic Words
title_short Computer Vision System for Expressing Texture Using Sound-Symbolic Words
title_sort computer vision system for expressing texture using sound-symbolic words
topic Psychology
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8529034/
https://www.ncbi.nlm.nih.gov/pubmed/34690855
http://dx.doi.org/10.3389/fpsyg.2021.654779
work_keys_str_mv AT yamagatakoichi computervisionsystemforexpressingtextureusingsoundsymbolicwords
AT kwonjinhwan computervisionsystemforexpressingtextureusingsoundsymbolicwords
AT kawashimatakuya computervisionsystemforexpressingtextureusingsoundsymbolicwords
AT shimodawataru computervisionsystemforexpressingtextureusingsoundsymbolicwords
AT sakamotomaki computervisionsystemforexpressingtextureusingsoundsymbolicwords