Cargando…
Computer Vision System for Expressing Texture Using Sound-Symbolic Words
The major goals of texture research in computer vision are to understand, model, and process texture and ultimately simulate human visual information processing using computer technologies. The field of computer vision has witnessed remarkable advancements in material recognition using deep convolut...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8529034/ https://www.ncbi.nlm.nih.gov/pubmed/34690855 http://dx.doi.org/10.3389/fpsyg.2021.654779 |
_version_ | 1784586379563368448 |
---|---|
author | Yamagata, Koichi Kwon, Jinhwan Kawashima, Takuya Shimoda, Wataru Sakamoto, Maki |
author_facet | Yamagata, Koichi Kwon, Jinhwan Kawashima, Takuya Shimoda, Wataru Sakamoto, Maki |
author_sort | Yamagata, Koichi |
collection | PubMed |
description | The major goals of texture research in computer vision are to understand, model, and process texture and ultimately simulate human visual information processing using computer technologies. The field of computer vision has witnessed remarkable advancements in material recognition using deep convolutional neural networks (DCNNs), which have enabled various computer vision applications, such as self-driving cars, facial and gesture recognition, and automatic number plate recognition. However, for computer vision to “express” texture like human beings is still difficult because texture description has no correct or incorrect answer and is ambiguous. In this paper, we develop a computer vision method using DCNN that expresses texture of materials. To achieve this goal, we focus on Japanese “sound-symbolic” words, which can describe differences in texture sensation at a fine resolution and are known to have strong and systematic sensory-sound associations. Because the phonemes of Japanese sound-symbolic words characterize categories of texture sensations, we develop a computer vision method to generate the phonemes and structure comprising sound-symbolic words that probabilistically correspond to the input images. It was confirmed that the sound-symbolic words output by our system had about 80% accuracy rate in our evaluation. |
format | Online Article Text |
id | pubmed-8529034 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-85290342021-10-22 Computer Vision System for Expressing Texture Using Sound-Symbolic Words Yamagata, Koichi Kwon, Jinhwan Kawashima, Takuya Shimoda, Wataru Sakamoto, Maki Front Psychol Psychology The major goals of texture research in computer vision are to understand, model, and process texture and ultimately simulate human visual information processing using computer technologies. The field of computer vision has witnessed remarkable advancements in material recognition using deep convolutional neural networks (DCNNs), which have enabled various computer vision applications, such as self-driving cars, facial and gesture recognition, and automatic number plate recognition. However, for computer vision to “express” texture like human beings is still difficult because texture description has no correct or incorrect answer and is ambiguous. In this paper, we develop a computer vision method using DCNN that expresses texture of materials. To achieve this goal, we focus on Japanese “sound-symbolic” words, which can describe differences in texture sensation at a fine resolution and are known to have strong and systematic sensory-sound associations. Because the phonemes of Japanese sound-symbolic words characterize categories of texture sensations, we develop a computer vision method to generate the phonemes and structure comprising sound-symbolic words that probabilistically correspond to the input images. It was confirmed that the sound-symbolic words output by our system had about 80% accuracy rate in our evaluation. Frontiers Media S.A. 2021-10-07 /pmc/articles/PMC8529034/ /pubmed/34690855 http://dx.doi.org/10.3389/fpsyg.2021.654779 Text en Copyright © 2021 Yamagata, Kwon, Kawashima, Shimoda and Sakamoto. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
spellingShingle | Psychology Yamagata, Koichi Kwon, Jinhwan Kawashima, Takuya Shimoda, Wataru Sakamoto, Maki Computer Vision System for Expressing Texture Using Sound-Symbolic Words |
title | Computer Vision System for Expressing Texture Using Sound-Symbolic Words |
title_full | Computer Vision System for Expressing Texture Using Sound-Symbolic Words |
title_fullStr | Computer Vision System for Expressing Texture Using Sound-Symbolic Words |
title_full_unstemmed | Computer Vision System for Expressing Texture Using Sound-Symbolic Words |
title_short | Computer Vision System for Expressing Texture Using Sound-Symbolic Words |
title_sort | computer vision system for expressing texture using sound-symbolic words |
topic | Psychology |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8529034/ https://www.ncbi.nlm.nih.gov/pubmed/34690855 http://dx.doi.org/10.3389/fpsyg.2021.654779 |
work_keys_str_mv | AT yamagatakoichi computervisionsystemforexpressingtextureusingsoundsymbolicwords AT kwonjinhwan computervisionsystemforexpressingtextureusingsoundsymbolicwords AT kawashimatakuya computervisionsystemforexpressingtextureusingsoundsymbolicwords AT shimodawataru computervisionsystemforexpressingtextureusingsoundsymbolicwords AT sakamotomaki computervisionsystemforexpressingtextureusingsoundsymbolicwords |