Cargando…
Improved Deep Neural Network for Cross-Media Visual Communication
Cross-media visual communication is an extremely complex task. In order to solve the problem of segmentation of visual foreground and background, improve the accuracy of visual communication scene reconstruction, and complete the task of visual real-time communication. We propose an improved generat...
Autor principal: | |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Hindawi
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9078775/ https://www.ncbi.nlm.nih.gov/pubmed/35535179 http://dx.doi.org/10.1155/2022/1556352 |
_version_ | 1784702411006279680 |
---|---|
author | Miao, Yubo |
author_facet | Miao, Yubo |
author_sort | Miao, Yubo |
collection | PubMed |
description | Cross-media visual communication is an extremely complex task. In order to solve the problem of segmentation of visual foreground and background, improve the accuracy of visual communication scene reconstruction, and complete the task of visual real-time communication. We propose an improved generative adversarial network. We take the generative adversarial network as the basis and add a combined codec package to the generator, while configuring the generator and discriminator as a cascade structure, preserving the feature upsampling and downsampling convolutional layers of visual scenes with different layers through correspondence. To classify features with different visual scene layers, we add a new auxiliary classifier based on convolutional neural networks. With the help of the auxiliary classifier, similar visual scenes with different feature layers have a more accurate recognition rate. In the experimental part, to better distinguish foreground and background in visual communication, we perform performance tests on foreground and background using separate datasets. The experimental results show that our method has good accuracy in both foreground and background in cross-media communication for real-time visual communication. In addition, we validate the efficiency of our method on Cityscapes, NoW, and Replica datasets, respectively, and experimentally demonstrate that our method performs better than traditional machine learning methods and outperforms deep learning methods of the same type. |
format | Online Article Text |
id | pubmed-9078775 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Hindawi |
record_format | MEDLINE/PubMed |
spelling | pubmed-90787752022-05-08 Improved Deep Neural Network for Cross-Media Visual Communication Miao, Yubo Comput Intell Neurosci Research Article Cross-media visual communication is an extremely complex task. In order to solve the problem of segmentation of visual foreground and background, improve the accuracy of visual communication scene reconstruction, and complete the task of visual real-time communication. We propose an improved generative adversarial network. We take the generative adversarial network as the basis and add a combined codec package to the generator, while configuring the generator and discriminator as a cascade structure, preserving the feature upsampling and downsampling convolutional layers of visual scenes with different layers through correspondence. To classify features with different visual scene layers, we add a new auxiliary classifier based on convolutional neural networks. With the help of the auxiliary classifier, similar visual scenes with different feature layers have a more accurate recognition rate. In the experimental part, to better distinguish foreground and background in visual communication, we perform performance tests on foreground and background using separate datasets. The experimental results show that our method has good accuracy in both foreground and background in cross-media communication for real-time visual communication. In addition, we validate the efficiency of our method on Cityscapes, NoW, and Replica datasets, respectively, and experimentally demonstrate that our method performs better than traditional machine learning methods and outperforms deep learning methods of the same type. Hindawi 2022-04-30 /pmc/articles/PMC9078775/ /pubmed/35535179 http://dx.doi.org/10.1155/2022/1556352 Text en Copyright © 2022 Yubo Miao. https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. |
spellingShingle | Research Article Miao, Yubo Improved Deep Neural Network for Cross-Media Visual Communication |
title | Improved Deep Neural Network for Cross-Media Visual Communication |
title_full | Improved Deep Neural Network for Cross-Media Visual Communication |
title_fullStr | Improved Deep Neural Network for Cross-Media Visual Communication |
title_full_unstemmed | Improved Deep Neural Network for Cross-Media Visual Communication |
title_short | Improved Deep Neural Network for Cross-Media Visual Communication |
title_sort | improved deep neural network for cross-media visual communication |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9078775/ https://www.ncbi.nlm.nih.gov/pubmed/35535179 http://dx.doi.org/10.1155/2022/1556352 |
work_keys_str_mv | AT miaoyubo improveddeepneuralnetworkforcrossmediavisualcommunication |