Cargando…

Incorporating Linguistic Knowledge for Learning Distributed Word Representations

Combined with neural language models, distributed word representations achieve significant advantages in computational linguistics and text mining. Most existing models estimate distributed word vectors from large-scale data in an unsupervised fashion, which, however, do not take rich linguistic kno...

Descripción completa

Detalles Bibliográficos
Autores principales: Wang, Yan, Liu, Zhiyuan, Sun, Maosong
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Public Library of Science 2015
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4395361/
https://www.ncbi.nlm.nih.gov/pubmed/25874581
http://dx.doi.org/10.1371/journal.pone.0118437
_version_ 1782366432213860352
author Wang, Yan
Liu, Zhiyuan
Sun, Maosong
author_facet Wang, Yan
Liu, Zhiyuan
Sun, Maosong
author_sort Wang, Yan
collection PubMed
description Combined with neural language models, distributed word representations achieve significant advantages in computational linguistics and text mining. Most existing models estimate distributed word vectors from large-scale data in an unsupervised fashion, which, however, do not take rich linguistic knowledge into consideration. Linguistic knowledge can be represented as either link-based knowledge or preference-based knowledge, and we propose knowledge regularized word representation models (KRWR) to incorporate these prior knowledge for learning distributed word representations. Experiment results demonstrate that our estimated word representation achieves better performance in task of semantic relatedness ranking. This indicates that our methods can efficiently encode both prior knowledge from knowledge bases and statistical knowledge from large-scale text corpora into a unified word representation model, which will benefit many tasks in text mining.
format Online
Article
Text
id pubmed-4395361
institution National Center for Biotechnology Information
language English
publishDate 2015
publisher Public Library of Science
record_format MEDLINE/PubMed
spelling pubmed-43953612015-04-21 Incorporating Linguistic Knowledge for Learning Distributed Word Representations Wang, Yan Liu, Zhiyuan Sun, Maosong PLoS One Research Article Combined with neural language models, distributed word representations achieve significant advantages in computational linguistics and text mining. Most existing models estimate distributed word vectors from large-scale data in an unsupervised fashion, which, however, do not take rich linguistic knowledge into consideration. Linguistic knowledge can be represented as either link-based knowledge or preference-based knowledge, and we propose knowledge regularized word representation models (KRWR) to incorporate these prior knowledge for learning distributed word representations. Experiment results demonstrate that our estimated word representation achieves better performance in task of semantic relatedness ranking. This indicates that our methods can efficiently encode both prior knowledge from knowledge bases and statistical knowledge from large-scale text corpora into a unified word representation model, which will benefit many tasks in text mining. Public Library of Science 2015-04-13 /pmc/articles/PMC4395361/ /pubmed/25874581 http://dx.doi.org/10.1371/journal.pone.0118437 Text en © 2015 Wang et al http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are properly credited.
spellingShingle Research Article
Wang, Yan
Liu, Zhiyuan
Sun, Maosong
Incorporating Linguistic Knowledge for Learning Distributed Word Representations
title Incorporating Linguistic Knowledge for Learning Distributed Word Representations
title_full Incorporating Linguistic Knowledge for Learning Distributed Word Representations
title_fullStr Incorporating Linguistic Knowledge for Learning Distributed Word Representations
title_full_unstemmed Incorporating Linguistic Knowledge for Learning Distributed Word Representations
title_short Incorporating Linguistic Knowledge for Learning Distributed Word Representations
title_sort incorporating linguistic knowledge for learning distributed word representations
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4395361/
https://www.ncbi.nlm.nih.gov/pubmed/25874581
http://dx.doi.org/10.1371/journal.pone.0118437
work_keys_str_mv AT wangyan incorporatinglinguisticknowledgeforlearningdistributedwordrepresentations
AT liuzhiyuan incorporatinglinguisticknowledgeforlearningdistributedwordrepresentations
AT sunmaosong incorporatinglinguisticknowledgeforlearningdistributedwordrepresentations