Cargando…

Semi-Supervised Learning of Statistical Models for Natural Language Understanding

Natural language understanding is to specify a computational model that maps sentences to their semantic mean representation. In this paper, we propose a novel framework to train the statistical models without using expensive fully annotated data. In particular, the input of our framework is a set o...

Descripción completa

Detalles Bibliográficos
Autores principales: Zhou, Deyu, He, Yulan
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Hindawi Publishing Corporation 2014
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4127215/
https://www.ncbi.nlm.nih.gov/pubmed/25152899
http://dx.doi.org/10.1155/2014/121650
_version_ 1782329998858780672
author Zhou, Deyu
He, Yulan
author_facet Zhou, Deyu
He, Yulan
author_sort Zhou, Deyu
collection PubMed
description Natural language understanding is to specify a computational model that maps sentences to their semantic mean representation. In this paper, we propose a novel framework to train the statistical models without using expensive fully annotated data. In particular, the input of our framework is a set of sentences labeled with abstract semantic annotations. These annotations encode the underlying embedded semantic structural relations without explicit word/semantic tag alignment. The proposed framework can automatically induce derivation rules that map sentences to their semantic meaning representations. The learning framework is applied on two statistical models, the conditional random fields (CRFs) and the hidden Markov support vector machines (HM-SVMs). Our experimental results on the DARPA communicator data show that both CRFs and HM-SVMs outperform the baseline approach, previously proposed hidden vector state (HVS) model which is also trained on abstract semantic annotations. In addition, the proposed framework shows superior performance than two other baseline approaches, a hybrid framework combining HVS and HM-SVMs and discriminative training of HVS, with a relative error reduction rate of about 25% and 15% being achieved in F-measure.
format Online
Article
Text
id pubmed-4127215
institution National Center for Biotechnology Information
language English
publishDate 2014
publisher Hindawi Publishing Corporation
record_format MEDLINE/PubMed
spelling pubmed-41272152014-08-24 Semi-Supervised Learning of Statistical Models for Natural Language Understanding Zhou, Deyu He, Yulan ScientificWorldJournal Research Article Natural language understanding is to specify a computational model that maps sentences to their semantic mean representation. In this paper, we propose a novel framework to train the statistical models without using expensive fully annotated data. In particular, the input of our framework is a set of sentences labeled with abstract semantic annotations. These annotations encode the underlying embedded semantic structural relations without explicit word/semantic tag alignment. The proposed framework can automatically induce derivation rules that map sentences to their semantic meaning representations. The learning framework is applied on two statistical models, the conditional random fields (CRFs) and the hidden Markov support vector machines (HM-SVMs). Our experimental results on the DARPA communicator data show that both CRFs and HM-SVMs outperform the baseline approach, previously proposed hidden vector state (HVS) model which is also trained on abstract semantic annotations. In addition, the proposed framework shows superior performance than two other baseline approaches, a hybrid framework combining HVS and HM-SVMs and discriminative training of HVS, with a relative error reduction rate of about 25% and 15% being achieved in F-measure. Hindawi Publishing Corporation 2014 2014-07-20 /pmc/articles/PMC4127215/ /pubmed/25152899 http://dx.doi.org/10.1155/2014/121650 Text en Copyright © 2014 D. Zhou and Y. He. https://creativecommons.org/licenses/by/3.0/ This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
spellingShingle Research Article
Zhou, Deyu
He, Yulan
Semi-Supervised Learning of Statistical Models for Natural Language Understanding
title Semi-Supervised Learning of Statistical Models for Natural Language Understanding
title_full Semi-Supervised Learning of Statistical Models for Natural Language Understanding
title_fullStr Semi-Supervised Learning of Statistical Models for Natural Language Understanding
title_full_unstemmed Semi-Supervised Learning of Statistical Models for Natural Language Understanding
title_short Semi-Supervised Learning of Statistical Models for Natural Language Understanding
title_sort semi-supervised learning of statistical models for natural language understanding
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4127215/
https://www.ncbi.nlm.nih.gov/pubmed/25152899
http://dx.doi.org/10.1155/2014/121650
work_keys_str_mv AT zhoudeyu semisupervisedlearningofstatisticalmodelsfornaturallanguageunderstanding
AT heyulan semisupervisedlearningofstatisticalmodelsfornaturallanguageunderstanding