Cargando…

Word2vec Skip-Gram Dimensionality Selection via Sequential Normalized Maximum Likelihood

In this paper, we propose a novel information criteria-based approach to select the dimensionality of the word2vec Skip-gram (SG). From the perspective of the probability theory, SG is considered as an implicit probability distribution estimation under the assumption that there exists a true context...

Descripción completa

Detalles Bibliográficos
Autores principales: Hung, Pham Thuc, Yamanishi, Kenji
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8391962/
https://www.ncbi.nlm.nih.gov/pubmed/34441136
http://dx.doi.org/10.3390/e23080997
_version_ 1783743395190013952
author Hung, Pham Thuc
Yamanishi, Kenji
author_facet Hung, Pham Thuc
Yamanishi, Kenji
author_sort Hung, Pham Thuc
collection PubMed
description In this paper, we propose a novel information criteria-based approach to select the dimensionality of the word2vec Skip-gram (SG). From the perspective of the probability theory, SG is considered as an implicit probability distribution estimation under the assumption that there exists a true contextual distribution among words. Therefore, we apply information criteria with the aim of selecting the best dimensionality so that the corresponding model can be as close as possible to the true distribution. We examine the following information criteria for the dimensionality selection problem: the Akaike’s Information Criterion (AIC), Bayesian Information Criterion (BIC), and Sequential Normalized Maximum Likelihood (SNML) criterion. SNML is the total codelength required for the sequential encoding of a data sequence on the basis of the minimum description length. The proposed approach is applied to both the original SG model and the SG Negative Sampling model to clarify the idea of using information criteria. Additionally, as the original SNML suffers from computational disadvantages, we introduce novel heuristics for its efficient computation. Moreover, we empirically demonstrate that SNML outperforms both BIC and AIC. In comparison with other evaluation methods for word embedding, the dimensionality selected by SNML is significantly closer to the optimal dimensionality obtained by word analogy or word similarity tasks.
format Online
Article
Text
id pubmed-8391962
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-83919622021-08-28 Word2vec Skip-Gram Dimensionality Selection via Sequential Normalized Maximum Likelihood Hung, Pham Thuc Yamanishi, Kenji Entropy (Basel) Article In this paper, we propose a novel information criteria-based approach to select the dimensionality of the word2vec Skip-gram (SG). From the perspective of the probability theory, SG is considered as an implicit probability distribution estimation under the assumption that there exists a true contextual distribution among words. Therefore, we apply information criteria with the aim of selecting the best dimensionality so that the corresponding model can be as close as possible to the true distribution. We examine the following information criteria for the dimensionality selection problem: the Akaike’s Information Criterion (AIC), Bayesian Information Criterion (BIC), and Sequential Normalized Maximum Likelihood (SNML) criterion. SNML is the total codelength required for the sequential encoding of a data sequence on the basis of the minimum description length. The proposed approach is applied to both the original SG model and the SG Negative Sampling model to clarify the idea of using information criteria. Additionally, as the original SNML suffers from computational disadvantages, we introduce novel heuristics for its efficient computation. Moreover, we empirically demonstrate that SNML outperforms both BIC and AIC. In comparison with other evaluation methods for word embedding, the dimensionality selected by SNML is significantly closer to the optimal dimensionality obtained by word analogy or word similarity tasks. MDPI 2021-07-31 /pmc/articles/PMC8391962/ /pubmed/34441136 http://dx.doi.org/10.3390/e23080997 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Hung, Pham Thuc
Yamanishi, Kenji
Word2vec Skip-Gram Dimensionality Selection via Sequential Normalized Maximum Likelihood
title Word2vec Skip-Gram Dimensionality Selection via Sequential Normalized Maximum Likelihood
title_full Word2vec Skip-Gram Dimensionality Selection via Sequential Normalized Maximum Likelihood
title_fullStr Word2vec Skip-Gram Dimensionality Selection via Sequential Normalized Maximum Likelihood
title_full_unstemmed Word2vec Skip-Gram Dimensionality Selection via Sequential Normalized Maximum Likelihood
title_short Word2vec Skip-Gram Dimensionality Selection via Sequential Normalized Maximum Likelihood
title_sort word2vec skip-gram dimensionality selection via sequential normalized maximum likelihood
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8391962/
https://www.ncbi.nlm.nih.gov/pubmed/34441136
http://dx.doi.org/10.3390/e23080997
work_keys_str_mv AT hungphamthuc word2vecskipgramdimensionalityselectionviasequentialnormalizedmaximumlikelihood
AT yamanishikenji word2vecskipgramdimensionalityselectionviasequentialnormalizedmaximumlikelihood