Cargando…

FedMed: A Federated Learning Framework for Language Modeling

Federated learning (FL) is a privacy-preserving technique for training a vast amount of decentralized data and making inferences on mobile devices. As a typical language modeling problem, mobile keyboard prediction aims at suggesting a probable next word or phrase and facilitating the human-machine...

Descripción completa

Detalles Bibliográficos
Autores principales: Wu, Xing, Liang, Zhaowang, Wang, Jianjia
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7412048/
https://www.ncbi.nlm.nih.gov/pubmed/32708152
http://dx.doi.org/10.3390/s20144048
_version_ 1783568518520766464
author Wu, Xing
Liang, Zhaowang
Wang, Jianjia
author_facet Wu, Xing
Liang, Zhaowang
Wang, Jianjia
author_sort Wu, Xing
collection PubMed
description Federated learning (FL) is a privacy-preserving technique for training a vast amount of decentralized data and making inferences on mobile devices. As a typical language modeling problem, mobile keyboard prediction aims at suggesting a probable next word or phrase and facilitating the human-machine interaction in a virtual keyboard of the smartphone or laptop. Mobile keyboard prediction with FL hopes to satisfy the growing demand that high-level data privacy be preserved in artificial intelligence applications even with the distributed models training. However, there are two major problems in the federated optimization for the prediction: (1) aggregating model parameters on the server-side and (2) reducing communication costs caused by model weights collection. To address the above issues, traditional FL methods simply use averaging aggregation or ignore communication costs. We propose a novel Federated Mediation (FedMed) framework with the adaptive aggregation, mediation incentive scheme, and topK strategy to address the model aggregation and communication costs. The performance is evaluated in terms of perplexity and communication rounds. Experiments are conducted on three datasets (i.e., Penn Treebank, WikiText-2, and Yelp) and the results demonstrate that our FedMed framework achieves robust performance and outperforms baseline approaches.
format Online
Article
Text
id pubmed-7412048
institution National Center for Biotechnology Information
language English
publishDate 2020
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-74120482020-08-25 FedMed: A Federated Learning Framework for Language Modeling Wu, Xing Liang, Zhaowang Wang, Jianjia Sensors (Basel) Article Federated learning (FL) is a privacy-preserving technique for training a vast amount of decentralized data and making inferences on mobile devices. As a typical language modeling problem, mobile keyboard prediction aims at suggesting a probable next word or phrase and facilitating the human-machine interaction in a virtual keyboard of the smartphone or laptop. Mobile keyboard prediction with FL hopes to satisfy the growing demand that high-level data privacy be preserved in artificial intelligence applications even with the distributed models training. However, there are two major problems in the federated optimization for the prediction: (1) aggregating model parameters on the server-side and (2) reducing communication costs caused by model weights collection. To address the above issues, traditional FL methods simply use averaging aggregation or ignore communication costs. We propose a novel Federated Mediation (FedMed) framework with the adaptive aggregation, mediation incentive scheme, and topK strategy to address the model aggregation and communication costs. The performance is evaluated in terms of perplexity and communication rounds. Experiments are conducted on three datasets (i.e., Penn Treebank, WikiText-2, and Yelp) and the results demonstrate that our FedMed framework achieves robust performance and outperforms baseline approaches. MDPI 2020-07-21 /pmc/articles/PMC7412048/ /pubmed/32708152 http://dx.doi.org/10.3390/s20144048 Text en © 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Wu, Xing
Liang, Zhaowang
Wang, Jianjia
FedMed: A Federated Learning Framework for Language Modeling
title FedMed: A Federated Learning Framework for Language Modeling
title_full FedMed: A Federated Learning Framework for Language Modeling
title_fullStr FedMed: A Federated Learning Framework for Language Modeling
title_full_unstemmed FedMed: A Federated Learning Framework for Language Modeling
title_short FedMed: A Federated Learning Framework for Language Modeling
title_sort fedmed: a federated learning framework for language modeling
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7412048/
https://www.ncbi.nlm.nih.gov/pubmed/32708152
http://dx.doi.org/10.3390/s20144048
work_keys_str_mv AT wuxing fedmedafederatedlearningframeworkforlanguagemodeling
AT liangzhaowang fedmedafederatedlearningframeworkforlanguagemodeling
AT wangjianjia fedmedafederatedlearningframeworkforlanguagemodeling