Cargando…
Differentially Private Singular Value Decomposition for Training Support Vector Machines
Support vector machine (SVM) is an efficient classification method in machine learning. The traditional classification model of SVMs may pose a great threat to personal privacy, when sensitive information is included in the training datasets. Principal component analysis (PCA) can project instances...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Hindawi
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8976603/ https://www.ncbi.nlm.nih.gov/pubmed/35378802 http://dx.doi.org/10.1155/2022/2935975 |
_version_ | 1784680611813785600 |
---|---|
author | Sun, Zhenlong Yang, Jing Li, Xiaoye |
author_facet | Sun, Zhenlong Yang, Jing Li, Xiaoye |
author_sort | Sun, Zhenlong |
collection | PubMed |
description | Support vector machine (SVM) is an efficient classification method in machine learning. The traditional classification model of SVMs may pose a great threat to personal privacy, when sensitive information is included in the training datasets. Principal component analysis (PCA) can project instances into a low-dimensional subspace while capturing the variance of the matrix A as much as possible. There are two common algorithms that PCA uses to perform the principal component analysis, eigenvalue decomposition (EVD) and singular value decomposition (SVD). The main advantage of SVD compared with EVD is that it does not need to compute the matrix of covariance. This study presents a new differentially private SVD algorithm (DPSVD) to prevent the privacy leak of SVM classifiers. The DPSVD generates a set of private singular vectors that the projected instances in the singular subspace can be directly used to train SVM while not disclosing privacy of the original instances. After proving that the DPSVD satisfies differential privacy in theory, several experiments were carried out. The experimental results confirm that our method achieved higher accuracy and better stability on different real datasets, compared with other existing private PCA algorithms used to train SVM. |
format | Online Article Text |
id | pubmed-8976603 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Hindawi |
record_format | MEDLINE/PubMed |
spelling | pubmed-89766032022-04-03 Differentially Private Singular Value Decomposition for Training Support Vector Machines Sun, Zhenlong Yang, Jing Li, Xiaoye Comput Intell Neurosci Research Article Support vector machine (SVM) is an efficient classification method in machine learning. The traditional classification model of SVMs may pose a great threat to personal privacy, when sensitive information is included in the training datasets. Principal component analysis (PCA) can project instances into a low-dimensional subspace while capturing the variance of the matrix A as much as possible. There are two common algorithms that PCA uses to perform the principal component analysis, eigenvalue decomposition (EVD) and singular value decomposition (SVD). The main advantage of SVD compared with EVD is that it does not need to compute the matrix of covariance. This study presents a new differentially private SVD algorithm (DPSVD) to prevent the privacy leak of SVM classifiers. The DPSVD generates a set of private singular vectors that the projected instances in the singular subspace can be directly used to train SVM while not disclosing privacy of the original instances. After proving that the DPSVD satisfies differential privacy in theory, several experiments were carried out. The experimental results confirm that our method achieved higher accuracy and better stability on different real datasets, compared with other existing private PCA algorithms used to train SVM. Hindawi 2022-03-26 /pmc/articles/PMC8976603/ /pubmed/35378802 http://dx.doi.org/10.1155/2022/2935975 Text en Copyright © 2022 Zhenlong Sun et al. https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. |
spellingShingle | Research Article Sun, Zhenlong Yang, Jing Li, Xiaoye Differentially Private Singular Value Decomposition for Training Support Vector Machines |
title | Differentially Private Singular Value Decomposition for Training Support Vector Machines |
title_full | Differentially Private Singular Value Decomposition for Training Support Vector Machines |
title_fullStr | Differentially Private Singular Value Decomposition for Training Support Vector Machines |
title_full_unstemmed | Differentially Private Singular Value Decomposition for Training Support Vector Machines |
title_short | Differentially Private Singular Value Decomposition for Training Support Vector Machines |
title_sort | differentially private singular value decomposition for training support vector machines |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8976603/ https://www.ncbi.nlm.nih.gov/pubmed/35378802 http://dx.doi.org/10.1155/2022/2935975 |
work_keys_str_mv | AT sunzhenlong differentiallyprivatesingularvaluedecompositionfortrainingsupportvectormachines AT yangjing differentiallyprivatesingularvaluedecompositionfortrainingsupportvectormachines AT lixiaoye differentiallyprivatesingularvaluedecompositionfortrainingsupportvectormachines |