Cargando…

Differentially Private Singular Value Decomposition for Training Support Vector Machines

Support vector machine (SVM) is an efficient classification method in machine learning. The traditional classification model of SVMs may pose a great threat to personal privacy, when sensitive information is included in the training datasets. Principal component analysis (PCA) can project instances...

Descripción completa

Detalles Bibliográficos
Autores principales: Sun, Zhenlong, Yang, Jing, Li, Xiaoye
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Hindawi 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8976603/
https://www.ncbi.nlm.nih.gov/pubmed/35378802
http://dx.doi.org/10.1155/2022/2935975
_version_ 1784680611813785600
author Sun, Zhenlong
Yang, Jing
Li, Xiaoye
author_facet Sun, Zhenlong
Yang, Jing
Li, Xiaoye
author_sort Sun, Zhenlong
collection PubMed
description Support vector machine (SVM) is an efficient classification method in machine learning. The traditional classification model of SVMs may pose a great threat to personal privacy, when sensitive information is included in the training datasets. Principal component analysis (PCA) can project instances into a low-dimensional subspace while capturing the variance of the matrix A as much as possible. There are two common algorithms that PCA uses to perform the principal component analysis, eigenvalue decomposition (EVD) and singular value decomposition (SVD). The main advantage of SVD compared with EVD is that it does not need to compute the matrix of covariance. This study presents a new differentially private SVD algorithm (DPSVD) to prevent the privacy leak of SVM classifiers. The DPSVD generates a set of private singular vectors that the projected instances in the singular subspace can be directly used to train SVM while not disclosing privacy of the original instances. After proving that the DPSVD satisfies differential privacy in theory, several experiments were carried out. The experimental results confirm that our method achieved higher accuracy and better stability on different real datasets, compared with other existing private PCA algorithms used to train SVM.
format Online
Article
Text
id pubmed-8976603
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Hindawi
record_format MEDLINE/PubMed
spelling pubmed-89766032022-04-03 Differentially Private Singular Value Decomposition for Training Support Vector Machines Sun, Zhenlong Yang, Jing Li, Xiaoye Comput Intell Neurosci Research Article Support vector machine (SVM) is an efficient classification method in machine learning. The traditional classification model of SVMs may pose a great threat to personal privacy, when sensitive information is included in the training datasets. Principal component analysis (PCA) can project instances into a low-dimensional subspace while capturing the variance of the matrix A as much as possible. There are two common algorithms that PCA uses to perform the principal component analysis, eigenvalue decomposition (EVD) and singular value decomposition (SVD). The main advantage of SVD compared with EVD is that it does not need to compute the matrix of covariance. This study presents a new differentially private SVD algorithm (DPSVD) to prevent the privacy leak of SVM classifiers. The DPSVD generates a set of private singular vectors that the projected instances in the singular subspace can be directly used to train SVM while not disclosing privacy of the original instances. After proving that the DPSVD satisfies differential privacy in theory, several experiments were carried out. The experimental results confirm that our method achieved higher accuracy and better stability on different real datasets, compared with other existing private PCA algorithms used to train SVM. Hindawi 2022-03-26 /pmc/articles/PMC8976603/ /pubmed/35378802 http://dx.doi.org/10.1155/2022/2935975 Text en Copyright © 2022 Zhenlong Sun et al. https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
spellingShingle Research Article
Sun, Zhenlong
Yang, Jing
Li, Xiaoye
Differentially Private Singular Value Decomposition for Training Support Vector Machines
title Differentially Private Singular Value Decomposition for Training Support Vector Machines
title_full Differentially Private Singular Value Decomposition for Training Support Vector Machines
title_fullStr Differentially Private Singular Value Decomposition for Training Support Vector Machines
title_full_unstemmed Differentially Private Singular Value Decomposition for Training Support Vector Machines
title_short Differentially Private Singular Value Decomposition for Training Support Vector Machines
title_sort differentially private singular value decomposition for training support vector machines
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8976603/
https://www.ncbi.nlm.nih.gov/pubmed/35378802
http://dx.doi.org/10.1155/2022/2935975
work_keys_str_mv AT sunzhenlong differentiallyprivatesingularvaluedecompositionfortrainingsupportvectormachines
AT yangjing differentiallyprivatesingularvaluedecompositionfortrainingsupportvectormachines
AT lixiaoye differentiallyprivatesingularvaluedecompositionfortrainingsupportvectormachines