Cargando…
An Improved Similarity-Based Clustering Algorithm for Multi-Database Mining
Clustering algorithms for multi-database mining (MDM) rely on computing [Formula: see text] pairwise similarities between n multiple databases to generate and evaluate [Formula: see text] candidate clusterings in order to select the ideal partitioning that optimizes a predefined goodness measure. Ho...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8144976/ https://www.ncbi.nlm.nih.gov/pubmed/33947081 http://dx.doi.org/10.3390/e23050553 |
Sumario: | Clustering algorithms for multi-database mining (MDM) rely on computing [Formula: see text] pairwise similarities between n multiple databases to generate and evaluate [Formula: see text] candidate clusterings in order to select the ideal partitioning that optimizes a predefined goodness measure. However, when these pairwise similarities are distributed around the mean value, the clustering algorithm becomes indecisive when choosing what database pairs are considered eligible to be grouped together. Consequently, a trivial result is produced by putting all the n databases in one cluster or by returning n singleton clusters. To tackle the latter problem, we propose a learning algorithm to reduce the fuzziness of the similarity matrix by minimizing a weighted binary entropy loss function via gradient descent and back-propagation. As a result, the learned model will improve the certainty of the clustering algorithm by correctly identifying the optimal database clusters. Additionally, in contrast to gradient-based clustering algorithms, which are sensitive to the choice of the learning rate and require more iterations to converge, we propose a learning-rate-free algorithm to assess the candidate clusterings generated on the fly in fewer upper-bounded iterations. To achieve our goal, we use coordinate descent (CD) and back-propagation to search for the optimal clustering of the n multiple database in a way that minimizes a convex clustering quality measure [Formula: see text] in less than [Formula: see text] iterations. By using a max-heap data structure within our CD algorithm, we optimally choose the largest weight variable [Formula: see text] at each iteration i such that taking the partial derivative of [Formula: see text] with respect to [Formula: see text] allows us to attain the next steepest descent minimizing [Formula: see text] without using a learning rate. Through a series of experiments on multiple database samples, we show that our algorithm outperforms the existing clustering algorithms for MDM. |
---|