Cargando…

Scalable deeper graph neural networks for high-performance materials property prediction

Machine-learning-based materials property prediction models have emerged as a promising approach for new materials discovery, among which the graph neural networks (GNNs) have shown the best performance due to their capability to learn high-level features from crystal structures. However, existing G...

Descripción completa

Detalles Bibliográficos
Autores principales: Omee, Sadman Sadeed, Louis, Steph-Yves, Fu, Nihang, Wei, Lai, Dey, Sourin, Dong, Rongzhi, Li, Qinyang, Hu, Jianjun
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Elsevier 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9122959/
https://www.ncbi.nlm.nih.gov/pubmed/35607621
http://dx.doi.org/10.1016/j.patter.2022.100491
Descripción
Sumario:Machine-learning-based materials property prediction models have emerged as a promising approach for new materials discovery, among which the graph neural networks (GNNs) have shown the best performance due to their capability to learn high-level features from crystal structures. However, existing GNN models suffer from their lack of scalability, high hyperparameter tuning complexity, and constrained performance due to over-smoothing. We propose a scalable global graph attention neural network model DeeperGATGNN with differentiable group normalization (DGN) and skip connections for high-performance materials property prediction. Our systematic benchmark studies show that our model achieves the state-of-the-art prediction results on five out of six datasets, outperforming five existing GNN models by up to 10%. Our model is also the most scalable one in terms of graph convolution layers, which allows us to train very deep networks (e.g., >30 layers) without significant performance degradation. Our implementation is available at https://github.com/usccolumbia/deeperGATGNN.