liu.seSearch for publications in DiVA
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Accelerating Molecular Graph Neural Networks via Knowledge Distillation
Linköping University, Faculty of Science & Engineering. Linköping University, Department of Computer and Information Science, The Division of Statistics and Machine Learning.ORCID iD: 0000-0003-4161-3631
Imperial College London, UK.
Technical University of Munich, Germany.
Google Research.
2023 (English)In: Advances in Neural Information Processing Systems 36 (NeurIPS 2023) / [ed] A. Oh, T. Naumann, A. Globerson, K. Saenko, M. Hardt, S. Levine, Neural Information Processing Systems Foundation Inc. (NeurIPS) , 2023Conference paper, Published paper (Refereed)
Abstract [en]

Recent advances in graph neural networks (GNNs) have enabled more comprehensive modeling of molecules and molecular systems, thereby enhancing the precision of molecular property prediction and molecular simulations. Nonetheless, as the field has been progressing to bigger and more complex architectures, state-of-the-art GNNs have become largely prohibitive for many large-scale applications. In this paper, we explore the utility of knowledge distillation (KD) for accelerating molecular GNNs. To this end, we devise KD strategies that facilitate the distillation of hidden representations in directional and equivariant GNNs, and evaluate their performance on the regression task of energy and force prediction. We validate our protocols across different teacher-student configurations and datasets, and demonstrate that they can consistently boost the predictive accuracy of student models without any modifications to their architecture. Moreover, we conduct comprehensive optimization of various components of our framework, and investigate the potential of data augmentation to further enhance performance. All in all, we manage to close the gap in predictive accuracy between teacher and student models by as much as 96.7% and 62.5% for energy and force prediction respectively, while fully preserving the inference throughput of the more lightweight models.

Place, publisher, year, edition, pages
Neural Information Processing Systems Foundation Inc. (NeurIPS) , 2023.
National Category
Computer Sciences
Identifiers
URN: urn:nbn:se:liu:diva-201969ISI: 001202273400026ISBN: 9781713899921 (electronic)OAI: oai:DiVA.org:liu-201969DiVA, id: diva2:1850797
Conference
37th Conference on Neural Information Processing Systems (NeurIPS), New Orleans, LA, dec 10-16, 2023
Available from: 2024-04-11 Created: 2024-04-11 Last updated: 2025-10-06Bibliographically approved
In thesis
1. Deep Learning for the Atomic Scale: Graph Neural Networks and Deep Generative Models with Some Applications to Materials and Molecules
Open this publication in new window or tab >>Deep Learning for the Atomic Scale: Graph Neural Networks and Deep Generative Models with Some Applications to Materials and Molecules
2025 (English)Doctoral thesis, comprehensive summary (Other academic)
Abstract [en]

The development of artificial intelligence, and in particular machine learning, has seen tremendous success in recent years. The use of machine learning, however, extends to vast application areas outside of those that we encounter in our day-to-day life. One such area is within the natural sciences, where research has shown promising results in using machine learning for modeling of systems of atoms. This is also the type of application for which the methods developed in this thesis are motivated. The thesis investigates and develops both predictive models that can predict properties and simulate these types of systems, and generative models that can propose new potential materials or molecules.

Particular emphasis is put on methods that model data as graphs, and the thesis starts with investigations of graph neural networks (GNNs) designed for predicting material and molecular properties. The performance of these models in the context of high-throughput screenings are put under scrutiny. The performance of GNNs when predicting properties of materials which are only hypothetical and structures which are not completely relaxed is investigated, and the insights are used to suggest a workflow that combines machine learning and conventional high-throughput methods. Additionally, an investigation of so-called knowledge distillation in the context of GNNs for systems of atoms has been performed. This study proposes some simple techniques for improving the performance of this type of GNNs, without sacrificing speed.

The generative modeling techniques developed in the thesis are both more generally applicable and specifically targeting the materials science domain. Among the general methods, the thesis investigates a type of generative autoregressive models where the generation order is a random variable, and develops discriminator guidance for such models. Additionally, a new sequential Monte Carlo algorithm, DDSMC, is developed for general Bayesian inverse problems. A dedicated materials science model, WyckoffDiff, is developed, utilizing a description of materials that explicitly encode information of their symmetries, with the aim of facilitating generation of materials with strict symmetrical properties.

While predictive and generative models can be useful on their own, the study on WyckoffDiff also highlights how they can be used together as parts of a materials discovery pipeline, with predictive models predicting the properties of the materials generated by a generative model like WyckoffDiff.

Place, publisher, year, edition, pages
Linköping: Linköping University Electronic Press, 2025. p. 74
Series
Linköping Studies in Science and Technology. Dissertations, ISSN 0345-7524 ; 2462
National Category
Artificial Intelligence Condensed Matter Physics
Identifiers
urn:nbn:se:liu:diva-218503 (URN)10.3384/9789181181852 (DOI)9789181181845 (ISBN)9789181181852 (ISBN)
Public defence
2025-11-07, Ada Lovelace, B-building, Campus Valla, Linköping, 13:15 (English)
Opponent
Supervisors
Note

Funding: This research was supported by the Excellence Center at Linköping–Lund in Information Technology (ELLIIT), the Swedish Research Council (VR) grant no. 2020-04122, 2024-05011, the Swedish Foundation for Strategic Research (SSF) Grant No. ICA16-0015, the Knut and Alice Wallenberg Foundation (KAW) via the Wallenberg AI, Autonomous Systems, and Software Program (WASP), the Wallenberg Initiative Material Science for Sustainability (WISE) through the joint WASP-WISE project Generative AI models for property to structure materials prediction, and KAW project 2020.0033. Much of the computations were enabled by the Berzelius resource provided by KAW at the National Supercomputer Centre and the Alvis resource provided by the National Academic Infrastructure for Supercomputing in Sweden (NAISS) at Chalmers e-Commons at Chalmers (C3SE) partially funded by the Swedish Research Council through grant agreement no. 2022-06725.

Available from: 2025-10-06 Created: 2025-10-06 Last updated: 2025-10-08Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Länk till konferensartikeln / Link to conference articleLänk till konferensen / Link to conference

Authority records

Ekström Kelvinius, Filip

Search in DiVA

By author/editor
Ekström Kelvinius, Filip
By organisation
Faculty of Science & EngineeringThe Division of Statistics and Machine Learning
Computer Sciences

Search outside of DiVA

GoogleGoogle Scholar

isbn
urn-nbn

Altmetric score

isbn
urn-nbn
Total: 145 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf