Hynes, Level 2, Room 206
DS02
The high-level goal of this tutorial is to demonstrate the versatility of graph neural networks for inverse design of materials. Different strategies for materials design will be explored.
The specific objectives of the tutorial are:
Introduction to Graph NN
Taylor Sparks, University of Utah
This section will include fundamentals of graph networks: components, directionality and tasks, molecules and crystals as graphs, matrix representation, graph connectivity and centralities, message passing, node/edge/graph representations and comments on advanced alternative GNNs (multigraphs, hierarchical graphs etc).
GNN Universal Interatomic Potential for Materials Design
Chi Chen, Microsoft Quantum
This section will include basics of computational materials and tools, running M3GNet for structural relaxation and property predictions and running M3GNet for molecular dynamics and diffusivity and conductivity calculations.
Zintl Phases and Hierarchical Graph NN
Prashun Gorai, Colorado School of Mines, NREL; Qian Yang, University of Connecticut
This section will include introduction to Zintl phases and structural motifs, hierarchical graph neural networks for crystals and examples of running code and results of trained model - visualization of automatically identified motifs for Zintl phases.
Reinforcement Learning with Graph NN
Peter St. John, National Renewable Energy Lab
This section will include introduction to the graph-env and rlcrystal packages, specification of an action space for a reduced crystal structure search, construction of a GNN policy model and running RL search for top-performing candidates.
The high-level goal of this tutorial is to introduce researchers to differentiable programming and to demonstrate how differentiable simulations can open qualitatively new avenues for research in materials science.
As described in the title, we will use JAX MD — a novel software library for differentiable molecular dynamics — as a platform for the tutorial. The entire tutorial will take the form of Julia notebooks (hosted on Google Colab) that will allow participants to interactively participate.
The tutorial has several specific goals.
Introduction
Samuel Schoenholz, Google Research
The introductory portion of the tutorial will describe AD and how it can be used to change how we think about atomistic simulations. It will include a short introduction to JAX and then it will introduce JAX MD.
Physical Quantities Made Easy
The next part of the tutorial will show how many quantities can be computed efficiently using AD by taking derivatives of the Hamiltonian. This will include forces, stress and pressure, elastic constants, and phonon spectra.
Neural Network Potentials
Amil Merchant, Stanford University
We will show how easy it is to combine state-of-the-art neural networks with atomistic simulations when everything has been built to support AD from the ground up. This will involve instantiating and (beginning) to train a state-of-the-art equivariant graph neural network. After this, we will demonstrate the usage of this network in several practical settings.
Carl Goodrich, IST Austria
To prepare for the final section of the tutorial on meta-optimization, we will see how primitive operations in molecular dynamics can be composed with JAX’s automatic vectorization to produce a wide range of simulation environments and tools. In particular, we will go through the construction of simulations with temperature gradients and the nudge elastic band method for identifying saddle points between optima.
Meta Optimization
Ella King, Harvard University