GenMS: Generative Hierarchical Materials Search

paper

Generative models trained at scale can now produce text, video, and more recently, scientific data such as crystal structures. In applications of generative approaches to materials science, and in particular to crystal structures, the guidance from the domain expert in the form of high-level instructions can be essential for an automated system to output candidate crystals that are viable for downstream research. In this work, we formulate end-to-end language-to-structure generation as a multi-objective optimization problem, and propose Generative Hierarchical Materials Search (GenMS) for controllable generation of crystal structures. GenMS consists of (1) a language model that takes high-level natural language as input and generates intermediate textual information about a crystal (e.g., chemical formulae), and (2) a diffusion model that takes intermediate information as input and generates low-level continuous value crystal structures. GenMS additionally uses a graph neural network to predict properties (e.g., formation energy) from the generated crystal structures. During inference, GenMS leverages all three components to conduct a forward tree search over the space of possible structures. Experiments show that GenMS outperforms other alternatives of directly using language models to generate structures both in satisfying user request and in generating low-energy structures. We confirm that GenMS is able to generate common crystal structures such as double perovskites, or spinels, solely from natural language input, and hence can form the foundation for more complex structure generation in near future.

Diffusion with Compact Crystal Representation

In order to support efficient tree search at inference time, we represent each crystal structure as a point cloud using an Ax4 tensor, where A represents the number of atoms in a crystal, and the inner 4 dimensions representing the location of an atom along with its atomic number. We then train a diffusion model to predict predict atom locations and atomic number as a continuous value, and round the continuous value to the closest atomic number. The diffusion model architecture with this compact crystal representation is shown below. Each atom undergoes blocks consisting of multi-layer perceptrons followed by orderinvariant self-attention. The MLP and self-attention blocks are repeated times where each repetition increases the dimension of the hidden units. The concatenation of skip connections are employed as in other U-Net architectures.

Evaluation

We first conduct end-to-end evaluation of generating crystal structure from natural language for three families of crystals. Generative Hierarchical Materials Search (GenMS) significantly outperforms LLM prompting baselines in producing unique and low-energy (predicted by GNN) structures that satisfy user request. We further conduct DFT calculation to compute formation energy in eV/atom averaged across structures generated by GenMS. Values before "/" in the row "(GNN/DFT)" represent GNN predicted formation energy, and after ``/'' represent DFT computed formation energy. We report formation energy from GNN prior to relaxation, and formation energy from DFT post relaxation. DFT calculations for baselines are eliminated as many structures from the baselines do not follow user instruction. N/A represents $E_f$ predicted by GNN falling outside of the reasonable range.

Furthermore, when we increase the amount of compute during inference (i.e., sample more formula and structure candidates), GenMS can find more stable crystal structures with lower formation energy:

Generated Structures

We test GenMS on a set of ad hoc language inputs to generate plausible examples from well-known crystal families. GenMS is able to search for the corresponding structures that satisfy user requests and have plausible initial geometries. Visualization provided by VESTA.

Citation


@article{yang2024generative,
	    title={Generative Hierarchical Materials Search},
            author={Yang, Sherry and Batzner, Simon and Gao, Ruiqi and Aykol, Muratahan and Gaunt, Alexander L and McMorrow, Brendan and Rezende, Danilo J and Schuurmans, Dale and Mordatch, Igor and Cubuk, Ekin D},
            journal={arXiv preprint arXiv:2409.06762},
	    year={2024}
	    }
1;95;0c