Feb. 8, 2024, 5:41 a.m. | Nate Gruver Anuroop Sriram Andrea Madotto Andrew Gordon Wilson C. Lawrence Zitnick Zachary Ulissi

cs.LG updates on arXiv.org arxiv.org

We propose fine-tuning large language models for generation of stable materials. While unorthodox, fine-tuning large language models on text-encoded atomistic data is simple to implement yet reliable, with around 90% of sampled structures obeying physical constraints on atom positions and charges. Using energy above hull calculations from both learned ML potentials and gold-standard DFT calculations, we show that our strongest model (fine-tuned LLaMA-2 70B) can generate materials predicted to be metastable at about twice the rate (49% vs 28%) of …

atom cond-mat.mtrl-sci constraints cs.lg data energy fine-tuning generate language language models large language large language models materials simple text

Lead Developer (AI)

@ Cere Network | San Francisco, US

Research Engineer

@ Allora Labs | Remote

Ecosystem Manager

@ Allora Labs | Remote

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote