What do Large Language Models know about materials?
By: Adrian Ehrenhofer, Thomas Wallmersperger, Gianaurelio Cuniberti
Potential Business Impact:
Computers learn about materials for engineering.
Large Language Models (LLMs) are increasingly applied in the fields of mechanical engineering and materials science. As models that establish connections through the interface of language, LLMs can be applied for step-wise reasoning through the Processing-Structure-Property-Performance chain of material science and engineering. Current LLMs are built for adequately representing a dataset, which is the most part of the accessible internet. However, the internet mostly contains non-scientific content. If LLMs should be applied for engineering purposes, it is valuable to investigate models for their intrinsic knowledge -- here: the capacity to generate correct information about materials. In the current work, for the example of the Periodic Table of Elements, we highlight the role of vocabulary and tokenization for the uniqueness of material fingerprints, and the LLMs' capabilities of generating factually correct output of different state-of-the-art open models. This leads to a material knowledge benchmark for an informed choice, for which steps in the PSPP chain LLMs are applicable, and where specialized models are required.
Similar Papers
34 Examples of LLM Applications in Materials Science and Chemistry: Towards Automation, Assistants, Agents, and Accelerated Scientific Discovery
Machine Learning (CS)
Helps scientists invent new materials faster.
Optimizing Data Extraction from Materials Science Literature: A Study of Tools Using Large Language Models
Digital Libraries
AI finds science facts in papers faster.
LLMs4All: A Review on Large Language Models for Research and Applications in Academic Disciplines
Computation and Language
AI helps study many school subjects better.