Abstract
Large Language Models have demonstrated impressive capabilities in natural language understanding and processing. However, as AI and LLMs continue to evolve, their ability to accurately and efficiently interpret data from scientific figures and plots remains obscure. In this study, we test and evaluate the ability of state-of-the-art large language models (LLMs) in analyzing and extracting information from scientific figures commonly found in materials science literature, including stress-strain curves, heatmaps, 3D plots, contour plots, and other visual representations. We utilize a benchmark dataset comprising figures derived from several published works to evaluate the model's performance in quantitative material property extraction. Preliminary results highlight both the potential and current limitations of LLMs in handling visual scientific content, pointing toward future opportunities for AI-assisted data extraction in materials informatics.
Supplementary weblinks
Title
Supplemental Info GitHub Repository
Description
This public GitHub Repository contains the data in tables, replicated and original figures, and the code used to generate plots for this study.
Actions
View 


![Author ORCID: We display the ORCID iD icon alongside authors names on our website to acknowledge that the ORCiD has been authenticated when entered by the user. To view the users ORCiD record click the icon. [opens in a new tab]](https://www.cambridge.org/engage/assets/public/coe/logo/orcid.png)