Skip to main content

Thank you for visiting nature.com. You are using a browser version with limited support for CSS. To obtain the best experience, we recommend you use a more up to date browser (or turn off compatibility mode in Internet Explorer). In the meantime, to ensure continued support, we are displaying the site without styles and JavaScript.

Advertisement

Communications Materials
  • View all journals
  • Search
  • My Account Login
  • Content Explore content
  • About the journal
  • Publish with us
  • Sign up for alerts
  • RSS feed
  1. nature
  2. communications materials
  3. articles
  4. article
Modular large language model agents for multi-task computational materials science
Download PDF
Download PDF
  • Article
  • Open access
  • Published: 26 March 2026

Modular large language model agents for multi-task computational materials science

  • Akshat Chaudhari1,
  • Janghoon Ock  ORCID: orcid.org/0009-0000-0370-42122,3 &
  • Amir Barati Farimani  ORCID: orcid.org/0000-0002-2952-85762,4,5,6 

Communications Materials , Article number:  (2026) Cite this article

  • 1237 Accesses

  • 2 Citations

  • Metrics details

We are providing an unedited version of this manuscript to give early access to its findings. Before final publication, the manuscript will undergo further editing. Please note there may be errors present which affect the content, and all legal disclaimers apply.

Subjects

  • Computational science
  • Theory and computation

Abstract

The integration of large language models (LLMs) with domain-specific computational tools provides a pathway to streamline and enhance materials science workflows. This paper introduces MatSciAgent, a multi-agent framework supporting tasks such as materials data retrieval, continuum simulation, crystal structure generation, and molecular dynamics simulation. At its core is a master agent that interprets user queries, identifies the task type, and delegates to task-specific agent(s) equipped with tools. Leveraging databases such as Materials Project and MatWeb, the framework retrieves and summarizes materials data with grounded, factual responses, addressing limitations of vanilla LLMs. When a target material is absent from databases, a generative agent can propose plausible crystal structures. For simulations, specialized agents extract parameters to perform continuum and molecular dynamics simulations using existing software or custom code. MatSciAgent demonstrates stability, with parameter extraction achieving 100% success across five runs and materials extraction consistent in 9 of 10 runs. Its modular design ensures seamless extensibility to evolve as new capabilities are integrated.

Similar content being viewed by others

A fine-tuned large language model based molecular dynamics agent for code generation to obtain material thermodynamic parameters

Article Open access 25 March 2025

A family of large language models for materials research with insights into model adaptability in continued pretraining

Article 27 February 2026

Enabling large language models for real-world materials discovery

Article 10 July 2025

Data availability

The case study results and reliability study results are provided as supplementary data.

Code availability

The necessary code used in this study can be accessed at https://github.com/cakshat/MatSci-LLM-Agents.

References

  1. Pollice, R. et al. Data-driven strategies for accelerated materials design. Acc. Chem. Res. 54, 849–860 (2021).

    Google Scholar 

  2. Kalidindi, S. R. & De Graef, M. Materials data science: current status and future outlook. Annu. Rev. Mater. Res. 45, 171–193 (2015).

    Google Scholar 

  3. Jain, A. et al. The materials project: a materials genome approach to accelerating materials innovation. APL Mater. 1, 011002 (2013).

    Google Scholar 

  4. LLC, M. Matweb: Online Materials Information Resource. http://www.matweb.com (2020).

  5. Kirklin, S. et al. The open quantum materials database (oqmd): assessing the accuracy of DFT formation energies. npj Comput. Mater. 1, 1–15 (2015).

    Google Scholar 

  6. Hellenbrandt, M. The inorganic crystal structure database (ICSD)-present and future. Crystallogr. Rev. 10, 17–22 (2004).

    Google Scholar 

  7. Tanaka, I., Rajan, K. & Wolverton, C. Data-centric science for materials innovation. MRS Bull. 43, 659–663 (2018).

    Google Scholar 

  8. Antunes, L. M., Butler, K. T. & Grau-Crespo, R. Crystal structure generation with autoregressive large language modeling. Nat. Commun. 15, 10570 (2024).

    Google Scholar 

  9. Xie, T., Fu, X., Ganea, O.-E., Barzilay, R. & Jaakkola, T. Crystal diffusion variational autoencoder for periodic material generation. arXiv https://arxiv.org/abs/2110.06197 (2022).

  10. Zeni, C. et al. A generative model for inorganic materials design. Nature 639, 624–632 (2025).

    Google Scholar 

  11. Himanen, L., Geurts, A., Foster, A. S. & Rinke, P. Data-driven materials science: status, challenges, and perspectives. Adv. Sci. 6, 1900808 (2019).

    Google Scholar 

  12. Vaswani, A. et al. Attention is all you need. In Guyon, I.et al. (eds.) Advances in Neural Information Processing Systems, vol. 30 (Curran Associates, Inc., 2017).

  13. Brown, T. et al. Language models are few-shot learners. Adv. neural Inf. Process. Syst. 33, 1877–1901 (2020).

    Google Scholar 

  14. Achiam, J. et al. Gpt-4 Technical Report. https://cdn.openai.com/papers/gpt-4.pdf (2023).

  15. Touvron, H. et al. Llama: Open and efficient foundation language models. arXiv https://arxiv.org/abs/2302.13971 (2024).

  16. Anthropic. Claude 3.7 Sonnet. https://www.anthropic.com (2025).

  17. Tshitoyan, V. et al. Unsupervised word embeddings capture latent knowledge from materials science literature. Nature 571, 95–98 (2019).

    Google Scholar 

  18. Ock, J., Guntuboina, C. & Barati Farimani, A. Catalyst energy prediction with CatBERTA: unveiling feature exploration strategies through large language models. ACS Catal. 13, 16032–16044 (2023).

    Google Scholar 

  19. Ock, J., Badrinarayanan, S., Magar, R., Antony, A. & Barati Farimani, A. Multimodal language and graph learning of adsorption configuration in catalysis. Nat. Mach. Intell. 6, 1501–1511 (2024).

  20. Cao, Z., Magar, R., Wang, Y. & Barati Farimani, A. Moformer: Self-supervised transformer model for metal-organic framework property prediction. J. Am. Chem. Soc. 145, 2958–2967 (2023).

    Google Scholar 

  21. Xu, C., Wang, Y. & Barati Farimani, A. Transpolymer: a transformer-based language model for polymer property predictions. npj Comput. Mater. 9, 64 (2023).

    Google Scholar 

  22. Kim, S., Mollaei, P., Antony, A., Magar, R. & Barati Farimani, A. Gpcr-bert: Interpreting sequential design of g protein-coupled receptors using protein language models. J. Chem. Inf. Modeling 64, 1134–1144 (2024).

    Google Scholar 

  23. Mollaei, P., Sadasivam, D., Guntuboina, C. & Barati Farimani, A. Idp-bert: Predicting properties of intrinsically disordered proteins using large language models. J. Phys. Chem. B 128, 12030–12037 (2024).

    Google Scholar 

  24. Badrinarayanan, S., Guntuboina, C., Mollaei, P. & Barati Farimani, A. Multi-peptide: Multimodality leveraged language-graph learning of peptide properties. J. Chem. Inf. Model. 65, 83–91 (2025).

    Google Scholar 

  25. Meda, R. S. & Farimani, A. B. Bapulm: Binding affinity prediction using language models. arXiv https://arxiv.org/abs/2411.04150 (2024).

  26. Chaudhari, A., Guntuboina, C., Huang, H. & Farimani, A. B. Alloybert: Alloy property prediction with large language models. Comput. Mater. Sci. 244, 113256 (2024).

    Google Scholar 

  27. Yeh, Y.-T., Ock, J. & Farimani, A. B. Text to band gap: Pre-trained language models as encoders for semiconductor band gap prediction. arXiv https://arxiv.org/abs/2501.03456 (2025).

  28. Pak, P. & Farimani, A. B. Additivellm: Large language models predict defects in additive manufacturing. arXiv https://arxiv.org/abs/2501.17784 (2025).

  29. Lu, C. et al. The ai scientist: Towards fully automated open-ended scientific discovery. arXiv https://arxiv.org/abs/2408.06292 (2024).

  30. Ock, J., Vinchurkar, T., Jadhav, Y. & Farimani, A. B. Adsorb-agent: Autonomous identification of stable adsorption configurations via large language model agent. arXiv https://arxiv.org/abs/2410.16658 (2024).

  31. Boiko, D. A., MacKnight, R., Kline, B. & Gomes, G. Autonomous chemical research with large language models. Nature 624, 570–578 (2023).

    Google Scholar 

  32. Bran, A. M. et al. Augmenting large language models with chemistry tools. Nat. Mach. Intell. 6, 525–535 (2024).

    Google Scholar 

  33. Jadhav, Y., Pak, P. & Farimani, A. B. LLM-3D print: Large language models to monitor and control 3d printing. arXiv https://arxiv.org/abs/2408.14307 (2024).

  34. Jadhav, Y. & Farimani, A. B. Large language model agent as a mechanical designer. arXiv https://arxiv.org/abs/2404.17525 (2024).

  35. Huang, L. et al. A survey on hallucination in large language models: Principles, taxonomy, challenges, and open questions. arXiv https://doi.org/10.1145/3703155 (2024).

  36. Perković, G., Drobnjak, A. & Botički, I. Hallucinations in llms: Understanding and addressing challenges. In 2024 47th MIPRO ICT and Electronics Convention (MIPRO) 2084–2088 (IEEE, 2024).

  37. Liu, F. et al. Exploring and evaluating hallucinations in LLM-powered code generation. arXiv https://arxiv.org/abs/2404.00971 (2024).

  38. Birhane, A., Kasirzadeh, A., Leslie, D. & Wachter, S. Science in the age of large language models. Nat. Rev. Phys. 5, 277–280 (2023).

    Google Scholar 

  39. Johnson, S. & Hyland-Wood, D. A primer on large language models and their limitations. arXiv https://arxiv.org/abs/2412.04503 (2024).

  40. Chiang, Y., Hsieh, E., Chou, C.-H. & Riebesell, J. Llamp: Large language model made powerful for high-fidelity materials knowledge retrieval and distillation. arXiv https://arxiv.org/abs/2401.17244 (2024).

  41. Zhang, H., Song, Y., Hou, Z., Miret, S. & Liu, B. Honeycomb: A flexible LLM-based agent system for materials science. arXiv https://arxiv.org/abs/2409.00135 (2024).

  42. Ghafarollahi, A. & Buehler, M. J. Atomagents: Alloy design and discovery through physics-aware multi-modal multi-agent artificial intelligence. arXiv https://arxiv.org/abs/2407.10022 (2024).

  43. Radford, A. & Narasimhan, K. Improving Language Understanding by Generative Pre-training. https://api.semanticscholar.org/CorpusID:49313245 (2018).

  44. Vaswani, A. Attention is all you need. In 31st Conference on Neural Information Processing Systems. (NIPS, 2017).

  45. Chase, H. Langchain. https://github.com/langchain-ai/langchain (2022).

  46. Ling, C. et al. Domain specialization as the key to make large language models disruptive: a comprehensive survey. arXiv https://arxiv.org/abs/2305.18703 (2024).

  47. Li, H. et al. Blade: Enhancing black-box large language models with small domain-specific models arXiv https://arxiv.org/abs/2403.18365 (2024).

  48. Mittelstadt, B., Wachter, S. & Russell, C. To protect science, we must use LLMs as zero-shot translators. Nat. Hum. Behav. 7, 1830–1832 (2023).

    Google Scholar 

  49. Fanfoni, M. & Tomellini, M. The Johnson-Mehl-Avrami-Khokhlov model: a brief review. Il Nuovo Cim. D. 20, 1171–1182 (1998).

    Google Scholar 

  50. Kinoshita, M., Okamoto, Y. & Hirata, F. First-principle determination of peptide conformations in solvents: Combination of Monte Carlo simulated annealing and RISM theory. J. Am. Chem. Soc. 120, 1855–1863 (1998).

    Google Scholar 

  51. Wolfram, S. Statistical mechanics of cellular automata. Rev. Mod. Phys. 55, 601 (1983).

    Google Scholar 

  52. Wolfram, S. Cellular automata as models of complexity. Nature 311, 419–424 (1984).

    Google Scholar 

  53. Metropolis, N. & Ulam, S. The Monte Carlo method. J. Am. Stat. Assoc. 44, 335–341 (1949).

    Google Scholar 

  54. Saito, Y. & Enomoto, M. Monte Carlo simulation of grain growth. ISIJ Int. 32, 267–274 (1992).

    Google Scholar 

  55. Radford, A. et al. Language Models are Unsupervised Multitask Learners. https://api.semanticscholar.org/CorpusID:160025533 (2019).

  56. Karpathy, A. NanoGPT. https://github.com/karpathy/nanoGPT (2022).

  57. Nomad: A distributed web-based platform for managing materials science research data. J. Open Source Softw. 8, 5388 (2023).

  58. Mavromaras, A. et al. Computational materials engineering: capabilities of atomic-scale prediction of mechanical, thermal, and electrical properties of microelectronic materials. In 2010 11th International Thermal, Mechanical & Multi-Physics Simulation, and Experiments in Microelectronics and Microsystems (EuroSimE), Bordeaux, 1–10 (IEEE, 2010).

  59. Eyert, V. et al. Atomistic simulations of microelectronic materials: prediction of mechanical, thermal, and electrical properties. In Molecular Modeling and Multiscaling Issues for Electronic Material Applications. (eds. Iwamoto, N., Yuen, M., Fan, H.) (Springer, 2012).

  60. Larsen, A. H. et al. The atomic simulation environment-a Python library for working with atoms. J. Phys. Condens. Matter 29, 273002 (2017).

    Google Scholar 

  61. Foiles, S., Baskes, M. & Daw, M. S. Embedded-atom-method functions for the fcc metals Cu, Ag, Au, Ni, Pd, Pt, and their alloys. Phys. Rev. B 33, 7983 (1986).

    Google Scholar 

  62. Lennard-Jones, J. E. On the forces between atoms and ions. Proc. A 109, 584–597 (1925).

    Google Scholar 

  63. Choy, T. C. Effective Medium Theory: Principles and Applications 2nd edn, Vol. 256 (Oxford University Press, 2015).

Download references

Acknowledgements

The authors gratefully acknowledge support from the H. Robert Sharbaugh Presidential Fellowship.

Author information

Authors and Affiliations

  1. Department of Material Science and Engineering, Carnegie Mellon University, 5000 Forbes Avenue, Pittsburgh, PA, USA

    Akshat Chaudhari

  2. Department of Chemical Engineering, Carnegie Mellon University, 5000 Forbes Avenue, Pittsburgh, PA, USA

    Janghoon Ock & Amir Barati Farimani

  3. Department of Chemical and Biomolecular Engineering, University of Nebraska–Lincoln, Lincoln, NE, USA

    Janghoon Ock

  4. Department of Mechanical Engineering, Carnegie Mellon University, 5000 Forbes Avenue, Pittsburgh, PA, USA

    Amir Barati Farimani

  5. Department of Biomedical Engineering, Carnegie Mellon University, 5000 Forbes Avenue, Pittsburgh, PA, USA

    Amir Barati Farimani

  6. Machine Learning Department, Carnegie Mellon University, 5000 Forbes Avenue, Pittsburgh, PA, USA

    Amir Barati Farimani

Authors
  1. Akshat Chaudhari
    View author publications

    Search author on:PubMed Google Scholar

  2. Janghoon Ock
    View author publications

    Search author on:PubMed Google Scholar

  3. Amir Barati Farimani
    View author publications

    Search author on:PubMed Google Scholar

Contributions

A.C. conceived the study, developed the methodology, implemented the software, curated data, performed analysis, prepared visualizations, and wrote the original draft. J.O. contributed to methodology, validation, investigation, data curation, and writing and editing. A.B.F. supervised the project, contributed to conceptualization, project administration, funding acquisition, and writing and editing.

Corresponding authors

Correspondence to Janghoon Ock or Amir Barati Farimani.

Ethics declarations

Competing interests

The authors declare no competing interests.

Peer review

Peer review information

Communications Materials thanks Federico Ottomano and the other, anonymous, reviewer(s) for their contribution to the peer review of this work. A peer review file is available.

Additional information

Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary information

Supplementary Information (download PDF )

Supplementary Data 1

Supplementary Data 2

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License, which permits any non-commercial use, sharing, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if you modified the licensed material. You do not have permission under this licence to share adapted material derived from this article or parts of it. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by-nc-nd/4.0/.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Chaudhari, A., Ock, J. & Barati Farimani, A. Modular large language model agents for multi-task computational materials science. Commun Mater (2026). https://doi.org/10.1038/s43246-025-00994-x

Download citation

  • Received: 05 May 2025

  • Accepted: 14 October 2025

  • Published: 26 March 2026

  • DOI: https://doi.org/10.1038/s43246-025-00994-x

Share this article

Anyone you share the following link with will be able to read this content:

Sorry, a shareable link is not currently available for this article.

Provided by the Springer Nature SharedIt content-sharing initiative

Download PDF

Advertisement

Explore content

  • Research articles
  • Reviews & Analysis
  • News & Comment
  • Collections
  • Subjects
  • Follow us on X
  • Sign up for alerts
  • RSS feed

About the journal

  • Aims & Scope
  • Journal Information
  • Open Access Fees and Funding
  • Editors
  • Editorial Board
  • Calls for Papers
  • Referees
  • Contact
  • Conferences
  • Editorial Values Statement
  • Editorial policies
  • Journal Metrics

Publish with us

  • For authors
  • Language editing services
  • Open access funding
  • Submit manuscript

Search

Advanced search

Quick links

  • Explore articles by subject
  • Find a job
  • Guide to authors
  • Editorial policies

Communications Materials (Commun Mater)

ISSN 2662-4443 (online)

nature.com footer links

About Nature Portfolio

  • About us
  • Press releases
  • Press office
  • Contact us

Discover content

  • Journals A-Z
  • Articles by subject
  • protocols.io
  • Nature Index

Publishing policies

  • Nature portfolio policies
  • Open access

Author & Researcher services

  • Reprints & permissions
  • Research data
  • Language editing
  • Scientific editing
  • Nature Masterclasses
  • Research Solutions

Libraries & institutions

  • Librarian service & tools
  • Librarian portal
  • Open research
  • Recommend to library

Advertising & partnerships

  • Advertising
  • Partnerships & Services
  • Media kits
  • Branded content

Professional development

  • Nature Awards
  • Nature Careers
  • Nature Conferences

Regional websites

  • Nature Africa
  • Nature China
  • Nature India
  • Nature Japan
  • Nature Middle East
  • Privacy Policy
  • Use of cookies
  • Legal notice
  • Accessibility statement
  • Terms & Conditions
  • Your US state privacy rights
Springer Nature

© 2026 Springer Nature Limited

Nature Briefing AI and Robotics

Sign up for the Nature Briefing: AI and Robotics newsletter — what matters in AI and robotics research, free to your inbox weekly.

Get the most important science stories of the day, free in your inbox. Sign up for Nature Briefing: AI and Robotics