Skip to main content

Thank you for visiting nature.com. You are using a browser version with limited support for CSS. To obtain the best experience, we recommend you use a more up to date browser (or turn off compatibility mode in Internet Explorer). In the meantime, to ensure continued support, we are displaying the site without styles and JavaScript.

Advertisement

npj Digital Medicine
  • View all journals
  • Search
  • My Account Login
  • Content Explore content
  • About the journal
  • Publish with us
  • Sign up for alerts
  • RSS feed
  1. nature
  2. npj digital medicine
  3. articles
  4. article
A robust vision language model for molecular status prediction and radiology report generation in adult-type diffuse gliomas
Download PDF
Download PDF
  • Article
  • Open access
  • Published: 02 April 2026

A robust vision language model for molecular status prediction and radiology report generation in adult-type diffuse gliomas

  • Yae Won Park1 na1,
  • Myeongkyun Kang2 na1,
  • Huiseung Ryu3,
  • Kyunghwa Han1,
  • Yongsik Sim4,
  • Ji Eun Park5,6,
  • Jong Hee Chang7,
  • Se Hoon Kim8,
  • Seung-Koo Lee1,
  • Sang Hyun Park3,9 &
  • …
  • Sung Soo Ahn1 

npj Digital Medicine , Article number:  (2026) Cite this article

We are providing an unedited version of this manuscript to give early access to its findings. Before final publication, the manuscript will undergo further editing. Please note there may be errors present which affect the content, and all legal disclaimers apply.

Subjects

  • Computational biology and bioinformatics
  • Medical research
  • Neurology
  • Neuroscience
  • Oncology

ABSTRACT

We aimed to establish a robust vision-language model (“Glio-LLaMA-Vision”) for molecular status prediction and radiology report generation (RRG) in adult-type diffuse gliomas. Multiparametric MRI data and paired radiology reports from 1001 patients with adult-type diffuse gliomas were included in the institutional training set. A vision-language model, Glio-LLaMA-Vision, was developed from LLaMA 3.1 pre-trained on 2.79 million biomedical image-text pairs from PubMed Central and further fine-tuned from the institutional training set. The performance was validated in 100 patients and 75 patients with paired MRI-radiology reports from an institutional validation set and another tertiary institution (AMC), and in 170 and 477 patients with MRI from TCGA and UCSF datasets, respectively. In terms of IDH mutation status prediction, Glio-LLaMA-Vision showed AUCs ranging from 0.85–0.95 in the internal validation and external datasets. In terms of RRG, the BLEU-1 and ROUGE-L scores were 0.50 and 0.49 in the internal validation, respectively, and 0.32 and 0.36 on the AMC dataset, respectively. Overall, 37.8% of generated reports were considered superior or equal to the original reports, while 91.0% of generated reports were considered clinically acceptable by neuroradiologists. In conclusion, Glio-LLaMA-Vision demonstrates promising performance in molecular status prediction and RRG in adult-type diffuse gliomas, showing potential for clinical assistance.

Data availability

The datasets analyzed during the current study are not publicly available due to institutional and ethical restrictions but are available from the corresponding author on reasonable request. Access to the data will be provided after review and approval by all authors and in compliance with applicable institutional policies and ethical guidelines.

Code availability

A valid OSI-approved open-source license, the MIT License, is applied to our code. The code repository includes a clear LICENSE file specifying that the code is released under the MIT License. Code is publicly available at (https://github.com/myeongkyunkang/Glio-LLaMA-Vision). Other information is available from the corresponding author upon request.

References

  1. Ostrom, Q. T. et al. CBTRUS statistical report: primary brain and other central nervous system tumors diagnosed in the United States in 2016-2020. Neuro Oncol. 25, iv1–iv99 (2023).

    Google Scholar 

  2. Louis, D. N. et al. The 2021 WHO classification of tumors of the central nervous system: a summary. Neuro Oncol. 23, 1231–1251 (2021).

    Google Scholar 

  3. Brat, D. J. et al. Comprehensive, integrative genomic analysis of diffuse lower-grade gliomas. N. Engl. J. Med. 372, 2481–2498 (2015).

    Google Scholar 

  4. Turkalp, Z., Karamchandani, J. & Das, S. IDH mutation in glioma: new insights and promises for the future. JAMA Neurol. 71, 1319–1325 (2014).

    Google Scholar 

  5. Sarkar, C. et al. Resource availability for CNS tumor diagnostics in the Asian Oceanian region: a survey by the Asian Oceanian Society of Neuropathology committee for Adapting Diagnostic Approaches for Practical Taxonomy in Resource-Restrained Regions (AOSNP-ADAPTR). Brain Pathol. e13329, https://doi.org/10.1111/bpa.13329 (2025).

  6. Vollmuth, P. et al. A Radiologist’s Guide to IDH-Wildtype Glioblastoma for Efficient Communication With Clinicians: PartI-Essential Information on Preoperative and Immediate Postoperative Imaging. Korean J. Radiol. 26, 246–248 (2025).

  7. Park, Y. W. et al. The 2021 WHO classification for gliomas and implications on imaging diagnosis: part 1-key points of the fifth edition and summary of imaging findings on adult-type diffuse gliomas. J. Magn. Reson. Imaging 58, 677–689 (2023).

    Google Scholar 

  8. Smith-Bindman, R. et al. Trends in use of medical imaging in US health care systems and in Ontario, Canada, 2000-2016. Jama 322, 843–856 (2019).

    Google Scholar 

  9. Kim, K. et al. Updated primer on generative artificial intelligence and large language models in medical imaging for medical professionals. Korean J. Radiol. 25, 224–242 (2024).

    Google Scholar 

  10. Nam, Y. et al. Multimodal large language models in medical imaging: current state and future directions. Korean J. Radiol. 26, 900–923 (2025).

    Google Scholar 

  11. Faghani, S., Park, Y. W. & Park, J. E. Uncover This Tech Term: Large Vision-Language Models in Radiology. Korean J. Radiol. 27, 375–378 (2026).

  12. Huang, W. et al. Enhancing representation in radiography-reports foundation model: a granular alignment algorithm using masked contrastive learning. Nat. Commun. 15, 7620 (2024).

    Google Scholar 

  13. Tanno, R. et al. Collaboration between clinicians and vision-language models in radiology report generation. Nat. Med. 31, 599–608 (2025).

    Google Scholar 

  14. Chen, Z. et al. A Vision-Language foundation model to enhance efficiency of chest x-ray interpretation. arXiv e-prints, arXiv: https://arxiv.org/abs/2401.12208 (2024).

  15. Dubey, A. et al. The llama 3 herd of models. arXiv preprint arXiv: https://arxiv.org/abs/2407.21783 (2024).

  16. Cluceru, J. et al. Improving the noninvasive classification of glioma genetic subtype with deep learning and diffusion-weighted imaging. Neuro Oncol. 24, 639–652 (2022).

    Google Scholar 

  17. van der Voort, S. R. et al. Combined molecular subtyping, grading, and segmentation of glioma using multi-task deep learning. Neuro Oncol. 25, 279–289 (2023).

    Google Scholar 

  18. Choi, Y. S. et al. Fully automated hybrid approach to predict the IDH mutation status of gliomas via deep learning and radiomics. Neuro Oncol. 23, 304–313 (2021).

    Google Scholar 

  19. Byeon, Y. et al. Interpretable multimodal transformer for prediction of molecular subtypes and grades in adult-type diffuse gliomas. npj Digit. Med. 8, 140 (2025).

    Google Scholar 

  20. Xiao, H. et al. A comprehensive survey of large language models and multimodal large language models in medicine. Inf. Fusion 117, 102888 (2025).

    Google Scholar 

  21. Li, C. et al. Llava-med: Training a large language-and-vision assistant for biomedicine in one day. Adv. Neural Inf. Process. Syst. 36, 28541–28564 (2023).

    Google Scholar 

  22. Tu, T. et al. Towards generalist biomedical AI. Nejm AI 1, AIoa2300138 (2024).

    Google Scholar 

  23. Hoopes, A., Butoi, V. I., Guttag, J. V. & Dalca, A. V. Voxelprompt: a vision-language agent for grounded medical image analysis. arXiv preprint arXiv: http://arxiv.org/abs/2410.08397 (2024).

  24. Tak, D. et al. A foundation model for generalized brain MRI analysis. medRxiv, https://doi.org/10.1101/2024.12.02.24317992 (2024).

  25. Gonzales, Ricardo A. et al. Metrics for Artificial Intelligence in Medicine: A Reference Resource. Radiol.: Artif. Intell. e260070 (2026).

  26. Park, S. H. & Kim, N. Challenges and proposed additional considerations for medical device approval of large language models beyond conventional AI. Radiology 312, e241703 (2024).

    Google Scholar 

  27. Yi, P. H. et al. Best practices for the safe use of large language models and other generative AI in radiology. Radiology 316, e241516 (2025).

    Google Scholar 

  28. Faghani, S. et al. Quantifying uncertainty in deep learning of radiologic images. Radiology 308, e222217 (2023).

    Google Scholar 

  29. Park, Y. W. et al. Differentiation of glioblastoma from solitary brain metastasis using deep ensembles: empirical estimation of uncertainty for clinical reliability. Comput. Methods Prog. Biomed. 254, 108288 (2024).

    Google Scholar 

  30. Louis, D. N. et al. cIMPACT-NOW update 1: not otherwise specified (NOS) and not elsewhere classified (NEC). Acta Neuropathol. 135, 481–484 (2018).

    Google Scholar 

  31. Gutman, D. A. et al. MR imaging predictors of molecular profile and survival: multi-institutional study of the TCGA glioblastoma data set. Radiology 267, 560–569 (2013).

    Google Scholar 

  32. Calabrese, E. et al. The University of California San Francisco preoperative diffuse glioma MRI dataset. Radio. Artif. Intell. 4, e220058 (2022).

    Google Scholar 

  33. Isensee, F. et al. Automated brain extraction of multisequence MRI using artificial neural networks. Hum. Brain Mapp. 40, 4952–4964 (2019).

    Google Scholar 

  34. Kickingereder, P. et al. Automated quantitative tumour response assessment of MRI in neuro-oncology with artificial neural networks: a multicentre, retrospective study. Lancet Oncol. 20, 728–740 (2019).

    Google Scholar 

  35. Zhang, S. et al. A multimodal biomedical foundation model trained from fifteen million image–text pairs. NEJM AI 2, AIoa2400640 (2025).

    Google Scholar 

  36. Yang, L. et al. Advancing multimodal medical capabilities of Gemini. arXiv preprint arXiv: http://arxiv.org/abs/2405.03162 (2024).

  37. Hasani, A. M. et al. Evaluating the performance of generative pre-trained transformer-4 (GPT-4) in standardizing radiology reports. Eur. Radiol. 34, 3566–3574 (2024).

    Google Scholar 

Download references

Acknowledgements

This work was supported by the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT) (RS-2025-00516124, RS-2025-00515423, and RS-2025-00515536). This study was also supported by a research grant from Yonsei University College of Medicine (4-2022-0828).

Author information

Author notes
  1. These authors contributed equally: Yae Won Park, Myeongkyun Kang.

Authors and Affiliations

  1. Department of Radiology and Research Institute of Radiological Science and Center for Clinical Imaging Data Science, Yonsei University College of Medicine, Seoul, Korea

    Yae Won Park, Kyunghwa Han, Seung-Koo Lee & Sung Soo Ahn

  2. Department of Robotics and Mechatronics Engineering, Daegu Gyeongbuk Institute of Science and Technology (DGIST), Daegu, Korea

    Myeongkyun Kang

  3. Graduate School of Artificial Intelligence, Pohang University of Science and Technology (POSTECH), Pohang, Korea

    Huiseung Ryu & Sang Hyun Park

  4. Department of Radiology, Samsung Medical Center, Sungkyunkwan University School of Medicine, Seoul, Korea

    Yongsik Sim

  5. Department of Radiology, Johns Hopkins University, Baltimore, MD, USA

    Ji Eun Park

  6. Department of Radiology and Research Institute of Radiology, Asan Medical Center, University of Ulsan College of Medicine, Seoul, Korea

    Ji Eun Park

  7. Department of Neurosurgery, Yonsei University College of Medicine, Seoul, Korea

    Jong Hee Chang

  8. Department of Pathology, Yonsei University College of Medicine, Seoul, Korea

    Se Hoon Kim

  9. Department of Computer Science and Engineering, Pohang University of Science and Technology (POSTECH), Pohang, Korea

    Sang Hyun Park

Authors
  1. Yae Won Park
    View author publications

    Search author on:PubMed Google Scholar

  2. Myeongkyun Kang
    View author publications

    Search author on:PubMed Google Scholar

  3. Huiseung Ryu
    View author publications

    Search author on:PubMed Google Scholar

  4. Kyunghwa Han
    View author publications

    Search author on:PubMed Google Scholar

  5. Yongsik Sim
    View author publications

    Search author on:PubMed Google Scholar

  6. Ji Eun Park
    View author publications

    Search author on:PubMed Google Scholar

  7. Jong Hee Chang
    View author publications

    Search author on:PubMed Google Scholar

  8. Se Hoon Kim
    View author publications

    Search author on:PubMed Google Scholar

  9. Seung-Koo Lee
    View author publications

    Search author on:PubMed Google Scholar

  10. Sang Hyun Park
    View author publications

    Search author on:PubMed Google Scholar

  11. Sung Soo Ahn
    View author publications

    Search author on:PubMed Google Scholar

Contributions

Y.W.P., K.H., Y.S., J.E.P., J.H.C., S.H.K., S.L., and S.S.A. collected and curated the data. Y.W.P., M.K., S.H.P., and S.S.A. conceptualized the study and developed the methodology. M.K. and H.R. developed the software. Y.W.P., M.K., H.R., K.H., Y.S., J.E.P., J.H.C., S.H.K., S.K., S.H.P., and S.S.A. performed the validation. Y.W.P., M.K., and H.R. performed the statistical analysis. S.H.P. and S.S.A. designed the study and supervised the research. Y.W.P. and M.K. wrote the original draft. All authors reviewed and approved the final manuscript.

Corresponding authors

Correspondence to Sang Hyun Park or Sung Soo Ahn.

Ethics declarations

Competing interests

The authors declare no competing interests.

Additional information

Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License, which permits any non-commercial use, sharing, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if you modified the licensed material. You do not have permission under this licence to share adapted material derived from this article or parts of it. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by-nc-nd/4.0/.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Park, Y.W., Kang, M., Ryu, H. et al. A robust vision language model for molecular status prediction and radiology report generation in adult-type diffuse gliomas. npj Digit. Med. (2026). https://doi.org/10.1038/s41746-026-02581-x

Download citation

  • Received: 06 December 2025

  • Accepted: 16 March 2026

  • Published: 02 April 2026

  • DOI: https://doi.org/10.1038/s41746-026-02581-x

Share this article

Anyone you share the following link with will be able to read this content:

Sorry, a shareable link is not currently available for this article.

Provided by the Springer Nature SharedIt content-sharing initiative

Download PDF

Associated content

Collection

Multimodal AI for Digital Medicine

Advertisement

Explore content

  • Research articles
  • Reviews & Analysis
  • News & Comment
  • Collections
  • Follow us on X
  • Sign up for alerts
  • RSS feed

About the journal

  • Aims and scope
  • Content types
  • Journal Information
  • About the Editors
  • Contact
  • Editorial policies
  • Calls for Papers
  • Journal Metrics
  • About the Partner
  • Open Access
  • Early Career Researcher Editorial Fellowship
  • Editorial Team Vacancies
  • News and Views Student Editor
  • Communication Fellowship

Publish with us

  • For Authors and Referees
  • Language editing services
  • Open access funding
  • Submit manuscript

Search

Advanced search

Quick links

  • Explore articles by subject
  • Find a job
  • Guide to authors
  • Editorial policies

npj Digital Medicine (npj Digit. Med.)

ISSN 2398-6352 (online)

nature.com footer links

About Nature Portfolio

  • About us
  • Press releases
  • Press office
  • Contact us

Discover content

  • Journals A-Z
  • Articles by subject
  • protocols.io
  • Nature Index

Publishing policies

  • Nature portfolio policies
  • Open access

Author & Researcher services

  • Reprints & permissions
  • Research data
  • Language editing
  • Scientific editing
  • Nature Masterclasses
  • Research Solutions

Libraries & institutions

  • Librarian service & tools
  • Librarian portal
  • Open research
  • Recommend to library

Advertising & partnerships

  • Advertising
  • Partnerships & Services
  • Media kits
  • Branded content

Professional development

  • Nature Awards
  • Nature Careers
  • Nature Conferences

Regional websites

  • Nature Africa
  • Nature China
  • Nature India
  • Nature Japan
  • Nature Middle East
  • Privacy Policy
  • Use of cookies
  • Legal notice
  • Accessibility statement
  • Terms & Conditions
  • Your US state privacy rights
Springer Nature

© 2026 Springer Nature Limited

Nature Briefing

Sign up for the Nature Briefing newsletter — what matters in science, free to your inbox daily.

Get the most important science stories of the day, free in your inbox. Sign up for Nature Briefing