Skip to main content

Thank you for visiting nature.com. You are using a browser version with limited support for CSS. To obtain the best experience, we recommend you use a more up to date browser (or turn off compatibility mode in Internet Explorer). In the meantime, to ensure continued support, we are displaying the site without styles and JavaScript.

  • Article
  • Published:

Automated classification of cellular expression in multiplexed imaging data with Nimbus

Abstract

Multiplexed imaging offers a powerful approach to characterize the spatial topography of tissues in both health and disease. To analyze such data, the specific combination of markers that are present in each cell must be enumerated to enable accurate phenotyping, a process that often relies on unsupervised clustering. We constructed the Pan-Multiplex (Pan-M) dataset containing 197 million distinct annotations of marker expression across 15 different cell types. We used Pan-M to create Nimbus, a deep learning model to predict marker positivity from multiplexed image data. Nimbus is a pretrained model that uses the underlying images to classify marker expression of individual cells as positive or negative across distinct cell types, from different tissues, acquired using different microscope platforms, without requiring any retraining. We demonstrate that Nimbus predictions capture the underlying staining patterns of the full diversity of markers present in Pan-M, and that Nimbus matches or exceeds the accuracy of previous approaches that must be retrained on each dataset. We then show how Nimbus predictions can be integrated with downstream clustering algorithms to robustly identify cell subtypes in image data. We have open-sourced Nimbus and Pan-M to enable community use at https://github.com/angelolab/Nimbus-Inference.

This is a preview of subscription content, access via your institution

Access options

Buy this article

Prices may be subject to local taxes which are calculated during checkout

Fig. 1: NIMBUS improves marker prediction for phenotyping in multiplex images.
Fig. 2: Data annotation.
Fig. 3: Qualitative evaluation.
Fig. 4: Quantitative evaluation.
Fig. 5: Bi-modality of Nimbus confidence scores.

Similar content being viewed by others

Data availability

The Pan-M dataset can be downloaded via huggingface at https://huggingface.co/datasets/JLrumberger/Pan-Multiplex. Interactive jupyter notebooks to enable easier viewing of multiplexed imaging data are available via GitHub at https://github.com/angelolab/Nimbus-Inference/blob/main/templates/3_interactive_viewer.ipynb and https://colab.research.google.com/drive/1LW0vHC3sKKA3TyvW_9FeIaHj3PonzhGS. Source data are provided with this paper.

Code availability

Lightweight and easy to use inference and fine-tuning code for Nimbus is available via GitHub at https://github.com/angelolab/Nimbus-Inference. Code for preparing the dataset, model training and evaluation is available via GitHub at https://github.com/angelolab/Nimbus and code for figure generation is available via GitHub at https://github.com/angelolab/publications/tree/main/2024-Rumberger_Greenwald_etal_Nimbus.

References

  1. Black, S. et al. CODEX multiplexed tissue imaging with DNA-conjugated antibodies. Nat. Protoc. 16, 3802–3835 (2021).

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  2. Liu, C. C. et al. Reproducible, high-dimensional imaging in archival human tissue by multiplexed ion beam imaging by time-of-flight (MIBI-TOF). Lab Invest. 102, 762–770 (2022).

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  3. Giesen, C. et al. Highly multiplexed imaging of tumor tissues with subcellular resolution by mass cytometry. Nat. Methods 11, 417–422 (2014).

    Article  CAS  PubMed  Google Scholar 

  4. Hickey, J. W. et al. Organization of the human intestine at single-cell resolution. Nature 619, 572–584 (2023).

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  5. Jackson, H. W. et al. The single-cell pathology landscape of breast cancer. Nature 578, 615–620 (2020).

    Article  CAS  PubMed  Google Scholar 

  6. McCaffrey, E. F. et al. The immunoregulatory landscape of human tuberculosis granulomas. Nat. Immunol. 23, 318–329 (2022).

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  7. Le, H. et al. Machine learning for cell type classification from single nucleus RNA sequencing data. PLoS ONE 17, e0275070 (2022).

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  8. Liu, C. C. et al. Robust phenotyping of highly multiplexed tissue imaging data using pixel-level clustering. Nat. Commun. 14, 4618 (2023).

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  9. Samusik, N., Good, Z., Spitzer, M. H., Davis, K. L. & Nolan, G. P. Automated mapping of phenotype space with single-cell data. Nat. Methods 13, 493–496 (2016).

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  10. Van Gassen, S. et al. FlowSOM: using self-organizing maps for visualization and interpretation of cytometry data. Cytom. A 87, 636–645 (2015).

    Article  Google Scholar 

  11. Zhang, W. et al. Identification of cell types in multiplexed in situ images by combining protein expression and spatial information using CELESTA. Nat. Methods 19, 759–769 (2022).

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  12. Levine, J. H. et al. Data-driven phenotypic dissection of AML reveals progenitor-like cells that correlate with prognosis. Cell 162, 184–197 (2015).

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  13. Traag, V., Waltman, L. & Van Eck, N. From Louvain to Leiden: guaranteeing well-connected communities. Sci. Rep. 9, 5233 (2019).

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  14. Fu, R. et al. clustifyr: an R package for automated single-cell RNA sequencing cluster classification. F1000Res. 9, 223 (2020).

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  15. Kang, J. B. et al. Efficient and precise single-cell reference atlas mapping with Symphony. Nat. Commun. 12, 5890 (2021).

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  16. Lotfollahi, M. et al. Mapping single-cell data to reference atlases by transfer learning. Nat. Biotechnol. 40, 121–130 (2022).

    Article  CAS  PubMed  Google Scholar 

  17. Greenwald, N. F. et al. Whole-cell segmentation of tissue images with human-level performance using large-scale data annotation and deep learning. Nat. Biotechnol. 40, 555–565 (2022).

    Article  CAS  PubMed  Google Scholar 

  18. Stringer, C., Wang, T., Michaelos, M. & Pachitariu, M. Cellpose: a generalist algorithm for cellular segmentation. Nat. Methods 18, 100–106 (2021).

    Article  CAS  PubMed  Google Scholar 

  19. Schmidt, U., Weigert, M., Broaddus, C. & Myers, G. Cell detection with star-convex polygons. In Proc. Medical Image Computing and Computer Assisted Intervention–MICCAI 2018: 21st International Conference, Part II (eds Frangi, A. F. et al.) 11, 265–273 (Springer, 2018).

  20. Israel, U. et al. A foundation model for cell segmentation. Preprint at bioRxiv https://doi.org/10.1101/2023.11.17.567630 (2023).

  21. Goldsborough, T. et al. A novel channel invariant architecture for the segmentation of cells and nuclei in multiplexed images using InstanSeg. Preprint at bioRxiv https://doi.org/10.1101/2024.09.04.611150 (2024).

  22. Geuenich, M. J. et al. Automated assignment of cell identity from single-cell multiplexed imaging and proteomic data. Cell Syst. 12, 1173–1186 (2021).

    Article  CAS  PubMed  Google Scholar 

  23. Brbić, M. et al. Annotation of spatially resolved single-cell data with STELLAR. Nat. Methods 19, 1411–1418 (2022).

    Article  PubMed  PubMed Central  Google Scholar 

  24. Shaban, M. et al. MAPS: pathologist-level cell type annotation from tissue images through machine learning. Nat. Commun. 15, 28 (2024).

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  25. Chen, R. et al. Single-frame deep-learning super-resolution microscopy for intracellular dynamics imaging. Nat. Commun. 14, 2854 (2023).

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  26. Laubscher, E. et al. Accurate single-molecule spot detection for image-based spatial transcriptomics with weakly supervised deep learning. Cell Syst. 15, 475–482 (2024).

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  27. Weigert, M. et al. Content-aware image restoration: pushing the limits of fluorescence microscopy. Nat. Methods 15, 1090–1097 (2018).

    Article  CAS  PubMed  Google Scholar 

  28. Kather, J. N. et al. Deep learning can predict microsatellite instability directly from histology in gastrointestinal cancer. Nat. Med. 25, 1054–1056 (2019).

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  29. Kather, J. N. et al. Pan-cancer image-based detection of clinically actionable genetic alterations. Nat. Cancer 1, 789–799 (2020).

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  30. Amitay, Y. et al. CellSighter: a neural network to classify cells in highly multiplexed images. Nat. Commun. 14, 4302 (2023).

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  31. Jumper, J. et al. Highly accurate protein structure prediction with AlphaFold. Nature 596, 583–589 (2021).

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  32. Avsec, Ž. et al. Base-resolution models of transcription-factor binding reveal soft motif syntax. Nat. Genet. 53, 354–366 (2021).

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  33. Greenbaum, S. et al. A spatially resolved timeline of the human maternal–fetal interface. Nature 619, 595–605 (2023).

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  34. Falk, T. et al. U-Net: deep learning for cell counting, detection, and morphometry. Nat. Methods 16, 67–70 (2019).

    Article  CAS  PubMed  Google Scholar 

  35. Isensee, F., Jaeger, P. F., Kohl, S. A., Petersen, J. & Maier-Hein, K. H. nnU-Net: a self-configuring method for deep learning-based biomedical image segmentation. Nat. Methods 18, 203–211 (2021).

    Article  CAS  PubMed  Google Scholar 

  36. Otsu, N. et al. A threshold selection method from gray-level histograms. Automatica 11, 23–27 (1975).

    Google Scholar 

  37. Harris, C. R. et al. Quantifying and correcting slide-to-slide variation in multiplexed immunofluorescence images. Bioinformatics 38, 1700–1707 (2022).

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  38. Lloyd, S. Least squares quantization in PCM. IEEE Trans. Inf. Theory 28, 129–137 (1982).

    Article  Google Scholar 

  39. Xiong, J. et al. GammaGateR: semi-automated marker gating for single-cell multiplexed imaging. Bioinformatics 40, btae356 (2024).

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  40. Wang, H., Xiao, R., Dong, Y., Feng, L. & Zhao, J. ProMix: combating label noise via maximizing clean sample utility. In Proc. International Joint Conference on Artificial Intelligence (ed. Elkind, E.) 4442–4450 (2023).

  41. Lee, M. Y. et al. CellSeg: a robust, pre-trained nucleus segmentation and pixel quantification software for highly multiplexed fluorescence images. BMC Bioinf. 23, 46 (2022).

    Article  Google Scholar 

  42. Greenwald, N. F. et al. Temporal and spatial composition of the tumor microenvironment predicts response to immune checkpoint inhibition. Preprint at bioRxiv https://doi.org/10.1101/2025.01.26.634557 (2025).

  43. Bankhead, P. et al. QuPath: open source software for digital pathology image analysis. Sci. Rep. 7, 16878 (2017).

    Article  PubMed  PubMed Central  Google Scholar 

  44. Zoph, B., Vasudevan, V., Shlens, J. & Le, Q. V. Learning transferable architectures for scalable image recognition. In Proc. IEEE Conference on Computer Vision and Pattern Recognition 8697–8710 (IEEE, 2018).

  45. Tan, M. & Le, Q. Efficientnetv2: smaller models and faster training. In Proc. International Conference on Machine Learning (eds Meila, M. & Zhang, T.) 10096–10106 (PMLR, 2021).

  46. He, K., Zhang, X., Ren, S. & Sun, J. Deep residual learning for image recognition. In Proc. IEEE Conference on Computer Vision and Pattern Recognition 770–778 (IEEE, 2016).

  47. Loshchilov, I. & Hutter, F. Decoupled weight decay regularization. Preprint at https://doi.org/10.48550/arXiv.1711.05101 (2019).

  48. Jung, A. B. et al. imgaug. GitHub https://github.com/aleju/imgaug (2020).

  49. Rumberger, J. L. et al. How shift equivariance impacts metric learning for instance segmentation. In Proc. IEEE/CVF International Conference on Computer Vision 7128–7136 (IEEE, 2021).

  50. Pedregosa, F. et al. Scikit-learn: machine learning in Python. J. Mach. Learn. Res. 12, 2825–2830 (2011).

    Google Scholar 

  51. Ross, B. C. Mutual information between discrete and continuous data sets. PLoS ONE 9, e87357 (2014).

    Article  PubMed  PubMed Central  Google Scholar 

  52. Schiemann, R., Gherardini, P. F., Kageyama, R., Travers, M. & Kitch, L. Mantis Viewer. Zenodo https://doi.org/10.5281/zenodo.4009579 (2020).

Download references

Acknowledgements

This work was supported by the IFI program of the German Academic Exchange Service (DAAD) (J.L.R.); grant no. NCI CA264307 (N.F.G.) and the Stanford Graduate Fellowship (N.F.G.); grant no. NIAID F31AI165180 (C.C.L.) and the Stanford Graduate Fellowship (C.C.L.); NIH grant nos. 5U54CA20997105 (M.A.), 1R01CA24063801A1 (M.A.), 5R01AG06827902 (M.A.), 5UH3CA24663303 (M.A.), 5R01CA22952904 (M.A.), 1U24CA22430901 (M.A.), 5R01AG05791504 (M.A.), 5R01AG05628705 (M.A.), 5U24CA22430903 (M.A.), 3U54HL165445-03S1 (M.A.), 5R01AG05628705 (M.A.), 5R01AG05791505 (M.A); the Department of Defense grant nos. W81XWH2110143 (M.A.), 5U54CA261719-05 (M.A.).

Author information

Authors and Affiliations

Authors

Contributions

J.L.R., N.F.G., D.K. and M.A. formulated the project. J.L.R. created the deep learning pipeline and trained the models. J.L.R., N.F.G., A.K., S.R.V., C.S. and C.C.L. wrote the software. J.S.R., H.P. and I.A. ran the cell phenotyping workflows. J.L.R., N.F.G., P.B. and C.W. provided manual annotations for the gold-standard dataset. R.V., I.N., M.K. and T.J.H. helped to generate training data. J.L.R. performed the analyses and generated the figures. J.F. revised the figures. J.L.R., N.F.G. and M.A. wrote the paper. X.J.W. and D.V.V. provided guidance. N.F.G. and M.A. supervised the work. All authors reviewed the paper and provided feedback.

Corresponding authors

Correspondence to Noah F. Greenwald or Michael Angelo.

Ethics declarations

Competing interests

M.A. is an inventor on patents related to MIBI technology (patent nos. US20150287578A1, WO2016153819A1 and US20180024111A1). N.F.G. is an advisor for CellFormatica. The remaining authors declare no competing interests.

Peer review

Peer review information

Nature Methods thanks Anniina Färkkilä and Yasmin Kassim for their contribution to the peer review of this work. Peer reviewer reports are available. Primary Handling Editor: Madhura Mukhopadhyay, in collaboration with the Nature Methods team.

Additional information

Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Extended data

Extended Data Fig. 1 Pan-M conversion matrices and NIMBUS accuracy comparison.

a-c, Conversion matrices used to transform the cell types in the MIBI-TOF Decidua, CODEX Colon and MIBI-TOF Breast to marker positivity estimates. d, Performance metrics of naïve vs. noise-robust training procedures. e, Performance metrics for models trained with additional nuclei and cell membrane channels (four channels) as input. f, Performance metrics for full resolution (20x) vs. quarter resolution (10x) input data. g, Performance metrics for different backbone architectures. h, Performance metrics vs. gold standard annotations of the final model split by the datasets in Pan-M. i, F1 score comparison between specialized models trained only on specific subsets of Pan-M with general Nimbus model trained on the full Pan-M dataset. j-k, F1 score comparison of FoV-specific and FoV-agnostic unsupervised benchmarking methods applied on the Codex Colon and MIBI Breast subset of Pan-M. Bars show averages and error bars indicate standard deviations across three independent training runs in h-k.

Extended Data Fig. 2 NIMBUS F1 score comparison for individual channels.

a-e, F1 scores for the individual channels of the data subsets of Pan-M. f, Receiver Operating Characteristic (ROC) curves for the individual datasets and combined Pan-M dataset. g, Inference time for Nimbus as a function of increasing image size. h, Comparison of inference time between Nimbus and integrated expression as a function of increasing image size. i, Comparison of inference time between Nimbus, MAPS, and STELLAR, with preprocessing (integrated expression calculation) included. Bars show averages and error bars indicate standard deviations across three independent training runs in a-e.

Extended Data Fig. 3 NIMBUS positivity score correlation with integrated expression across datasets.

a-e, Correlation of Nimbus scores (x-axis) with integrated expression (y-axis) across the five datasets in Pan-M. Scores are colored based on their annotation as positive or negative from the gold-standard evaluations. f-g, Robustness analysis demonstrating the impact of modifying the radius for sparsity (f) and density (g) calculations.

Extended Data Fig. 4 NIMBUS positivity score correlation with integrated expression in specific channels.

a-e, Correlation of Nimbus scores (x-axis) with integrated expression (y-axis) across the five datasets in Pan-M, with representative channels shown for each. Scores are colored based on their annotation as positive or negative from the gold-standard evaluations.

Extended Data Fig. 5 Example of individual channels of the Codex Colon subset of Pan-M and their predicted Nimbus scores.

Codex colon markers that bind to the membrane are shown in a. b, Codex colon markers that bind to the nucleus. c, Codex colon markers binding to the cell membrane.

Extended Data Fig. 6 Example of individual channels of the Codex Colon and Vectra Colon subset of Pan-M and their predicted Nimbus scores.

Codex Colon markers that bind to the membrane are shown in a. b, Vectra Colon markers that bind to the nucleus. c, Vectra colon markers binding to the cell membrane.

Extended Data Fig. 7 Example of individual channels of the Vectra Pancreas and MIBI Breast subset of Pan-M and their predicted Nimbus scores.

Vectra Pancreas markers that bind to the membrane are shown in a. b, MIBI Breast markers that bind to the nucleus. c, MIBI Breast extracellular markers and d, markers binding to the cell membrane.

Extended Data Fig. 8 Example of individual channels of the MIBI Breast and MIBI Decidua subset of Pan-M and their predicted Nimbus scores.

MIBI Breast markers that bind to the membrane are shown in a. b, MIBI Decidua extracellular markers. c, MIBI Decidua markers that bind to the cell membrane.

Extended Data Fig. 9 Data Annotation.

a-c, Image samples with corresponding gold standard labels (positive cells marked with a yellow star) for the imaging platforms Vectra, CODEX and MIBI-TOF, respectively. d, overview of available, included and excluded markers for each individual subset of the Pan-M dataset.

Supplementary information

Reporting Summary

Peer Review File

Supplementary Table 1

Overview of dataset subsets contributing to the Pan-M dataset. Summary of multiplexed imaging datasets included in the Pan-M dataset, spanning multiple imaging platforms (MIBI-TOF, CODEX and Vectra), tissue types (breast, decidua, colon and pancreas) and segmentation and clustering methodologies.

Supplementary Table 2

Accuracy metrics across a range of different model designs, backbones and training setups.

Source data

Source Data Fig. 2

Dataset metrics.

Source Data Fig. 4

Performance benchmark.

Source Data Fig. 5

Phenotyping experiments.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Rumberger, J.L., Greenwald, N.F., Ranek, J.S. et al. Automated classification of cellular expression in multiplexed imaging data with Nimbus. Nat Methods 22, 2161–2170 (2025). https://doi.org/10.1038/s41592-025-02826-9

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue date:

  • DOI: https://doi.org/10.1038/s41592-025-02826-9

Search

Quick links

Nature Briefing AI and Robotics

Sign up for the Nature Briefing: AI and Robotics newsletter — what matters in AI and robotics research, free to your inbox weekly.

Get the most important science stories of the day, free in your inbox. Sign up for Nature Briefing: AI and Robotics