Skip to main content

Thank you for visiting nature.com. You are using a browser version with limited support for CSS. To obtain the best experience, we recommend you use a more up to date browser (or turn off compatibility mode in Internet Explorer). In the meantime, to ensure continued support, we are displaying the site without styles and JavaScript.

Advertisement

Scientific Reports
  • View all journals
  • Search
  • My Account Login
  • Content Explore content
  • About the journal
  • Publish with us
  • Sign up for alerts
  • RSS feed
  1. nature
  2. scientific reports
  3. articles
  4. article
A multi-task deep learning and radiomics framework for fetal anatomical structure detection and classification in ultrasound imaging
Download PDF
Download PDF
  • Article
  • Open access
  • Published: 02 March 2026

A multi-task deep learning and radiomics framework for fetal anatomical structure detection and classification in ultrasound imaging

  • Xuan Zhou1 na1,
  • Jie Wan1 na1,
  • Fengjie Sun1,
  • Ruxin Wang2,
  • Yafei Yan1,
  • Pin Li1 &
  • …
  • Cuihua Wang1 

Scientific Reports , Article number:  (2026) Cite this article

  • 858 Accesses

  • Metrics details

We are providing an unedited version of this manuscript to give early access to its findings. Before final publication, the manuscript will undergo further editing. Please note there may be errors present which affect the content, and all legal disclaimers apply.

Subjects

  • Anatomy
  • Computational biology and bioinformatics
  • Health care
  • Mathematics and computing
  • Medical research

Abstract

To develop a comprehensive deep learning and radiomics-based multi-task pipeline for the detection and classification of key fetal anatomical structures in first-trimester ultrasound images, using a diverse multi-center dataset to ensure high variability, reproducibility, and generalizability. A total of 4,532 fetal ultrasound scans (gestational age 11–14 weeks), retrospectively collected from nine medical centers, were included in this study. Two detection models, You Only Look Once version 11 (YOLOv11) and shifted window transformer (Swin Transformer), were trained to localize nine fetal brain and craniofacial structures. From each detected region, 215 radiomic features and 1,792 deep features were extracted. Feature stability was ensured through intra-class correlation coefficient (ICC) filtering (threshold ≥ 0.75), Pearson correlation analysis, and least absolute shrinkage and selection operator (LASSO) regression. The selected features were then used to train a Transformer-based model for tabular data (TabTransformer) to classify fetal anatomical structures into clinically defined categories based on their sonographic appearance. Model performance was evaluated using Accuracy, area under the receiver operating characteristic curve (AUC), and Sensitivity across training, internal validation, and external test datasets. Fusion models integrating radiomic and deep features consistently outperformed single-modality models in both detection and classification. On the external test set, classification accuracy reached 96.1%, with AUCs up to 96.89%, and sensitivity exceeding 95% for key anatomical structures. Swin Transformer achieved superior localization performance compared to YOLOv11, with Intersection over Union (IoU) values up to 0.97 and F1-scores ≥ 0.94. Feature reproducibility remained above 75% across centers. The TabTransformer classifier demonstrated strong generalization and robustness, effectively leveraging the fused feature space for high-precision classification. This study presents the fully integrated, multi-task framework for fetal anatomical structure detection and classification using multi-center ultrasound data. The proposed approach demonstrates high reproducibility and diagnostic performance, offering strong clinical potential for early and objective fetal anomaly screening in the first trimester.

Data availability

The datasets used and/or analyzed during the current study are available from the corresponding author on reasonable request.

Abbreviations

AUC:

Area under the receiver operating characteristic curve

CI:

Confidence interval

CM:

Cisterna magna

CNN:

Convolutional neural network

DL:

Deep learning

F1-score:

Harmonic mean of precision and recall

ICC:

Intraclass correlation coefficient

IoU:

Intersection over Union

IT:

Intracranial translucency

LASSO:

Least absolute shrinkage and selection operator

mAP:

Mean average precision

NT:

Nuchal translucency

PCA:

Principal component analysis

ROC:

Receiver operating characteristic

ROI:

Region of interest

SD:

Standard deviation

SE:

Standard error

SMOTE:

Synthetic minority over-sampling technique

ViT:

Vision transformer

YOLOv11:

You only look once, Version 11

Swin transformer:

Shifted window transformer

TabTransformer:

Transformer model for tabular data

IBSI:

Image biomarker standardisation initiative

SGD:

Stochastic gradient descent

GPU:

Graphics processing unit

MRI:

Magnetic resonance imaging

2D:

Two-dimensional

3D:

Three-dimensional

References

  1. Sriraam, N., Chinta, B., Suresh, S. & Sudharshan, S. Ultrasound imaging-based recognition of prenatal anomalies: a systematic clinical engineering review. Prog Biomed. Eng. 6 (2), 023002 (2024).

    Google Scholar 

  2. Pozza, A. et al. Utility of fetal cardiac resonance imaging in prenatal clinical practice: current state of the art. Diagnostics (Basel). 13 (23), 3523 (2023).

    Google Scholar 

  3. Pelayo-Delgado, I., Gómez-Montes, E. & Álvaro-Navidad, M. Update on second trimester ultrasound scanning in pregnancy. Clínica e Investigación en Ginecología y Obstetricia 52(1), 100997 (2025).

    Google Scholar 

  4. Sepulveda, W., Wong, A. E., Ximenes, R. & Meagher, S. The first-trimester fetal anatomy scan. In: Obstetric Imaging: Fetal Diagnosis and Care-E-Book. 24 (2025).

  5. Puerto, B., Azumendi, P., Corrales, C. & Azumendi, G. How to perform a fetal neurosonography: Key points. Clin. Invest. Ginecol. Obstet. 52, 101050 (2025).

    Google Scholar 

  6. Gupta, N., Hiremath, S. B., Gauthier, I., Wilson, N. & Miller, E. Pediatric neurosonography: Comprehensive review and systematic approach. Can. Assoc. Radiol. J. 76(3), 519–533 (2025).

    Google Scholar 

  7. Kim, R. et al. Artificial intelligence based automatic classification, annotation, and measurement of the fetal heart using HeartAssist. Sci. Rep. 15 (1), 13055 (2025).

    Google Scholar 

  8. Qi, Y. et al. Multi-Center study on deep learning-assisted detection and classification of fetal central nervous system anomalies using ultrasound imaging. https://arXiv.org/abs/250102000. (2025).

  9. Salini, Y., Mohanty, S. N., Ramesh, J. V. N., Yang, M. & Chalapathi, M. M. V. Cardiotocography data analysis for fetal health classification using machine learning models. IEEE Access. 12, 26005–26022 (2024).

    Google Scholar 

  10. Yin, Y. & Bingi, Y. Using machine learning to classify human fetal health and analyze feature importance. BioMedInformatics 3 (2), 280–298 (2023).

    Google Scholar 

  11. Mushtaq, G. & Veningston, K. AI driven interpretable deep learning based fetal health classification. SLAS Technol. 29 (6), 100206 (2024).

    Google Scholar 

  12. Montin, E. et al. Radiomics for precision diagnosis of FAI: How close are we to clinical translation? A Multi-Center Validation of a Single-Center Trained Model. J. Clin. Med. 14(12), 4042 (2025).

    Google Scholar 

  13. Krishna, N. S. et al. Generalizability of AI-based image segmentation and centering estimation algorithm: A multi-region, multi-center, and multi-scanner study. Radiat. Prot. Dosimetry 201(6), 441–449 (2025).

    Google Scholar 

  14. Yin, S., Ming, J., Chen, H., Sun, Y. & Jiang, C. Integrating deep learning and radiomics for preoperative glioma grading using multi-center MRI data. Sci. Rep. 15(1), 36756 (2025).

    Google Scholar 

  15. Richter, M. et al. Generalizability of clinical prediction models in mental health. Mol. Psychiatry 1–8. (2025).

  16. Degtiar, I. & Rose, S. A review of generalizability and transportability. Annu. Rev. Stat. Its Appl. 10 (1), 501–524 (2023).

    Google Scholar 

  17. Maleki, F. et al. Generalizability of machine learning models: Quantitative evaluation of three methodological pitfalls. Radiol. Artif. Intell. 5(1), e220028 (2022).

    Google Scholar 

  18. Ambsdorf, J. et al. General methods make great domain-specific foundation models: A case-study on fetal ultrasound. In: International Conference on Medical Image Computing and Computer-Assisted Intervention. 271–281 (Springer, 2025).

  19. Fiorentino, M. C., Migliorelli, G., Villani, F. P., Frontoni, E. & Moccia, S. Contrastive prototype federated learning against noisy labels in fetal standard plane detection. Int. J. Comput. Assist. Radiol. Surg. https://doi.org/10.1007/s11548-025-03400-6 (2025).

    Google Scholar 

  20. Wang, F. et al. Fusing radiomic features with deep representations for gestational age estimation in fetal ultrasound images. In: International Conference on Medical Image Computing and Computer-Assisted Intervention. 230–240 (Springer, 2025).

  21. Lasala, A., Fiorentino, M. C., Micera, S., Bandini, A. & Moccia, S. Exploiting class activation mappings as prior to generate fetal brain ultrasound images with GANs. Annu. Int. Conf. IEEE Eng. Med. Biol. Soc. 2023, 1–4 (2023).

    Google Scholar 

  22. Huang, S. et al. Semi-supervised fetal brain parcellation via hierarchical learning framework. Med. Image Anal. 103835. (2025).

  23. Islam, U. et al. Fetal-Net: enhancing Maternal-Fetal ultrasound interpretation through Multi-Scale convolutional neural networks and Transformers. Sci. Rep. 15 (1), 25665 (2025).

    Google Scholar 

  24. Khanam, R. & Hussain, M. YOLOv11: An overview of the key architectural enhancements. https://arXiv.org/abs/241017725 (2024).

  25. Wei, W. et al. YOLOv11-based multi-task learning for enhanced bone fracture detection and classification in X-ray images. J. Radiat. Res. Appl. Sci. 18(1), 101309 (2025).

    Google Scholar 

  26. Cao, H. et al. Swin-UNet: Unet-like pure transformer for medical image segmentation. In: European Conference on Computer Vision. 205–218 (Springer, 2022).

  27. Liu, Z. et al. Swin Transformer: Hierarchical vision transformer using shifted windows. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. 10012–10022 (2021).

  28. Scapicchio, C. et al. A deep look into radiomics. Radiol. Med. 126(10), 1296–1311 (2021).

    Google Scholar 

  29. Yip, S. S. F. & Aerts, H. J. W. L. Applications and limitations of radiomics. Phys. Med. Biol. 61 (13), R150 (2016).

    Google Scholar 

  30. Rezaeijo, S. M. et al. Neighboring tissues as diagnostic windows: Neighborhood effects in radiomic detection of pancreatic ductal adenocarcinoma. Comput. Methods Programs Biomed. https://doi.org/10.1016/j.cmpb.2025.109056 (2025).

    Google Scholar 

  31. Khan, A. et al. A survey of the vision transformers and their CNN-transformer based variants. Artif. Intell. Rev. 56(Suppl 3), 2917–2970 (2023).

    Google Scholar 

  32. Pereira, G. A. & Hussain, M. A review of transformer-based models for computer vision tasks: Capturing global context and spatial relationships. https://arXiv.org/abs/240815178 (2024).

  33. Wang, J. et al. A machine learning model based on placental magnetic resonance imaging and clinical factors to predict fetal growth restriction. BMC Pregnancy Childbirth. 25 (1), 325 (2025).

    Google Scholar 

  34. Lai, H. et al. Radiomics-based correlation analysis of fetal brain MRI features and children’s neurodevelopmental outcomes in monochorionic twins. BMC Pregnancy Childbirth. 25 (1), 1040 (2025).

    Google Scholar 

  35. Zuo, M. et al. A nomogram based on MR radiomics and MR sign score for prenatal diagnosis of placenta accreta spectrum disorders and risk assessment of adverse clinical outcomes. Abdom. Radiol. 1–12. (2025).

  36. Xu, F. et al. Prediction of clinical pregnancy after frozen embryo transfer based on ultrasound radiomics: An analysis based on the optimal periendometrial zone. BMC Pregnancy Childbirth 25(1), 391 (2025).

    Google Scholar 

  37. Koo, T. K. & Li, M. Y. A guideline of selecting and reporting intraclass correlation coefficients for reliability research. J. Chiropr. Med. 15(2), 155–163 (2016).

    Google Scholar 

  38. Kim, Y. & Kim, J. Gradient LASSO for feature selection. In: Proceedings of the Twenty-First International Conference on Machine Learning, p. 60. (2004).

  39. Gomes, R., Pham, T., He, N., Kamrowski, C. & Wildenberg, J. Analysis of Swin-UNet vision transformer for Inferior Vena Cava filter segmentation from CT scans. Artif. Intell. Life Sci. 4, 100084 (2023).

    Google Scholar 

  40. Liang, J. et al. SwinIR: Image restoration using Swin transformer. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. 1833–44 (2021).

  41. Liu, Z. et al. Swin Transformer V2: Scaling up capacity and resolution. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 12009–19 (2022).

  42. Hatamizadeh, A. et al. Swin UNETR: Swin transformers for semantic segmentation of brain tumors in MRI images. In: International MICCAI Brainlesion Workshop. 272–284 (Springer, 2021).

  43. Kotthapalli, M., Ravipati, D. & Bhatia, R. YOLOv1 to YOLOv11: A Comprehensive Survey of Real-Time Object Detection Innovations and Challenges. https://arXiv.org/abs/250802067 (2025).

  44. Tan, M., Pang, R. & Le, Q. V. EfficientDet: Scalable and efficient object detection. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 10781–90 (2020).

  45. Powers, D. M. W. Evaluation: from precision, recall and F-measure to ROC, informedness, markedness and correlation. https://arXiv.org/abs/201016061 (2020).

  46. Wang, X. et al. Advanced network intrusion detection with TabTransformer. J. Theory Pract. Eng. Sci. 4(03), 191–8 (2024).

    Google Scholar 

  47. Huang, X., Khetan, A., Cvitkovic, M. & Karnin, Z. TabTransformer: Tabular data modeling using contextual embeddings. https://arXiv.org/abs/201206678 (2020).

  48. Insalata, B., Schmidt, F. & Vlassov, V. Multimodal survival prediction using TabTransformer and BioClinicalBERT on MIMIC-III. In: 2024 IEEE International Conference on Big Data (BigData). 1986–92 (IEEE, 2024).

  49. Cichosz, P. Assessing the quality of classification models: Performance measures and evaluation procedures. Cent. Eur. J. Eng. 1 (2), 132–158 (2011).

    Google Scholar 

  50. Du, Y. et al. Ultrasound-based radiomics technology in fetal lung texture analysis prediction of neonatal respiratory morbidity. Sci. Rep. 12(1), 12747 (2022).

    Google Scholar 

  51. Drukker, L. et al. Clinical workflow of sonographers performing fetal anomaly ultrasound scans: Deep-learning‐based analysis. Ultrasound Obstet. Gynecol. 60(6), 759–65 (2022).

    Google Scholar 

  52. Krishna, T. B. & Kokil, P. Standard fetal ultrasound plane classification based on stacked ensemble of deep learning models. Expert Syst. Appl. 238, 122153 (2024).

    Google Scholar 

  53. Gofer, S., Haik, O., Bardin, R., Gilboa, Y. & Perlman, S. Machine learning algorithms for classification of first-trimester fetal brain ultrasound images. J. Ultrasound Med. 41 (7), 1773–1779 (2022).

    Google Scholar 

  54. Hesse, L. S. et al. Subcortical segmentation of the fetal brain in 3D ultrasound using deep learning. Neuroimage 254, 119117 (2022).

    Google Scholar 

  55. Coronado-Gutiérrez, D. et al. Automatic deep learning-based pipeline for automatic delineation and measurement of fetal brain structures in routine mid-trimester ultrasound images. Fetal Diagn. Ther. 50 (6), 480–490 (2023).

    Google Scholar 

  56. Prieto, J. C. et al. An automated framework for image classification and segmentation of fetal ultrasound images for gestational age estimation. In Medical Imaging 2021: Image Processing 453–462 (SPIE, 2021).

    Google Scholar 

  57. Ghabri, H. et al. Transfer learning for accurate fetal organ classification from ultrasound images: A potential tool for maternal healthcare providers. Sci Rep 13(1), 17904 (2023).

    Google Scholar 

Download references

Author information

Author notes
  1. Xuan Zhou and Jie Wan contributed equally to this article and should be considered co-first authors.

Authors and Affiliations

  1. Department of Ultrasound Diagnosis, Affiliated Hospital of Hebei University, Baoding City, 071000, Hebei Province, China

    Xuan Zhou, Jie Wan, Fengjie Sun, Yafei Yan, Pin Li & Cuihua Wang

  2. Department of Radiology, Affiliated Hospital of Hebei University, Baoding City, 071000, Hebei Province, China

    Ruxin Wang

Authors
  1. Xuan Zhou
    View author publications

    Search author on:PubMed Google Scholar

  2. Jie Wan
    View author publications

    Search author on:PubMed Google Scholar

  3. Fengjie Sun
    View author publications

    Search author on:PubMed Google Scholar

  4. Ruxin Wang
    View author publications

    Search author on:PubMed Google Scholar

  5. Yafei Yan
    View author publications

    Search author on:PubMed Google Scholar

  6. Pin Li
    View author publications

    Search author on:PubMed Google Scholar

  7. Cuihua Wang
    View author publications

    Search author on:PubMed Google Scholar

Contributions

Conceptualization: Xuan Zhou, Jie Wan, and Pin Li. Data Curation: Fengjie Sun and Ruxin Wang. Formal Analysis: Xuan Zhou, Jie Wan, and Yafei Yan. Methodology: Pin Li, Cuihua Wang, and Xuan Zhou. Resources: Cuihua Wang and Pin Li. Supervision: Cuihua Wang and Pin Li. Validation: Jie Wan, Fengjie Sun, and Yafei Yan. Visualization: Xuan Zhou and Jie Wan. Writing—Original Draft: Xuan Zhou and Jie Wan. Writing—Review & Editing: Pin Li and Cuihua Wang.

Corresponding authors

Correspondence to Pin Li or Cuihua Wang.

Ethics declarations

Competing interests

The authors declare no competing interests.

Ethics approval and consent to participate

This study was conducted in accordance with the Declaration of Helsinki. Ethical approval and the requirement for informed consent were waived by the Institutional Review Board of the Affiliated Hospital of Hebei University (Baoding, China) because the study involved retrospective analysis of fully anonymized data and posed no risk to participants.

Additional information

Publisher’s note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary Information

Below is the link to the electronic supplementary material.

Supplementary Material 1 (download DOCX )

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License, which permits any non-commercial use, sharing, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if you modified the licensed material. You do not have permission under this licence to share adapted material derived from this article or parts of it. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by-nc-nd/4.0/.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Zhou, X., Wan, J., Sun, F. et al. A multi-task deep learning and radiomics framework for fetal anatomical structure detection and classification in ultrasound imaging. Sci Rep (2026). https://doi.org/10.1038/s41598-026-41635-8

Download citation

  • Received: 15 December 2025

  • Accepted: 22 February 2026

  • Published: 02 March 2026

  • DOI: https://doi.org/10.1038/s41598-026-41635-8

Share this article

Anyone you share the following link with will be able to read this content:

Sorry, a shareable link is not currently available for this article.

Provided by the Springer Nature SharedIt content-sharing initiative

Keywords

  • Deep learning
  • Fetal anatomy
  • First trimester
  • Machine learning
  • Radiomics
  • Ultrasound imaging
  • Visual transformer
Download PDF

Advertisement

Explore content

  • Research articles
  • News & Comment
  • Collections
  • Subjects
  • Follow us on Facebook
  • Follow us on X
  • Sign up for alerts
  • RSS feed

About the journal

  • About Scientific Reports
  • Contact
  • Journal policies
  • Guide to referees
  • Calls for Papers
  • Editor's Choice
  • Journal highlights
  • Open Access Fees and Funding

Publish with us

  • For authors
  • Language editing services
  • Open access funding
  • Submit manuscript

Search

Advanced search

Quick links

  • Explore articles by subject
  • Find a job
  • Guide to authors
  • Editorial policies

Scientific Reports (Sci Rep)

ISSN 2045-2322 (online)

nature.com footer links

About Nature Portfolio

  • About us
  • Press releases
  • Press office
  • Contact us

Discover content

  • Journals A-Z
  • Articles by subject
  • protocols.io
  • Nature Index

Publishing policies

  • Nature portfolio policies
  • Open access

Author & Researcher services

  • Reprints & permissions
  • Research data
  • Language editing
  • Scientific editing
  • Nature Masterclasses
  • Research Solutions

Libraries & institutions

  • Librarian service & tools
  • Librarian portal
  • Open research
  • Recommend to library

Advertising & partnerships

  • Advertising
  • Partnerships & Services
  • Media kits
  • Branded content

Professional development

  • Nature Awards
  • Nature Careers
  • Nature Conferences

Regional websites

  • Nature Africa
  • Nature China
  • Nature India
  • Nature Japan
  • Nature Middle East
  • Privacy Policy
  • Use of cookies
  • Legal notice
  • Accessibility statement
  • Terms & Conditions
  • Your US state privacy rights
Springer Nature

© 2026 Springer Nature Limited

Nature Briefing AI and Robotics

Sign up for the Nature Briefing: AI and Robotics newsletter — what matters in AI and robotics research, free to your inbox weekly.

Get the most important science stories of the day, free in your inbox. Sign up for Nature Briefing: AI and Robotics