Abstract
Deep learning tools based on computer vision have emerged as alternative methods for assessing radiographic image patterns. These approaches have been explored for various forensic applications, including sex and age estimation. This study aimed to evaluate the diagnostic accuracy of a Convolutional Neural Network (CNN) in classifying radiographic images by sex and age, focusing on the nasal aperture as the morphological feature of interest. The sample comprised 9,349 radiographs annotated for the nasal aperture region. A CNN architecture based on the You Only Look Once series—specifically the intermediate version 11 for object classification (YOLO11m-cls)—was implemented, with training performed using 5-fold cross-validation. The overall accuracy rate was 74% (ranging from 61% to 88%), and the area under the Receiver Operating Characteristic (ROC) curve was 0.74. Correct classification rates were 73% for males and 75.17% for females. Accuracy varied with age, showing a 10% decrease among younger individuals compared to older ones. The study confirmed the reduced expression of sexually dimorphic traits in younger individuals and supported existing recommendations against performing sex estimation in subadults. Within the present methodological framework, the nasal aperture demonstrated limited applicability for sex estimation in the studied sample, with an accuracy rate corresponding to approximately one misclassification out of every four predictions.
Similar content being viewed by others
Introduction
In forensic science, human identification often relies on skeletal and dental morphological features to reconstruct a postmortem biological profile1. This profile can be composed of the estimated sex, age, population affinity, and stature of an individual2. Modern forensic anthropology has benefited from computer vision and artificial intelligence technologies to enhance and optimize its practices3. Some of these resources rely on the virtual analysis of the human anatomy by means of three-dimensional (3D) and bidimensional (2D) medical imaging, such as computed tomography4,5 and traditional radiography6,7 respectively. While the former allows realistic and detailed visualization of bones and teeth,8 the latter offers the advantages of faster data processing and transfer, as well as the ability to acquire and store large samples in small spaces. These advantages are especially useful when operating computer vision solutions via Convolutional Neural Network (CNN). This is because CNNs are deep learning models designed to analyze visual data through successive layers that detect image patterns,9 enabling complex feature classification and detection. Recent forensic science studies employing CNN to detect and classify anatomical features in radiographs have demonstrated applicability of artificial intelligence for sex and age estimation10,11,12,13,14,15. Among the improvements foreseen for forensic practice, reduction in operator interventions and faster processing are expected7.
The nasal aperture is a maxillofacial region of interest when it comes to reconstructing a biological profile given its potential populational-16 and sex-specific19,20 variations. For instance, authors19 have demonstrated statistically significant differences between males and females based on the morphometric analyses of the nasal width and height in computed tomography scans. However, the mean accuracy rates were below 65%.19 Sexual dimorphism has also been reported in the scientific literature when measurements are taken directly from dry human skulls20. In this context, nasal height has been highlighted as a sexually dimorphic anatomical feature, being greater in males than in females20. Regarding shape, authors have found an absence of sexual dimorphism in the nasal aperture in specific populations, namely Black and White South African skulls21. Anthropological studies such as these have contributed to the current body of knowledge in the field. However, subsequent investigations can be proposed to further expand the scientific understanding of the nasal aperture’s applicability for sex estimation. To this end, radiographs of the viscerocranium may serve as valuable resources to facilitate large-scale data collection and analytical compatibility with CNN models.
Based on the foregoing, this study aimed to test the diagnostic accuracy of the nasal aperture for sex estimation using a semi-automated CNN approach applied to radiographic analysis.
Materials and methods
Study design and ethical aspects
In this study, a diagnostic accuracy test was designed to compare the performance of a CNN (index test) in estimating the sex of children, adolescents and young adults through analysis of the nasal aperture on panoramic radiographs. The assessment of medical images was conducted retrospectively using an existing database of patient records. All radiographs were acquired exclusively for diagnostic, therapeutic or dental treatment follow-up purposes, ensuring that no patient was exposed to ionizing radiation solely for research. The outlined investigation protocol received approval from the Institutional Committee of Ethics in Human Research at the Faculdade São Leopoldo Mandic (Protocol No. 76809023.9.0000.5374) and was reported in accordance with the Standards for Reporting of Diagnostic Accuracy Studies (STARD)22, while addressing current key considerations regarding dental artificial intelligence research23. The images utilized in this study constituted secondary data sourced from an established radiology database (Center of Oral Radiology and Imaging). Access to the data was authorized through informed permission granted by the database’s legal custodian. Given the retrospective design and the use of anonymized imaging data, the Institutional Committee of Ethics in Human Research at the Faculdade São Leopoldo Mandic formally waived the requirement for direct informed consent from individual patients.
Participants
The sample consisted of panoramic radiographs (n = 9349) from Brazilian males (n = 4375, 46.8%) and females (n = 4974, 53.2%), aged between 6 and 22.9 years (Table 1). The inclusion criteria comprised individuals with at least one radiograph with known date of image acquisition, date of birth and recorded sex. The exclusion criteria included radiographs showing nose piercings, evidence of trauma or surgery in the middle third of the face (e.g., orthopedic fixation devices) or other metallic apparatus in the viscerocranium, or visible signs of skeletal deformity. Further dataset partitioning was performed at the participant level, meaning that only one image from each patient was assigned to a single split and not to another (i.e. train or validation), thereby preventing data leakage and promoting unbiased evaluation of model generalization.
Analysis
Image anonymization was performed by cropping out the radiographic frames containing the patients’ age and sex and image side indicator (left/right). Subsequently, each radiograph was assigned an alphanumeric code to facilitate further de-identification. All radiographs were originally similar, as they were obtained from a single oral radiology clinic. However, to ensure higher standardization, they were pre-processed to preserve their size, image detail, spatial resolution, and quality. Image annotation was performed by five trained forensic odontologists experienced in annotations on panoramic radiographs6,7,10,14 using Darwin V7 software package (Darwin V7 Labs, London, UK) with its native bounding-box tool. The bounding-box enabled selection of the region of interest (ROI) on panoramic radiographs by manually dragging a rectangular outline over the image. In the present study, the ROI was the nasal aperture (Fig. 1), enabling a margin of 8–10% to preserve the nasal aperture contour and best fit its anatomic context. The images were resized to 224 × 224 pixels, converted to 3 channels (replication of the grayscale), scaled to the [0,1] range, and min–max normalized per image. Image augmentation was applied to the training dataset and included the following transformations with the indicated probabilities: random horizontal flip (p = 0.5), rotation of ± 7° (p = 0.5), translation of ± 6% (p = 0.3), zoom ranging from 0.9 to 1.1 (p = 0.3), brightness/contrast variation of ± 10% (p = 0.3), Gaussian noise with σ = 0.01 (p = 0.2) and mild sharpening (p = 0.2). Aggressive cropping was avoided to prevent truncation of the superior border of the ROI (nasal aperture region, often located close to the upper limit of the panoramic radiograph).
Workflow of a convolutional neural network (CNN) applied to panoramic radiographs for the estimation of sex and age. A total of 9349 radiographs from individuals aged from 6 to 22.99 years old (y. o.) were used. Images, sex, and chronological age were provided as input features. Data were split into five folds for cross-validation, with training (blue) and validation (yellow) sets. Image preprocessing was applied prior to CNN feature extraction through convolution and pooling layers. Fully connected layers integrated extracted features to predict two outcomes: sex (male [♂] vs. female [♀]) and age (> 15 years old vs. ≤ 15 years old).
A deep learning architecture based on the YOLO11m-cls model24,25 was trained over 100 epochs (Table 2). Recent studies in the scientific literature have demonstrated the potential of this model series for medical imaging applications and have specifically supported its use in two-dimensional radiographic assessments26. This version of the model incorporates advanced architectural components, including the introduction of the C3k2 (Cross Stage Partial with kernel size 2) block, Spatial Pyramid Pooling – Fast (SPPF), and C2PSA (Convolutional block with Parallel Spatial Attention), which contribute to improving the model’s performance in several ways, such as enhanced feature extraction.
Categorical cross-entropy loss and L2 regularization (weight decay = 5 × 10⁻⁴) were used in the training, implemented in the stochastic gradient descent optimizer with a learning rate of 0.0125, momentum of 0.937, 100 epochs, and a batch size of 16. Model evaluation was performed with 5-fold cross-validation,27,28 where in each iteration approximately 20% of the images (n ≈ 1,869–1,870) were retained as an external test set, while from the remaining ~ 80% (n ≈ 7,479–7,480), about 10% was reserved exclusively for monitoring the training process. Early stopping was not applied; instead, at the end of 100 epochs, the checkpoint with the lowest monitoring loss was selected and subsequently evaluated on the corresponding test fold. We reported the average performance across all five folds. The choice of k = 5 represented a balance between computational cost and robustness: increasing k linearly raises the training cost (e.g., k = 10 would double the computational burden without proportionally improving precision), while with n = 9,349, each fold provided a sufficiently large test set to yield stable estimates and a training set large enough to preserve generalization. This arrangement ensured that all images were used once as test data, allowed confidence interval estimation from the distribution of fold scores, and provided a practically robust yet computationally feasible evaluation strategy. The computer vision analysis was performed by two experienced engineers.
Test methods
The reference standards to which the CNN was compared were the individuals’ documented sex (binary: male or female) and their chronological age (obtained between the date of birth and date of image acquisition). YOLO11m-cls was tested based on its diagnostic performance to classify individuals according to sex after analyzing the nasal aperture in all the radiographs (combined sample). Binary classifications considered the decision cutoff of 0.5. Next, separate analyses were conducted to assess the diagnostic accuracy of the CNN by sex (correct classification of males and females) and by age group. In a subsequent experimental procedure, the CNN’s performance was exclusively evaluated for classifying individuals based on age (> 15 or ≤ 15 years). Multiclass task decisions considered the highest predicted probability. In this phase, the sample distribution was balanced with 4,403 individuals over 15 years and 4,946 individuals age 15 years younger. A subsequent age-based analysis was performed separately for males and females. Sample distribution was as follows: males over 15 years (n = 1,955), females over 15 years (n = 2,448), males aged 15 years or younger (n = 2,420), and females aged 15 years or younger (n = 2,526). It should be noted that this approach used the 15 years as the cut-off to distinguish younger and older individuals because the bony framework of the nose is estimated to grow more noticeably until around 15 years of age in males, and at an earlier age in females29. Moreover, this study acknowledges the significant limitations of sex estimation in subadult individuals and emphasizes that it should not be recommended in practice, particularly when using the nasal aperture as the evaluated parameter. Instead, the analysis of morphological features of viscerocranium and their differences between sexes across age groups was proposed and presented as an educational and exploratory approach, aiming to enhance basic anatomical understanding and supporting further research in the field.
To quantify the diagnostic accuracy of the nasal aperture in classifying individuals by sex and age, this study used metrics commonly employed to assess deep learning models in forensic computer-vision: accuracy, precision, recall, sensitivity and specificity. The outcomes were tabulated and visually presented using confusion matrices, Receiver Operating Characteristic (ROC) curves with their Area Under the Curve (AUC), and Gradient-weighted Class Activation Mapping (Grad-CAM). To account for variability across different datasets, mitigate overfitting, and ensure that performance metrics were not biased towards any specific part of the dataset, this study calculated the average of each metric across all five folds to obtain an overall measure of model performance. Computations were performed on a Linux machine running Ubuntu 22.04, equipped with an AMD Ryzen 9 7950X processor, 2 Nvidia™ RTX A5500 24 GB GPUs, and 128 GB of DDR5 RAM. All models were developed using TensorFlow API30 version 2.18. Python 3.8.10 was employed for algorithm implementation and data wrangling31.
Results
When sex was considered a binary outcome for the CNN’s performance in analyzing the nasal aperture on panoramic radiographs, accuracy, precision, recall, sensitivity, and specificity all reached approximately 74%. For the combined sample, accuracy rates ranged from 61% to 88%.
In males, accuracy rates ranged from 50% to 90%, while in females they ranged from 54% to 86% (Table 3). The correct classification rate was 73% for males and 75.17% for females (Fig. 2). The area under the ROC curve was 0.74 (Fig. 3).
When evaluating the CNN’s performance in classifying individuals as below or above 15 years, the correct classification rate was 83% for those older than 15 years and 89.5% for those aged 15 years or younger (Fig. 4). The correct classification rate for males older than 15 years was 80.5% compared to 76.5% for females. Among individuals aged 15 years or younger, the correct classification rate was 72.06% for males and 84.18% for females (Fig. 5).
Confusion matrix comparing real (chronological) and predicted age of individuals divided into groups > 15 years and ≤ 15 years separately for males (M) and females (F). The correct classification rates of females and males > 15 years were 76.5% and 80.5%, respectively, while for females and males ≤ 15 years the results were 84.18% and 72.6%, respectively.
Loss function analysis for sex, age and combined predictions demonstrated a pattern of relevant learning up to 50 epochs, with a plateau and subsequent divergence between training and validation curves (Fig. 6).
Grad-CAM analysis revealed stronger signals originating from the central mineralized tissue of the nasal aperture, including the nasal septum, as well as from the upper portion of the nasal aperture (Fig. 7).
Nasal aperture (A), its region of annotation on panoramic radiograph (B) and the Gradient-weighted Class Activation Mapping (Grad-CAM) showing in red (C) regions of higher activation signal of the convolutional neural network to identify potential sex-related patterns during the classification task.
Discussion
Among the various morphological features of the viscerocranium, the nasal aperture is of special interest due to its potential applications in anthropological assessments of population affinity, sex, and age32. To date, studies on this topic have been conducted using morphometric19,20,33,35 and morphoscopic analyses,21 through direct examination of dry skulls20,21,33 or via medical imaging19,35. The latter has been conducted using imaging modalities that are either 3D,19,34 such as computed tomography scans, or 2D,36 such as extraoral radiography.
The present study revisited the topic using artificial intelligence solutions—specifically, CNN-based computer vision. A distinctive feature of the current methodological design was the sample composition, which included subadults (children and adolescents) as well as young adults. It is important to highlight that sex assessment in individuals under 12 years of age is generally not recommended due to a lack of reliable methods36. The rationale for the sample is based on the understanding that, while a sample composed entirely of mature specimens would yield more accurate sex estimates, including immature individuals in this diagnostic accuracy experiment allows for investigation of how age may influence the expression of morphological features between males and females. By doing so, the reduced expression of sexually dimorphic features at a young age was confirmed. For example, when the CNN’s sex estimation performance is analyzed separately for individuals under 12 years, the mean accuracy rate becomes 67%, representing a 10% decrease compared to individuals aged 12 years and older (Table 3). This decrease was even more pronounced among males, with a 14% drop. Interestingly, the accuracy rate observed using the 12-year threshold remained consistent when the threshold was raised to 15 years. In other words, accuracy rates remain considerably low even when the bony framework of the nose is expected to be more stable—after the age of 15 years, at least—30 compared to earlier immature phases. This finding underscores the importance of considering age37 when planning sex assessment and supports a previous study21 that suggested a possible absence of sexual dimorphism in the nasal aperture.
Authors have demonstrated morphological variance of the nasal aperture between populations21. In the present study, radiographs were sampled from an existing image database of the Central-West region, which likely included Brazilians with diverse populational affinities. A previous radiographic study with a Brazilian sample analyzed 97 individuals and found highly significant differences between males and females, with males exhibiting greater height, width, and area of the nasal aperture35. Our findings may differ based on methodology, including the larger sample size in the present study, the use of different extraoral imaging modalities, and the application of deep learning and image pattern analysis rather than the morphometric assessment of height, width, and area used by the previous authors35.
Compared to other maxillofacial features used for sex assessment, the nasal aperture has demonstrated inferior diagnostic accuracy. A study, applying landmarking to computed tomography scans, demonstrated up to 95% of sex classification accuracy after the analysis of adult human mandibles through machine learning. High accuracy rates have also been observed for statistical models based on the combination of cranial measurements – including the nasal aperture38,39. This phenomenon can be justified firstly by sample characteristics, covering only the adult age range, where sexual dimorphism can be more pronounced. Secondly, by the comprehensive approach of the human skull integrating several anatomic features that may express sexual differences. The mandible is an example of a bone that undergoes modeling and remodeling influenced by the surrounding musculature. Strong muscles, such as the masseters, insert onto the mandible and generate traction vectors in different directions. These biomechanical forces contribute to morphological changes and introduce variables that can be closely associated with sexual dimorphism. Other examples extend also to the posterior region of the skull, such as the sternocleidomastoid muscle and its influence on modeling the mastoid process region, possibly leading to differences37 between males and females.
To our knowledge, among the available studies assessing the nasal aperture, the present work is the first to apply CNN within an AI-based computer vision framework and includes the largest sample to date. As a result, a robust deep learning model was trained, contributing valuable insights to the scientific literature. Given the unsatisfactory accuracy rates and the high risk of misclassification, this study does not recommend using the nasal aperture as a reliable sexually dimorphic feature, especially considering population-specific variations. Moreover, careful interpretation is warranted when considering the study’s methodological approach to testing the CNN’s classification of age, which divided the sample into groups below or above the 15-year threshold. At first glance, the observed correct classification rates may seem promising; however, this setup posed a relatively simple task for the CNN—distinguishing radiographs of children as young as 6 years from young adults up to 22.9 years old using 15 years as the cutoff. Therefore, moderate to high correct classification rates were expected but offer limited practical applicability. Given the suboptimal correct classification rates and the frequency of misclassifications, alternative methods should be preferred to the nasal aperture for age assessment. In this regard, assessing permanent tooth development using CNN has shown to be a useful approach7,10.
In addition to the presented limitations, the type of imaging modality used in this study should also be considered. Panoramic radiographs, while common in extraoral imaging and often available in large datasets, capture the nasal aperture with the upper region typically positioned near the superior edge of the image. This positioning can cause the outline to appear smoothed or less defined. Moreover, the rotational acquisition technique employed in panoramic radiography provides a broad view of the maxillofacial structures but can also introduce40. To mitigate these limitations, the images were collected from an oral radiology center that follows standardized protocols. Additionally, pure morphometric analyses based on linear measurements—which can be biased in panoramic radiographs—were avoided by employing a computer vision approach that assesses image patterns. Future radiological studies should consider imaging modalities with fewer distortions, such as posteroanterior Caldwell radiography35,41 or Cone Beam Computed Tomography (CBCT). The latter allows for realistic assessment of the viscerocranium and dentomaxillofacial structures, along with 3D navigation39.
Another input for future studies is increasing the sample size and exploring not only a single CNN, but also alternative models. This is especially relevant because the training curves for all tasks showed a progressive reduction in training loss across epochs, indicating that the models were learning patterns from the radiographic data. However, the validation loss exhibited a distinct plateau followed by mild to moderate divergence from the training curve. This reflects a certain degree of overfitting, meaning that while the models continued to improve on the training data, their performance on unseen validation data stabilized or even worsened slightly around fifty epochs. Such behavior can be expected in deep learning applications with relatively limited sample sizes, where model capacity can exceed the available variability in the dataset. This finding corroborates the need for larger and alternative (focusing on sex estimation of adults, for instance) datasets to support stronger generalization, and testing additional architectures, regularization strategies, or hyperparameter adjustments in future research. Despite this, the models were still able to capture meaningful trends and provide results that support the feasibility of computer-aided approaches in forensic odontology, while making clear that expert confirmation remains indispensable.
In addition to these methodological perspectives, future research should also incorporate external validation to confirm the generalizability of our findings. External validation using independent datasets from different clinical centers is therefore important to test whether the patterns detected in the present study persist across populations, imaging protocols, and equipment. Such validation would not only demonstrate reproducibility and address potential biases linked to local features, but also clarify the extent to which computer-aided analysis of the nasal aperture can contribute to forensic applications, reinforcing that this structure alone is not a robust tool for definitive sex or age assessment.
Conclusion
Under the present methodological conditions, the nasal aperture demonstrated limited discriminative power for sex classification, with performance metrics indicating an accuracy rate equivalent to one misclassification per four cases. A secondary, confirmatory finding was the notable drop in accuracy among younger individuals compared to older (mature) ones.
Therefore, the diagnostic accuracy metrics of the nasal aperture assessed from extraoral radiographs can be considered unsatisfactory to support its use as the sole anatomical feature for radiographic sex assessment via CNN.
Data availability
The data supporting this study’s findings are available from the project supervisor, Prof. Ademir Franco, upon reasonable request and with permission from the Center of Oral Radiology and Imaging.
References
Adserias-Garriga, J., Thomas, C., Ubelaker, D. H. & Zapico, S. C. When forensic odontology Met biochemistry: multidisciplinary approach in forensic human identification. Arch. Oral Biol. 87, 7–14. https://doi.org/10.1016/j.archoralbio.2017.12.001 (2018).
Cunha, E. Aging the death: the importance of having better methods for age at death Estimation of old individuals. Ann. Med. 53 (sup1), S1. https://doi.org/10.1080/07853890.2021.1893522 (2021).
Thurzo, A. et al. Use of advanced artificial intelligence in forensic medicine, forensic anthropology and clinical anatomy. Healthc. (Basel). 9 (11), 1545. https://doi.org/10.3390/healthcare9111545 (2021).
Lo, M., Mariconti, E., Nakhaeizadeh, S. & Morgan, R. M. Preparing computed tomography images for machine learning in forensic and virtual anthropology. Forensic Sci. Int. Synergy. 6, 100319. https://doi.org/10.1016/j.fsisyn.2023.100319 (2023).
Lye, R. et al. Deep learning versus human assessors: forensic sex Estimation from three-dimensional computed tomography scans. Sci. Rep. 14 (1), 30136. https://doi.org/10.1038/s41598-024-81718-y (2024).
Franco, A. et al. Diagnostic performance of convolutional neural networks for dental sexual dimorphism. Sci. Rep. 12 (1), 17279. https://doi.org/10.1038/s41598-022-21294-1 (2022).
Franco, A. et al. Binary decisions of artificial intelligence to classify third molar development around the legal age thresholds of 14, 16 and 18 years. Sci. Rep. 14 (1), 4668. https://doi.org/10.1038/s41598-024-55497-5 (2024).
Zhang, M. Forensic imaging: a powerful tool in modern forensic investigation. Forensic Sci. Res. 7 (3), 385–392. https://doi.org/10.1080/20961790.2021.2008705 (2022).
Ghosh, A., Sufian, A., Sultana, F., Chakrabarti, A. & De, D. Fundamental concepts of convolutional neural network. In: Balas VE, Kumar R, Srivastava R. Recent trends and advances in artificial intelligence and internet of things. Intelligent Systems Reference Library. Switzerland: Springer; (2020). https://doi.org/10.1007/978-3-030-32644-9 pp. 519–567.
Murray, J. et al. Applying artificial intelligence to determination of legal age of majority from radiographic data. Morphologie 108 (360), 100723. https://doi.org/10.1016/j.morpho.2023.100723 (2024).
Nonthasaen, P., Mahikul, W., Chobpenthai, T. & Achararit, P. Sex Estimation from Thai hand radiographs using convolutional neural networks. Forensic Sci. Int. Rep. 8, 100332. https://doi.org/10.1016/j.fsir.2023.100332 (2023).
Khazaei, M., Mollabashi, V., Khotanlou, H. & Farhadian, M. Sex determination from lateral cephalometric radiographs using an automated deep learning convolutional neural network. Imaging Sci. Dent. 52 (3), 239–244. https://doi.org/10.5624/isd.20220016 (2022).
Bu, W. et al. Automatic sex Estimation using deep convolutional neural network based on orthopantomogram images. Forensic Sci. Int. 348, 111704. https://doi.org/10.1016/j.forsciint.2023.111704 (2023).
Franco, A. et al. Radiographic morphology of canines tested for sexual dimorphism via convolutional-neural-network-based artificial intelligence. Morphologie 108 (362), 100772. https://doi.org/10.1016/j.morpho.2024.100772 (2024).
Ulubaba, H. E., Atik, İ., Çiftçi, R., Eken, Ö. & Aldhahi, M. I. Deep learning for gender Estimation using hand radiographs: a comparative evaluation of CNN models. BMC Med. Imaging. 25 (1), 260. https://doi.org/10.1186/s12880-025-01809-8 (2025).
Plemons, A. & Hefner, J. T. Ancestry Estimation using macromorphoscopic traits. Acad. Forensic Pathol. 6 (3), 400–412. https://doi.org/10.23907/2016.041 (2016).
Hefner, J. T. & Ousley, S. D. Statistical classification methods for estimating ancestry using morphoscopic traits. J. Forensic Sci. 59, 883–890. https://doi.org/10.1111/1556-4029.12421 (2014).
Mbonani, T., L’Abbé, E., Chen, D. G. & Ridel, A. Population affinity Estimation in forensic anthropology: a South African perspective. Int. J. Legal Med. https://doi.org/10.1007/s00414-025-03529-8 (2025).
Scendoni, R. et al. Anthropometric analysis of orbital and nasal parameters for sexual dimorphism: new anatomical evidences in the field of personal identification through a retrospective observational study. PLoS One. 18 (5), e0284219. https://doi.org/10.1371/journal.pone.0284219 (2023).
López, M. C., Galdames, I. C. S., Matamala, D. A. Z. & Smith, R. L. Sexual dimorphism determination by piriform aperture morphometric analysis in Brazilian human skulls. Int. J. Morphol. 27 (2), 327–331. https://doi.org/10.4067/s0717-95022009000200007 (2009).
McDowell, J. L., L’Abbé, E. N. & Kenyhercz, M. W. Nasal aperture shape evaluation between black and white South Africans. Forensic Sci. Int. 222 (1–3). https://doi.org/10.1016/j.forsciint.2012.06.007 (2012). 397.e1–6.
Bossuyt, P. M. et al. STARD 2015: an updated list of essential items for reporting diagnostic accuracy studies. Radiology 277 (3), 826–832. https://doi.org/10.1148/radiol.2015151516 (2015).
Uribe, S. E. et al. Evaluating dental AI research papers: key considerations for editors and reviewers. J. Dent. https://doi.org/10.1016/j.jdent.2025.105867 (2025).
Jocher, G. & Qiu, J. Ultralytics YOLO11 [software]. Version 11.0.0. (2024). Available from: https://github.com/ultralytics/ultralytics
Wang, C-Y., Bochkovskiy, A. & Liao, H-Y-M. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In: 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR); 2023; Vancouver, Canada. pp. 7464–75. https://doi.org/10.1109/CVPR52729.2023.00721
Tariq, M. & Choi, K. YOLO11-driven deep learning approach for enhanced detection and visualization of wrist fractures in X-ray images. Mathematics 13 (9), 1419. https://doi.org/10.3390/math13091419 (2025).
Kohavi, R. A study of cross-validation and bootstrap for accuracy estimation and model selection. In: Proceedings of the 14th International Joint Conference on Artificial Intelligence (IJCAI); pp. 1137–45. (1995).
Hastie, T., Tibshirani, R. & Friedman, J. The Elements of Statistical Learning: Data mining, Inference and Prediction (Springer, 2009).
Akgüner, M., Barutçu, A. & Karaca, C. Adolescent growth patterns of the bony and cartilaginous framework of the nose: a cephalometric study. Ann. Plast. Surg. 41 (1), 66–69. https://doi.org/10.1097/00000637-199807000-00012 (1998).
Abadi, M. et al. TensorFlow: Large-scale machine learning on heterogeneous systems; 2015. Software available from https://tensorflow.org.
Van Rossum, G. & Drake, F. L. Python 3 Reference Manual (CreateSpace, 2009).
Ominde, B. S., Ikubor, J. E., Jaiyeoba-Ojigho, J. E., Omoro, O. F. & Igbigbi, P. S. Morphometric assessment of the piriform aperture and its clinical and forensic applications. Mustansiriya Med. J. 23 (2), 78–83. https://doi.org/10.4103/mj.mj_67_23 (2024).
Asghar, A., Dixit, A. & Rani, M. Morphometric study of nasal bone and piriform aperture in human dry skull of Indian origin. J. Clin. Diagn. Res. 10 (1), AC05–7. https://doi.org/10.7860/JCDR/2016/15677.7148 (2016).
Yüzbaşioğlu, N., Yilmaz, M. T., Çicekcibasi, A. E., Şeker, M. & Sakarya, M. E. The evaluation of morphometry of nasal bone and pyriform aperture using multidetector computed tomography. J. Craniofac. Surg. 25 (6), 2214–2219. https://doi.org/10.1097/SCS.0000000000001063 (2014).
Prado, F. B. et al. Piriform aperture morfometry and nasal bones morphology in Brazilian population by postero-anterior caldwell radiographys. Int. J. Morphol. 29 (2), 393–398. https://doi.org/10.4067/S0717-95022011000200014 (2011).
Scientific Working Group for Forensic Anthropology (SWGANTH). Sex assessment. (2010). Available at: https://www.nist.gov/system/files/documents/2018/03/13/swganth_sex_assessment.pdf
Deitos, A. R. & Cunha, E. Estimativa do Sexo Em antropologia forense. In: (eds Machado, C. E. P., Deitos, A. R., Velho, J. A. & Cunha, E.) Tratado De Antropologia Forense. Campinas, SP: Millennium; p. 343-360. (2022). [Portuguese].
Toneva, D. H. et al. Data mining for sex Estimation based on cranial measurements. Forensic Sci. Int. 315, 110441. https://doi.org/10.1016/j.forsciint.2020.110441 (2020).
Toneva, D. et al. Sex Estimation based on mandibular measurements. Anthropol. Anz. 81 (1), 19–42. https://doi.org/10.1127/anthranz/2023/1733 (2024).
Farman, A. G. & Scarfe, W. C. The basics of maxillofacial cone beam computed tomography. Semin Orthod. 15 (1), 2–13. https://doi.org/10.1053/j.sodo.2008.09.001 (2009).
Camargo, J. R. et al. The frontal sinus morphology in radiographs of Brazilian subjects: its forensic importance. Braz J. Morphol. Sci. 24, 1–5 (2007).
Acknowledgements
None.
Funding
This study was financed in part by the Coordination for the Improvement of Higher Education Personnel – Brazil (CAPES) – Finance Code 001. This study was financed in part by the National Council for Scientific and Technological Development (CNPq).
Author information
Authors and Affiliations
Contributions
Leonardo SCAVASSINI ( [leonardo.scavassini@hotmail.com](mailto: leonardo.scavassini@hotmail.com) ): made substantial contributions to the conception of the work; drafted the work; approved the version to be published; agreed to be accountable for all aspects of the work; Rianne SILVA ( [riannekeith@yahoo.com.br](mailto: riannekeith@yahoo.com) ), Amber KHAN ( [amber1993khan@gmail.com](mailto: amber1993khan@gmail.com) ), Wahaj ANEES ( [wahajanees@live.com](mailto: wahajanees@live.com) ), and Jared MURRAY ( [forensicodont@protonmail.com](mailto: forensicodont@protonmail.com) ): made substantial contributions to data acquisition and design of the work; drafted the work; approved the version to be published; agreed to be accountable for all aspects of the work; Nikolaos ANGELAKOPOULOS (nikolaos.angelakopoulos@unibe.ch) and Marianna ARAKELYAN ( [mariastom87@inbox.ru](mailto: mariastom87@inbox.ru) ): made substantial contributions to the design of the work; revised it critically for important intellectual content; approved the version to be published; agreed to be accountable for all aspects of the work; Lucas PORTO ( [lporto@gmail.com](mailto: lporto@gmail.com) ) and André ABADE( [andreabade@gmail.com](mailto: andreabade@gmail.com) ): made substantial contributions to the conception, design, data analysis and software used in the work; revised it critically for important intellectual content; approved the version to be published; agreed to be accountable for all aspects of the work; Ademir FRANCO ( [franco.gat@gmail.com](mailto: franco.gat@gmail.com) ): made substantial contributions to the conception, design and data interpretation in the work; drafted and revised it critically for important intellectual content; approved the version to be published; agreed to be accountable for all aspects of the work.
Corresponding author
Ethics declarations
Compliance with ethical standards
The study was carried out following the ethical standards of the Declaration of Helsinki (Finland).
Competing interests
The authors declare no competing interests.
Competing interests
The authors declare that they have no competing interests.
Ethical approval
The outlined investigation protocol received approval from the Institutional Committee of Ethics in Human Research at the Faculdade São Leopoldo Mandic (Protocol No. 76809023.9.0000.5374).
Informed consent
The images utilized in this study constituted secondary data sourced from an established radiology database (Center of Oral Radiology and Imaging). Access to the data was authorized through informed permission granted by the database’s legal custodian, and subsequently approved by the relevant ethics committee.
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary Information
Below is the link to the electronic supplementary material.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Scavassini, L., Silva, R., Khan, A. et al. YOLO11m-cls applied to sex and age classification based on the radiographic analysis of the nasal aperture. Sci Rep 15, 40784 (2025). https://doi.org/10.1038/s41598-025-24593-5
Received:
Accepted:
Published:
Version of record:
DOI: https://doi.org/10.1038/s41598-025-24593-5









