Background & Summary

Dental diseases are a significant public health concern worldwide. According to the Global Oral Health Status Report1, oral diseases impacted 3.5 billion people in 2019. This is 1 billion more than the total number of cases of the five major non-communicable diseases (NCDs), which include cardiovascular disease, diabetes mellitus, chronic respiratory diseases, mental disorders and cancers2. Untreated dental diseases could lead to several health complications such as tooth loss, pain and discomfort, which could also ultimately affect the quality of life3.

Early detection of dental diseases is crucial for effective management4. Thus, the clinical examination of the patient will be followed by appropriate investigations in order to arrive at a diagnosis of dental diseases 5.

Dental imaging is one such type of investigation that would facilitate the diagnosis, enhancing the accuracy of the findings achieved during clinical examination. These diagnostic imaging includes intraoral or extraoral plain x-rays and three-dimensional imaging modalities such as cone beam computed tomography (CBCT)5. Among intraoral radiographs, bitewing radiographs, particularly vertical bitewings, are often used to assess crest bone levels, as they provide an accurate depiction of the distance between the Cemento-Enamel junction (CEJ) and the alveolar crest, especially in cases with shallow periodontal pockets ( < 5 mm). However, in clinical situations involving deeper or irregular pocketing ( > 5 mm), bitewing radiographs do not adequately capture the full extent of bone loss. In such cases, periapical radiographs are considered more suitable, as they offer a greater vertical field of view. Intra-oral periapical (IOPA) radiographs, are essential for diagnosing a range of dental conditions, including assessing tooth structure and periodontal health, detecting periapical infections and bone abnormalities, monitoring endodontic status and identifying unerupted or impacted teeth6. IOPA radiographs are relatively simple and less expensive compared to other diagnostic imaging methods such as dental panoramic tomogram (DPT), making them a cost-effective option for dental check-ups7. Moreover, IOPA radiographs provide high-resolution images of the teeth and surrounding structures.

Artificial intelligence (AI) and computer vision-based systems capable of analysing the radiographs could be of significant assistance to dentists by reducing errors while enhancing the efficiency and accuracy of diagnoses8,9,10. In some cases, AI shows promising potential in diagnosing dental-related issues with reliability, transparency, and precision that could aid dentists in specific scenarios11. The use of computer vision and convolutional neural networks in dental research has been continuously growing since the first published application in 201512. To obtain these benefits, having a large and diverse dataset of radiographs is crucial. Although datasets of dental radiographs are available, most of these datasets contain panoramic radiographs with teeth segmentation only. Table 1 provides an overview of these datasets. This new data set includes IOPA radiographs with annotations of important landmarks along with tooth segmentation. With a sufficiently robust dataset, both machine learning and deep learning models could be developed and trained to automatically detect dental diseases, allowing early diagnosis and improved treatment outcomes.

Table 1 Overview of available dental radiography datasets by type: dental panoramic radiograph, periapical radiograph.

Our dataset, DenPAR, consists of 1000 IOPA radiographs with detailed annotations, including tooth segmentation masks, Cemento-Enamel Junction (CEJ) points, apex points, and alveolar crestal bone levels, all annotated using the LabelBox annotation tool13. Sample images and annotations are shown in Figs. 1 and 2. Each image is also accompanied by metadata on age and sex.

Fig. 1
figure 1

Sample images from the dataset, following anonymization, without annotations. (a) lower posterior, (b) upper posterior, (c) lower anterior, and (d) upper anterior.

Fig. 2
figure 2

Sample annotations. Cemento-Enamel junctions (CEJ) points (red) are highlighted with red circles, apex points (yellow) are highlighted with yellow circles, alveolar bone-level lines with green lines in gray-highlighted areas, and teeth are shown as blue masks.

This comprehensive annotated dataset of IOPA radiographs together with the relevant metadata can be used for advancements in automated technology in analysis of dental radiographs, which could improve the efficiency of dental diagnosis. The dataset and annotations can be utilized for training various models, including teeth segmentation models, multi-class or single-class tooth segmentation models, keypoint detection (CEJ and apex) models, and alveolar crestal bone level identification models. These can be further developed for advanced clinical decision support systems, such as alveolar bone loss detection using IOPA radiographs. Overall, this dataset holds significant potential for reuse and has the capability to make a meaningful impact on clinical practice, research, and technological advancements in dentistry.

Methods

The data set comprises 1000 IOPA radiographs acquired for routine dental treatments and diagnosis in the Radiology Unit of the Dental Hospital Peradeniya, Sri Lanka during the period from January 2020 to January 2023.

The sample includes radiographs of 440 male and 560 female patients, with an age range of 9 to 81 years. The age range for male patients is between 9 and 81 years, while in female patients it ranges from 12 to 79 years. The age and sex distribution of the patients is shown in Fig. 3, and the majority of the male and female patients were in the age range of 31-50 years.

Fig. 3
figure 3

Patient distribution by age and sex.

Out of the 1000 radiographs, 590 are from the lower arch and 410 are from the upper arch. In terms of anatomical orientation, 398 radiographs are from the right side, 380 from the left side, and 222 from the anterior region.

Fig. 4
figure 4

Overview of dataset structure. Contents are shown for the training folder; validation and testing folders follow the same structure.

Data Privacy

Ethical approval for this study was granted by the Ethics Review Committee of the Faculty of Dental Sciences, University of Peradeniya, Sri Lanka, under the Ethics clearance certificate containing the number ERC/FDS/UOP/2023/45. As part of the hospital’s standard registration process, patients provide general informed consent (blanket) that allows the use of their clinical records, including images, for research and educational purposes, provided that personal identifiers are removed. In this study, all data were fully de-identified prior to the analysis, in order to protect patient privacy. The use of this de-identified data for this research was also approved by the institution’s ethics committee. The original images had several details labeled, such as the patient’s name, age, and sex. Two steps were taken to anonymize the images. First, all images were renamed using random integer numbers to remove identifiable information. Second, an algorithm was employed to eliminate all personal details embedded within the images. Figure 1 displays examples of images following the anonymization process.

Data annotation

The annotation software used to create the dataset is Labelbox13. The annotations included CEJ points, apex points, alveolar crestal bone levels, and individual teeth. The CEJ was identified as the end point of the enamel, and it was marked on the proximal surfaces (mesial and distal) of each tooth. The tip of the root, as seen on the radiograph, was marked as the root apex. In multi-rooted teeth, the apex was marked for each root. The alveolar crestal bone levels were annotated based on the highest point of the crestal bone. All clearly identifiable CEJ points, apex points, and alveolar crestal bone levels in any given radiograph were annotated.

The annotations were performed collaboratively by two experienced specialists (MD in Restorative Dentistry), each with significant expertise in radiographic interpretation related to periodontal bone loss. Both experts followed standardized diagnostic criteria and reached consensus on each case to minimize subjectivity and bias in the annotation process. Figure 2 illustrates a sample of an image following Labelbox annotations.

Following annotation, all data were blindly reviewed and approved by two experienced specialists (MD in Oral Surgery, and MD in Restorative Dentistry) from the Department of Oral Medicine and Periodontology, Faculty of Dental Sciences, University of Peradeniya, Sri Lanka, both of whom have substantial expertise in dental radiographic interpretation.

Data Records

The compressed dataset is available on Zenodo (https://doi.org/10.5281/zenodo.16645076)14 under Creative Commons Attribution 4.0 license. The uncompressed folder consists of three directories: Training, Validation, and Testing, which can be used as official splits for training, validation, and testing machine learning models. Additionally, the folder includes one metadata spreadsheet file named “characteristics of radiographs included”.

Each directory contains five sub-directories, namely, Images, Masks (Radiograph-wise), Masks (Tooth-wise), Bone Level Annotations, and Key Points Annotations. The metadata spreadsheet file provides detailed information about the arch type, anatomical orientation, and FDI notation of each tooth in every radiograph.

Images and Masks (Radiograph-wise) subdirectories contain corresponding IOPA radiographs in .JPG format and teeth mask images for radiographs in .PNG format, respectively.

Key Points Annotations subdirectory contains key point annotation JSON files. Figure 5, illustrates the visualization of information from a key point annotation JSON file. In this visualization, the CEJ key points are highlighted with red circles, and the apex points are highlighted with blue circles. Each tooth is enclosed within an individual rectangular bounding box derived from its corresponding tooth mask. The bounding box coordinates are represented as [x-minimum, y-minimum, x-maximum, y-maximum], where x-minimum and y-minimum correspond to the top-left corner of the box, and x-maximum and y-maximum correspond to the bottom-right corner of the bounding box. These bounding boxes are included to meet the input requirements of certain keypoint detection models, which utilize bounding box constraints to enhance the localization and precision of keypoint detection.

Fig. 5
figure 5

(a) The original radiograph image, (b) representation of information in keypoint annotation JSON file. CEJ points (red) are highlighted with red circles, apex points (blue) with blue circles, and bounding boxes (yellow) enclose each tooth containing keypoint information.

Bone Level Annotations subdirectory contains alveolar crestal bone-level annotation JSON files along with a Common Objects in Context (COCO) format annotation file. In the alveolar crestal bone-level annotation file, the alveolar crestal bone levels are represented as separate X and Y coordinates for each alveolar crestal bone level because, in 2D images, a bone level is a line, which is made up of multiple points. Radiograph-wise images and .JSON files are renamed with the corresponding name of the IOPA radiographs.

Masks (Tooth-wise) subdirectory contains sub-folders with the name of the corresponding IOPA radiographs and a COCO format annotation file. Each subfolder includes mask images of individual teeth in .PNG format.

The overview of the dataset structure is presented in Fig. 4. The figure illustrates the content structure of the “Training” folder, while the “Validation” and “Testing” folders follow the same structure.

Radiograph-wise masks help in tasks such as identifying all teeth in the radiograph, while tooth-wise masks help in tasks such as identifying each individual tooth within the radiograph. The COCO format annotation files can be used directly or with the pre-processing, depending on the requirements of machine learning models. This format is highly versatile and widely used because it allows for the creation of customized annotation files, making it adaptable to different computer vision tasks.

Figure 6 presents the distribution of image resolutions across the training, validation, and test sets. The average resolutions of the training, validation, and test sets are 1059 × 960 pixels, 1084 × 944 pixels, and 1066 × 953 pixels, respectively. For the training dataset, the resolutions range from a minimum of 549 × 717 pixels to a maximum of 1542 × 1537 pixels. For the validation dataset, the resolutions range from 627 × 698 pixels to 1366 × 1533 pixels. For the test dataset, the resolutions range from 567 × 685 pixels to 1542 × 1370 pixels.

Fig. 6
figure 6

Distribution of image resolutions across training, validation, and test sets.

Technical Validation

The dataset was divided into 65%, 15%, and 20% for training, validation, and testing, respectively, based on statistical analysis to ensure an even distribution of images, facilitating effective use with machine learning models. Since the primary goal of this dataset is alveolar bone loss detection, each of the following attributes was considered equally important for ensuring a fair division:

i) Number of teeth in each radiograph: This refers to the total number of teeth visible in each radiograph image. ii) Radiograph arch (mandibular or maxillary): Each radiograph represents either the mandibular or maxillary arch. Ensuring a fair division of images from both arches is important for balanced dataset representation. iii) Number of alveolar crestal bone levels marked in a radiograph: These levels, marked by dentists, play a central role in the dataset’s purpose. These levels help identify alveolar bone loss. iv) Type of roots: Radiographs may include images of only single-rooted teeth, only multi-rooted teeth, or a combination of both. The dataset division ensured a fair representation of all these categories. v) Number of apex points in each radiograph: This refers to the total number of apex points visible in each radiograph image. Similarly, vi) Number of CEJ points in each radiograph: This refers to the total number of CEJ points visible in each radiograph image.

The dataset division, as shown in Fig. 7, ensures a fair distribution across six attributes. This balance is important for developing effective machine learning models, enabling them to train, validate, and test properly across all attributes.

Fig. 7
figure 7

Distribution of Six Attributes in Training, Validation, and Test Sets: Number of Teeth, Root Structure Classification, Number of CEJ Points, Number of APEX Points, Arch Type (Mandibular/Maxillary), and Alveolar Crestal Bone Level Count.

The YOLOv8 pose15 model was used to experiment with keypoint detection. YOLOv8 pose is a variant of YOLOv816 designed for detecting key points, such as joints in human pose estimation and other object keypoint tasks. YOLOv816 simultaneously detects key points associated with objects while localizing the objects using rectangular bounding boxes.

In our study, the key points of interest are the CEJ and the apex, while the objects are the individual teeth associated with these key points. Table 2 shows experimental results of detecting CEJ and apex key points separately, as well as the bounding boxes for their corresponding tooth objects. The output results for CEJ and apex key points are presented in Figs. 8 and 9, respectively, along with their ground truths. Ground truth key points and tooth bounding boxes are shown as red points and yellow boxes, while predicted key points and tooth bounding boxes are represented as blue points and red boxes.

Table 2 Experimental results with the YOLOv8 pose key-point detection model.
Fig. 8
figure 8

Two different sets of radiograph images, shown from left to right: original image, ground truth CEJ points and teeth bounding boxes overlaid on original image, YOLOv8 pose output overlaid on original image.

Fig. 9
figure 9

Two different sets of radiograph images, shown from left to right: original image, ground truth apex points and teeth bounding boxes overlaid on original image, YOLOv8 pose output overlaid on original image.

This dataset is suitable for both teeth semantic segmentation and tooth instance segmentation. For semantic segmentation, masks from the ‘Masks (Radiograph-wise)’ folder were used. U-Net17 and U-Net with a ResNet34 encoder, leveraging the residual learning framework introduced by18, were employed in the experiments. Table 3 presents the experimental results of teeth semantic segmentation with our dataset, along with ground truth and output examples shown in Fig. 10.

Table 3 Experimental results with semantic segmentation models for teeth segmentation.
Fig. 10
figure 10

Two different sets of radiograph images, shown from left to right: original image, ground truth mask, U-Net model predicted mask, and U-Net with ResNet34 encoder predicted mask.

For instance segmentation, the masks in the ‘Masks (Tooth-wise)’ folder and the COCO format annotation file included in this folder were used. YOLOv519 and Mask R-CNN20 were used in the experiments. Table 4 presents the experimental results of tooth instance segmentation with our dataset, along with ground truth and output results shown in Fig. 11.

Table 4 Experimental results with instance segmentation models for tooth segmentation.
Fig. 11
figure 11

Two different sets of radiograph images, shown from left to right: original image, ground truth mask with original image, YOLOv5 output with original image, and Mask R-CNN output with original image.

For alveolar crestal bone level detection, bone levels were first converted into masks with a specified thickness of 3 pixels. UNet++21 and MultiResUNet22 were then used in the experiments. Table 5 presents the experimental results of alveolar bone level detection using our dataset, with ground truth shown as green lines and output results as red lines in Fig. 12.

Table 5 Experimental results with bone level detection.
Fig. 12
figure 12

Two different sets of radiograph images, shown from left to right: original image, ground truth bone level mask with original image, UNet++ output with original image, and MultiResUNet output with original image.

Usage Notes

The researchers can download this dataset from zenodo14.

For those using machine learning, this dataset can help build and test models that detect and measure bone loss. For example, researchers could use TensorFlow or PyTorch to create models that identify areas of bone loss with respect to each individual tooth, measure how much bone is missing, or predict bone health status based on image patterns. They can also apply models trained on this data to other, similar datasets, helping make the models more reliable.

Finally, combining this dataset with other public data could help researchers understand how bone loss patterns vary in different groups. We encourage researchers to share any code, models, or tools they develop, as this could help others in the dental field improve the detection and treatment of alveolar bone loss.