Clin Oral Investig. 2025 Jan 31;29(2):101. doi: 10.1007/s00784-025-06156-0.
ABSTRACT
OBJECTIVE: This study aimed to apply the DeepLabv3 + model and compare it with the U-Net model in terms of detecting and segmenting apical lesions on panoramic radiography.
METHODS: 260 panoramic images that contain apical lesions in different regions were collected and randomly divided into training and test datasets. All images were manually annotated for apical lesions using Computer Vision Annotation Tool software by two independent dental radiologists and a master reviewer. The DeepLabv3 + model, one of the state-of-the-art deep semantic segmentation models, was utilized using Python programming language and the TensorFlow library and applied to the prepared datasets. The model was compared with the U-Net model applied to apical lesions and other medical image segmentation problems in the literature.
RESULTS: The DeepLabv3 + and U-Net models were applied to the same datasets with the same hyper-parameters. The AUC and recall results of the DeepLabv3 + were 29.96% and 61.06% better than the U-Net model. However, the U-Net model gets 69.17% and 25.55% better precision and F1-score results than the DeepLabv3 + model. The difference in the IoU results of the models was not statistically significant.
CONCLUSIONS: This paper comprehensively evaluated the DeepLabv3 + model and compared it with the U-Net model. Our experimental findings indicated that DeepLabv3 + outperforms the U-Net model by a substantial margin for both AUC and recall metrics. According to those results, for detecting apical lesions, we encourage researchers to use and improve the DeepLabv3 + model.
CLINICAL RELEVANCE: The DeepLabv3 + model has the poten tial to improve clinical diagnosis and treatment planning and save time in the clinic.
PMID:39888441 | DOI:10.1007/s00784-025-06156-0