Tehran University of Medical Sciences

Science Communicator Platform

Stay connected! Follow us on X network (Twitter):
Share this content! On (X network) By
Artificial Intelligence Based Vision Transformer Application for Grading Histopathological Images of Oral Epithelial Dysplasia: A Step Towards Ai-Driven Diagnosis Publisher Pubmed



Hadilou M1 ; Mahdavi N2 ; Keykha E3 ; Ghofrani A4 ; Tahmasebi E3 ; Arabfard M5
Authors
Show Affiliations
Authors Affiliations
  1. 1. Research Center for Prevention of Oral and Dental Diseases, Baqiyatallah University of Medical Sciences, Tehran, Iran
  2. 2. Department of Oral and Maxillofacial Pathology, School of Dentistry, Tehran University of Medical Sciences, Tehran, Iran
  3. 3. Research Center for Prevention of Oral and Dental Diseases, School of Dentistry, Baqiyatallah University of Medical Sciences, Tehran, Iran
  4. 4. Student Research Committee, Baqiyatallah University of Medical Sciences, Tehran, Iran
  5. 5. Chemical Injuries Research Center, Systems Biology and Poisonings Institute, Baqiyatallah University of Medical Sciences, Tehran, Iran

Source: BMC Cancer Published:2025


Abstract

Background: This study aimed to classify dysplastic and healthy oral epithelial histopathological images, according to WHO and binary grading systems, using the Vision Transformer (ViT) deep learning algorithm—a state-of-the-art Artificial Intelligence (AI) approach and compare it with established Convolutional Neural Network models (VGG16 and ConvNet). Methods: A total of 218 histopathological slide images were collected from the Department of Oral and Maxillofacial Pathology at Tehran University of Medical Sciences archive and combined with two online databases. Two oral pathologists independently labeled the images based on the 2022 World Health Organization (WHO) grading system (mild, moderate and severe), the binary grading system (low risk and high risk), including an additional normal tissue class. After preprocessing, the images were fed to the ViT, VGG16 and ConvNet models. Results: Image preprocessing yielded 2,545 low-risk, 2,054 high-risk, 726 mild, 831 moderate, 449 severe, and 937 normal tissue patches. The proposed ViT model outperformed both CNNs with the accuracy of 94% (VGG16:86% and ConvNet: 88%) in 3-class scenario and 97% (VGG16:79% and ConvNet: 88%) in 4-class scenario. Conclusions: The ViT model successfully classified oral epithelial dysplastic tissues with a high accuracy, paving the way for AI to serve as an adjunct or independent tool alongside oral and maxillofacial pathologists for detecting and grading oral epithelial dysplasia. © The Author(s) 2025.