Logo
Middle East J Rehabil Health Stud

Image Credit:Middle East J Rehabil Health Stud

Diagnostic and Prognostic Performance of Artificial Intelligence Models in Detecting Odontogenic Keratocysts from Histopathologic Images: A Systematic Review and Meta-Analysis

Author(s):
Reyhaneh ShoorgashtiReyhaneh ShoorgashtiReyhaneh Shoorgashti ORCID1,*, Farnaz JafariFarnaz JafariFarnaz Jafari ORCID2, Asma YazdanfarAsma Yazdanfar1, Sarah Sadat EhsaniSarah Sadat EhsaniSarah Sadat Ehsani ORCID3, Sana BaghizadehSana BaghizadehSana Baghizadeh ORCID3, Mohadeseh AlimohammadiMohadeseh AlimohammadiMohadeseh Alimohammadi ORCID3, Simin LesanSimin LesanSimin Lesan ORCID1, Donia SadriDonia SadriDonia Sadri ORCID4
1Department of Oral and Maxillofacial Medicine, School of Dentistry, Islamic Azad University of Medical Sciences, Tehran, Iran
2Oral and Dental Diseases Research Center, Kerman University of Medical Sciences, Kerman, Iran
3Faculty of Dentistry, Tehran Medical Sciences, Islamic Azad University, Tehran, Iran
4Department of Oral and Maxillofacial Pathology, School of Dentistry, Tehran Medical Sciences, Islamic Azad University, Tehran, Iran


Middle East Journal of Rehabilitation and Health Studies:Vol. 12, issue 3; e158082
Published online:Apr 29, 2025
Article type:Systematic Review
Received:Nov 17, 2024
Accepted:Apr 21, 2025
How to Cite:Reyhaneh ShoorgashtiFarnaz JafariAsma YazdanfarSarah Sadat EhsaniSana BaghizadehMohadeseh AlimohammadiSimin LesanDonia Sadriet al.Diagnostic and Prognostic Performance of Artificial Intelligence Models in Detecting Odontogenic Keratocysts from Histopathologic Images: A Systematic Review and Meta-Analysis.Middle East J Rehabil Health Stud.2025;12(3):e158082.https://doi.org/10.5812/mejrh-158082.

Abstract

Context:

Odontogenic keratocysts (OKCs) are aggressive jaw cysts characterized by a high recurrence rate, making accurate diagnosis critical for effective treatment. Recent advances in artificial intelligence (AI) have demonstrated potential for enhancing diagnostic accuracy in histopathology. However, the effectiveness of AI in diagnosing OKCs has not yet been systematically reviewed.

Objectives:

This study aims to evaluate the diagnostic and prognostic performance of AI models in detecting OKCs in histopathologic images.

Methods:

This systematic review was conducted in accordance with the preferred reporting items for systematic reviews and meta-analyses (PRISMA) guidelines. A comprehensive literature search was performed across PubMed, Scopus, Embase, Google Scholar, and ScienceDirect to identify studies that utilized AI models for diagnosing OKCs from histopathologic images. Studies were eligible for inclusion if they addressed the PICO (patient/population, intervention, comparison, and outcomes) framework, specifically investigating whether AI models (I) can enhance diagnostic and prognostic accuracy (O) for OKCs in histopathologic images (P). A meta-analysis was performed to pool the diagnostic performance of AI models across studies, and Egger’s test was conducted to assess publication bias.

Results:

A total of eight studies were included in the review. The risk of bias (ROB) across the included studies was generally low, with a few exceptions. The pooled area under the curve (AUC) for AI models in diagnosing OKCs was 0.967 (95% CI: 0.957 - 0.978). The pooled sensitivity ranged from 0.89 to 0.92, and the pooled specificity ranged from 0.88 to 0.94. The summary receiver operating characteristic (sROC) curve demonstrated an AUC of 0.93. Egger’s test for publication bias yielded a P-value of 0.522, indicating no significant evidence of publication bias. The review also highlighted several limitations, including small sample sizes, lack of external validation, and limited interpretability of the AI models.

Conclusions:

Artificial intelligence models, particularly deep learning architectures, demonstrate high diagnostic accuracy in detecting OKCs from histopathologic images.

1. Background

Odontogenic keratocysts (OKCs) are distinct cystic lesions of the jaw characterized by their aggressive nature, high recurrence rates, and potential for malignant transformation (1, 2). Accurate diagnosis of OKCs is critical, as they often share overlapping histopathological features with other odontogenic cysts, such as dentigerous and radicular cysts, making differentiation challenging (1, 3). Misdiagnosis can lead to suboptimal treatment strategies, resulting in higher recurrence rates, increased surgical interventions, and prolonged patient morbidity. Given these clinical implications, there is a pressing need for more precise and reliable diagnostic approaches (1, 4).

Traditional diagnostic methods rely on manual examination of histopathologic slides, which, although considered the gold standard, have inherent limitations. These include interobserver variability, subjectivity, and diagnostic inconsistencies, particularly in cases with subtle morphological differences (5-10). Moreover, manual evaluation can be time-consuming and labor-intensive, further complicating timely and accurate diagnoses in high-volume pathology settings. Delays in diagnosis can negatively impact treatment planning, potentially leading to disease progression and an increased healthcare burden (5, 7, 10-13). These limitations underscore the need for more objective, efficient, and reproducible diagnostic tools.

Artificial intelligence (AI), particularly machine learning (ML) techniques, has emerged as a transformative solution for overcoming these challenges (14-16). Deep learning models, including convolutional neural networks (CNNs), have demonstrated exceptional capabilities in analyzing medical imaging data, including histopathologic slides (17-19). Unlike traditional methods, AI-driven models can rapidly process large datasets, identify complex histopathological patterns with high precision, and minimize diagnostic discrepancies among pathologists. These models have also shown the potential to enhance reproducibility, reduce diagnostic turnaround time, and improve workflow efficiency, addressing key limitations of conventional approaches (19-21).

Notably, deep learning models can detect subtle histological features that may be overlooked by human observers, resulting in superior diagnostic accuracy. Additionally, AI-driven analysis significantly accelerates the diagnostic workflow by processing thousands of images in a fraction of the time required for manual examination, making it a valuable tool for high-throughput pathology applications. Successes in various medical fields, such as cancer detection, dermatopathology, and gastrointestinal pathology, further underscore the potential of AI to improve diagnostic accuracy and efficiency (19, 20, 22). However, the application of AI to odontogenic lesions, including OKCs, has faced both successes and limitations. While AI-based approaches have shown promise for diagnosing odontogenic lesions in radiographic images, their performance in analyzing histopathologic slides has been less extensively investigated and remains an area requiring further exploration.

Several studies have applied AI to the diagnosis of odontogenic lesions, with varying degrees of success (21, 23-26). For example, AI has achieved high diagnostic accuracy in detecting odontogenic lesions in radiographic and cone beam computed tomographic (CBCT) imaging. In the study by Fedato Tobias et al. (27), AI models demonstrated high sensitivity and specificity in detecting and classifying lesions such as OKCs, ameloblastomas, and dentigerous cysts from radiographic images. Similarly, the study by Shrivastava et al. (28) yielded comparable results through the analysis of radiographic and CBCT images.

However, the application of AI to the analysis of histopathologic images of OKCs has been less successful, largely due to the scarcity of high-quality annotated datasets. The limited availability of diverse and well-curated histopathologic images restricts model generalizability and increases the risk of overfitting, particularly when deep learning models are trained on small sample sizes. Inconsistent staining protocols, imaging artifacts, and variations in histological interpretation further contribute to these challenges, making it difficult to develop robust AI models that perform consistently across different datasets (21, 26, 27).

To address these challenges, techniques such as data augmentation, transfer learning, and generative adversarial networks (GANs) have been explored to improve model robustness (15, 19, 29, 30). Data augmentation artificially expands training datasets by applying transformations such as rotation, flipping, and contrast adjustments, thereby enhancing model adaptability to real-world variations. Transfer learning enables AI models to leverage pre-trained knowledge from larger medical imaging datasets, reducing the reliance on extensive OKC-specific data. Generative adversarial networks have also been utilized to generate synthetic histopathologic images, providing additional training data to enhance model performance. Implementing these strategies may help mitigate the limitations posed by small sample sizes, improve generalizability, and increase diagnostic accuracy (19, 30-32). Moreover, AI models can be trained not only to diagnose OKCs but also to assess prognostic markers, enabling clinicians to predict recurrence risk and tailor personalized treatment plans. However, while AI has been used to predict recurrence risks in some odontogenic lesions, studies specifically addressing its prognostic capabilities for OKCs remain scarce, highlighting an important gap in current research (33, 34). Understanding AI’s role in both diagnosis and prognosis is essential for integrating these models into clinical workflows and ultimately improving patient outcomes.

This systematic review and meta-analysis aim to address this gap by evaluating the diagnostic and prognostic performance of AI models in detecting OKCs from histopathologic images. While previous systematic reviews, including recent publications by Fedato Tobias et al. (27) and Shrivastava et al. (28), have explored the use of AI in diagnosing odontogenic lesions, their focus has primarily been on radiographic and CBCT imaging.

This study differs significantly by providing a comprehensive analysis of AI’s diagnostic and prognostic performance in histopathologic image analysis — an area that has received comparatively less attention. By synthesizing data from previous studies, this work aims to provide a critical assessment of the current state of AI applications in this domain, highlight existing limitations, and propose directions for future research and clinical implementation. Ultimately, this review seeks to bridge the gap between AI research and clinical practice, offering valuable insights for pathologists, researchers, and healthcare professionals interested in integrating AI into odontogenic pathology.

2. Methods

2.1. Protocol and Registration

The preferred reporting items for systematic reviews and meta-analyses (PRISMA) flow diagram used to track the study selection process is presented in Figure 1. This systematic review adheres to the guidelines outlined in the PRISMA, specifically the extension for diagnostic test accuracy studies (PRISMA-DTA) (35). The protocol for this review has been registered with the international prospective register of systematic reviews (PROSPERO) under the registration number CRD42024607673.

The preferred reporting items for systematic reviews and meta-analyses (PRISMA) flowchart
Figure 1.

The preferred reporting items for systematic reviews and meta-analyses (PRISMA) flowchart

2.2. Search Strategy

The search strategy for this systematic review involved conducting a comprehensive literature search across PubMed, Scopus, Embase, Google Scholar, and ScienceDirect, covering publications from their inception to May 5, 2024, with an update performed on January 9, 2025. The selection of search terms was based on a combination of the study’s keywords, corresponding medical subject headings (MeSH) terms in PubMed, and insights drawn from existing AI-based systematic reviews in related fields. Additionally, expert opinion was sought to optimize and refine the keyword list. Several key strategies were employed to ensure comprehensive coverage of relevant studies. Table 1 provides detailed information on the search queries used for each database.

Table 1.The Exact Search Query for Each Database
DatasetSearch QueryResults
PubMed("artificial intelligence" OR "artificial intelligence"[MeSH Terms] OR "deep learning" OR "deep learning"[MeSH Terms] OR "machine learning" OR "machine learning"[MeSH Terms] OR “convolutional neural networks” OR “convolutional neural networks"[MeSH Terms]) AND (“odontogenic keratocyst” OR “odontogenic keratocyst”[MeSH Terms] OR “jaw cysts” OR “jaw cysts”[MeSH Terms] OR “Odontogenic cysts” OR “Odontogenic cysts”[MeSH Terms])41
ScopusTITLE-ABS-KEY (["artificial intelligence" OR "deep learning" OR "machine learning" OR "convolutional neural networks"] AND ["odontogenic keratocyst" OR "jaw cysts" OR "Odontogenic cysts"])51
ScienceDirect("artificial intelligence" OR "deep learning" OR "machine learning" OR "convolutional neural networks") AND (“odontogenic keratocyst” OR “jaw cysts” OR “Odontogenic cysts”)85
Embase('artificial intelligence'/exp OR 'artificial intelligence' OR 'deep learning'/exp OR 'deep learning' OR 'machine learning'/exp OR 'machine learning' OR 'convolutional neural networks') AND ('odontogenic keratocyst'/exp OR 'odontogenic keratocyst' OR 'jaw cysts'/exp OR 'jaw cysts' OR 'odontogenic cysts'/exp OR 'odontogenic cysts')68
Google ScholarAllintitle: ("artificial intelligence" OR "deep learning" OR "machine learning" OR "convolutional neural networks") AND (“odontogenic keratocyst” OR “jaw cysts” OR “Odontogenic cysts”)7

The Exact Search Query for Each Database

No language or time restrictions were applied during the search. Additionally, the reference lists of the included articles were screened to identify any additional relevant studies.

2.3. Inclusion and Exclusion Criteria

Studies were eligible for inclusion if they addressed the PICO (patient/population, intervention, comparison, and outcomes) framework by evaluating whether AI models (I) can enhance diagnostic and prognostic accuracy (O) for OKCs in histopathologic images (P): Population (P): Histopathologic images of OKCs; intervention (I): Use of AI or MI models for diagnostic purposes; comparison and outcomes (C, O): Studies were required to report at least one performance metric, such as accuracy, sensitivity, specificity, or area under the curve (AUC); study design: This review included original research articles employing experimental, observational, retrospective, or prospective study designs.

Case reports, reviews, editorials, and commentaries were excluded. Additionally, studies that did not report diagnostic performance metrics or used only panoramic or CBCT images for diagnosis were excluded.

2.4. Study Selection

EndNote X9 was used to manage citations. After eliminating duplicates, three reviewers independently screened the titles and abstracts of the identified studies. Full-text articles were retrieved for studies deemed potentially eligible by the lead reviewer. Any discrepancies among the three reviewers were resolved through discussion, and, if necessary, with the input of a fourth reviewer.

2.5. Data Extraction

Two trained researchers and dentists independently extracted data from the included studies using a predefined data extraction form. The extracted data were verified by two reviewers to reduce the risk of any possible bias. The collected data include study characteristics such as author, year, sample size, inclusion and exclusion criteria, artificial intelligence model (types of models like convolutional neural networks, support vector machines), architecture, training, and testing datasets, diagnostic performance including sensitivity, specificity, accuracy, and area under the curve, data preprocessing methods such as data augmentation and image normalization.

Additionally, we recorded details regarding dataset size, the number of histopathologic images used for training and validation, and whether studies employed single-institution or multi-institutional data sources. Studies that did not specify dataset characteristics or validation procedures were noted. Discrepancies in data extraction were resolved through consultation with the reviewer.

2.6. Quality Assessment

The risk of bias in the included studies was independently assessed by three reviewers using the quality assessment of diagnostic accuracy studies tool. This tool evaluates four domains: patient selection, index test, reference standard, and flow and timing. Each study was rated as having a low, high, or unclear risk of bias in each domain.

To specifically assess the reliability of the datasets used in artificial intelligence model training, we examined whether studies reported dataset size, data augmentation techniques, external validation, and measures to prevent overfitting, such as cross-validation and transfer learning. Studies with small, single-institution datasets or unclear validation procedures were flagged as having potential concerns regarding generalizability.

Any disagreements among the reviewers were resolved through discussion led by the lead reviewer.

2.7. Data Synthesis

The primary outcome measured was the AUC, while secondary outcomes included sensitivity, specificity, and accuracy. A random-effects model was employed to account for heterogeneity among the studies, which was assessed using the I² statistic. The pooled diagnostic accuracy was represented as a summary receiver operating characteristic (sROC) curve. In this meta-analysis, the weight of each study was determined based on its sample size, with larger studies contributing more to the pooled estimates.

Publication bias was evaluated using Egger’s test. All statistical analyses were conducted using Python, employing the NumPy package for data manipulation, Matplotlib for visualization, and scikit-learn (sklearn) for ROC curve analysis. A P-value of less than 0.05 was considered statistically significant.

3. Results

The key findings from the studies are summarized in Table 2.

Table 2.The Summary of Findings of Included Studies
Authors (Year)Study’s AimSample PrevalenceData SourcesData Set (Training, Validation, Testing)Inclusion and Exclusion CriteriaLabeling ProcedureAI Model TaskPre-processing (P) Augmentation (A)Model (Algorithm, Architecture)Outcome
Cai et al. (2024) (5)Develop AI models for diagnosis and prognosis of OKC519 cases (OKC: 400, OOC: 90, GC: 29)Single-institutionDiagnostic: Training (70% - 363 cases), testing (30% - 156 cases) prognostic: 1688 WSIs, training (280 cases), testing (120 cases)Excluded cases with unclear or faded H&E stainingNAClassificationP: Macenko method (tiles of 512 × 512 pixels from WSIs, white background removal using AI, color normalization), Z-score normalization A: Random horizontal and vertical flippingSVM, random forest, extra trees, XGBoost, LightGBM, MLPDiagnostic AUC = 0.935 (95% CI: 0.898-0.973) prognostic AUC = 0.840 (95% CI: 0.751-0.930) prognostic accuracy = 67.5%, sensitivity = 92.9% (multi-slide)
Kim et al. (2024) (36)Evaluate the agreement between clinical diagnoses and histopathological outcomes for OKCs and odontogenic tumors by clinicians, ChatGPT-4, and ORAD623 specimens (OKC: 321, odontogenic tumors: 302)Single-institutionNAExcluded non-odontogenic, metastatic, adjacent spread, and non-OKC cystsNAClassificationNAChatGPT-4 (language model) ORAD (Bayesian algorithm) ChatGPT-4 concordance rates: 41.4% ORAD concordance rates: 45.6%
Mohanty et al. (2023) (7)Automate risk stratification of OKC using WSIs48 WSIs (508 tiled images)Multi-institutionalTraining (70%), validation (10%), testing (20%)Excluded blurry images and those with poor staining qualityManual labeling by pathologistsClassificationP: Entropy/variance calculation to remove white tiles A: Image rotation, flippingAttention-based image sequence analyzer (ABISA), Vision Transformer, LSTMABISA: Accuracy = 98%, Sensitivity = 100%, AUC = 0.98 VGG16: Accuracy = 80%, AUC = 0.82 VGG19: Accuracy = 73%, AUC = 0.77 inception V3: Accuracy = 82%, AUC = 0.91
Mohanty et al. (2023) (6)Build automation pipeline for diagnostic classification of OKCs and non-KCs using WSIs48 OKC slides, 20 DC slides, 37 RC slides, 6069 OKC tiles, 5967 non-KC tilesMulti-institutionalTraining (80%), validation (20%)Excluded blurry tiles, white tiles, and those with non-important information using OTSU thresholdingManual labeling by pathologistsClassificationP: Tile generation (2048 × 2048), white tile removal, OTSU thresholding A: Rotation, shifting, zoom, flippingP-C-reliefF model (PCA + reliefF), VGG16, VGG19, inception V3, standard CNNP-C-reliefF: Accuracy = 97%, AUC = 0.99 VGG16: Accuracy = 97%, AUC = 0.93 VGG19: Accuracy = 96%, AUC = 0.93 standard CNN: Accuracy = 96%, AUC = 0.93
Rao et al. (2022) (11)Develop ensemble deep-learning-based prognostic and prediction algorithm for OKC recurrence1660 digital slide images (1216 non-recurring, 444 recurring OKC)Single-institutionTraining (70%), testing (15%), validation (remaining)Sporadic OKC with 5-year follow-up; excluded syndromic OKC and radical treatment casesLabeling based on histopathological features (subepithelial hyalinization, incomplete epithelial lining, corrugated surface)ClassificationA: Rotation, shifting, shear, flippingDenseNet-121, inception-V3, inception-ResNet-V2, novel ensemble modelDenseNet-121: Accuracy = 93%, AUC = 0.9452 inception-V3: Accuracy = 92%, AUC = 0.9653 novel ensemble: Accuracy = 97%, AUC = 0.98
Rao et al. (2021) (37)Develop a deep learning-based system for diagnosing OKCs and non-OKCs2657 images (54 OKCs, 23 DCs, 20 RCs)Single-institutionTraining (70%), validation (15%), test (15%)Excluded inflamed keratocysts and inadequate biopsiesManual labeling by pathologistsClassificationA: Shear, rotation, zooming, flipping; region-of-interest isolationVGG16, DenseNet-169; DenseNet-169 trained on full image and epithelium-only datasetVGG16: Validation accuracy = 89.01%, Test accuracy = 62% DenseNet-169 (Exp II): Validation accuracy = 89.82%, Test accuracy = 91%, AUC = 0.9597 DenseNet-169 (Exp III): Test accuracy = 91%, AUC = 0.9637 Combined model (Exp IV): Accuracy = 93%
Florindo et al. (2017) (38)Classify OKC and radicular cysts150 images (65 sporadic OKCs, 40 syndromic OKCs, 45 RCs)NARandom 10-fold cross-validationExcluded images with artifacts or poor qualityManual labeling by histopathologistsClassificationP: Grey-level conversion, segmentationBouligand-Minkowski descriptors + LDA72% (OKC vs. RC vs. syndromic OKC) 98% (OKC vs. radicular) 68% (sporadic vs. syndromic OKC)
Frydenlund et al. (2014) (39)Automated classification of four types of odontogenic cysts73 images (20 DC, 20 LPC, 20 OKC, 13 GOC)NATraining (38 images), validation (37 images), testing (39 images)Excluded images with poor quality or resolution issuesManual labeling by pathologistsClassificationP: Color standardization, smoothing, segmentationSVM, bagging with logistic regression (BLR)SVM: 83.8% (validation), 92.3% (testing) BLR: 95.4% (testing)
Eramian et al. (2011) (40)Segmentation of epithelium in H&E-stained odontogenic cysts38 training images, 35 validation imagesNATraining: 38 images (10 dentigerous cysts, 10 odontogenic keratocysts, 10 lateral periodontal cysts, 8 glandular odontogenic cysts), validation: 35 images (same distribution)Excluded inflammatory odontogenic cystsManually identified epithelium and stroma pixelsSegmentationP: Luminance and chrominance standardizationGraph-based segmentationSensitivity: 91.5% specificity: 85.1% accuracy: 85%

The Summary of Findings of Included Studies

3.1. Study Selection and Inclusion Process

A total of 252 studies were identified through database searches, with one additional record found through manual searching of reference lists. After removing 86 duplicates, 166 studies remained for title and abstract screening. Following this step, 152 studies were excluded based on irrelevance, leaving 14 studies for full-text review. After full-text evaluation, 6 studies were excluded for not meeting the inclusion criteria (e.g., lack of analysis of OKCs, use of non-histopathologic images, or lack of access to full text). This process resulted in the final inclusion of 9 studies (8 from database searches and 1 from manual search) in this systematic review. Among these, 5 studies provided sufficient data to be included in the meta-analysis.

3.2. Study Characteristics and Data Quality

Nine studies published between 2011 and 2024 met the inclusion criteria, demonstrating progressive advancements in the application of AI for OKC analysis (5-7, 11, 36-40).

3.2.1. Sample Prevalence, Data Sources, and Eligibility Criteria

The sample sizes in the studies varied substantially, ranging from 48 whole slide images (WSIs) in Mohanty et al. (7) to over 1,500 images in other studies (11, 37). The prevalence of OKC cases within these datasets ranged from 20 in Frydenlund et al. (39) and Eramian et al. (40) to 400 cases in Cai et al. (5).

Data sources were diverse, including both single-institution and multi-institution datasets, with variations in image resolution, staining protocols, and acquisition equipment. Most studies relied on single-institution datasets (5, 11, 36, 37), while some incorporated multi-institutional data (6, 7). Additionally, three studies did not specify their data sources (38-40), which may impact dataset diversity and overall generalizability.

The eligibility criteria among the included studies varied, but generally ensured that only high-quality images and relevant cases were included for AI-based analysis. Studies commonly excluded cases with unclear or faded hematoxylin and eosin staining, non-odontogenic or metastatic cysts, and images with poor resolution or artifacts.

3.2.2. Data Preprocessing and Labeling Procedure

To enhance model robustness, most studies implemented comprehensive preprocessing pipelines. Common techniques included color normalization (e.g., the Macenko method), white background removal, and tile generation from WSIs. Additionally, data augmentation strategies — such as random flipping, rotation, shifting, zooming, and shear transformations — were routinely applied to expand the effective size and diversity of the training sets. These preprocessing steps were critical for ensuring consistent image quality and mitigating the effects of dataset heterogeneity, thereby contributing to the high diagnostic performance observed.

In most included studies, samples were labeled by experienced pathologists, ensuring high-quality annotations. Labels were assigned based on characteristic histopathological features of OKCs, including subepithelial hyalinization, incomplete epithelial lining, and surface corrugation.

3.2.3. Training Data Set and Training Process of Models

The number of images in the training sets and the training processes varied significantly across the studies. Training datasets ranged from as few as 38 images in the study by Eramian et al. (40) to more than 1,500 images in the studies by Rao et al. (11, 37). Among the deep learning architectures, DenseNet-169 — employed in the studies by Rao et al. (11, 37) — was trained using the Adam optimizer (learning rate = 0.001, batch size = 12 or higher) and incorporated data augmentation techniques such as shear, rotation, and zoom transformations. DenseNet-169 achieved high classification performance, with reported accuracy exceeding 91% and an AUC ranging from 0.9597 to 0.9653 (11, 37).

Similarly, Inception-V3 — used by Mohanty et al. (6) and Cai et al. (5) — demonstrated robust classification performance, with RMSprop as the optimizer and a learning rate of 0.0001 (5, 6). Additionally, Cai et al. (5) optimized the model using fivefold cross-validation and grid-search hyperparameter tuning, with a batch size of 32 and a dropout rate of 0.5 to prevent overfitting. An advanced hybrid deep learning approach, attention-based image sequence analyzer (ABISA), was also evaluated. This model, proposed by Mohanty et al. (7), integrates convolutional and recurrent layers with an attention mechanism for enhanced interpretability. The ABISA was trained using the Adam optimizer with a cyclical learning rate schedule (0.0001 to 0.001) and extensive data augmentation techniques, including contrast adjustment and histogram equalization (7).

Florindo et al. (38) introduced a morphological classification framework using Bouligand-Minkowski fractal descriptors to differentiate OKCs from other cysts. This method, applied to 150 histological images, demonstrated a classification accuracy of 98%, highlighting the potential of fractal-based feature extraction as a complementary approach to deep learning models.

The study by Eramian et al. (40) further contributed by developing an automated epithelial segmentation algorithm using binary graph cuts. The segmentation model was trained on 38 histological images, achieving a mean sensitivity of 91.5% and specificity of 85.1%, with the highest accuracy observed in OKCs (96.1% sensitivity, 98.7% specificity). Their color standardization and edge detection preprocessing steps were crucial in refining AI-driven histopathological segmentation.

3.2.4. Model Diagnostic and Prognostic Performance

Recent studies (2021 - 2024) exhibited superior diagnostic performance, with accuracy rates consistently exceeding 90% (5, 7, 11, 37). The ABISA achieved the highest diagnostic accuracy (98%, AUC = 0.98) (7), while the P-C-ReliefF model demonstrated comparable performance (97% accuracy, AUC = 0.99) (6). Traditional MI approaches showed varying but promising results, with LightGBM and XGBoost achieving AUCs of 0.935 and 0.93, respectively (5). Deep learning architectures, particularly DenseNet-169 and Inception-V3, demonstrated robust performance with AUCs ranging from 0.9597 to 0.9653 (11, 37), while ensemble approaches further improved accuracy up to 97% (11). Notably, two recent studies specifically addressed OKC recurrence prediction, with Rao et al. (11) achieving 97% accuracy (AUC = 0.98) using a novel ensemble model, while Cai et al. (5) reported a prognostic AUC of 0.840 (95% CI: 0.751 - 0.930). Most studies implemented comprehensive preprocessing pipelines, including color normalization, white background removal, and tile generation from WSIs (5-7, 11, 37). Data augmentation strategies, reported in five studies (5-7, 11, 37), predominantly utilized random flipping, rotation, and shear transformations. Early approaches (2011 - 2017) focused on basic classification tasks using traditional MI methods, achieving moderate success (72 - 98% accuracy range) (38-40), while recent studies employed more sophisticated architectures and addressed complex tasks such as recurrence prediction (5, 7, 11, 37). This temporal progression reflects both methodological advancement and improved performance metrics, with modern approaches consistently achieving AUC values above 0.90 and providing more robust diagnostic and prognostic capabilities (5, 7, 11, 37).

Notably, one study incorporated large language models (LLMs) such as Chat-GPT4, demonstrating comparable diagnostic accuracy to clinicians in diagnosing OKCs, with a concordance rate of 41% and a kappa value of 0.14 (36). While Chat-GPT4 showed strengths in processing textual clinical data, it exhibited slightly lower specificity (90.4%) and sensitivity (30.1%) compared to clinicians (specificity: 95.7%, sensitivity: 32.9%).

3.3. Risk of Bias and Applicability

The ROB across the included studies was generally low, with a few exceptions. Four of the nine studies (44.4%) demonstrated low risk across all four domains (5, 7, 11), whereas five studies (55.6%) exhibited concerns in one or more areas. Specifically, in the patient selection domain, 55.6% of studies were rated as low risk, 33.3% as unclear, and 11.1% as high risk. For the index test domain, 66.7% were rated as low risk and 33.3% as unclear, while all studies (100%) showed low risk in the reference standard domain. In the flow and timing domain, 77.8% of studies were classified as low risk, and 22.2% as unclear (Figure 2 and Table 3).

The results of the risk of bias (ROB) assessment.
Figure 2.

The results of the risk of bias (ROB) assessment.

Table 3.The Risk of Bias Assessment and Applicability Results
Authors (Year)ROB AssessmentApplicability Assessment
Patient/Sample SelectionIndex TestReference StandardFlow and TimingPatient/Sample SelectionIndex TestReference Standard
Cai et al. (2024) (5)LowLowUnclearLowLowLowLow
Kim et al. (2024) (36)UnclearLowLowLowLowLowLow
Mohanty et al. (2023) (7)LowLowLowLowLowLowLow
Mohanty et al. (2023) (6)LowLowLowLowLowLowLow
Rao et al. (2022) (11)LowLowLowLowLowLowLow
Rao et al. (2021) (37)LowUnclearLowUnclearLowLowLow
Florindo et al. (2017) (38)UnclearUnclearLowLowLowLowLow
Frydenlund et al. (2014) (39)UnclearLowLowUnclearLowLowLow
Eramian et al. (2011) (40)HighUnclearLowLowLowLowLow

The Risk of Bias Assessment and Applicability Results

3.4. Meta-Analysis

The meta-analysis was conducted to assess the pooled diagnostic accuracy of the included studies. The pooled analysis of five studies that reported the AUC yielded a value of 0.967 (95% CI: 0.957 – 0.978), indicating strong overall diagnostic accuracy. Additionally, heterogeneity among the studies was minimal (I² = 0%), suggesting consistent performance across various AI models and datasets (Figure 3A).

A, The pooled analysis of five studies that reported area under the curve (AUC) with a value of 0.967 (95% CI: 0.957 - 0.978) and low heterogeneity (I<sup>2</sup> = 0.0%); B, the pooled sensitivity and specificity across the studies further demonstrated the robust performance of the artificial intelligence (AI) models, the pooled sensitivity was 0.90 (95% CI: 0.89 - 0.92) and the pooled specificity was 0.91 (95% CI: 0.88 - 0.94), the forest plots for both sensitivity and specificity displayed relatively narrow confidence intervals, indicating consistent diagnostic performance across the studies included; C, the summary receiver operating characteristic (sROC) curve had an AUC of approximately 0.93, further highlighting the diagnostic strength of the AI models.
Figure 3.

A, The pooled analysis of five studies that reported area under the curve (AUC) with a value of 0.967 (95% CI: 0.957 - 0.978) and low heterogeneity (I2 = 0.0%); B, the pooled sensitivity and specificity across the studies further demonstrated the robust performance of the artificial intelligence (AI) models, the pooled sensitivity was 0.90 (95% CI: 0.89 - 0.92) and the pooled specificity was 0.91 (95% CI: 0.88 - 0.94), the forest plots for both sensitivity and specificity displayed relatively narrow confidence intervals, indicating consistent diagnostic performance across the studies included; C, the summary receiver operating characteristic (sROC) curve had an AUC of approximately 0.93, further highlighting the diagnostic strength of the AI models.

Forest plots for sensitivity and specificity showed that the pooled sensitivity ranged from 0.89 to 0.92, while the pooled specificity ranged from 0.88 to 0.94 (Figure 3B). The sROC curve, which assesses overall diagnostic performance by combining sensitivity and specificity, demonstrated an AUC of 0.93. This result further confirms the high diagnostic accuracy of AI models across studies. Most data points were concentrated near the upper-left corner of the sROC plot, indicating a strong balance between sensitivity and specificity (Figure 3C).

Egger’s test for publication bias produced a P-value of 0.522, suggesting that there was no significant evidence of publication bias.

4. Discussion

This systematic review and meta-analysis explored the diagnostic and prognostic performance of AI models in detecting OKCs from histopathological images. The findings confirm the promising potential of AI — particularly deep learning models — in achieving high diagnostic accuracy, as demonstrated by a pooled AUC of 0.967. The pooled sensitivity of 0.90 suggests that AI systems were effective in identifying positive OKC cases, while a pooled specificity of 0.91 indicates a robust capability to reduce false positives by correctly identifying negative cases. These results emphasize the role AI can play in enhancing clinical decision-making for OKCs, especially in scenarios where access to expert pathologists is limited.

AI models excel at processing intricate histopathological images and autonomously learning hierarchical features (19, 30, 41, 42). Compared to conventional diagnostic methods, AI-driven models significantly reduce interobserver variability and offer a more standardized approach to pathology assessment. Specifically, DenseNet-169, which achieved an impressive AUC of 0.9872, exemplifies the strengths of AI in handling high-resolution medical images due to its efficient feature extraction mechanisms (29, 37, 43). Unlike traditional MI models such as support vector machines (SVMs), deep learning approaches eliminate the need for manual feature extraction, minimizing potential biases and allowing for more adaptive analysis of complex datasets (44-47). This adaptability is particularly beneficial in distinguishing OKCs from histologically similar cystic lesions.

Several studies included in the meta-analysis demonstrated that AI models achieve diagnostic accuracy comparable to, or even surpassing, that of human pathologists (5, 7). The pooled sensitivity (0.89 - 0.92) and specificity (0.88 - 0.94) reflect the high diagnostic precision of AI, which also offers additional benefits such as the ability to process large volumes of data consistently without fatigue (48, 49). Moreover, AI can help mitigate inter-observer variability, a common issue in histopathological evaluations (31, 49-51). This aspect is particularly relevant in diagnosing OKCs, where histopathological features may overlap with those of other odontogenic cysts, leading to potential misclassification.

4.1. Conclusions

In conclusion, AI models represent a significant advancement in diagnosing OKCs, offering high accuracy, sensitivity, and specificity in analyzing histopathologic images. These systems hold great potential for enhancing diagnostic workflows, improving concordance rates, and reducing variability among pathologists. However, several challenges must be addressed to fully realize this potential. A key priority is the external validation of AI models using diverse, large-scale datasets to ensure generalizability across populations. The development of explainable AI tools is equally critical to facilitate trust and adoption among clinicians. Additionally, standardized evaluation methods are essential to enable meaningful cross-study comparisons and to guide clinical integration efforts.

4.2. Comparison with Previous Studies

To the best of our knowledge, no systematic review and meta-analysis have been conducted to date to assess the diagnostic and prognostic performance of AI models in detecting OKCs from histopathologic images. While previous reviews, such as those by Fedato Tobias et al. (27) and Shrivastava et al. (28), have explored the role of AI in diagnosing odontogenic cysts and tumors, their focus was primarily on radiographic and CBCT imaging. In contrast, our systematic review provides a comprehensive evaluation of AI performance in histopathologic image analysis — an area that has received comparatively less attention.

Our findings align with those of Fedato Tobias et al. (27), who reported high sensitivity and specificity of AI models in odontogenic cyst classification. However, our study expands on this by emphasizing the impact of preprocessing techniques, model architecture, and dataset diversity on AI performance. Similarly, Shrivastava et al. (28) highlighted the need for AI validation across multiple imaging modalities, which aligns with our recommendation for external dataset testing. Additionally, Shi et al. (52) focused solely on radiographic images without conducting a meta-analysis or systematic review, further underscoring the novelty and significance of our approach.

4.3. Generalizability and Model Evaluation on External Data

One of the key limitations of current AI models is their limited generalizability due to reliance on single-institution datasets and the lack of external validation. Many of the studies included in this review trained their models on internal datasets (5, 11, 36, 37), raising concerns about their applicability across different populations, imaging settings, and clinical environments.

Potential solutions to these concerns can be considered in future research. Conducting multi-center validation studies will help ensure that AI models generalize well across different clinical settings. Utilizing independent test datasets, instead of relying solely on internal cross-validation, can improve robustness and detect overfitting. Standardizing imaging protocols and annotation methods will reduce inconsistencies in histopathologic image analysis, leading to more reliable AI predictions. Collaborative efforts with leading organizations, including the international organization for standardization (ISO) and the Institute of Electrical and Electronics Engineers (IEEE), can facilitate the creation of these guidelines. Implementing common data models (CDMs) and standardized ontologies, such as the observational medical outcomes partnership (OMOP) and the unified medical language system (UMLS), will harmonize data representation across institutions. Employing domain adaptation techniques can enhance model adaptability to different imaging sources and patient populations. Additionally, federated learning approaches allow AI models to be trained on diverse datasets from multiple institutions without transferring raw patient data, preserving privacy while improving generalizability (11, 37-40).

4.4. Challenges in of Artificial Intelligence-Based Odontogenic Keratocysts Diagnosis

One of the primary objectives of this study was to examine the challenges associated with AI-driven diagnosis, particularly in detecting OKCs. One challenge is the small sample sizes used for training, testing, and validating the model. of Artificial intelligence models trained on datasets with disproportionately low numbers of OKC cases may exhibit bias toward more prevalent cyst types, leading to suboptimal diagnostic performance (53, 54). To mitigate these biases, we recommend using data augmentation, class weighting, and synthetic data generation techniques — such as GANs — to create more balanced datasets (32, 53, 54).

Variability in histopathologic image preprocessing, including color normalization, tile selection, and contrast adjustment, significantly affects AI model accuracy. Inconsistent preprocessing pipelines across studies can lead to performance discrepancies, highlighting the need for standardized methodologies to ensure reproducibility and reliability. Deep learning models often function as "black boxes," limiting the interpretability of their decision-making processes (55-57). Explainability techniques such as Grad-CAM and SHAP values can improve transparency, fostering trust among clinicians and facilitating the integration of AI into diagnostic workflows. Grad-CAM generates heatmaps highlighting the regions of histopathological images that contributed most to the model’s decision, aiding pathologists in understanding why certain areas were classified as OKC. SHAP values, on the other hand, quantify the contribution of each feature in a model’s decision-making process, providing numerical insights into AI predictions. In clinical practice, these techniques can enhance the interpretability of AI-driven diagnoses by offering visual and quantitative explanations, increasing clinician trust, and supporting AI-assisted decision-making (55).

4.5. Practical Applications of Artificial Intelligence Models in Diagnostic Environments

AI has significant potential for integration into real-world diagnostic workflows in hospitals and clinics. By incorporating AI models into pathology laboratories, clinicians can improve diagnostic speed and accuracy, particularly in settings with high case volumes. Potential applications include: (1) Automated screening of histopathologic slides to assist pathologists in prioritizing complex cases; (2) AI-assisted second opinions to reduce diagnostic variability and enhance confidence in challenging cases; and (3) predictive modeling for recurrence risk assessment, enabling personalized treatment planning (15, 28, 52).

For successful integration of AI models into clinical workflows, several structured steps should be considered. First, AI models should be validated on external datasets and multi-center studies to ensure robustness across different clinical environments. Second, AI systems must be seamlessly integrated into existing laboratory information systems (LIS) and digital pathology platforms to facilitate real-time analysis of histopathological slides. Third, AI-driven decision support tools should be designed to provide interpretable outputs, such as probability scores, heatmaps, or case prioritization, to assist rather than replace pathologists in diagnostic decision-making. Fourth, comprehensive training programs should be implemented to familiarize clinicians and laboratory staff with AI-based tools, ensuring their appropriate use and interpretation. Fifth, regulatory approval and compliance with healthcare data protection laws (e.g., HIPAA, GDPR) should be ensured before deployment. Finally, continuous model monitoring and updates based on real-world clinical feedback are essential to maintain accuracy and address evolving diagnostic challenges. These steps will facilitate the responsible and effective implementation of AI in pathology laboratories, ultimately improving diagnostic efficiency and accuracy.

However, practical challenges remain, particularly concerning how AI can be seamlessly incorporated into existing diagnostic systems. To facilitate this, AI tools must be user-friendly and compatible with current diagnostic infrastructure, allowing pathologists to easily interact with and interpret AI results. Additionally, training programs should be developed to familiarize clinicians with the capabilities and limitations of AI tools, ensuring they are used appropriately to augment human expertise.

4.6. Ethical Considerations and Future Recommendations

The integration of AI in pathology necessitates careful ethical consideration to ensure responsible implementation. Adopting the framework proposed by Rokhshad et al. (58), key ethical principles — transparency, fairness, privacy protection, accountability, and equitable access — must guide AI deployment.

4.6.1. Transparency

Transparent AI development, including clear documentation of training data, algorithms, and decision-making processes, is essential to foster trust among clinicians and patients. Active participation from diverse stakeholders, such as pathologists, ethicists, and policymakers, is crucial for maintaining ethical AI applications in pathology (58, 59).

4.6.2. Fairness

Addressing bias in AI models is imperative, as imbalanced training datasets may lead to disparities in diagnostic performance. Ensuring diverse representation in training data and prioritizing equitable access — particularly in low-resource settings — can enhance the fairness of AI applications in pathology (58-60).

4.6.3. Privacy Protection

Given AI’s reliance on sensitive medical data, compliance with data protection regulations (e.g., GDPR, HIPAA) is critical. Techniques like federated learning and anonymization can help maintain privacy while enabling AI models to learn securely from diverse datasets (58).

4.6.4. Accountability and Equitable Access

Artificial intelligence should function as a supportive tool rather than a replacement for human expertise. Clinicians must retain responsibility for final diagnostic decisions, necessitating targeted training programs to improve digital literacy and ensure appropriate AI integration into clinical workflows (58, 59, 61).

AI development should consider long-term sustainability, particularly its environmental impact. Computational efficiency and energy-conscious AI model optimization can mitigate concerns related to high resource consumption and carbon emissions (58, 59, 61).

To facilitate responsible AI adoption, regulatory frameworks should define legal responsibilities for AI developers and healthcare providers. Independent AI oversight committees and interdisciplinary collaborations are essential for standardizing AI integration in diagnostic workflows while addressing ethical concerns (58, 61).

4.7. Environmental Impacts of Developing Artificial Intelligence Models

Deploying AI models in clinical settings offers significant benefits but also raises environmental concerns, primarily due to high energy consumption, increased carbon emissions, and electronic waste. Training sophisticated AI models requires substantial computational power, contributing to rising electricity usage and carbon footprints — particularly if data centers rely on fossil fuels. Additionally, frequent hardware updates generate electronic waste, further impacting the environment. To mitigate these effects, several strategies can be implemented. Developing energy-efficient AI models through techniques such as model quantization and pruning can reduce computational demands. Optimizing data storage and management, along with utilizing renewable energy sources for data centers, can significantly lower AI’s carbon footprint. Implementing green computing practices, such as dynamic power management and hardware–software optimization, can further enhance energy efficiency. Regular monitoring, maintenance, and lifecycle assessments of AI systems can support long-term sustainability while reducing the need for resource-intensive retraining. By adopting these strategies, healthcare institutions can balance the benefits of AI advancements with environmental responsibility, ensuring that AI-driven clinical applications align with broader sustainability goals (62-64).

4.8. Study Limitations

This study acknowledges certain limitations. A key limitation is the small sample sizes in several studies (38-40), which heightens the risk of overfitting. Overfitting occurs when AI models perform exceptionally well on training data but fail to generalize to unseen cases (53, 54). To mitigate this risk, several strategies can be employed. First, data augmentation techniques — such as rotation, flipping, contrast adjustment, and stain normalization — can artificially expand the dataset, improving the model’s ability to generalize to new cases. Second, transfer learning, where AI models are pre-trained on large-scale histopathological datasets before fine-tuning on smaller OKC-specific datasets, can enhance performance and reduce overfitting. Third, cross-validation methods, particularly k-fold cross-validation, should be used to ensure that the model is evaluated on different subsets of data, thereby reducing bias from reliance on a single training set. Fourth, regularization techniques such as dropout, batch normalization, and L2 regularization can help prevent the model from overfitting to features that may not generalize well. Fifth, leveraging synthetic data generation using GANs can supplement real training samples, particularly for rare cases like OKCs. Lastly, multi-center collaborations and federated learning approaches should be pursued to train AI models on more diverse datasets without compromising patient privacy.

Another limitation was that few studies reported data suitable for meta-analysis, including AUC, sensitivity, specificity, and accuracy. Additionally, in the meta-analysis, the weight of each study was determined by its sample size, reflecting the relative contribution of each study's data volume to the pooled estimate. While this approach is straightforward, it does not account for differences in study precision or variability, which may limit the robustness of the pooled results.

Finally, one study had to be excluded due to the lack of access to its full text, which prevented a comprehensive evaluation of its methodology and results.

Footnotes

References

  • 1.
    Mirzaee S, Shoorgashti R, Sadri D, Farhadi S. Comparison of EGFR Expression in Ameloblastoma and Odontogenic Keratocyst. J Res Dent Maxillofacial Sci. 2023;8:274-9. https://doi.org/10.61186/jrdms.8.4.274.
  • 2.
    Shoorgashti R, Alimohammadi M, Baghizadeh S, Radmard B, Ebrahimi H, Lesan S. Artificial Intelligence Models Accuracy for Odontogenic Keratocyst Detection From Panoramic View Radiographs: A Systematic Review and Meta-Analysis. Health Sci Rep. 2025;8(4). e70614. [PubMed ID: 40165928]. [PubMed Central ID: PMC11956212]. https://doi.org/10.1002/hsr2.70614.
  • 3.
    Motaleb L, Zakai D, Stocker J. Fourteen-year study of the management of the odontogenic keratocyst. Are adjunctive therapies all they are cut out to be? Br J Oral Maxillofac Surg. 2022;60(2):105-12. [PubMed ID: 35067413]. https://doi.org/10.1016/j.bjoms.2021.01.021.
  • 4.
    Shoorgashti R, Sadri D, Farhadi S. Expression of Epidermal Growth Factor Receptor by Odontogenic Cysts: A comparative Study of Dentigerous Cyst and Odontogenic Keratocyst. J Res Dent Sci. 2020;17(2):127-36. https://doi.org/10.29252/jrds.17.2.127.
  • 5.
    Cai X, Zhang H, Wang Y, Zhang J, Li T. Digital pathology-based artificial intelligence models for differential diagnosis and prognosis of sporadic odontogenic keratocysts. Int J Oral Sci. 2024;16(1):16. [PubMed ID: 38403665]. [PubMed Central ID: PMC10894880]. https://doi.org/10.1038/s41368-024-00287-y.
  • 6.
    Mohanty S, Shivanna DB, Rao RS, Astekar M, Chandrashekar C, Radhakrishnan R, et al. Building Automation Pipeline for Diagnostic Classification of Sporadic Odontogenic Keratocysts and Non-Keratocysts Using Whole-Slide Images. Diagnostics (Basel). 2023;13(21). [PubMed ID: 37958281]. [PubMed Central ID: PMC10648794]. https://doi.org/10.3390/diagnostics13213384.
  • 7.
    Mohanty S, Shivanna DB, Rao RS, Astekar M, Chandrashekar C, Radhakrishnan R, et al. Development of Automated Risk Stratification for Sporadic Odontogenic Keratocyst Whole Slide Images with an Attention-Based Image Sequence Analyzer. Diagnostics (Basel). 2023;13(23). [PubMed ID: 38066780]. [PubMed Central ID: PMC10706752]. https://doi.org/10.3390/diagnostics13233539.
  • 8.
    Shoorgashti R, Moshiri A, Lesan S. Evaluation of Oral Mucosal Lesions in Iranian Smokers and Non-smokers. Niger J Clin Pract. 2024;27(4):467-74. [PubMed ID: 38679769]. https://doi.org/10.4103/njcp.njcp_702_23.
  • 9.
    Mirhosseini N, Shoorgashti R, Lesan S. The evaluation of clinical factors affecting oral health impacts on the quality of life of Iranian elderly patients visiting dental clinics: A cross-sectional study. Spec Care Dentist. 2024;44(4):1219-27. [PubMed ID: 38430466]. https://doi.org/10.1111/scd.12980.
  • 10.
    Shoorgashti R, Khalilirad M, Haji Molla Hoseini M, Lesan S. Evaluation of the relationship between nitric oxide and candida albicans-associated denture stomatitis: A cross-sectional study. Caspian J Dent Res. 2023;12(2):82-8. https://doi.org/10.22088/cjdr.12.2.82.
  • 11.
    Rao RS, Shivanna DB, Lakshminarayana S, Mahadevpur KS, Alhazmi YA, Bakri MMH, et al. Ensemble Deep-Learning-Based Prognostic and Prediction for Recurrence of Sporadic Odontogenic Keratocysts on Hematoxylin and Eosin Stained Pathological Images of Incisional Biopsies. J Pers Med. 2022;12(8). [PubMed ID: 35893314]. [PubMed Central ID: PMC9332803]. https://doi.org/10.3390/jpm12081220.
  • 12.
    Ghasemian Gorji R, Shoorgashti R, Lotfali E, Farhadi S. MIC Determination of Silver Nanoparticles on Nystatin-Resistant Candida albicans in Patients with Denture Stomatitis: An In Vitro Study. Middle East J Rehabil Health Stud. 2023;11(1). https://doi.org/10.5812/mejrh-138030.
  • 13.
    Shoorgashti R, Havakhah S, Nowroozi S, Ghadamgahi B, Mehrara R, Oroojalian F. Evaluation of the antibacterial and cytotoxic activities of Ag/ZnO nanoparticles loaded polycaprolactone/chitosan composites for dental applications. Nanomed J. 2023;10(1):68-76. https://doi.org/10.22038/nmj.2022.69755.1744.
  • 14.
    Shafi S, Parwani AV. Artificial intelligence in diagnostic pathology. Diagn Pathol. 2023;18(1):109. [PubMed ID: 37784122]. [PubMed Central ID: PMC10546747]. https://doi.org/10.1186/s13000-023-01375-z.
  • 15.
    Zhang DY, Venkat A, Khasawneh H, Sali R, Zhang V, Pei Z. Implementation of Digital Pathology and Artificial Intelligence in Routine Pathology Practice. Lab Invest. 2024;104(9):102111. [PubMed ID: 39053633]. https://doi.org/10.1016/j.labinv.2024.102111.
  • 16.
    Jain E, Patel A, Parwani AV, Shafi S, Brar Z, Sharma S, et al. Whole Slide Imaging Technology and Its Applications: Current and Emerging Perspectives. Int J Surg Pathol. 2024;32(3):433-48. [PubMed ID: 37437093]. https://doi.org/10.1177/10668969231185089.
  • 17.
    Rokhshad R, Mohammad-Rahimi H, Price JB, Shoorgashti R, Abbasiparashkouh Z, Esmaeili M, et al. Artificial intelligence for classification and detection of oral mucosa lesions on photographs: a systematic review and meta-analysis. Clin Oral Investig. 2024;28(1):88. [PubMed ID: 38217733]. https://doi.org/10.1007/s00784-023-05475-4.
  • 18.
    Mahmood H, Shaban M, Indave BI, Santos-Silva AR, Rajpoot N, Khurram SA. Use of artificial intelligence in diagnosis of head and neck precancerous and cancerous lesions: A systematic review. Oral Oncol. 2020;110:104885. [PubMed ID: 32674040]. https://doi.org/10.1016/j.oraloncology.2020.104885.
  • 19.
    Pirayesh Z, Mohammad-Rahimi H, Ghasemi N, Motamedian SR, Sadeghi TS, Koohi H, et al. Deep Learning-Based Image Classification and Segmentation on Digital Histopathology for Oral Squamous Cell Carcinoma: A Systematic Review and Meta-Analysis. J Oral Pathol Med. 2024;53(9):551-66. [PubMed ID: 39256895]. https://doi.org/10.1111/jop.13578.
  • 20.
    Malhotra M, Shaw AK, Priyadarshini SR, Metha SS, Sahoo PK, Gachake A. Diagnostic Accuracy of Artificial Intelligence Compared to Biopsy in Detecting Early Oral Squamous Cell Carcinoma: A Systematic Review and Meta Analysis. Asian Pac J Cancer Prev. 2024;25(8):2593-603. [PubMed ID: 39205556]. [PubMed Central ID: PMC11495466]. https://doi.org/10.31557/APJCP.2024.25.8.2593.
  • 21.
    El-Bouzaidi YEI, Abdoun O. Advances in artificial intelligence for accurate and timely diagnosis of COVID-19: A comprehensive review of medical imaging analysis. Scientific African. 2023;22. e01961. https://doi.org/10.1016/j.sciaf.2023.e01961.
  • 22.
    Rokhshad R, Nasiri F, Saberi N, Shoorgashti R, Ehsani SS, Nasiri Z, et al. Deep learning for age estimation from panoramic radiographs: A systematic review and meta-analysis. J Dent. 2025;154:105560. [PubMed ID: 39826609]. https://doi.org/10.1016/j.jdent.2025.105560.
  • 23.
    Kiran N, Sapna FNU, Kiran FNU, Kumar D, Raja FNU, Shiwlani S, et al. Digital Pathology: Transforming Diagnosis in the Digital Age. Cureus. 2023;15(9). e44620. https://doi.org/10.7759/cureus.44620.
  • 24.
    Raissi Dehkordi N, Raissi Dehkordi N, Karimi Toudeshki K, Farjoo MH. Artificial Intelligence in Diagnosis of Long QT Syndrome: A Review of Current State, Challenges, and Future Perspectives. Mayo Clin Proc Digit Health. 2024;2(1):21-31. [PubMed ID: 40206686]. [PubMed Central ID: PMC11975689]. https://doi.org/10.1016/j.mcpdig.2023.11.003.
  • 25.
    Karalis VD. The Integration of Artificial Intelligence into Clinical Practice. Appl Biosci. 2024;3(1):14-44. [PubMed ID: doi:10.3390/applbiosci3010002].
  • 26.
    Fatima M, Pachauri P, Akram W, Parvez M, Ahmad S, Yahya Z. Enhancing retinal disease diagnosis through AI: Evaluating performance, ethical considerations, and clinical implementation. Inform Health. 2024;1(2):57-69. https://doi.org/10.1016/j.infoh.2024.05.003.
  • 27.
    Fedato Tobias RS, Teodoro AB, Evangelista K, Leite AF, Valladares-Neto J, de Freitas Silva BS, et al. Diagnostic capability of artificial intelligence tools for detecting and classifying odontogenic cysts and tumors: a systematic review and meta-analysis. Oral Surg Oral Med Oral Pathol Oral Radiol. 2024;138(3):414-26. [PubMed ID: 38845306]. https://doi.org/10.1016/j.oooo.2024.03.004.
  • 28.
    Shrivastava PK, Hasan S, Abid L, Injety R, Shrivastav AK, Sybil D. Accuracy of machine learning in the diagnosis of odontogenic cysts and tumors: a systematic review and meta-analysis. Oral Radiol. 2024;40(3):342-56. [PubMed ID: 38530559]. https://doi.org/10.1007/s11282-024-00745-7.
  • 29.
    Xu Y, Quan R, Xu W, Huang Y, Chen X, Liu F. Advances in Medical Image Segmentation: A Comprehensive Review of Traditional, Deep Learning and Hybrid Approaches. Bioengineering (Basel). 2024;11(10). [PubMed ID: 39451409]. [PubMed Central ID: PMC11505408]. https://doi.org/10.3390/bioengineering11101034.
  • 30.
    Naseri S, Shukla S, Hiwale KM, Jagtap MM, Gadkari P, Gupta K, et al. From Pixels to Prognosis: A Narrative Review on Artificial Intelligence’s Pioneering Role in Colorectal Carcinoma Histopathology. Cureus. 2024;16(4). e59171. https://doi.org/10.7759/cureus.59171.
  • 31.
    McGenity C, Clarke EL, Jennings C, Matthews G, Cartlidge C, Freduah-Agyemang H, et al. Artificial intelligence in digital pathology: a systematic review and meta-analysis of diagnostic test accuracy. NPJ Digit Med. 2024;7(1):114. [PubMed ID: 38704465]. [PubMed Central ID: PMC11069583]. https://doi.org/10.1038/s41746-024-01106-8.
  • 32.
    Massa E, Mozeika A, Coolen ACC. Replica analysis of overfitting in regression models for time to event data: the impact of censoring. J Physics A: Mathematical Theoretical. 2024;57(12). https://doi.org/10.1088/1751-8121/ad2e40.
  • 33.
    Larrain C, Torres-Hernandez A, Hewitt DB. Artificial Intelligence, Machine Learning, and Deep Learning in the Diagnosis and Management of Hepatocellular Carcinoma. Livers. 2024;4(1):36-50. https://doi.org/10.3390/livers4010004.
  • 34.
    Dixit S, Kumar A, Srinivasan K. A Current Review of Machine Learning and Deep Learning Models in Oral Cancer Diagnosis: Recent Technologies, Open Challenges, and Future Research Directions. Diagnostics (Basel). 2023;13(7). [PubMed ID: 37046571]. [PubMed Central ID: PMC10093759]. https://doi.org/10.3390/diagnostics13071353.
  • 35.
    McInnes MDF, Moher D, Thombs BD, McGrath TA, Bossuyt PM, and the PD, et al. Preferred Reporting Items for a Systematic Review and Meta-analysis of Diagnostic Test Accuracy Studies: The PRISMA-DTA Statement. JAMA. 2018;319(4):388-96. [PubMed ID: 29362800]. https://doi.org/10.1001/jama.2017.19163.
  • 36.
    Kim P, Seo B, De Silva H. Concordance of clinician, Chat-GPT4, and ORAD diagnoses against histopathology in Odontogenic Keratocysts and tumours: a 15-Year New Zealand retrospective study. Oral Maxillofac Surg. 2024;28(4):1557-69. [PubMed ID: 39060850]. https://doi.org/10.1007/s10006-024-01284-5.
  • 37.
    Rao RS, Shivanna DB, Mahadevpur KS, Shivaramegowda SG, Prakash S, Lakshminarayana S, et al. Deep Learning-Based Microscopic Diagnosis of Odontogenic Keratocysts and Non-Keratocysts in Haematoxylin and Eosin-Stained Incisional Biopsies. Diagnostics (Basel). 2021;11(12). [PubMed ID: 34943424]. [PubMed Central ID: PMC8700488]. https://doi.org/10.3390/diagnostics11122184.
  • 38.
    Florindo JB, Bruno OM, Landini G. Morphological classification of odontogenic keratocysts using Bouligand-Minkowski fractal descriptors. Comput Biol Med. 2017;81:1-10. [PubMed ID: 27992735]. https://doi.org/10.1016/j.compbiomed.2016.12.003.
  • 39.
    Frydenlund A, Eramian M, Daley T. Automated classification of four types of developmental odontogenic cysts. Comput Med Imaging Graph. 2014;38(3):151-62. [PubMed ID: 24411103]. https://doi.org/10.1016/j.compmedimag.2013.12.002.
  • 40.
    Eramian M, Daley M, Neilson D, Daley T. Segmentation of epithelium in H&E stained odontogenic cysts. J Microsc. 2011;244(3):273-92. [PubMed ID: 21974807]. https://doi.org/10.1111/j.1365-2818.2011.03535.x.
  • 41.
    Archana R, Jeevaraj P. Deep learning models for digital image processing: a review. Artificial Intelligence Rev. 2024;57(1). https://doi.org/10.1007/s10462-023-10631-z.
  • 42.
    Rayed ME, Islam S, Niha SI, Jim JR, Kabir MM, Mridha MF. Deep learning for medical image segmentation: State-of-the-art advancements and challenges. Inform Med Unlocked. 2024;47. https://doi.org/10.1016/j.imu.2024.101504.
  • 43.
    Uddin AH, Chen YL, Akter MR, Ku CS, Yang J, Por LY. Colon and lung cancer classification from multi-modal images using resilient and efficient neural network architectures. Heliyon. 2024;10(9). e30625. [PubMed ID: 38742084]. [PubMed Central ID: PMC11089372]. https://doi.org/10.1016/j.heliyon.2024.e30625.
  • 44.
    Feng B, Zhang M, Zhu H, Wang L, Zheng Y. MRI Image Segmentation Model with Support Vector Machine Algorithm in Diagnosis of Solitary Pulmonary Nodule. Contrast Media Mol Imaging. 2021;2021:9668836. [PubMed ID: 34377105]. [PubMed Central ID: PMC8318753]. https://doi.org/10.1155/2021/9668836.
  • 45.
    Noguchi K, Saito I, Namiki T, Yoshimura Y, Nakaguchi T. Reliability of non-contact tongue diagnosis for Sjogren's syndrome using machine learning method. Sci Rep. 2023;13(1):1334. [PubMed ID: 36693892]. [PubMed Central ID: PMC9872069]. https://doi.org/10.1038/s41598-023-27764-4.
  • 46.
    Lu G, Wang D, Qin X, Muller S, Wang X, Chen AY, et al. Detection and delineation of squamous neoplasia with hyperspectral imaging in a mouse model of tongue carcinogenesis. J Biophotonics. 2018;11(3). [PubMed ID: 28921845]. [PubMed Central ID: PMC5839941]. https://doi.org/10.1002/jbio.201700078.
  • 47.
    Lam V, Nguyen T, Bui V, Chung BM, Chang LC, Nehmetallah G, et al. Quantitative scoring of epithelial and mesenchymal qualities of cancer cells using machine learning and quantitative phase imaging. J Biomed Opt. 2020;25(2):1-17. [PubMed ID: 32072775]. [PubMed Central ID: PMC7026523]. https://doi.org/10.1117/1.JBO.25.2.026002.
  • 48.
    Kaczmarczyk R, Wilhelm TI, Martin R, Roos J. Evaluating multimodal AI in medical diagnostics. NPJ Digit Med. 2024;7(1):205. [PubMed ID: 39112822]. [PubMed Central ID: PMC11306783]. https://doi.org/10.1038/s41746-024-01208-3.
  • 49.
    Liu X, Faes L, Kale AU, Wagner SK, Fu DJ, Bruynseels A, et al. A comparison of deep learning performance against health-care professionals in detecting diseases from medical imaging: a systematic review and meta-analysis. Lancet Digit Health. 2019;1(6):e271-97. [PubMed ID: 33323251]. https://doi.org/10.1016/S2589-7500(19)30123-2.
  • 50.
    Chanda T, Hauser K, Hobelsberger S, Bucher TC, Garcia CN, Wies C, et al. Dermatologist-like explainable AI enhances trust and confidence in diagnosing melanoma. Nat Commun. 2024;15(1):524. [PubMed ID: 38225244]. [PubMed Central ID: PMC10789736]. https://doi.org/10.1038/s41467-023-43095-4.
  • 51.
    Acs B, Rantalainen M, Hartman J. Artificial intelligence as the next step towards precision pathology. J Intern Med. 2020;288(1):62-81.
  • 52.
    Shi YJ, Li JP, Wang Y, Ma RH, Wang YL, Guo Y, et al. Deep learning in the diagnosis for cystic lesions of the jaws: a review of recent progress. Dentomaxillofac Radiol. 2024;53(5):271-80. [PubMed ID: 38814810]. [PubMed Central ID: PMC11211683]. https://doi.org/10.1093/dmfr/twae022.
  • 53.
    Ying X. An Overview of Overfitting and its Solutions. J Physics: Conference Series. 2019;1168. https://doi.org/10.1088/1742-6596/1168/2/022022.
  • 54.
    Sheikh M, Coolen AC. Analysis of overfitting in the regularized Cox model. J Physics A: Mathematical Theoretical. 2019;52(38). https://doi.org/10.1088/1751-8121/ab375c.
  • 55.
    Salahuddin Z, Woodruff HC, Chatterjee A, Lambin P. Transparency of deep neural networks for medical image analysis: A review of interpretability methods. Comput Biol Med. 2022;140:105111. [PubMed ID: 34891095]. https://doi.org/10.1016/j.compbiomed.2021.105111.
  • 56.
    Hassija V, Chamola V, Mahapatra A, Singal A, Goel D, Huang K, et al. Interpreting Black-Box Models: A Review on Explainable Artificial Intelligence. Cognitive Computation. 2023;16(1):45-74. https://doi.org/10.1007/s12559-023-10179-8.
  • 57.
    Qamar T, Bawany NZ. Understanding the black-box: towards interpretable and reliable deep learning models. PeerJ Comput Sci. 2023;9. e1629. [PubMed ID: 38077598]. [PubMed Central ID: PMC10702969]. https://doi.org/10.7717/peerj-cs.1629.
  • 58.
    Rokhshad R, Ducret M, Chaurasia A, Karteva T, Radenkovic M, Roganovic J, et al. Ethical considerations on artificial intelligence in dentistry: A framework and checklist. J Dent. 2023;135:104593. [PubMed ID: 37355089]. https://doi.org/10.1016/j.jdent.2023.104593.
  • 59.
    Chauhan C, Gullapalli RR. Ethics of AI in Pathology: Current Paradigms and Emerging Issues. Am J Pathol. 2021;191(10):1673-83. [PubMed ID: 34252382]. [PubMed Central ID: PMC8485059]. https://doi.org/10.1016/j.ajpath.2021.06.011.
  • 60.
    Varsha PS. How can we manage biases in artificial intelligence systems – A systematic literature review. Int J Inform Management Data Insights. 2023;3(1). https://doi.org/10.1016/j.jjimei.2023.100165.
  • 61.
    Schwendicke F, Samek W, Krois J. Artificial Intelligence in Dentistry: Chances and Challenges. J Dent Res. 2020;99(7):769-74. [PubMed ID: 32315260]. [PubMed Central ID: PMC7309354]. https://doi.org/10.1177/0022034520915714.
  • 62.
    Katirai A. The Environmental Costs of Artificial Intelligence for Healthcare. Asian Bioeth Rev. 2024;16(3):527-38. [PubMed ID: 39022383]. [PubMed Central ID: PMC11250743]. https://doi.org/10.1007/s41649-024-00295-4.
  • 63.
    Moyano-Fernandez C, Rueda J, Delgado J, Ausin T. May Artificial Intelligence take health and sustainability on a honeymoon? Towards green technologies for multidimensional health and environmental justice. Glob Bioeth. 2024;35(1):2322208. [PubMed ID: 38476503]. [PubMed Central ID: PMC10930144]. https://doi.org/10.1080/11287462.2024.2322208.
  • 64.
    Bratan T, Heyen NB, Hüsing B, Marscheider-Weidemann F, Thomann J. Hypotheses on environmental impacts of AI use in healthcare. J Climate Change Health. 2024;16:100299. https://doi.org/10.1016/j.joclim.2024.100299.
comments

Leave a comment here


Crossmark
Crossmark
Checking
Share on
Cited by
Metrics

Purchasing Reprints

  • Copyright Clearance Center (CCC) handles bulk orders for article reprints for Brieflands. To place an order for reprints, please click here (   https://www.copyright.com/landing/reprintsinquiryform/ ). Clicking this link will bring you to a CCC request form where you can provide the details of your order. Once complete, please click the ‘Submit Request’ button and CCC’s Reprints Services team will generate a quote for your review.
Search Relations

Author(s):

Related Articles