AIM: Videomics, the application of deep learning (DL) to endoscopic video, enables real-time tissue segmentation and anatomical recognition. Within endoscopic endonasal approaches, these methods may improve intraoperative visualization, tumor delineation, and surgical precision. Despite growing interest, its translation into routine clinical practice is still limited and not yet fully characterized. This systematic review aimed to synthesize current evidence on DL-based segmentation in endoscopic endonasal surgery, focusing on model architectures, segmentation targets, and reported outcomes. METHODS: This review was conducted according to Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) 2020 guidelines. A systematic search of PubMed, Scopus, and Web of Science was performed on 12 January 2025, and updated on 5 June 2025. Studies published between 2018 and 2025 were included, as no eligible studies were available prior to 2018. Studies were included if they involved human endoscopic endonasal procedures and applied DL techniques to endoscopic video for segmentation purposes. Data extraction included sample size, image resolution, annotated datasets, DL architectures, segmentation targets, and model performance metrics. Study quality was assessed using the Newcastle-Ottawa Scale, and descriptive statistics were used to summarize findings. RESULTS: Out of 223 screened articles, 28 studies met the inclusion criteria, encompassing 154,989 patients and 1,028,440 annotated images. The most common segmentation targets included nasal polyps (25%), nasopharyngeal carcinoma (21.4%), and pituitary adenomas (7.14%). ResNet and YOLO architectures were each used in 5 studies (17.9%), while transformer-based models such as Swin Transformer, NasVLM, and NaMA-Mamba were increasingly utilized in recent years. Performance metrics were high across studies: area under the receiver operating characteristic curve (AUC-ROC) ranged from 87.4% to 99.2%, mean intersection over union [IoU] (mIoU) from 61.2% to 81.7%, and mean average precision (mAP) [0.50] from 53.4% to 94.9%. Inference times varied from 0.14 ms to 100 ms per image. However, only 35.7% of studies reported segmentation tools, and dataset heterogeneity was common. CONCLUSIONS: DL-based videomics demonstrates high segmentation accuracy across various pathologies and anatomical targets in endoscopic endonasal surgery. Models such as Swin Transformer and YOLO show potential for real-time surgical support. However, translation into clinical practice remains limited by dataset heterogeneity and variability in reporting.

Deep Learning-Based Videomics for Automatic Segmentation in Endoscopic Endonasal Surgery

Agosti, Edoardo;Rampinelli, Vittorio;Fiorindi, Alessandro;Panciani, Pier Paolo;Fontanella, Marco Maria
2026-01-01

Abstract

AIM: Videomics, the application of deep learning (DL) to endoscopic video, enables real-time tissue segmentation and anatomical recognition. Within endoscopic endonasal approaches, these methods may improve intraoperative visualization, tumor delineation, and surgical precision. Despite growing interest, its translation into routine clinical practice is still limited and not yet fully characterized. This systematic review aimed to synthesize current evidence on DL-based segmentation in endoscopic endonasal surgery, focusing on model architectures, segmentation targets, and reported outcomes. METHODS: This review was conducted according to Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) 2020 guidelines. A systematic search of PubMed, Scopus, and Web of Science was performed on 12 January 2025, and updated on 5 June 2025. Studies published between 2018 and 2025 were included, as no eligible studies were available prior to 2018. Studies were included if they involved human endoscopic endonasal procedures and applied DL techniques to endoscopic video for segmentation purposes. Data extraction included sample size, image resolution, annotated datasets, DL architectures, segmentation targets, and model performance metrics. Study quality was assessed using the Newcastle-Ottawa Scale, and descriptive statistics were used to summarize findings. RESULTS: Out of 223 screened articles, 28 studies met the inclusion criteria, encompassing 154,989 patients and 1,028,440 annotated images. The most common segmentation targets included nasal polyps (25%), nasopharyngeal carcinoma (21.4%), and pituitary adenomas (7.14%). ResNet and YOLO architectures were each used in 5 studies (17.9%), while transformer-based models such as Swin Transformer, NasVLM, and NaMA-Mamba were increasingly utilized in recent years. Performance metrics were high across studies: area under the receiver operating characteristic curve (AUC-ROC) ranged from 87.4% to 99.2%, mean intersection over union [IoU] (mIoU) from 61.2% to 81.7%, and mean average precision (mAP) [0.50] from 53.4% to 94.9%. Inference times varied from 0.14 ms to 100 ms per image. However, only 35.7% of studies reported segmentation tools, and dataset heterogeneity was common. CONCLUSIONS: DL-based videomics demonstrates high segmentation accuracy across various pathologies and anatomical targets in endoscopic endonasal surgery. Models such as Swin Transformer and YOLO show potential for real-time surgical support. However, translation into clinical practice remains limited by dataset heterogeneity and variability in reporting.
File in questo prodotto:
File Dimensione Formato  
Videomics_Automatic_Segmentation_Endonasal_Surgery.pdf

accesso aperto

Tipologia: Full Text
Licenza: Non specificato
Dimensione 1.94 MB
Formato Adobe PDF
1.94 MB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11379/642125
 Attenzione

Attenzione! I dati visualizzati non sono stati sottoposti a validazione da parte dell'ateneo

Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus ND
  • ???jsp.display-item.citation.isi??? ND
social impact