Development and validation of a recommended checklist for assessment of surgical videos quality: the LAParoscopic surgery Video Educational GuidelineS (LAP-VEGaS) video assessment tool

Introduction There has been a constant increase in the number of published surgical videos with preference for open-access sources, but the proportion of videos undergoing peer-review prior to publication has markedly decreased, raising questions over quality of the educational content presented. Th...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Surgical endoscopy 2021-03, Vol.35 (3), p.1362-1369
Hauptverfasser: Celentano, Valerio, Smart, Neil, Cahill, Ronan A., Spinelli, Antonino, Giglio, Mariano Cesare, McGrath, John, Obermair, Andreas, Pellino, Gianluca, Hasegawa, Hirotoshi, Lal, Pawanindra, Lorenzon, Laura, De Angelis, Nicola, Boni, Luigi, Gupta, Sharmila, Griffith, John P., Acheson, Austin G., Cecil, Tom D., Coleman, Mark G.
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:Introduction There has been a constant increase in the number of published surgical videos with preference for open-access sources, but the proportion of videos undergoing peer-review prior to publication has markedly decreased, raising questions over quality of the educational content presented. The aim of this study was the development and validation of a standard framework for the appraisal of surgical videos submitted for presentation and publication, the LAParoscopic surgery Video Educational GuidelineS (LAP-VEGaS) video assessment tool. Methods An international committee identified items for inclusion in the LAP-VEGaS video assessment tool and finalised the marking score utilising Delphi methodology. The tool was finally validated by anonymous evaluation of selected videos by a group of validators not involved in the tool development. Results 9 items were included in the LAP-VEGaS video assessment tool, with every item scoring from 0 (item not presented in the video) to 2 (item extensively presented in the video), with a total marking score ranging from 0 to 18. The LAP-VEGaS video assessment tool resulted highly accurate in identifying and selecting videos for acceptance for conference presentation and publication, with high level of internal consistency and generalisability. Conclusions We propose that peer review in adherence to the LAP-VEGaS video assessment tool could enhance the overall quality of published video outputs. Graphic Abstract
ISSN:0930-2794
1432-2218
DOI:10.1007/s00464-020-07517-4