Abstract
Pediatric surgery training is challenged by declining clinical exposure and limited access to high-quality educational resources. Advances in generative artificial intelligence (AI) allow enhancement of educational videos without professional studio infrastructure, but evidence supporting their educational value remains limited. To evaluate the educational impact and cost efficiency of locally deployed AI-enhanced lecture videos compared with standard faculty-recorded videos in pediatric surgery training. A quasi-experimental, non-randomized, post-test-only study conducted at Al-Azhar University, Cairo, Egypt between December 2024 and September 2025. Medical students and pediatric surgery residents viewed standardized lecture videos on five core pediatric surgery topics. Identical scripts and narration were used in both groups; the intervention differed only by AI-based background enhancement using open-source, locally deployed tools. Primary outcomes were perceived educational clarity (five-point Likert scale) and knowledge acquisition assessed by a 40-item multiple-choice examination. Secondary outcomes included visual comfort, perceived professionalism, and willingness to recommend the format. Between-group comparisons and effect sizes were calculated. A cost analysis compared AI-enhanced production with traditional studio-based methods. A total of 240 participants were included (120 per group). AI-enhanced videos were associated with higher educational clarity scores (mean 4.52 vs. 3.88; p < 0.001) and higher knowledge scores (76.8% vs. 69.5%; p < 0.001). Secondary outcomes consistently favored AI-enhanced videos, including visual comfort and perceived professionalism. Locally deployed AI production reduced per-video costs by 72–94% compared with traditional studio production. Locally deployed AI-enhanced videos improved short-term educational outcomes and substantially reduced production costs while preserving presenter authenticity. This approach represents a practical, scalable strategy for enhancing pediatric surgery education, particularly in resource-limited settings. Further studies are needed to assess long-term retention and clinical impact.
Introduction
Pediatric surgery education faces significant challenges in contemporary medical training. General surgery residents complete only 33.8 designated pediatric cases (3.7% of total operative volume)1, with the ACGME requiring only 20 cases for graduation2. Nearly 60% of pediatric surgery experience concentrates in inguinal and umbilical hernia repairs3. Graduating residents demonstrate variable competence in pediatric procedures, with 97.3–97.6% competent for umbilical hernia repairs but only 79.5% competent for inguinal hernia repair in infants under 6 months4. Pediatric surgery case volume for general surgery residents has decreased by approximately 30.7% over the past decade5.
The COVID-19 pandemic exacerbated these challenges. Pediatric surgery residents experienced 46–57% reduction in operative procedures, with 70% reporting decreased surgical confidence6. These disruptions highlighted the need for innovative educational approaches that can supplement limited operative exposure.
Contemporary medical education increasingly embraces Education 4.0 principles emphasizing student-centered, technology-enhanced learning. Video-based education has emerged as a valuable tool in surgical training. Meta-analyses demonstrate equivalent effectiveness to traditional teaching7. Studies demonstrate significant normalized learning gains from video-based education, with interactive videos achieving 49–70% gains depending on format8. Video-based learning shows large effects on knowledge acquisition in dentistry (Cohen’s d = 2.18) and moderate effects in medicine (Cohen’s d = 0.67)9. These findings support video as an effective educational modality, though quality and production values vary considerably.
The Cognitive Theory of Multimedia Learning suggests that learners process visual and verbal information through limited cognitive channels, and that well-designed multimedia presentations can reduce extraneous cognitive load and support comprehension10. Enhanced visual presentation may improve engagement and learning outcomes without altering content delivery.
However, video production in non-studio settings requires optimization of set, lighting, sound, and video equipment, with few formal resources existing to guide educators for high-quality video production11. High-quality video production typically requires professional studios with substantial costs, including studio rental, equipment, and videographer services. Green screen chroma keying presents substantial technical challenges: uneven lighting causing shadows and hot spots, color spill requiring extensive correction, shadows interfering with keying, six-foot separation requirements, wardrobe restrictions, even diffused lighting across entire screen, and extensive post-production color matching. Green screen infrastructure requires dedicated studio space, professional costly lighting equipment, and technical expertise, with periodic replacement needed to avoid student disengagement.
Artificial intelligence (AI) offers new possibilities for educational video enhancement. Generative AI can transform educational content into multimodal formats, potentially improving comprehension12. However, current AI applications in surgical education face limitations including bias in algorithms, concerns about transparency, and risks of over-reliance13.
Concerns also exist regarding the use of generative AI in medical education, particularly related to anatomical accuracy and visual realism. Fully synthetic AI-generated presenters may introduce distortions or implausible features, underscoring the importance of controlled and limited AI application. Current generative AI models can produce “hallucinations,” and text-to-video models may generate anatomically incorrect motion which is defined as convincing yet false outputs14.
Online AI video enhancement platforms often produce color mismatches and inconsistent results. Local deployment with selective masking offers superior control. No published studies directly compare locally deployed AI-enhanced videos to standard faculty-recorded videos for pediatric surgical teaching. This represents a critical research opportunity, particularly given declining exposure and need for accessible, cost-effective solutions in resource-limited settings.
Our objective was to evaluate the impact of locally deployed AI-enhanced video presentation on educational clarity and knowledge acquisition in pediatric surgical teaching compared with standard recordings, and to assess learner satisfaction and production costs. We hypothesized that AI-enhanced videos would improve educational outcomes while substantially reducing costs.
Methods
Study design and participants
This quasi-experimental, non-randomized, educational evaluation used a post-test-only design conducted at Faculty of Medicine, Al-Azhar University, Cairo, Egypt. (December 2024–September 2025). Participation was voluntary, and informed consent was obtained electronically or verbally before participation in the study. All study procedures were conducted in accordance with the ethical standards of the institutional research committee and with the principles of the Declaration of Helsinki and its later amendments. Figure 1 presents the study flow diagram.
Participation was voluntary. Eligible participants included clinical medical students (years 3–5) and pediatric surgery residents (postgraduate years ≥ 1). Exclusion criteria were prior exposure to study videos or pediatric surgery fellowship enrollment. Participants were allocated non-randomly based on scheduled teaching sessions, which were predetermined by the academic timetable and independent of individual learner characteristics. Sessions were separated by ≥ 48 h to reduce contamination between groups.
Flow diagram of the study.
Participants were informed during the consent process that the study was evaluating different educational video formats for pediatric surgery teaching, but they were not told which format they would view or the specific hypothesis being tested. After viewing their assigned video, participants could potentially recognize differences in visual presentation quality between formats. This lack of complete blinding may have influenced subjective ratings, particularly for perceived professionalism and willingness to recommend.
Educational content
Five topics were selected based on clinical relevance: functional constipation, undescended testes, bowel management programs, hypospadias, and ureteropelvic junction obstruction. Standardized 6–10 min lecture scripts were developed by board-certified pediatric surgeons. Identical scripts, narration, and illustrations were used in both groups to isolate visual enhancement effects.
Video production
Recording protocol Videos were recorded using smartphone camera (4 K) on tripod with faculty against plain walls. No specialized studio or green screen was required. Audio was captured using low-cost lavalier microphone ($20–50), with professional quality achieved through Adobe Podcast’s free web-based enhancement tool.
Standard video production Standard videos underwent basic editing including trimming silences, removing pauses, color adjustment, and overlay of expert-curated illustrations using DaVinci Resolve (1–2 h per video, common to both conditions).
AI-Enhanced video production AI processing was performed exclusively on still frame images extracted from source video. All video assembly occurred in DaVinci Resolve. This frame-based approach provided precise control over processed regions (background) versus preserved regions (presenter), avoiding color mismatches common with automated platforms. Figure 2 demonstrates the ComfyUI inpainting workflow.
AI inpainting workflow in ComfyUI18.
The workflow consisted of: (1) movement analysis; (2) frame extraction at HD resolution (1920 × 1080); (3) masking creation delineating presenter regions (preserved) from background regions (replaced); (4) AI background replacement using FLUX.1-Fill (inpainting) and FLUX.1-dev (development) models15—two distinct models where Fill handles inpainting tasks and dev provides fine control and quality—with outpainting techniques extending spatial boundaries to create expansive studio impressions; (5) frame import as overlay in DaVinci Resolve; (6) central masking to reveal original presenter; (7) gradient feathering (50–100 pixels) for natural blending; (8) frame-by-frame review; (9) hand gesture correction via layer duplication where movements extended into AI background; (10) color matching (inherently well-matched through AI inpainting, eliminating extensive correction typical of green screen); and (11) illustration overlay (0.5–1 h per video).
AI processing was strictly limited to background enhancement. No alterations were made to presenter’s facial features, body proportions, gestures, voice, or speech content. This avoided the risk of anatomical distortions (extra fingers, distorted features) that occur in fully AI-generated videos where presenters are synthetic16,17. All authentic presenter features remained entirely unmodified. Figure 3 and supplementary Video 1 demonstrate the transition from plain wall recording to studio-enhanced view.
Transformation of a plain wall (A) into a well-decorated studio setting (B) achieved through AI-based inpainting. Background clutter such as windows and doors was removed, with seamless color matching and realistic integration of decorative elements and light reflections on the wall. Panels A and B feature the first author during the video capture process. Panels C and F depict two different constipation-related pathologies, each presented in simplified conceptual formats. Panel D illustrates the gastrocolic reflex. Panel E provides an additional fully AI-generated caricature showing a doctor humorously chasing the cause of constipation. Panels B, C, D, E, and F were AI-manipulated or fully created using our workflow and FLUX.1 models in ComfyUI.18.
Technical implementation AI enhancement used FLUX.1-Fill and FLUX.1-dev models via ComfyUI18, an open-source workflow interface, with models loaded using GGUF Q8 quantization19reducing computational requirements to enable operation on consumer-grade laptop (Lenovo Legion 5 Pro with 6GB VRAM). GGUF Q8 quantization was selected for publication-grade quality. Inpainting used 20–30 sampling steps. Individual frame generation required 1–2 min. AI processing time ranged from 30 min to 3 h per video (iterative generation for aesthetic quality, as learner perception of professionalism was a study outcome). Post-AI processing in DaVinci Resolve required approximately 1 h. All AI-generated visual elements underwent board-certified pediatric surgeon review.
Outcome measures
Outcome assessment was performed immediately after viewing using REDCap20. Primary outcomes were educational clarity (three-item, five-point Likert scale assessing presentation clarity, visual support effectiveness, and confidence in understanding) and knowledge acquisition (percentage correct on a 40-item multiple-choice examination).
The knowledge test items were mapped to predefined learning objectives for each topic and reviewed by board-certified pediatric surgeons for content relevance and clarity prior to administration. The clarity scale was designed to capture key dimensions of instructional quality (presentation clarity, visual support, and learner confidence).
Secondary outcomes included visual comfort, perceived professionalism, willingness to recommend, and perceived improvement compared with textbook-based learning.
Statistical analysis
Continuous variables were summarized as mean ± SD. Between-group comparisons used independent-samples t-tests; categorical outcomes used chi-square tests. Effect sizes were calculated using Cohen’s d. Multivariable linear regression assessed independent associations adjusting for training level, age, prior exposure, and baseline familiarity. Significance was p < 0.05.
Economic analysis
Cost comparison estimated per-video production costs including studio rental, equipment, services, hardware amortization, faculty time, and illustrations. Faculty time was valued at $50/hour for base analysis.
Results
Participant characteristics
Between December 2024 and September 2025, 256 trainees were screened. Sixteen (6.3%) were excluded: twelve (4.7%) had prior exposure, four (1.6%) were in fellowship. Total 240 participants were included: 120 (50%) per condition, with zero attrition. The cohort consisted predominantly of undergraduate medical students (n = 210, 87.5%), with smaller proportions of junior residents PGY 1–2 (n = 20, 8.3%) and senior residents PGY ≥ 3 (n = 10, 4.2%). Mean age was 23.7 years (SD 2.2, range 21–29). Baseline characteristics were well balanced (Table 1). Prior clinical exposure: 87 (72.5%) standard versus 88 (73.3%) AI-enhanced (χ²=0.02, p = 0.89). Baseline familiarity: mean 2.8 (SD 1.1) standard versus 2.9 (SD 1.2) AI-enhanced (t = 0.43, p = 0.67).
Primary educational outcomes
Educational clarity scores were significantly higher among AI-enhanced video participants (Table 2). Mean overall clarity was 4.52 (SD 0.46) AI-enhanced versus 3.88 (SD 0.58) standard, representing difference of 0.64 points (95% CI 0.51–0.77; t = 11.37, p < 0.001) and very large effect size (Cohen’s d = 1.24; 95% CI 1.00–1.48.00.48). All three scale components demonstrated superiority: presentation clarity (4.56 vs. 3.82, p < 0.001), visual support effectiveness (4.51 vs. 3.89, p < 0.001), and confidence in understanding (4.49 vs. 3.93, p < 0.001).
Knowledge acquisition was significantly higher in AI-enhanced group. Mean scores were 76.8% (SD 9.2) AI-enhanced versus 69.5% (SD 10.4) standard, representing difference of 7.3% points (95% CI 4.9–9.7; t = 5.93, p < 0.001) and large effect size (Cohen’s d = 0.76; 95% CI 0.53–0.99)—equivalent to approximately three additional correct answers per participant. Proportion achieving ≥ 80% was significantly greater in AI-enhanced group (46/120, 38.3% vs. 26/120, 21.7%; χ²=8.23, p = 0.004, OR = 2.25, 95% CI 1.28–3.97). Fewer AI-enhanced participants scored < 60% (7/120, 5.8% vs. 19/120, 15.8%; χ²=6.14, p = 0.013, OR = 0.33, 95% CI 0.13–0.82).
Secondary outcomes and subgroup analyses
Secondary outcomes consistently favored AI-enhanced videos. Participants reported greater visual comfort (mean 4.48 vs. 3.91; t = 8.16, p < 0.001; d = 1.07) and professionalism (mean 4.51 vs. 3.89; t = 8.62, p < 0.001; d = 1.13). They were significantly more likely to recommend AI-enhanced format (105/120, 87.5% vs. 76/120, 63.3%; χ²=18.82, p < 0.001, OR = 4.02) and perceive it as superior to textbook learning (101/120, 84.2% vs. 71/120, 59.2%; χ²=18.22, p < 0.001, OR = 3.68).
Knowledge improvements were observed across all five topics. Mean differences consistently favored AI-enhanced condition with significant gains: functional constipation (7.6% points, p < 0.001, d = 0.71), undescended testes (7.3% points, p < 0.001, d = 0.72), bowel management (5.4% points, p = 0.001, d = 0.68), hypospadias (8.7% points, p < 0.001, d = 0.84), and ureteropelvic junction obstruction (7.4% points, p < 0.001, d = 0.78). No single topic disproportionately drove overall effect.
Educational benefits were consistent across training levels. Among undergraduate students (n = 210), AI-enhanced videos showed higher clarity (difference 0.66 points, 95% CI 0.53–0.79; p < 0.001; d = 1.27) and knowledge (difference 7.6% points, 95% CI 5.1–10.1; p < 0.001; d = 0.79). Among junior residents (n = 20), AI-enhanced videos showed higher clarity (difference 0.52 points, 95% CI 0.01–1.03; p = 0.048; d = 0.99) and knowledge (difference 5.7% points, p = 0.028; d = 0.56), though power was limited.
After adjustment for training level, age, prior exposure, and baseline familiarity, AI-enhanced videos remained independently associated with higher clarity (adjusted estimate 0.62 points, 95% CI 0.49–0.75, p < 0.001) and knowledge (7.1% points, 95% CI 4.7–9.5, p < 0.001). No significant interaction between video condition and training level was observed for clarity (F = 0.84, p = 0.43) or knowledge (F = 0.65, p = 0.52).
Economic analysis
Cost comparison demonstrated substantial economic advantage for AI-enhanced production (Table 3). Professional studio production typically requires studio rental ($500-2,000/day), professional equipment ($1,000–5,000/session), and videographer/editing services ($500-1,500/video), totaling $2,000–8,500 per video. AI-enhanced approach required one-time hardware investment (~$1,200, amortized to $21–33 per video over three years assuming 36 videos) and faculty time. Total active faculty time was 7–10 h per video at $50/hour ($350–500). Total per-video cost was $371–533 without illustrations or $396–583 with illustrations, representing 72–94% cost reduction. Break-even was reached after 1–2 videos. For five-video series, total savings ranged from $8,000 to $40,000. Cost advantage persists even valuing faculty time at institutional rates ($100–200/hour), as total AI-enhanced costs ($700-2,000/video) remain below professional studio ($2,000–8,500/video).
Discussion
This study demonstrates that locally deployed AI-enhanced educational videos significantly improve perceived educational clarity and objective knowledge acquisition in pediatric surgical teaching compared with standard faculty-recorded videos, with consistent improvements across learner levels and topics, while substantially reducing costs. The approach eliminates green screen infrastructure and its technical challenges.
The improvement observed addresses a critical gap. Pediatric surgery exposure remains severely limited, with residents completing only 33.8 designated cases (3.7% of volume), down from 47.7 cases, with ACGME requiring just 20 cases1,2. In many medical schools, pediatric surgery is taught within general surgery curriculum for undergraduates, making video-based resources essential. Video-based education has become essential as Education 4.0 emphasizes technology-enhanced, student-centered learning.
The knowledge improvement magnitude (7.3% points, Cohen’s d = 0.76) compares favorably with prior meta-analyses showing large effects in dentistry (d = 2.18) and moderate effects in medicine (d = 0.67)9. Studies demonstrate significant normalized learning gains from video-based education, with interactive videos achieving 49–70% gains depending on format8. Our results suggest AI enhancement provides additive value beyond standard recording. The very large effect size for perceived clarity (d = 1.24) likely reflects isolation of visual presentation as the sole manipulated variable, with identical content across groups.
This aligns with Cognitive Theory of Multimedia Learning, which posits that multimedia learning engages separate channels for processing words and pictures, with limited working memory requiring appropriate cognitive processing10. Three important goals are reducing extraneous processing, managing essential processing, and fostering generative processing. Enhanced visual presentation may improve engagement without altering content delivery, consistent with student-centered principles emphasizing learner autonomy and technology integration. Dual coding theory suggests content presented through verbal and visual channels improves information processing. An AI Educational Video Assistant designed using these principles received positive evaluations across engagement, organization, clarity, and usability21.
Findings are particularly relevant for programs lacking professional media resources. Locally deployed, open-source AI tools offer pragmatic solution preserving faculty autonomy, avoiding recurring subscriptions, and maintaining institutional control. The 72–94% cost reduction has particular relevance for resource-limited settings where traditional infrastructure is unavailable. The most expensive component of video-based education modules is time cost, which is highly variable depending on the level of clinical seniority of the individuals involved22.
This aligns with Education 4.0 goals. Practical solutions include recording sessions onto DVDs, reaching over 140 surgical trainees across four countries23. The UN Global Surgery Learning Hub provides e-learning courses to over 11,000 learners from 190 countries, with 55% from low- or lower-middle income countries24. Our approach complements these efforts by enabling local production without studios or green screen infrastructure.
Professional-quality enhancement was achieved using consumer-grade hardware through GGUF quantization techniques19. GGUF Q8 quantization reduced computational requirements, enabling operation on standard consumer laptops with modest VRAM (6GB), making this technology particularly accessible for pediatric surgery education in low- and middle-income countries. Our frame-based workflow, in which AI processing was restricted to background regions, preserved presenter authenticity while avoiding the technical limitations of green screen recording. This approach reduced lighting and color-matching constraints and enabled visually consistent backgrounds without specialized studio infrastructure.
Additional innovations included outpainting to extend spatial boundaries, creating expansive studios from confined spaces; HD-resolution AI processing with 4 K output; elimination of green screen infrastructure and replacement costs; and AI audio enhancement using free tools. These collectively reduce barriers to high-quality production. Commercially available technology allows filming high-quality content at minimal cost using smartphone cameras25. Our method bridges this gap by combining low-cost recording with AI-based enhancement.
The American College of Surgeons multidisciplinary consensus establishes that AI methods and AI-enabled metrics should have strong evidence of validity to facilitate adoption by surgeons (90% agreement), and that benchmarking against expert assessments is essential for skill assessment (90% agreement), though this is susceptible to biases inherent in subjective human opinions (85% agreement)26. Only 44% of published AI models in pediatric surgery were interpretable, just 6% were both interpretable and externally validated, and 40% had high risk of bias27. Our approach limits AI processing to background replacement while preserving all authentic presenter features through selective masking, with board-certified pediatric surgeon validation.
In this study, AI functioned strictly as a visual enhancement tool rather than a content generator. By restricting processing to background elements, anatomical accuracy and presenter authenticity were preserved, with all materials reviewed by board-certified pediatric surgeons. This controlled use of AI aligns with current consensus recommendations for responsible adoption in surgical education.
From practical implementation perspective, AI-enhanced production required more faculty time (7–10 vs. 2–4 h) but eliminated professional service costs ($2,000–8,500) and green screen infrastructure. The high cost of producing good-quality video-films for teaching has historically hindered the use of this method28. This trade-off favors AI enhancement when faculty time is institutionally subsidized, making the approach attractive for academic centers where faculty contributions are valued but external budgets are limited. Even valuing faculty time at institutional rates ($100–200/hour), cost advantages persist.
Key cost ingredients for video-based education include faculty time (median >$170 per physician), equipment (median >$170), and learner time (median $525 per physician)29. Programs prioritizing efficiency could reduce time by accepting functional rather than refined backgrounds. The primary barrier to effective AI integration is not technological sophistication, but critical infrastructure deficiencies including institutional implementation structures, sustainable funding, and rigorous research methodologies30. Our approach addresses technological and cost barriers while acknowledging that successful implementation requires institutional support structures.
The transition from language models to multimodal visual AI represents an important paradigm shift12. Multimodal AI utilizing contrastive language-image pretraining (CLIP) enables sophisticated manipulation of educational media previously requiring professional studios. Generative AI can transform educational content into multimodal formats, potentially improving comprehension through alignment with Cognitive Theory of Multimedia Learning13. The FLUX.1-Fill and FLUX.1-dev models represent state-of-the-art inpainting technology, with Fill handling inpainting tasks and dev providing fine control and quality optimization. This enables precise background replacement while preserving presenter authenticity, critical requirements that would be compromised by fully automated approaches15.
AI holds significant promise for personalizing health professional education, with an increasing trend in AI applications, particularly in surgical education31. AI’s affordances—expressivity, interactivity, multimodality—map to educational experiences such as interactive tutoring, though meaningful use demands critical awareness of AI’s operational limits, ethical boundaries, and technical mechanisms.
Although video-based education and AI-assisted learning have each demonstrated educational value7, the specific use of locally deployed AI for visual enhancement of pediatric surgery lecture videos has not previously been evaluated. The present findings therefore provide preliminary evidence for this application.
The greatest promise lies in hybrid models where AI augments rather than replaces faculty mentorship. AI has the potential to be a powerful adjunct to surgical education if deployed thoughtfully and ethically, complementing rather than replacing the vital educator-learner relationship32. AI cannot replace the critical role of educators and must be used as complementary tool with robust governance frameworks.
Limitations
This study has several limitations. First, the quasi-experimental, non-randomized, post-test-only design introduces potential selection bias and limits causal inference, although baseline characteristics were well balanced and multivariable adjustment was performed. Second, outcomes were assessed immediately after video viewing; therefore, long-term knowledge retention and transfer to clinical performance were not evaluated. Third, although the knowledge assessment instrument was mapped to learning objectives and reviewed by faculty, formal psychometric validation was not performed. This may limit interpretation of absolute score levels, although relative between-group differences remain informative. Fourth, the observed improvements in perceived clarity may partially reflect enhanced production quality and novelty effects rather than instructional superiority alone. Fifth, the study was conducted at a single institution with a predominance of undergraduate learners, which may limit generalizability to other training environments or learner populations. Finally, the intervention evaluated lecture-based educational videos and does not address operative or procedural video instruction, where different educational dynamics may apply.
Conclusions
Locally deployed AI-enhanced educational videos were associated with significantly improved perceived clarity and short-term knowledge acquisition compared with standard faculty-recorded videos, while achieving substantial reductions in production costs. By restricting AI use exclusively to environmental enhancement and preserving all authentic presenter features, this approach addresses key concerns regarding anatomical accuracy and educational integrity. These findings suggest that practical, open-source AI tools can enhance the accessibility and professionalism of pediatric surgery educational content, particularly in resource-limited settings. Further multicenter, randomized studies with longitudinal follow-up are needed to assess knowledge retention, clinical impact, and broader generalizability across surgical training contexts.
Data availability
The datasets generated and analyzed during the current study are available from the corresponding author upon reasonable request.
References
Gow, K. W., Drake, F. T., Aarabi, S. & Waldhausen, J. H. The ACGME case log: general surgery resident experience in pediatric surgery. J. Pediatr. Surg. 48 (8), 1643–1649. https://doi.org/10.1016/j.jpedsurg.2012.09.027 (2013).
Talutis, S., McAneny, D., Chen, C., Doherty, G. & Sachs, T. Trends in pediatric surgery operative volume among residents and fellows: improving the experience for all. J. Am. Coll. Surg. 222 (6), 1082–1088. https://doi.org/10.1016/j.jamcollsurg.2015.11.034 (2016).
Esparaz, J. R. et al. The diminishing experience in pediatric surgery for general surgery residents in the united States. J. Pediatr. Surg. 56 (7), 1219–1221. https://doi.org/10.1016/j.jpedsurg.2021.03.042 (2021).
Moreci, R. et al. Readiness of graduating general surgery residents to perform common pediatric surgery procedures. J. Surg. Educ. 82 (1), 103318. https://doi.org/10.1016/j.jsurg.2024.103318 (2025).
Tuma, F., Malgor, R. D., Kapila, N. & Kamel, M. K. Trends of core versus subspecialty operative procedures completed in general surgery residency training programs. Postgrad. Med. J. 98 (1164), 772–777. https://doi.org/10.1136/postgradmedj-2021-140503 (2022).
Lerendegui, L. et al. Impact of the COVID-19 pandemic on surgical skills training in pediatric surgery residents. Pediatr. Surg. Int. 37 (10), 1415–1420. https://doi.org/10.1007/s00383-021-04961-8 (2021).
Mao, B. P. et al. Is online Video-Based education an effective method to teach basic surgical skills to students and surgical trainees? A systematic review and Meta-analysis. J. Surg. Educ. 79 (6), 1536–1545. https://doi.org/10.1016/j.jsurg.2022.07.016 (2022).
Robson, A., Scantling-Birch, Y., Morton, S., Anbu, D. & Border, S. Assessing the impact of interactive educational videos and screencasts within Pre-clinical microanatomy and medical physiology teaching. Adv. Exp. Med. Biol. 1356, 319–343. https://doi.org/10.1007/978-3-030-87779-8_14 (2022).
Morgado, M. et al. Full title: Video-based approaches in health education: a systematic review and meta-analysis. Sci. Rep. 14 (1), 23651. https://doi.org/10.1038/s41598-024-73671-7 (2024).
Mayer, R. E. Applying the science of learning to medical education. Med. Educ. 44 (6), 543–549. https://doi.org/10.1111/j.1365-2923.2010.03624.x (2010).
Castillo, S. et al. Production processes for creating educational videos. CBE Life Sci. Educ. 20 (2), es7. https://doi.org/10.1187/cbe.20-06-0120 (2021).
Boscardin, C. K., Gin, B., Golde, P. B. & Hauer, K. E. ChatGPT and generative artificial intelligence for medical education: potential impact and opportunity. Acad. Med. 99 (1), 22–27. https://doi.org/10.1097/ACM.0000000000005439 (2024).
Boscardin, C. K., Abdulnour, R. E. & Gin, B. C. Macy foundation innovation report part I: current landscape of artificial intelligence in medical education. Acad. Med. 100 (9S Suppl 1), S15–S21. https://doi.org/10.1097/ACM.0000000000006107 (2025).
Menz, B. D., Modi, N. D., Sorich, M. J. & Hopkins, A. M. Health disinformation use case highlighting the urgent need for artificial intelligence vigilance: weapons of mass disinformation. JAMA Intern. Med. 184 (1), 92–96. https://doi.org/10.1001/jamainternmed.2023.5947 (2024).
Forest Labs B et al. FLUX.1 kontext: flow matching for In-Context image generation and editing in latent space. ArXiv e-prints. arXiv:2506.15742 https://doi.org/10.48550/arXiv.2506.15742 (2025).
Mohamed, A. A. & Lucke-Wold, B. Text-to-video generative artificial intelligence: Sora in neurosurgery. Neurosurg. Rev. 47 (1), 272. https://doi.org/10.1007/s10143-024-02514-w (2024).
Robinson, J. R. et al. Generative artificial intelligence in academic surgery: ethical implications and transformative potential. J. Surg. Res. 307, 212–220. https://doi.org/10.1016/j.jss.2024.12.059 (2025).
ComfyUI [Computer software]. December (2024). https://github.com/comfyanonymous/ComfyUI Accessed.
City96. ComfyUI-GGUF: GGUF Quantization support for native ComfyUI models. GitHub repository. December (2024). https://github.com/city96/ComfyUI-GGUF Accessed.
Harris, P. A. et al. Research electronic data capture (REDCap)--a metadata-driven methodology and workflow process for providing translational research informatics support. J. Biomed. Inf. 42 (2), 377–381. https://doi.org/10.1016/j.jbi.2008.08.010 (2009).
Doherty, C. An investigation into the relationship between multimedia lecture design and learners’ engagement behaviours using web log analysis. PLoS One. 17 (8), e0273007. https://doi.org/10.1371/journal.pone.0273007 (2022).
Axelsson, C. G. S. et al. Applying a novel cost-evaluation framework to assess video-based neurosurgery education. Am. J. Surg. 220 (3), 604–609. https://doi.org/10.1016/j.amjsurg.2020.01.018 (2020).
Hadley, G. P. & Mars, M. e-Education in paediatric surgery: a role for recorded seminars in areas of low bandwidth in sub-Saharan Africa. Pediatr. Surg. Int. 27 (4), 407–410. https://doi.org/10.1007/s00383-010-2806-z (2011).
O’Flynn, E. et al. Advancing Open-Access education for the surgical team worldwide: the development and rollout of the united nations global surgery learning hub (SURGhub). World J. Surg. 49 (7), 1700–1707. https://doi.org/10.1002/wjs.12661 (2025).
van Kesteren, J. et al. Technical perspective for video based assessment of surgeries in Low-Resource settings. J. Surg. Educ. 80 (4), 495–498. https://doi.org/10.1016/j.jsurg.2023.01.003 (2023).
Vedula, S. S. et al. Artificial intelligence methods and artificial intelligence-Enabled metrics for surgical education: A multidisciplinary consensus. J. Am. Coll. Surg. 234 (6), 1181–1192. https://doi.org/10.1097/XCS.0000000000000190 (2022).
Elahmedi, M., Sawhney, R., Guadagno, E., Botelho, F. & Poenaru, D. The state of artificial intelligence in pediatric surgery: A systematic review. J. Pediatr. Surg. 59 (5), 774–782. https://doi.org/10.1016/j.jpedsurg.2024.01.044 (2024).
Lewis, L., Jones, J. & Haynes, E. Low-cost video-films in the teaching of undergraduate and postgraduate medical students. J. Telemed Telecare. 6 (Suppl 2), S45–S47. https://doi.org/10.1258/1357633001935545 (2000).
Cook, D. A., Wilkinson, J. M. & Foo, J. Costs of physician continuous professional development: A systematic review. Acad. Med. 97 (10), 1554–1563. https://doi.org/10.1097/ACM.0000000000004805 (2022).
Turner, L., Zhou, C. & Burk-Rafel, J. It takes more than enthusiasm: the missing infrastructure to unlock ai’s potential in medical education. Acad. Med. 100 (9S Suppl 1), S34–S8. https://doi.org/10.1097/ACM.0000000000006104 (2025).
Almansour, M. & Alfhaid, F. M. Generative artificial intelligence and the personalization of health professional education: A narrative review. Med. (Baltim). 103 (31), e38955. https://doi.org/10.1097/MD.0000000000038955 (2024).
Kewalramani, D. et al. Position statement from the society of university surgeons, surgical education committee: artificial intelligence in surgical training for medical students, residents, and fellows. Surgery 190, 109849. https://doi.org/10.1016/j.surg.2025.109849 (2026).
Funding
Open access funding provided by The Science, Technology & Innovation Funding Authority (STDF) in cooperation with The Egyptian Knowledge Bank (EKB).
Author information
Authors and Affiliations
Contributions
A.A. conceived the original idea and developed the core technique introduced in this study. He also led the conceptualization, course design, media production, literature review, synthesis, and manuscript drafting. M.A. contributed to the AI framework, technical solutions, and related editorial work. N.E., R.B., and S.S. contributed to data collection and media revision. H.A., O.A., and M.A. conducted the economic and cost analysis, contributed to the critical discussion of the findings, and provided editorial input during the final revisions. All authors approved the final manuscript.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Ethics approval and consent to participate
Participation was voluntary, and informed consent was obtained from all participants prior to enrollment in the study. Participants were informed about the study purpose, procedures, voluntary nature of participation, and their right to withdraw. Consent was documented either electronically or verbally, as approved by the institutional research committee. All study procedures were conducted in accordance with the ethical standards of the institutional research committee and with the principles of the Declaration of Helsinki and its later amendments.
Consent for publication
The individual appearing in Fig. 3 is the first author during the video recording. Written informed consent for publication of the image was obtained. No patients are identifiable in any image. No names, facial identifiers, or personal data are included in any figures, tables, or text. No blurring or masking was used as a means of anonymization.
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary Information
Below is the link to the electronic supplementary material.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Abdelmohsen, A., Elshehawy, N., Abdel Razek, M. et al. Educational impact and cost efficiency of AI-enhanced videos in pediatric surgery training: a quasi-experimental study. Sci Rep 16, 8952 (2026). https://doi.org/10.1038/s41598-026-39961-y
Received:
Accepted:
Published:
Version of record:
DOI: https://doi.org/10.1038/s41598-026-39961-y


