Artificial Intelligence Across the Surgical Oncology Continuum: Decision Support, Operative Intelligence, and a Translation-First Roadmap

Matthew Abikenari1,2* ORCiD, Iman Enayati3*, Kimia Mohseni4, Vivek Sanker1, Faina Ablyazova5, Luis Vargas6, Vratko Himic7, Yijiang Chen8, Alisher Baibussinov9, Sri Polkampally1, Shirley Liu1, James Poe1, Claire Kim1, Danial Davani10, Aryan Jain1, Rene Freichel11, Derek Abikenari12, Ahmed Kerwan13 and Riaz Agha14
1. Department of Neurosurgery, Stanford University School of Medicine, Stanford, CA, USA Research Organization Registry (ROR)
2. Nuffield Department of Clinical Neurosciences, University of Oxford, Oxford, UK
3. Department of Orthopaedic Surgery, University of California, Los Angeles, CA, USA
4. Department of Medicine, University of California, Los Angeles, CA, USA
5. Department of Neurosurgery, Northwell, Zucker School of Medicine at Hofstra/Northwell, New Hyde Park, NY, USA
6. Department of Neurosurgery, University of Buffalo Jacobs School of Medicine and Biomedical Sciences, Buffalo, NY, USA
7. Department of Neurological Surgery, University of Miami Miller School of Medicine, FL, USA
8. Department of Radiation Oncology, Stanford University School of Medicine, Stanford, CA, USA
9. Department of Veteran Affairs, University of California, Los Angeles, CA, USA
10. Viterbi School of Engineering, University of Southern California, CA, USA
11. Department of Psychology, University of Amsterdam, Amsterdam, The Netherlands
12. Department of Philosophy, California State Polytechnic University, Pomona, CA, USA
13. Department of Public Health, Harvard TH Chan School of Public Health, Harvard University, MA, USA
14. Premier Science, London, UK
*Indicates Co-first authorship
Correspondence to: Matthew Abikenari, MS mattabi@stanford.edu

Premier Journal of Science

Additional information

  • Ethical approval: N/a
  • Consent: N/a
  • Funding: No industry funding
  • Conflicts of interest: Disclosure: No pertinent conflicts of interest relevant to this manuscript.
  • Author contribution: Matthew Abikenari – Writing – review & editing, Writing – original draft, Visualization, Validation, Supervision, Data curation, Conceptualization
  • Guarantor: Matthew Abikenari
  • Provenance and peer-review: Unsolicited and externally peer-reviewed
  • Data availability statement: N/a

Keywords: Artificial intelligence, Machine learning, Surgical oncology, Perioperative decision-making, Preoperative planning, Intraoperative navigation, Robotic surgery, Clinical translation.

Peer Review
Received: 16 January 2026
Last revised: 1 March 2026
Accepted: 3 March 2026
Version accepted: 3
Published: 9 March 2026

Plain Language Summary Infographic
Artificial Intelligence Across the Surgical Oncology Continuum: Decision Support, Operative Intelligence, and a Translation-First Roadmap” illustrating how AI and machine learning support surgical oncology across the perioperative workflow, including clinical decision support for risk prediction and outcome forecasting, preoperative imaging analysis for tumor detection and anatomical segmentation, and intraoperative navigation with robotic assistance, augmented reality, and tissue tracking, while highlighting translational requirements such as data provenance, external validation, safety assurance, uncertainty calibration, and robustness for reliable clinical AI deployment.
Abstract

Artificial intelligence (AI) and machine learning (ML) methods are increasingly being developed for surgical oncology, a domain in which perioperative decision-making must account for uncertainty and substantial variability in patient anatomy, tumor extent, imaging protocols, devices, and operative workflows. Although proof-of-concept performance has been reported across multiple tasks, there remain limited data regarding the extent to which these systems generalize across institutions and whether they confer measurable clinical utility under routine conditions. In parallel with broader advances in medical AI and surgical robotics, there has been a growing effort to operationalize AI/ML across the perioperative continuum. The aim of this narrative review is to define the role of AI/ML in surgical oncology and to delineate the evidentiary and methodological requirements for translation into reliable clinical systems.

For clarity, we structure the field into three domains that align with the perioperative workflow: (1) clinical translation and decision support, (2) preoperative planning, and (3) intraoperative navigation with robotic assistance/control. Clinical translation centers on decision support, including perioperative risk stratification, outcome prediction, and workflow-integrated recommendations. Preoperative planning is dominated by imaging-based methods for lesion detection and classification, anatomic segmentation, and image registration.

Intraoperative navigation and robotic assistance/control emphasize perception and guidance, such as localization, tissue and instrument tracking, geometric reconstruction, and augmented reality visualization, alongside emerging autonomy-enabling approaches, such as learning from demonstration and reinforcement learning, implemented within supervised human–robot interaction paradigms. Across these domains, we evaluate how data provenance, endpoint selection, external and prospective validation, robustness to dataset shift, uncertainty calibration, interpretability, and safety assurance constrain performance and determine translational readiness.

Introduction

Surgical oncology integrates technically demanding resection and reconstruction with perioperative decision-making constrained by biologic uncertainty and substantial heterogeneity in patient anatomy, tumor extent, prior treatment response, and operative environment. Even when intent is curative, the operative plan is frequently refined intraoperatively based on unanticipated patterns of invasion, occult metastatic disease, tissue planes distorted by neoadjuvant therapy, and evolving physiologic tolerance for blood loss or prolonged anesthesia. In this setting, the margin for error is narrow: oncologic adequacy must be balanced against preservation of function and minimization of morbidity, and these trade-offs must be negotiated using information that is incomplete, distributed across modalities, and variably reliable.

Artificial intelligence (AI) refers broadly to computational methods that perform tasks typically requiring human intelligence; machine learning (ML) denotes a subset of AI in which models learn statistical structure from data to generate predictions or recommendations without explicit rule-based programming. Contemporary AI/ML in perioperative medicine is dominated by data-driven approaches, particularly deep learning, applied to high-dimensional signals such as cross-sectional imaging, digital pathology, endoscopic video, and structured and unstructured clinical data. In surgical oncology, preoperative applications have been shaped by imaging-centered tasks, including lesion detection and characterization, risk stratification, treatment response assessment, and selection of candidates for neoadjuvant or adjuvant strategies. Radiomics and related quantitative imaging methods exemplify this trajectory by converting medical images into features that can be integrated with clinical and molecular variables for prognostic and predictive modeling.1

Preoperative planning has also increasingly incorporated anatomic reconstruction and simulation, motivated by the need to anticipate anatomic variants, determine feasible resection planes, and select minimally invasive trajectories that preserve critical structures. Three-dimensional (3D) reconstruction workflows, often combined with AI-assisted segmentation and registration, have been particularly prominent in thoracic surgery, where complex bronchovascular branching and limited operative exposure create dependence on preoperative imaging for safe dissection.2 Across surgical oncology more broadly, these imaging-based approaches have progressed from offline analysis toward workflow-integrated tools that aim to provide interpretable anatomic abstractions and patient-specific risk estimates at the point of care.

The intraoperative transition from open surgery to endoscopic and robot-assisted platforms has further expanded the volume of machine-readable data while simultaneously reducing direct tactile feedback. Within this environment, computer vision-enabled perception (eg, tissue and instrument localization), workflow modeling (eg, phase recognition), and context-aware guidance have been proposed as mechanisms to reduce operator dependence and improve the consistency of oncologic technique. A systematic review of AI integration in robotic cancer surgery underscores both the breadth of proof-of-concept tasks, ranging from instrument tracking and skill assessment to intraoperative decision support, and the methodological heterogeneity that still characterizes the literature, including variable dataset provenance, inconsistent endpoint definitions, and limited external validation.3 Parallel advances in multimodal AI seek to fuse complementary signals (eg, imaging, text, and video) to support perioperative decision-making, documentation, and surveillance; these approaches are conceptually attractive in oncologic surgery because key decisions often hinge on concordance across modalities rather than any single data stream.4

Despite technical progress, translation of AI/ML systems into reliable clinical support requires evidence that extends beyond retrospective performance within a single institution. Models intended to influence surgical oncology decisions must demonstrate calibration and robustness to dataset shift (systematic differences between training and deployment distributions), integrate uncertainty in ways that are actionable for clinicians, and evaluated against clinically meaningful endpoints under realistic workflow constraints. Importantly, many published systems are derived from retrospective, single-institution datasets and are evaluated using internal validation strategies, limiting inference about transportability and clinical impact.3

Clinical implementation has been reported but remains comparatively uncommon to the volume of proof-of-concept studies. A large-scale implementation study in colorectal cancer surgery deployed an AI-based prediction model as part of perioperative decision support and evaluated outcomes using routine clinical data capture, illustrating both the potential clinical impact of workflow-integrated prediction and the interpretive constraints of non-randomized implementation designs.5 Complementary evidence from international external validation of an ML-based risk prediction model for 90-day mortality after gastrectomy for cancer highlights the methodological necessity of testing transportability across institutions and case-mix distributions before widespread adoption.6

Clinical deployment also raises domain-specific ethical and epistemic considerations. In surgical oncology, AI-informed recommendations may affect treatment selection, extent of resection, and tolerance for operative risk, thereby shaping patient outcomes- and values-sensitive trade-offs. These realities sharpen questions of equipoise, liability, transparency, and informed consent, particularly when models are embedded in proprietary devices or closed robotic ecosystems and when failure modes are difficult to characterize a priori.7 Consequently, translation must couple algorithmic performance with safety assurance practices that explicitly define intended use, identify hazards, and evaluate human–system interaction under conditions that approximate routine clinical care.

Intraoperative navigation and robotic assistance/control constitute a distinct translational domain in which performance is constrained by dynamic anatomy, occlusion, deformation, and real-time computational requirements. Representative work includes real-time vascular anatomical image navigation for laparoscopic surgery, in which deep learning is used to recognize key vascular structures in endoscopic video as a substrate for context-aware guidance.8 Optical stereotactic navigation has also been explored in rectosigmoid cancer surgery using deep learning, supported by 3D modeling to improve localization and support more standardized dissection strategies.9 In robot-assisted settings, safety-oriented perception has been approached through AI-based hazard detection methods that attempt to identify high-risk configurations of instruments and tissue interaction, reflecting a shift toward proactive risk mitigation rather than post hoc error attribution.10 There are still many unresolved issues. One of the major surgical complications, with higher occurrence in cancer patients, is intraoperative hemorrhages, which if detected early, can be more efficiently controlled.

Aim: This paper proposes a hazard detection system which incorporates the advantages of both Artificial Intelligence (AI For oncologic tasks in which decision thresholds are tightly coupled to margins, uncertainty-aware modeling is increasingly recognized as critical; uncertainty estimation has been evaluated, for example, in intraoperative margin detection using mass spectrometry to better characterize confidence in tissue classification.11 In parallel, prospective evaluations of non-contact, AI-assisted intraoperative 3D navigation technologies in lung cancer surgery illustrate continued interest in reducing friction in intraoperative data acquisition while maintaining surgeon autonomy.12 Collectively, the contemporary literature suggests that AI/ML methods can enable consistent extraction of clinically relevant signals from heterogeneous perioperative data, but also that translational readiness is principally determined by methodological rigor and deployment context rather than by headline accuracy.

Surgical oncology provides a stringent test bed for clinical AI: models must generalize across variable anatomy and imaging protocols, support decisions that are safety-critical and time-constrained, and remain reliable when confronted with unexpected intraoperative conditions. In this narrative review, we define the role of AI/ML in surgical oncology and delineate the evidentiary and methodological requirements for translation into reliable clinical systems. For clarity, we structure the field into three domains aligned with the perioperative workflow:

  • clinical translation and decision support,
  • preoperative planning, and
  • intraoperative navigation with robotic assistance/control.

Across these domains, we evaluate how data provenance, endpoint selection, external and prospective validation, robustness to dataset shift, uncertainty calibration, interpretability, and safety assurance constrain performance and determine translational readiness. Lastly, to maintain the surgical oncology focus of this review, the remainder of this review schematizes the capabilities of ML systems against perioperative cancer-related decisions and actions such as oncologic surgery candidacy and risk tailoring, imaging-based characterization /planning for tumor and critical structure relationships, and intraoperative perception/guidance for margin assessment and complication avoidance. Where clinical AI frameworks are cited in general terms (eg, reporting, risk of bias, device control), they have been included insofar as they provide translation-relevant constraints for perioperative oncologic systems.

Methodological Transparency

This review was designed with the goal of translational readiness in mind rather than maximal topical relevance. We conducted a structured literature review according to the following methods. We searched PubMed/MEDLINE, Embase, IEEE Xplore, and the Cochrane Library for English-language publications from January 2010 to the most recent date before submission. Search concepts were designed as follows: (“surgical oncology” or procedure- or organ-specific cancer surgery) AND (“artificial intelligence,” “machine learning,” “deep learning,” “computer vision,” “decision support,” “robotic surgery,” “augmented reality,” “navigation,” “risk prediction”). We also searched reference lists of relevant systematic reviews and landmark studies to ensure that any pertinent literature would not be missed. Studies were eligible for inclusion if they developed, validated, or clinically used AI/ML systems that were intended to affect or were used for perioperative decision-making in cancer surgery, preoperative planning, or intraoperative navigation in cancer surgery. We did not include non-oncologic applications unless they provided a generalizable translational framework that is pertinent to cancer surgery.

Two reviewers screened the abstracts for relevance. The abstracts were then used to screen the full text for inclusion based on the study’s clear intention to use the AI/ML system for the aforementioned applications. Data were abstracted for dataset source and representativeness, labeling strategy and reference standards, internal or external or prospective validation, calibration and uncertainty estimation, decision-analytic evaluation, workflow integration, and safety features. To minimize bias in the review process and to improve the interpretability of the study results for the reader in terms of the applicability to the development of AI/ML systems for cancer surgery, the review is designed to assess the applicability and the risk of bias in the study results according to the domains in the PROBAST/PROBAST+AI criteria (participants, predictors, outcomes, analysis) while focusing on the reporting criteria in the TRIPOD+AI, CONSORT-AI, SPIRIT-AI, DECIDE-AI criteria as applicable.

Lastly, to make the evidence assembly process explicit and reproducible, we provide a summary of identification, screening, eligibility, and inclusion criteria, including details of sources searched, de-duplication strategy, inclusion and exclusion criteria, and final included items by domain. For ease of reading and conciseness of the main text, we have placed expanded framework tables, such as comprehensive standards tables and expanded failure mode tables, throughout the manuscript.

Clinical Translation and Decision Support

To help evidence maturity become clearer and avoid over-reliance on narrative emphasis, we provide domain-specific evidence maps that provide a summary of representative studies in surgical oncology (Tables 1–4). Each table includes information on dataset provenance (single-site vs multi-site, registry vs electronic health record [EHR], imaging), validation approach (internal, external, internal–external, prospective), calibration and handling of uncertainty, decision-analytic evaluation, and, most importantly, what specific clinical action is intended to be triggered by model output (eg, pathway enrollment, increased surveillance, changes in surgery approach, increased margins, sampling).

Table 1: Surgical oncology AI operational taxonomy with translation maturity rubric.
Operational TierModule or TaskPrimary Clinical ProductTypical Model FamilyTranslation Maturity Rubric for This ModuleMinimum Translation BarDominant Harm If It Fails
Decision SupportRisk prediction, complication prediction, therapy guidanceDecision-relevant outputs for tumor board and pre-op planningSupervised ML, deep nets, survival modelsRetrospective discrimination → internal validation → external validation → prospective workflow study → impact study with patient or process endpoints → monitoringExternal validation and prospective assessment should become routine; interface and workflow integration are scientific outputsSilent distribution-shift failure, false reassurance, inappropriate plan changes
Pre-op PlanningClassificationActionable probabilities for diagnosis or risk phenotypeCNN classifiers, multimodal modelsBenchmark performance → multi-site generalizability → calibration and uncertainty → clinical impact on planning decisionsOutputs must align with decisions rather than outcomes; handle missing and imbalanced dataOverconfident wrong probabilities bias operative aggressiveness
Pre-op PlanningDetectionLesion and landmark localization that reduces search burdenDetection networks, regression headsPhantom or single-site tests → robustness across scanners and protocols → error propagation testing into segmentation and registrationMust be treated as upstream dependency that can cascade into later modulesCascading localization error creates systematic downstream overlay error
Pre-op PlanningSegmentationOperative geometry: tumor margins, organs at risk, vessels, corridorsU-Net variants, encoder–decoder netsAccuracy on curated sets → robustness to label heterogeneity → clinically relevant geometry metricsAnnotation scalability is the binding constraint; weak and self-supervised are necessitiesWrong geometry yields unsafe distances, corridors, or resection maps
Pre-op PlanningRegistrationAlignment across modalities, timepoints, and pre-op model to intra-op viewDeformable registration, learned deformation fieldsRetrospective registration error → stress tests under deformation → downstream overlay safety testsConfident subtle error is worse than no system; overlays require trustConfident wrong overlay guides the surgeon toward danger
Intra-op NavigationShape instantiationRecover usable 3D geometry from partial or 2D views3D reconstruction, surface modelsLab or phantom → robustness under deformation and occlusion → failure-behavior characterizationShould be evaluated by modes and speed of failure, not frame accuracyWrong 3D model leads to wrong collision avoidance or distance-to-critical-structure
Intra-op NavigationEndoscopic navigationDepth, pose, mapping in non-rigid spacesDepth estimation, odometry, SLAM, self-supervisedPhantom or limited datasets → domain shift robustness → OOD detection and safe degradationSystems must be tested for failure behavior across environmentsCatastrophic navigation drift with false confidence
Intra-op NavigationTissue trackingContinuous lock on regions of interest under motion and deformationTracking-by-detection, online updatingControlled settings → robustness in real OR conditions → governance for non-stationary updatingTight constraints during low confidence; governance challenges for online learningTracker drift causes unsafe AR and automation bias
Intra-op NavigationAugmented reality overlaysIntuitive overlay of pre-op models onto operative fieldAR overlay pipelinesLab alignment → non-rigid organ overlay validation → human factors evaluationPeriodically incorrect overlay is dangerous; uncertainty must be representedSurgeon overtrusts an intermittently wrong overlay
Robotics and AutonomyInstrument segmentation and trackingTool localization as prerequisite for safe automationU-Net tools, tracking smoothingBenchmark → robustness under blood, smoke, occlusion → integrated safety testsPerception is prerequisite for bounded autonomyMislocalized tool leads to unsafe action
Robotics and AutonomyLearning from demonstrationSurgical primitives and reusable motion policiesImitation learning, primitive decompositionBenchmark datasets → OOD detection → safe handoffMust detect edge conditions and transfer control swiftly to humanUnrecognized edge case leads to unsafe continuation
Robotics and AutonomyReinforcement learningNarrow task-level autonomy via policiesRL with sim-to-realSimulation → transfer learning → robust safety constraintsSim-to-real transfer is a major practical concern; autonomy is boundedPolicy fails at rare events and creates instantaneous harm
Robotics and AutonomyHuman–robot interfaceCommand-intention detection and safe interactionMultimodal intent modelsLab tests → OR noise and occlusion validation → extremely low tolerance error auditsLow tolerance for errors; single wrong command can be dangerousErroneous command triggers dangerous actuation
Table 2: Integrity-chain audit table: how upstream failures propagate to downstream harms.
Upstream ModuleDownstream ModuleDependency You ClaimHow Failure PropagatesConcrete Harm ExampleTranslation Mitigation You Advocate
DetectionSegmentationLandmark and ROI constraintsMissed or shifted detection biases segmentation seed and geometryIncorrect tumor margin near critical structureEvaluate cascade explicitly, not modules in isolation
SegmentationRegistrationGeometry informs alignment and overlaysWrong anatomy geometry biases deformation alignmentAR overlay is systematically shifted despite looking plausibleTreat overlay trustworthiness as safety-critical
RegistrationAR overlaysOverlay depends on correct alignmentSmall registration error becomes high-confidence visual guidanceConfident wrong overlay increases surgeon confidence wronglyReport uncertainty; fail safely when confidence is low
TrackingAR overlaysOverlay stability depends on trackingTracker drift makes overlay intermittently correct and intermittently wrongSurgeon trusts overlay during a wrong intervalHuman factors evaluation and uncertainty-first outputs
Shape InstantiationRobotics constraintsCollision avoidance requires usable 3D geometryIncorrect surface model breaks distance-to-structure constraintsInstrument trajectory violates safe boundaryRequire multimodal sensing and defined low-confidence behavior
Decision SupportPre-op planningPlanning decisions depend on risk and advisabilityMiscalibrated risk pushes corridor choice or aggressivenessOverly aggressive plan in fragile patientExternal and prospective validation; workflow-native interfaces
Pre-op Planning PipelineIntra-op navigationPre-op model is used as intra-op referenceNormalized pre-op outputs become wrong under deformationNavigation uses outdated geometryClosed-loop consistency across operating suite
Intra-op NavigationBounded autonomyPerception is prerequisite for actionPerception errors produce unsafe actuationAutonomous subtask continues under OOD conditionsOOD detection and swift transfer of control
Any ModuleDownstream clinical workflowOptional systems get droppedPoor integration reduces adoption and increases misuseTool ignored or used incorrectly under time pressureInterface and workflow are primary scientific outputs
Table 3: Translation-grade evaluation crosswalk to reporting and staged-evidence frameworks.
Evaluation ElementWhat to Report in Surgical Oncology AI PapersApplies Most toGuideline and Staged-Evidence Crosswalk
Cohort Definition and Target DecisionDefine the decision the model supports, not just the outcomeDecision support, classificationTRIPOD+AI for prediction model reporting; DECIDE-AI for decision-support evaluation framing
Missing Data and Imbalance HandlingExplicit handling strategies and sensitivity analysesDecision supportTRIPOD+AI and DECIDE-AI emphasize transparent data handling and evaluation design
Internal vs External Validation SeparationDistinguish internal validation from true external testingDecision support, all modulesTRIPOD+AI; IDEAL encourages staged evidence as systems mature
Prospective Evaluation in Real WorkflowProspective studies with workflow endpoints, time to intervention, concordanceDecision support, intra-op systemsDECIDE-AI specifically targets early-stage prospective evaluation; CONSORT-AI applies when randomized trials exist; IDEAL later stages
Interface and Human FactorsDescribe interface, tumor board integration, AR human factors testingDecision support, ARDECIDE-AI highlights integration into practice; CONSORT-AI requires reporting AI-specific trial elements when trialed
Uncertainty as First-Class OutputCalibration, confidence, abstention rules, and defined fail behaviorAll modules, especially AR and roboticsDECIDE-AI and TRIPOD+AI support transparent uncertainty and safety behavior reporting
Failure-Mode CharacterizationEvaluate modes and speed of failure, not only average accuracyIntra-op navigation, trackingIDEAL style staged evaluation aligns with progressively harsher real-world testing
Cascade and System-Level EvaluationTest integrity chain and upstream–downstream error propagationPre-op to intra-op to ARIDEAL supports system maturation; DECIDE-AI supports system evaluation rather than isolated metrics
Dataset Strategy for Domain ShiftBuild datasets with domain shift, label quality, temporal changesAll modulesTRIPOD+AI encourages transparent dataset and modeling description; staged evidence expects multi-site robustness
Post-Deployment MonitoringDefine monitoring signals, drift detection, auditabilityDeployed systemsIDEAL long-term evaluation concepts; AI trial reporting literature emphasizes this as necessary for trustworthy adoption
Table 4: Risk-of-harm and failure-mode matrix mapped to mitigations and what is actually tested.
Failure ModeWhere It ManifestsWhy It HappensWhy It Is Dangerous in SurgeryMitigation Strategy You ProposeWhat the Literature Commonly Tests vs Rarely Tests
Distribution ShiftDecision support, classificationMulti-site variation in patients, imaging, practiceQuiet failure with plausible outputsRoutine external validation and prospective evaluationCommon: retrospective discrimination; Rare: prospective impact and multi-site robustness
Confident Wrong OverlaysRegistration and ARSubtle misalignment, non-rigid deformation, tracking driftElevates surgeon confidence toward wrong actionUncertainty-first overlays; defined low-confidence behavior; multimodal sensingCommon: offline overlay accuracy; Rare: human factors harm testing under realistic conditions
Cascading Upstream ErrorDetection to segmentation to registrationPipeline dependencyDownstream harm scales from small upstream biasExplicit cascade testing and integrity-chain evaluationCommon: module benchmarks; Rare: end-to-end safety tests
Tracker DriftIntra-op tracking and AROcclusion, lighting, deformation, online updatingIntermittent wrong guidance at critical momentsConstraints during low confidence; governance for online updatesCommon: controlled tracking benchmarks; Rare: governance and non-stationary behavior audits
OOD Edge CasesRobotics autonomyRare bleeds, unusual anatomy, unexpected complicationsPolicy performs worst at the edge where patients actually areOOD detection and swift control transfer to humanCommon: simulation success; Rare: validated OOD detection and safe handoff metrics
Sim-to-Real BrittlenessReinforcement learningTraining in simulation with domain gapUnsafe actuation when real-world dynamics differConservative bounded autonomy; transfer learning validationCommon: simulation results; Rare: robust sim-to-real safety demonstration
Command-Intention MisfireHuman–robot interfaceOR noise, occlusion, cognitive loadSingle wrong command can be dangerousExtremely low tolerance error audits; multimodal redundancyCommon: lab HRI demos; Rare: OR-grade validation under noise and stress
Dataset Label HeterogeneitySegmentation, detectionAnnotation cost, heterogeneity, weak labelsGeometry errors appear precise and persuasiveWeak or self-supervised scaling plus robust evaluationCommon: single-site curated labels; Rare: heterogeneity stress tests
Security and System BreakdownDeployed systemsCybersecurity gaps, device failuresInstantaneous harm in high-stakes environmentSecure, auditable systems and governanceCommon: privacy discussion; Rare: engineering-grade failure accountability

Scope and Clinical Tasks

In surgical oncology, clinical translation denotes the pathway by which AI and ML methods move from retrospective proof-of-concept toward deployment as workflow-integrated tools that measurably improve patient care under routine conditions.5,13 This translational interface is dominated by clinical decision support (CDS), defined here as computer-based systems that provide patient-specific assessments or recommendations intended to aid clinical decision-making.14,15 Decisions such as operative candidacy, extent and timing of resection, perioperative optimization, postoperative triage, and surveillance planning directly shape both short-term morbidity and long-term oncologic benefit.5,16,17 At this stage, model outputs must be not only accurate but also actionable, interpretable, and deliverable at the moment a decision is made, properties repeatedly associated with CDS effectiveness in broader health systems research and explicitly emphasized in surgical risk calculator design.14,18

Contemporary perioperative decision-making is anchored by traditional risk stratification tools (eg, comorbidity indices and general surgical risk calculators), yet these instruments can be imperfectly calibrated to specific oncologic procedures, patient populations, and evolving perioperative pathways.18–21 ML-based CDS has therefore concentrated on three clinically actionable tasks:

  • preoperative prediction of postoperative morbidity, mortality, and resource use (eg, prolonged length of stay);
  • perioperative process-of-care decisions (eg, escalation of monitoring intensity or thromboprophylaxis); and
  • oncologic outcome prediction (eg, early recurrence, metastasis, survival) intended to tailor adjuvant therapy counseling and surveillance strategies.5,17,22–24

Data Substrates and Modeling Pradigms

Decision-support systems in surgical oncology are typically trained on multimodal clinical data rather than imaging alone.22,25,26 Common substrates include structured EHR variables (demographics, comorbidities, laboratory values, medications), perioperative process data (procedure type, operative duration, transfusion, enhanced recovery pathway adherence), and longitudinal physiologic measurements.22,27,28 Unstructured text has particular value because key prognostic and procedural descriptors are often documented in operative notes, pathology reports, and radiology narratives.25,26 In ovarian cancer surgery, for example, natural language processing (NLP) applied to preoperative computed tomography (CT) reports improved prediction of morbidity and mortality beyond structured variables, illustrating how extraction of clinically salient detail from free text can enrich perioperative risk models without requiring new data collection.26

Patient-generated health data are an emerging input modality for post-discharge decision support.29 Remote telemonitoring platforms capturing symptoms and physiologic parameters have been evaluated for the prediction of postoperative complications after cancer surgery, addressing a recognized vulnerability in surgical pathways: many complications manifest after discharge, when traditional in-hospital monitoring has ended, and clinical contact is intermittent.29 These studies suggest that integration of patient-reported and sensor-derived data may extend the temporal window of actionable risk detection, but they also raise implementation requirements (device adherence, data governance, and clinical response capacity) that must be incorporated into translational evaluation rather than treated as downstream operational details.

Lastly, the notion of regulatory readiness for perioperative AI now extends to include a lifecycle plan for controlled updates and performance management, not merely pre-market performance and accuracy. In the case of the United States, the Food and Drug Administration (FDA) has issued guidance for AI-enabled medical devices that supports iterative modification via the Predetermined Change Control Plan (PCCP), which addresses what changes can occur and how they can be assessed to provide reasonable assurance of safety and effectiveness. At the same time, FDA cybersecurity guidance also addresses security-by-design within the quality system and the integration of cybersecurity documents within the device submission package for those devices with cybersecurity risk.

In the EU, the Medical Device Regulation (MDR 2017/745) has also placed greater responsibility on manufacturers for post-market surveillance (Articles 83–86) and for software IT security within general safety and performance requirements. In practice, we would advocate the alignment of OR-related quality and cybersecurity processes (secure health software lifecycle processes IEC 81001-5-1) and the establishment of triggers for escalation and rollback based upon statistically significant performance degradation on previously determined clinical endpoints (including calibration), sustained increases in out-of-distribution (OOD) rates and tracking and registration failures beyond control limits, safety events plausibly linked to the AI model performance, or significant cybersecurity events; mitigation should include rollback to the last known good model performance, silent mode revalidation, and auditable change history consistent with the PCCP.15,29

Methodologically, most decision-support applications remain supervised prediction models trained on retrospective cohorts, benchmarking logistic regression or regularized regression against tree-based ensembles and, less commonly, neural networks.22,30–32 Several comparative evaluations suggest that increased model complexity does not guarantee improved performance.22,30 For example, in a population-based analysis of mortality prediction using administrative diagnosis codes, boosted trees did not outperform logistic regression, emphasizing that the added computational and interpretability costs of complex models should be justified by demonstrable gains in calibration, discrimination, or clinical utility relevant to the intended decision.14,22,33

Perioperative Risk Stratification and Workflow-Integrated Recommendations

A central translational premise of perioperative CDS is to identify individuals at elevated risk for adverse outcomes and to trigger targeted pathways that reduce preventable morbidity without disproportionate resource consumption.5,14 In colorectal cancer surgery, Rosen et al. reported a registry-based ML model predicting 1-year mortality, deployed as a clinical decision support tool coupled to personalized perioperative care pathways; the implementation was associated with reductions in postoperative complications, improved “textbook” outcomes, shorter length of stay, and lower costs, illustrating an end-to-end translational pattern from prediction to workflow-embedded intervention and measurable system-level impact.5 This kind of coupling between risk prediction and standardized, protocolized responses is also consistent with broader CDS literature showing that systems are more likely to change practice when they deliver recommendations (or pathways) within routine workflow rather than standalone risk estimates.14,15

Across oncologic procedures, ML-based models have been developed to predict composite postoperative morbidity as well as specific complications that motivate discrete prophylactic or monitoring interventions.21,28,31,33,34 Examples include models for postoperative complications after liver surgery, early postoperative complications after radical gastrectomy, and complication risk stratification after lung cancer surgery.21,31,33 In a thoracic oncology cohort, an elastic net model predicting postoperative complications after lung cancer surgery achieved good discrimination and improved risk prediction compared with the Charlson Comorbidity Index, demonstrating that procedure-specific perioperative models can outperform general comorbidity summaries when appropriately developed and temporally validated.19,21 Similarly, in head and neck cancer surgery, ML models have been used to predict prolonged length of stay and were compared against the ACS-NSQIP risk calculator and conventional statistical models, highlighting a recurring translational theme: models must be evaluated not only in isolation but also against the clinical baselines that actually shape care decisions.18,20

Risk prediction for procedure-specific events illustrates both the opportunity and the translational constraints of CDS. Anastomotic complications after esophagectomy (leak, stricture) are high-impact outcomes with potential for algorithmically guided monitoring and early intervention.23,35,36 Recent studies have developed models combining clinical variables with imaging-derived features to predict anastomotic leakage, reporting discrimination and calibration metrics, as well as threshold-based evaluation intended to support clinical decision-making.23,35 In parallel, automated ML approaches have been applied to predict post-esophagectomy strictures, suggesting that algorithmic pipelines can be designed for scalability when feature extraction and model selection are automated, although automation does not obviate the need for external validation, drift monitoring, and transparent reporting.36,37

Dynamic decision support aligned to perioperative trajectories is increasingly emphasized because physiologic and laboratory changes often precede clinically recognized deterioration.27,28 In a multicenter gastrectomy cohort, time-sequential ML models using serial laboratory and vital-sign measurements improved early prediction of postoperative complications compared with baseline models, supporting the concept that CDS should be designed as a longitudinal process rather than a single preoperative score.28 Complementary evidence comes from dynamic modeling of postoperative venous thromboembolism risk after colorectal cancer surgery, where repeated updates to risk estimates reflected evolving clinical states and supported individualized prophylaxis and monitoring decisions in a multicenter context.27

Prospective validation is particularly important for perioperative CDS because apparent performance in retrospective data can be inflated by outcome misclassification, missing data mechanisms, and site-specific practice patterns.13,37,38 Notably, prospective evaluation has begun to appear in some surgical oncology prediction studies.28,34 In lung cancer surgery, Chen et al. developed and prospectively validated an explainable ML model for postoperative pulmonary complications, reporting discrimination, calibration, decision curve analysis, and feature attribution via SHapley Additive exPlanations (SHAP), an unusually comprehensive suite of translational evaluation components for a single procedure-focused model.34 Together, these studies illustrate a maturation of perioperative CDS from static, internally validated scores toward longitudinal, interpretable, and prospectively assessed systems aligned with intervention thresholds and workflow integration.27,34,37

Oncologic Outcomes and Treatment-Tailoring Decision Support

Decision support in surgical oncology must also address endpoints that determine cancer control and long-term survivorship, particularly early recurrence and metastasis after ostensibly curative resection.17,39 In esophageal cancer, Rahman et al. developed ML models to predict early recurrence after surgery using multinational data and applied internal–external validation across centers, providing an instructive template for transportability assessment when multicenter external validation is not yet feasible.39 In pancreatic cancer surgery, an interpretable ML model was developed to predict early liver metastasis after resection and was externally validated with calibration and decision-curve analyses; notably, the model was also implemented as an accessible application, reflecting an explicit translation intent. These examples underscore that oncologic CDS often targets events (recurrence, metastasis) that are clinically meaningful precisely because they can change downstream actions such as adjuvant therapy selection, surveillance intensity, and enrollment into clinical trials.15,39

In glioblastoma, for instance, a quintessential conundrum in oncology, the same reasoning applies, in that despite ‘maximal safe’ resection, early failure on imaging and failure patterns at the edge of resection are frequent and directly inform intensification of treatment, trial participation, and follow-up intensity.40,41 A CDS approach using preoperative magnetic resonance imaging (MRI) radiomics, intraoperative factors such as extent of resection, and molecular factors such as MGMT promoter methylation or IDH mutation could risk-stratify for early failure and inform postoperative treatment and follow-up.42 Notably, in glioblastoma, this CDS system would need to directly address issues related to uncertainties in the differentiation of true failure from treatment-associated radiation changes and pseudoprogression.

Certain CDS applications bridge the boundary between decision support and intraoperative perception by producing staging or resectability assessments that directly influence operative strategy.24,32 The artificial intelligence laparoscopic exploration system (AiLES) is an AI system designed to recognize intra-abdominal metastasis during laparoscopic gastric cancer surgery using real-time video, with evaluation against surgeon assessments; this form of intraoperative decision support can alter the surgical plan and downstream therapy selection when occult metastatic disease is detected.24 In parallel, perioperative personalized decision support reviews have emphasized the need to integrate educational and feedback components with prediction outputs so that CDS improves not only prediction but also clinical action and decision quality.15,32

Evidence Maturity, External Validation, and Clinical Utility

A consistent finding across systematic reviews is that much of the surgical oncology AI literature remains concentrated in retrospective development studies, with heterogeneous endpoints and limited prospective evaluation demonstrating improved patient outcomes or workflow efficiency.13,37 Bektaş et al., in a systematic review of ML applications in upper gastrointestinal cancer surgery, noted substantial heterogeneity in outcomes and methodological approaches, limiting comparative synthesis and emphasizing the need for standardized reporting and validation.37 Similarly, in rectal cancer surgery, a systematic review of AI models predicting surgical difficulty highlighted variability in definitions, imaging features, and validation strategies, illustrating that even seemingly bounded decision support targets (eg, “difficulty”) can suffer from endpoint ambiguity that impairs translation.37

Generalizability across institutions and time is a core translational constraint in perioperative CDS because case mix, perioperative protocols, coding practices, and follow-up capture differ across sites and evolve over time.13 Several surgical oncology studies have begun to address this through explicit external validation designs.6,27,29,39 Dal Cero et al. performed an international external validation of a machine learning model for 90-day mortality after gastrectomy, providing direct evidence that transportability must be tested under clinical and geographic shifts rather than assumed from internal performance alone.6 Complementary internal–external validation strategies, as used in multinational recurrence prediction after esophageal cancer surgery, similarly represent pragmatic approaches for early transportability assessment when full external validation cohorts are not yet available.37,39

From a clinical translation perspective, discrimination is necessary but insufficient.43,44 Poorly calibrated models can misestimate absolute risk and therefore misdirect threshold-based decisions (eg, intensive monitoring, admission to higher-acuity units, or initiation of prophylaxis).18,43 Consequently, calibration assessment and decision-analytic evaluation should be considered minimum requirements for CDS systems intended to change clinical actions.43,44 This emphasis is reflected both in methodological guidance and in several recent surgical oncology studies reporting calibration curves, Brier scores, and decision curve analysis (DCA) alongside discrimination.23,34,35,40 The reporting of DCA is particularly relevant to decision support because it quantifies net benefit across thresholds, connecting model outputs to the clinical consequences of false positives and false negatives rather than relying solely on rank-based discrimination.34,44

Human Factors, Equity, and Reporting Standards

Even when predictive performance is favorable, CDS adoption depends on usability, clinician trust, and integration at the time and location of decision-making.14,45 A systematic review of clinical decision support systems identified features associated with improved clinical practice, including automatic provision within workflow and delivery of recommendations rather than assessments alone, design principles that are directly relevant to AI/ML tools intended for perioperative oncology care.14 Qualitative research on decision support interventions targeting shared decision-making similarly underscores clinician concerns about appropriateness of recommendations, time costs, and alignment with clinical judgment, anticipating barriers to AI/ML CDS if outputs are poorly contextualized or not coupled to feasible actions.45 For patient-centered surgical decisions, evidence from randomized evaluation of conversation aids in breast cancer surgery further illustrates that the quality of decisions depends on aligning clinical options with patient values and understanding; prediction outputs are therefore only one component of the translational goal of improving decision quality.15,46

Equity considerations are central to translational readiness because prediction models trained on historical healthcare data may encode structural inequities in access, treatment, and documentation.13,47 In a widely cited example outside oncology, Obermeyer et al. demonstrated substantial racial bias in a commercial risk prediction algorithm arising from the use of healthcare costs as a proxy for health needs; analogous proxy-label and documentation biases are plausible in surgical oncology when outcomes, comorbidities, and follow-up are differentially captured across groups.25,37,47 Systematic review evidence in cancer pathways also indicates inconsistent reporting of subgroup performance and limited attention to equity in prospective evaluations, reinforcing the need to treat fairness assessment as a core translational requirement rather than an optional post hoc analysis.13,38,42

Finally, reporting and evaluation standards are prerequisites for reliable translation.42,48–50 TRIPOD and the updated TRIPOD+AI statement provide minimum reporting items for prediction model studies using regression or ML methods, including a transparent description of data sources, outcome definitions, handling of missingness, and full reporting of performance metrics.48,49 PROBAST and the updated PROBAST+AI tool support structured assessment of risk of bias and applicability, offering a shared framework for judging whether a model is likely to generalize and whether reported performance is credible.42,50 For prospective and interventional evaluation of AI-enabled decision support, CONSORT-AI and SPIRIT-AI extend trial reporting and protocol guidance, while DECIDE-AI provides a reporting guideline for early-stage clinical evaluation of decision support systems driven by AI, an appropriate framework for perioperative CDS that may be introduced first in limited workflows before broader rollout.13,15,51

Human factors readiness of perioperative AI systems for use will demand interface design elements that support safety in situations of uncertainty, drift, and incorrectness. For example, at the UI level, uncertainty-first design principles must be implemented through quantification of probabilities and explicit threshold bands with prespecified actions, insufficient evidence states when probabilities are too low, drifting detection and notification, and hard constraints against providing confident guidance through overlays and trajectories when upstream uncertainty bands are breached.

At the workflow level, each AI system will need a compact failure mode playbook that maps common types of failures (eg, detection miss, segmentation bias, registration offset, tracking drift/OOD, cybersecurity issues) to standard responses such as confirmation steps, escalation protocols, fail-safe transitions to manual or alternative modes of imaging, documentation requirements, and review triggers after failure events. In parallel, equity assessment will need to be addressed as a translational deliverable through a simple audit template that reports stratified performance and calibration across clinically meaningful subgroups and settings (eg, by sex, age, race/ethnicity if available, comorbidity burden, language and insurance status if applicable), subgroup-specific net benefit analysis through DCA, and prespecified mitigation approaches such as targeted data enrichment and reweighting and domain adaptation, subgroup-dependent thresholds with clinically justified trade-offs and monitoring with rollback criteria if disparate errors are seen (Figure 1).

Fig 1 | An uncertainty-first operative intelligence stack for AI-enabled surgical care This diagram represents the end-to-end, dependency-conscious approach to the systematic integration of AI across the surgical continuum of care, from decision support to planning and navigation with bounded autonomy. (1) Decision support: Intra-operatively and retrospectively validated models are used to make inferences regarding risk assessment, complication, and therapy, which are quantitatively paired with model-fit or solution-space uncertainties to facilitate decision-making for clinicians. (2) Preoperatively, patient-specific models of anatomy and geometry are created via the use of pipelines of classification, detection, segmentation, and registration, which incorporate operative conditions to provide the patient with a surgical model. (3) Intraoperatively, navigation and robotics provide the operative intelligence stack that incorporates techniques for endoscopic navigation and SLAM, tissue tracking, and augmented reality overlays, which utilize supervised, bounded autonomy.
Figure 1: An uncertainty-first operative intelligence stack for AI-enabled surgical care This diagram represents the end-to-end, dependency-conscious approach to the systematic integration of AI across the surgical continuum of care, from decision support to planning and navigation with bounded autonomy. (1) Decision support: Intra-operatively and retrospectively validated models are used to make inferences regarding risk assessment, complication, and therapy, which are quantitatively paired with model-fit or solution-space uncertainties to facilitate decision-making for clinicians. (2) Preoperatively, patient-specific models of anatomy and geometry are created via the use of pipelines of classification, detection, segmentation, and registration, which incorporate operative conditions to provide the patient with a surgical model. (3) Intraoperatively, navigation and robotics provide the operative intelligence stack that incorporates techniques for endoscopic navigation and SLAM, tissue tracking, and augmented reality overlays, which utilize supervised, bounded autonomy.

Note the failure modes depicted in the figure: detection failure, segmentation bias, registration offset, and tracking drift with out-of-distribution (OOD) inputs, which propagate to confident but erroneous guidance. As the dependency rail makes manifest, the reliability of the former models of intelligence conditionally bounds the latter model’s autonomy to emphasize the imperative for safe AI applications. For each domain, translational readiness is constrained by: data provenance and representativeness; endpoint definition and clinical action linkage; internal, external, and prospective validation; robustness to dataset shift; calibration and uncertainty communication; human factors and workflow integration; and lifecycle governance (safety, monitoring, and cybersecurity).

The framework emphasizes that headline accuracy is insufficient when downstream actions are irreversible, time-constrained, and safety-critical. An accompanying domain-level evidence map should make the action linkage explicit (what decision changes, by whom, and at what threshold) and align evaluation to that decision rather than headline accuracy. Minimum reporting should include external and prospective validation (including silent trials) plus calibration and clinical-utility analyses (reliability curves and decision-curve analysis) to justify operating points and fail-safe triggers.

Regulatory, Safety, and Lifecycle Governance for Perioperative Oncology AI

When ML systems are used to provide patient-specific information for the guidance of perioperative cancer care, they are likely to be categorized as Software as a Medical Device (SaMD), where the intention is to diagnose, treat, mitigate, or in any way affect or influence the care provided to the patient. For such ML systems to be translated for practical use in the field of medicine, they must align with the lifecycle approach that covers the major regulatory regions or the regulatory harmonization process (FDA, EMA, IMDRF), as opposed to the retrospective approach that is based purely on the system’s historical performance.

For the ML system to be translated for practical use in the field of medicine, the system must be designed to meet the regulatory requirements through the intended use, the definition of the system’s role in the clinical setting (whether assistive or advisory), the hazards associated with the system’s use, as well as the associated mitigations in line with the traditional approach to risk management. This means that the system must meet the requirements for the software development lifecycle (IEC 62304), usability engineering for the system’s interfaces (IEC 62366), as well as the systematic approach to risk management (ISO 14971) that applies to the field of surgical oncology, where time pressure, irreversible actions, and the consequences associated with errors are the norm. The system’s governance must not only be limited to the pre-deployment validation but must extend to the post-deployment monitoring to ensure that the system is performing as required in the real-world setting.52–56

Collectively, the evidence to date indicates that clinical translation of ML decision support in surgical oncology will depend less on incremental improvements in discrimination and more on rigorous alignment of model endpoints with actionable decisions, transparent reporting, robust external and prospective validation, user-centered workflow integration, and ongoing monitoring for calibration drift and inequitable performance.13,15,49,50

Preoperative Planning

Preoperative planning in surgical oncology refers to computational and procedural methods that transform pre-treatment data, most commonly cross-sectional imaging, into patient-specific representations intended to guide operative strategy, define resection targets, anticipate technical constraints, and coordinate adjunct technologies (eg, navigation, fluorescence, or robotic assistance).1,2,51,52 As deployed in current clinical workflows, planning systems typically emphasize geometric and spatial reasoning, what anatomy is present, where the tumor is related to critical structures, and how operative goals map onto feasible dissection planes, rather than solely providing probabilistic risk estimates.2,57 Across organ systems, the dominant technical primitives underlying these workflows remain lesion characterization (including staging-relevant inference), anatomic segmentation and reconstruction, and multimodal fusion/registration.1,2,53–56,58

Imaging-Derived Characterization and Staging

Radiologic staging and anatomic definition remain central determinants of operability and approach selection; accordingly, AI/ML methods have been applied to extract structured staging signals from imaging, radiology narratives, and complementary tissue sources.1,59–62 Radiomics-based strategies, whether relying on engineered feature families or learned representations, have been positioned as a means to quantify tumor phenotype and context in ways that may complement conventional human interpretation, while also amplifying sensitivity to imaging protocol variability and preprocessing choices.1 In breast oncology, convolutional neural network analysis of multiparametric MRI has been evaluated for preoperative prediction of axillary lymph node metastasis, directly targeting a staging component that may influence the extent of axillary surgery and systemic therapy sequencing.59 In gastric cancer, multimodal approaches that combine CT-derived information with digital pathology features (whole-slide imaging [WSI]) have been reported for staging-related prediction, illustrating how “preoperative planning” increasingly extends beyond imaging alone when pre-treatment tissue is available (eg, endoscopic biopsy).60

Natural language processing (NLP) has also been explored to convert narrative radiology into structured, decision-relevant staging information. Automated esophageal cancer staging from free-text radiology reports has been evaluated using large language models (LLMs), supporting the feasibility of extracting standardized staging descriptors from routine documentation and potentially reducing manual abstraction burden in multidisciplinary workflows.61 Because many planning decisions hinge on a synthesis of imaging, histology, and clinical narrative, LLM-centered approaches may be attractive for workflow integration; however, their translational value depends on rigorous evaluation under realistic variability in report style, missingness, and institutional terminology, as well as safeguards against unsupported inference, and values that integrate restorative ventures for at risk social groups in healthcare.61,63

Planning-Oriented Targets Beyond Categorical Stage

Planning outputs extend beyond the categorical stage to intermediate, decision-relevant targets such as anticipated technical difficulty, required exposure, and margin risk.41,62 In rectal cancer surgery, AI models applied to preoperative MRI have been studied for predicting surgical difficulty, an endpoint that can influence the choice of minimally invasive vs open approach, need for specialized instrumentation, and staffing/experience requirements.41 In oral squamous cell carcinoma, pathomics-based models have been studied for predicting positive surgical margins, illustrating an emerging paradigm in which pre-treatment or perioperative tissue-derived features are translated into margin-focused planning inputs.62 These examples underscore a broader trend: planning systems often seek to forecast “actionable constraints” (eg, limited working space, high-risk dissection planes) rather than only long-term outcomes, but such targets require careful operationalization and validation because they can be sensitive to surgeon factors, institutional practice patterns, and differences in operative technique.1,41

Segmentation, Quantitative Anatomy, and Patient-Specific Modeling

Most preoperative planning workflows depend on delineation of tumors, organs-at-risk, and critical structures, making anatomic segmentation an enabling task for 3D reconstruction, simulation, and downstream guidance.2,54–56,58,64–67 In thoracic oncology, semantic segmentation of chest CT has been used to recognize patient-specific pulmonary vessel variants relevant to resection planning, illustrating how segmentation can be directed toward clinically consequential anatomic variability rather than organ boundary extraction alone.58 In rectal cancer, 3D reconstruction has been described as a means to improve diagnosis and surgical planning, reflecting sustained interest in translating pelvic MRI into spatially coherent representations that can support operative strategy and multidisciplinary discussion.55

Segmentation also supports simulation-oriented planning that seeks to anticipate intraoperative views and dissection trajectories.64–66 An AI-driven 3D simulation system for gastric cancer surgery has been reported as a retrospective validation study, consistent with an early translational pathway in which automated model construction and anatomy recognition are first evaluated for feasibility and concordance with expert interpretation before the impact on procedural decisions is tested prospectively.64 Similarly, AI-based technology to generate a 3D model for rectal cancer surgery planning from MRI has been reported as a step toward preoperative simulation, highlighting an approach in which imaging-derived models are used for rehearsal and spatial understanding rather than direct automation of operative steps.65 At a more granular anatomic scale, automated segmentation of male pelvic floor soft tissues have been proposed for preoperative simulation and morphologic assessment in lower rectal cancer surgery, reflecting the planning value of extracting patient-specific pelvic anatomy that may influence exposure and dissection strategy.66

Planning applications have also been described in hepatobiliary contexts using intelligent image segmentation approaches, aligning with long-standing clinical needs for patient-specific visualization of vascular and biliary anatomy and for volumetric assessment in resection planning.67 Although organ-specific reviews emphasize potential benefits of AI/ML for liver cancer surgery, the evidentiary standard for preoperative planning systems remains high because errors in anatomic delineation or spatial relationships can directly affect operative decisions.52

Quantitative Biomarkers Derived from Segmentation

Quantitative biomarkers derived from segmentation can inform preoperative optimization and approach selection when measurement performance is adequate, and the clinical mapping from measurement to action is explicit.68–70 Automated body composition assessment from routine CT has been evaluated as a scalable method to measure muscle and adipose compartments, including validation-focused studies of deep learning derived measurements from standard-of-care CT examinations.68,69 Related work has linked imaging-derived sarcopenic obesity with prognostic outcomes, supporting the premise that quantitative imaging may identify phenotypes relevant to prehabilitation, nutritional optimization, and candidacy assessments.70 These approaches may be particularly attractive because they leverage imaging already obtained for staging, but translational use requires transparent reporting of measurement error, robustness across scanners and protocols, and clear articulation of how biomarker thresholds would change management.1,68,70

Image Fusion, Registration, and Deformation-Aware Planning

Preoperative planning frequently requires integration across modalities (eg, multiphase CT and MRI), across timepoints, and across coordinate frames that differ from the intraoperative configuration; therefore, image registration and fusion are recurrent bottlenecks for constructing coherent 3D models that can be inspected by clinicians and, in some settings, transferred to guidance systems.2,53,54,56 Reviews of 3D reconstruction and organ modeling emphasize that static preoperative models can be undermined by soft-tissue deformation and physiologic motion (eg, respiration), motivating deformation-aware planning and, where feasible, updating or reconciling models with intraoperative sensing.2,54 The urologic robotics literature similarly situates preoperative modeling within a broader imaging-robotics ecosystem, highlighting both the promise of patient-specific reconstructions and the technical challenges of aligning preoperative images with intraoperative anatomy.53

Simulation, Rehearsal, and Operator-Facing Visualization

Because a preoperative plan must be interpretable and actionable for surgeons and operative teams, many planning systems emphasize operator-facing 3D visualization and rehearsal rather than prediction outputs alone.2,54–56,58,64,65 Work on 3D lung model development for minimally invasive lung cancer surgery has explicitly framed a progression from static reconstructions toward real-time or dynamic modeling, consistent with the clinical observation that thoracoscopic and robotic approaches increase dependence on accurate spatial representations when tactile feedback is limited.54 This emphasis on visualization is also apparent in gastrointestinal oncology planning efforts that use AI-enabled reconstruction to support preoperative simulation in gastric and rectal cancer operations.64,65

Virtual reality and 3D-printed models have been systematically reviewed as planning tools in image-guided, robot-assisted nephron-sparing surgery, providing evidence synthesis on how patient-specific 3D representations may support surgeon understanding, rehearsal, and intra-team communication.57 Complementary discussions of robotic partial nephrectomy in the era of 3D virtual reconstructions similarly reflect continued interest in operationalizing preoperative 3D models as part of routine planning.71 While these technologies are conceptually aligned with surgical oncology goals, improving spatial understanding and reducing unanticipated anatomy, the translational question is whether such tools measurably improve decision quality or outcomes when implemented at scale, accounting for the time and expertise required for model generation and review.57,71

Selecting and Planning Adjunct Imaging and Guidance Technologies

Preoperative planning also includes selection and orchestration of adjunct intraoperative sensing modalities and tracers, with plans often specifying timing of administration, expected contrast behavior, and decision thresholds.56,72,73 Reviews of precision cancer surgery describe fluorescence-guided approaches that seek to improve tumor margin identification and lymphatic mapping, while emphasizing constraints imposed by probe pharmacokinetics, tissue optical properties, and imaging hardware.72 A randomized translational study of protein- and peptide-based probes illustrates that probe selection and deployment strategy are empirically tractable components of the planning process, but also highlights that clinical utility depends on matching probe behavior to a defined decision task (eg, margin assessment or lesion localization).73 Robotic- and image-guided workflows have likewise been described for fluorescence detection, reflecting ongoing efforts to integrate targeted imaging into operative systems.74

Lymphatic mapping provides a complementary example in which tracer workflow and imaging protocol determine intraoperative localization targets. Indocyanine green (ICG)-based lymph node mapping has been described in cancer surgery contexts, and preoperative sentinel node mapping work in gynecologic oncology demonstrates how imaging protocol design and tracer deployment can be codified into planning pathways that influence operative navigation targets.75,76 Optical coherence tomography (OCT) adds a further dimension by providing microstructural imaging that can, in principle, support margin-focused decision-making; systematic and scoping reviews describe OCT in intraoperative tissue assessment and AI-enabled OCT for disease detection and diagnosis, respectively, emphasizing both potential and the need for clinically grounded validation.77–79

Methodological and Evidentiary Requirements for Translation

For AI/ML-enabled preoperative planning, translational readiness depends on demonstrating that outputs are reliable under routine imaging heterogeneity, that errors are bounded and detectable, and that any claimed benefit is linked to decision-relevant endpoints rather than surrogate technical metrics alone.1,2,54,56,68 Validation studies in segmentation and quantitative imaging (eg, body composition measurement) illustrate the importance of reporting agreement with reference standards and clarifying conditions under which manual correction remains necessary.68,69 Because planning models often function as upstream dependencies for downstream navigation, simulation, or robotic workflows, failure modes should be characterized not only at the level of segmentation overlap but also in terms of clinically meaningful spatial errors (eg, mislocalization relative to vessels, planes, or margins).2,54,56,58

Surgeon-facing generative systems and automated staging tools introduce additional requirements beyond conventional imaging models, including provenance-aware synthesis, safeguards against hallucinated or non–evidence-based assertions, and evaluation under realistic documentation variability.61 Current single-center retrospective planning studies, whether focused on 3D reconstruction, simulation, or prompt-driven assistance, should therefore be interpreted as feasibility signals; subsequent work must incorporate external validation, prospectively defined endpoints, and, where possible, decision-impact study designs that quantify how planning outputs change operative choices and whether those changes improve patient-centered outcomes.1,54,57,61–65

Intraoperative Navigation with Robotic Assistance/Control

Intraoperative navigation with robotic assistance/control comprises computational systems that interpret operative data streams in real time and translate those interpretations into actionable guidance, visualization, or robot-mediated behaviors during surgery.3,53,80,81 In surgical oncology, this domain has expanded alongside the adoption of minimally invasive and robotic approaches across urologic, colorectal, foregut, and pancreatic procedures, which standardize endoscopic viewpoints and yield high-fidelity digital signals (eg, video and robotic platform telemetry).3,53,80,81 Contemporary frameworks described in systematic and specialty-focused reviews generally position intraoperative AI/ML as progressing from perception and measurement to navigation and decision support, and only later (and more cautiously) toward shared autonomy or control-constraining behaviors.3,53,80,81

Data Modalities and the Primacy of Intraoperative Perception

Across robotic oncologic surgery, the dominant substrate for intraoperative ML is surgical video, reflecting the ubiquity and standardization of endoscopic and robotic camera feeds and the relative ease of rendering outputs as overlays within existing display pipelines.3,53,80 Proposed enabling architectures for AI-enhanced processing of da Vinci robot-assisted video emphasize that translational feasibility is conditioned by system-level constraints, deterministic latency, robust video ingestion, and reliable handling of large data streams, which can be as decisive as model accuracy for operating room deployment.82

Video-based perception has been used to recognize anatomic structures directly tied to oncologic safety and postoperative function. Deep learning approaches for vessel and tissue recognition during third-space endoscopy illustrate a generalizable pattern: models are trained to detect structures that may be visually subtle yet clinically consequential when injured, with outputs intended to support safer dissection in minimally invasive contexts.83 In urologic oncology, an AI-based intraoperative nerve recognition system has been described for nerve sparing during robotic prostatectomy, exemplifying a navigation-adjacent use case in which “recognize-and-overlay” can provide moment-to-moment spatial cues during technically demanding dissection.84 These approaches align with broader reviews of robotic oncologic surgery that highlight computer vision-based recognition of anatomy and “danger zones” as a near-term, clinically plausible tier of intraoperative AI support.3,53,80,81

Intraoperative perception has also been extended to targets that are explicitly oncologic rather than purely anatomic. In laparoscopic gastric cancer surgery, the AiLES was developed for real-time recognition of intra-abdominal metastasis during exploration, directly addressing a clinically consequential failure mode in which small or occult lesions are overlooked, and staging is thereby misclassified.24 Reviews of intraoperative imaging tools in gastrointestinal oncology likewise emphasize that the clinical value of intraoperative perception increases when outputs are tied to specific intraoperative actions (eg, escalation to biopsy/frozen section, targeted inspection, or altered resection strategy) rather than solely to descriptive visualization.56

Functional state estimation provides a complementary intraoperative objective because tissue perfusion and viability influence anastomotic strategy and can interact with oncologic extent through resection planning.56,85,86 AI-based real-time microcirculation imaging has been evaluated for assessing colonic perfusion status, representing a class of intraoperative systems that infer physiologic states from subtle imaging cues that are difficult to quantify reliably by human vision alone.85 In addition, intraoperative near-infrared (NIR) functional imaging has been studied in colorectal cancer settings, reflecting efforts to quantify perfusion-related and function-related signals that could be integrated into navigation workflows and assessed against endpoints such as leak risk, complication rates, or re-intervention.86

Geometry-Aware Navigation: Localization, Registration, and Coordinate Consistency

Navigation systems must establish spatial relationships between recognized structures, instruments, and (when applicable) preoperative models, often under conditions of soft-tissue deformation, camera motion, and variable insufflation.3,53,80 Real-time vascular anatomical image navigation has been demonstrated for laparoscopic surgery as an approach to aligning vascular anatomy with intraoperative views to support orientation and safer dissection around critical vessels.8 In rectosigmoid oncology, the feasibility of optical stereotactic navigation supported by deep learning based 3D modeling has been reported, highlighting the clinical interest in “coordinate-consistent” guidance for tumor localization and margin-conscious dissection within anatomically constrained pelvic spaces.9

Evidence from thoracic oncology similarly indicates active development of geometry-aware navigation strategies. A prospective application study of non-contact, AI-assisted intraoperative 3D navigation in lung cancer surgery demonstrates a pragmatic approach in which spatial information is acquired and reconstructed intraoperatively to mitigate the mismatch between preoperative imaging and intraoperative anatomy.12 Across these demonstrations, the translational challenge is not merely producing a geometric solution but maintaining accuracy under clinically routine perturbations (eg, deformation, retractors, motion, variable insufflation), and predefining failure detection and escalation workflows when navigation assumptions no longer hold.3,9,12,53

Visualization is a central component of intraoperative navigation because guidance must be interpretable at the point of action. Reviews of extended reality and “metaverse” concepts in surgery discuss how augmented visualization environments could provide spatial context and training ecosystems, while emphasizing limited clinical evidence for outcome benefit and the nontrivial challenges of integrating head-mounted or mixed reality interfaces into sterile workflows.87 In robotic radical prostatectomy, specialty-focused reviews similarly frame advanced visualization (including potential augmented guidance) as an emerging direction that will require rigorous clinical validation and human factors evaluation before routine adoption.81

Imaging-Augmented Guidance in Robotic Oncology

Optical contrast agents and intraoperative imaging can function as “signal amplifiers” for tumor, lymphatics, or function, and ML methods are increasingly used to standardize interpretation and generate overlays that can be integrated into robotic visualization pipelines.53,56 Robotic assistance can stabilize imaging geometry and enable instrument-mounted sensing, as illustrated by robotic “click-on” fluorescence detection using surgical instruments to characterize molecular tissue aspects, a design pattern that can support localized assessment within an otherwise video-dominated workflow.74

Deep learning enabled fluorescence imaging has been developed for surgical guidance tasks aligned with oncologic decision-making. For oral cancer, in silico trained deep learning methods have been described for fluorescence-based depth quantification and for fluorescence-guided margin classification in preclinical models, representing an approach that seeks to convert fluorescence signal into explicitly interpretable, decision-relevant outputs rather than qualitative visualization alone.88,89 More broadly, reviews of precision cancer surgery emphasize that the translational value of imaging-augmented guidance depends on linking the imaging signal (and any ML-derived interpretation) to prespecified intraoperative decisions (eg, additional resection, targeted sampling, or altered dissection plane) and demonstrating benefit against clinically meaningful endpoints.56

Near-infrared fluorescence and functional imaging also support navigation decisions beyond margin assessment, including lymphatic mapping and perfusion-oriented guidance.56,86,90 In upper gastrointestinal oncology, indocyanine green (ICG) lymph node mapping has been described for cancer surgery, and prospective mapping of lymphatic drainage patterns with NIR fluorescence has been studied during robotic-assisted minimally invasive Ivor Lewis esophagectomy.75,91 These studies illustrate that intraoperative guidance is often probabilistic and pathway-dependent (eg, drainage patterns and node basins) and therefore benefits from clear protocols describing how intraoperative findings are intended to modify the operative plan.56

Label-Free Tissue Assessment and Uncertainty-Aware Inference

Label-free tissue characterization provides an alternative approach to intraoperative guidance by aiming to classify tissue state directly from biophysical signatures, potentially enabling margin assessment without exogenous tracers.11,92,93 Perioperative tissue assessment using combined mass spectrometry and histopathology imaging has been described as a framework for integrating molecular and morphologic information to inform intraoperative decisions in cancer surgery.93 Within this domain, uncertainty estimation for surgical margin detection using mass spectrometry has been reported as a strategy to calibrate confidence and support decision thresholds suited to high-stakes intraoperative use.11

Representation-learning strategies have also been applied to reduce dependence on exhaustive labeling while preserving clinically relevant performance, exemplified by image-driven self-supervised learning for mass spectrometry-based tissue assessment.92 Although these methodological advances may improve scalability, systematic syntheses in robotic cancer surgery converge on the need for external validation and explicit specification of action policies when model outputs are uncertain, particularly when outputs are used to trigger irreversible actions such as widening margins, converting operative plans, or escalating to additional resection.3,53 Accordingly, translational protocols should predefine how uncertainty will be communicated to the surgical team, how clinicians should respond to uncertain outputs (eg, confirmatory pathology), and how performance will be monitored over time once deployed.3,11,53,92

Microscopic Intraoperative Imaging and AI-Enabled Interpretation

Microscopic intraoperative imaging modalities such as OCT have been evaluated as tools for margin-focused guidance and tissue characterization when macroscopic imaging is insufficient to resolve microstructural features at the resection interface.77,79 A systematic review of OCT in oncologic surgery summarizes evidence for intraoperative tissue assessment and highlights translational barriers, including limited standardization, variable acquisition conditions, and uncertainty about how best to integrate OCT-derived information into operative workflows.79 A scoping review focused on AI in advancing OCT further reinforces that algorithmic interpretation is a key enabler for converting high-dimensional OCT data into actionable intraoperative outputs, while also underscoring the need for robust validation and clinically grounded study designs.77

Hybrid systems that fuse OCT with ML inference have been reported for breast cancer surgical margin visualization, illustrating a design pattern in which high-resolution imaging is paired with automated classification to generate decision-supportive outputs intraoperatively.78 Translation of such systems into routine oncologic practice requires demonstration of robustness to blood, motion, and specular artifacts; prespecified uncertainty-aware operating points; and empirical evidence that OCT-based guidance improves margin-related outcomes without disproportionate resection or operative burden.77–79

Robotic Assistance, Safety Monitoring, and the Boundary Between Guidance and Control

Robotic platforms create opportunities for continuous performance monitoring and algorithmic safety layers that detect and mitigate hazards, but the evidentiary threshold increases substantially when systems move from passive guidance to action-constraining control.3,53,80 AI-based hazard detection in robotic-assisted single-incision oncologic surgery exemplifies a safety-oriented direction by targeting the identification of hazardous states in a modality characterized by constrained instrument motion, limited triangulation, and heightened collision risk.10

Robotic data streams can also be used to evaluate and standardize operative performance, with potential downstream implications for oncologic outcomes and complication risk. A study associating skill and errors with outcomes in robotic rectal cancer surgery illustrates how intraoperative performance metrics can be linked to clinical endpoints, supporting the concept that intraoperative analytics may provide actionable feedback for quality improvement.80,94 At the same time, outcome-linked analytics emphasize the need to avoid overinterpreting surrogate performance measures unless they are validated against patient-centered endpoints and contextualized by case complexity, anatomy, and oncologic extent.3,53,80,81,94,95

Engineering frameworks for AI-enhanced processing of da Vinci robot-assisted video reinforce that translation requires not only accurate models but also reliable system-level integration, including secure data handling, deterministic latency, and compatibility with the robotic platform’s data interfaces.82 Systematic synthesis of AI integration into robotic cancer surgery indicates that most published systems remain at the level of feasibility or retrospective validation, with limited evidence from prospective studies and limited clarity regarding how ML outputs are incorporated into standardized intraoperative actions.3,53 This evidence profile supports a conservative near-term posture: prioritize decision support, safety monitoring, and uncertainty-aware guidance before pursuing control-constraining behaviors that would require substantially higher safety assurance.3,10,53

Translational Requirements: Real-Time Operation, Robustness, and Clinically Testable Utility

Across intraoperative navigation and robotic assistance/control, translational readiness depends on demonstrating reliable real-time operation under routine variability, explicit handling of uncertainty, and measurable benefit in decision-relevant endpoints.3,53,56 Feasibility studies in vascular and stereotactic navigation demonstrate that geometry-aware guidance can be operationalized in oncologic procedures, while prospective application of AI-assisted 3D navigation in lung surgery shows that intraoperative navigation can be evaluated in forward-looking designs rather than solely retrospective settings.8,9,12 Prospective demonstration of real-time metastasis recognition during exploration further illustrates how intraoperative perception can be tied to immediate oncologic decisions, providing a model for future studies that specify workflow integration and decision thresholds a priori.24

For imaging-enhanced guidance (eg, fluorescence and NIR functional imaging), translation requires separating device- and protocol-specific variability from biologically meaningful signal and validating ML-derived overlays across acquisition settings, patient factors, and institutions.56,86,88,89 For label-free modalities (eg, mass spectrometry and OCT), uncertainty-aware inference and robustness to intraoperative artifacts are central because both false reassurance and false alarms can produce clinically meaningful harm (eg, inadequate margins vs unnecessary tissue loss).11,77–79 These requirements are emphasized across systematic and specialty-focused reviews, which report that the clinical evidence for many AI-enabled intraoperative systems remains dominated by retrospective analyses and early feasibility work, with relatively few prospective evaluations tied to patient-centered outcomes.3,53,80,81,95 Prospective application studies in navigation (eg, non-contact 3D guidance) and real-time disease detection (eg, AiLES) provide useful templates for study designs that connect real-time inference to standardized intraoperative actions and measurable outcomes.12,24 Finally, because robotic surgery continues to expand across complex oncologic procedures (including pancreatic surgery), the bar for safe, clinically effective integration of intraoperative AI should remain anchored in workflow-aware validation, transparent uncertainty handling, and rigorous prospective evaluation rather than technical performance alone.3,10,53

In addition, outside the perioperative realm, there have been equal advancements in AI for psychiatry and mental health applications, including risk assessment (suicide attempts, relapse, and acute decompensation), symptom profiling, and treatment recommendation with multimodal data sources like longitudinal electronic health records, wearable and smartphone-based digital phenotyping, speech/language processing, and neuroimaging.96–98 In precision oncology, the combination of radiology, digital pathology, and genomic profiling enabled by AI has facilitated tumor profiling, prediction of treatment response (including immunotherapy), and computational triage for molecular analyses, thereby propelling the transition of high-dimensional biomarkers into decision-making.99–101 These parallel applications collectively point to a shift towards data-driven, personalized decision support systems, while also pointing out common issues in fairness, generalization, explainability, and validation that are just as applicable in surgical oncology.

Conclusion

Across surgical oncology, AI/ML research increasingly spans the full perioperative continuum, with maturation that varies substantially by domain and by proximity to real-world clinical integration.53,95 In clinical translation and decision support, recent work demonstrates that deployment and monitoring of prediction systems in routine colorectal cancer surgical workflows are feasible, but such implementation-level evidence remains uncommon to the volume of retrospective development studies.5,53,95 Consistent with broader concerns in clinical prediction and medical AI, the primary constraint on translation is not proof-of-concept discrimination but the combination of generalizability, calibration, and clinically meaningful utility under routine conditions.42–44,48–50

Preoperative planning remains dominated by imaging-based methods, particularly segmentation and localization, where performance may appear strong in curated datasets yet degrade under cross-site variation in acquisition protocols and label practices. The translational requirement is therefore not solely architectural sophistication but demonstrable robustness to dataset shift, transparent reporting of data provenance, and external validation that reflects target deployment heterogeneity.42–44,48–50 In parallel, perioperative risk prediction models for thoracic and esophageal oncologic surgery increasingly integrate imaging with clinical variables, and emerging reports emphasize explainability; however, interpretability claims must be matched to calibrated risk estimates and decision-relevant evaluations to support counseling, triage, and perioperative planning.23,34,43,44

Intraoperative navigation and robotic assistance/control are advancing in tandem with surgical robotics and training-focused AI, yet the evidentiary base remains weighted toward feasibility demonstrations and educational applications rather than prospective, outcome-linked clinical utility.53,81 The methodological gap is well captured by contemporary guidance: early-stage evaluations should explicitly report human–AI interaction, workflow integration, safety mitigations, and error–case behavior (DECIDE-AI), while later-stage interventional studies should adhere to AI-specific trial and protocol standards (CONSORT-AI, SPIRIT-AI).15,38,51

Future progress in surgical oncology AI will depend on aligning model development with deployment realities: multicenter and temporally robust evaluation, explicit calibration assessment and updating strategies, decision-analytic utility assessment at clinically relevant thresholds, and systematic appraisal of bias and subgroup performance when outcomes and labels reflect structural differences in access and care pathways. These requirements are increasingly codified in updated reporting and risk-of-bias instruments (TRIPOD+AI, PROBAST+AI) and in good ML practice guidance emphasizing lifecycle governance, monitoring, and safety assurance for medical-device ML systems.49,50 When these evidentiary expectations are met, AI/ML systems in surgical oncology are more likely to transition from high in silico performance to reproducible, safe, and clinically useful tools that improve perioperative decision-making rather than merely predicting outcomes.

List of Abbreviations
  • AI: Artificial intelligence
  • AiLES: Artificial intelligence Laparoscopic Exploration System
  • CDS: Clinical Decision Support
  • CT: Computed Tomography
  • DCA: Decision Curve Analysis
  • FDA: Food and Drug Administration
  • HER: Electronic Health Record
  • LLM: Large Language Model
  • MDR: Medical Device Regulation
  • ML: Machine Learning
  • MRI: Magnetic Resonance Imaging
  • NIR: Near-Infrared
  • NLP: Natural language processing
  • OCT: Optical coherence tomography
  • PCCP: Predetermined Change Control Plan
  • WSI: Whole-Slide Imaging

References
  1. Williams TL, Saadat LV, Gonen M, Wei A, Do RK, Simpson AL. Radiomics in surgical oncology: applications and challenges. Comput Assist Surg (Abingdon). 2021;26(1):85–96. https://doi.org/10.1080/24699322.2021.1994014
  2. Lin G, Li R, Li X, Wang D, Chen X. Advances in the application of three-dimensional reconstruction in thoracic surgery: a comprehensive review. Thorac Cancer. 2025;16(17):e70159. https://doi.org/10.1111/1759-7714.70159
  3. Leszczyńska A, Obuchowicz R, Strzelecki M, Seweryn M. The integration of artificial intelligence into robotic cancer surgery: a systematic review. J Clin Med. 2025;14(17):6181. https://doi.org/10.3390/jcm14176181
  4. Buck L, Kohler J, Risch J, Incesu RB, Hügelmann K, Weiss ML, et al. The emerging role of multimodal artificial intelligence in urological surgery. Curr Oncol. 2025;32(12):665. https://doi.org/10.3390/curroncol32120665
  5. Rosen AW, Ose I, Gögenur M, Andersen LP, Bojesen RD, Vogelsang RP, et al.; AID-SURG study group. Clinical implementation of an AI-based prediction model for decision support for patients undergoing colorectal cancer surgery. Nat Med. 2025;31(11):3737–48. https://doi.org/10.1038/s41591-025-03942-x
  6. Dal Cero M, Gibert J, Grande L, Gimeno M, Osorio J, Bencivenga M, et al.; On behalf of the Spanish Eurecca Esophagogastric Cancer Group and the European Gastrodata Study Group. International external validation of risk prediction model of 90-day mortality after gastrectomy for cancer using machine learning. Cancers (Basel). 2024;16(13):2463. https://doi.org/10.3390/cancers16132463
  7. Vakili-Ojarood M, Naseri A, Shirinzadeh-Dastgiri A, Saberi A, HaghighiKian SM, Rahmani A, et al. Ethical considerations and equipoise in cancer surgery. Indian J Surg Oncol. 2024;15(S3 Suppl 3):363–73. https://doi.org/10.1007/s13193-024-02023-8
  8. Kitaguchi D, Takeshita N, Matsuzaki H, Igaki T, Hasegawa H, Kojima S, et al. Real-time vascular anatomical image navigation for laparoscopic surgery: experimental study. Surg Endosc. 2022;36(8):6105–12. https://doi.org/10.1007/s00464-022-09384-7
  9. Ten Brink R, Schram R, Kats-Ugurlu G, Kwee T, Hemmer P, Havenga K, et al. Feasibility of optical stereotactic navigation for rectosigmoid cancer with deep learning-supported 3D modelling. Eur J Surg Oncol. 2025;51(11):110397. https://doi.org/10.1016/j.ejso.2025.110397
  10. Rus G, Andras I, Vaida C, Crisan N, Gherman B, Radu C, et al. Artificial intelligence-based hazard detection in robotic-assisted single-incision oncologic surgery. Cancers (Basel). 2023;15(13):3387. https://doi.org/10.3390/cancers15133387
  11. Fooladgar F, Jamzad A, Connolly L, Santilli A, Kaufmann M, Ren K, et al. Uncertainty estimation for margin detection in cancer surgery using mass spectrometry. Int J Comput Assist Radiol Surg. 2022;17(12):2305–13. https://doi.org/10.1007/s11548-022-02764-3
  12. Li C, Song J, Yan R, Zhang B, Wang B, Fan J, et al. Non-contact artificial intelligence-assisted intraoperative 3D navigation technology prospective application study in lung cancer surgery. J Thorac Dis. 2025;17(11):9610–21. https://doi.org/10.21037/jtd-2025-1136
  13. Macheka S, Ng PY, Ginsburg O, Hope A, Sullivan R, Aggarwal A. Prospective evaluation of artificial intelligence (AI) applications for use in cancer pathways following diagnosis: a systematic review. BMJ Oncol. 2024;3(1):e000255. https://doi.org/10.1136/bmjonc-2023-000255
  14. Kawamoto K, Houlihan CA, Balas EA, Lobach DF. Improving clinical practice using clinical decision support systems: a systematic review of trials to identify features critical to success. BMJ. 2005;330(7494):765. https://doi.org/10.1136/bmj.38398.500764.8F
  15. Vasey B, Nagendran M, Campbell B, Clifton DA, Collins GS, Denaxas S, et al.; DECIDE-AI expert group. Reporting guideline for the early-stage clinical evaluation of decision support systems driven by artificial intelligence: DECIDE-AI. BMJ. 2022;28(5):924–33. https://doi.org/10.1136/bmj-2022-070904
  16. Alselaim NA, Alsemari MA, Alyabsi M, Al-Mutairi AM. Factors associated with 30-day mortality and morbidity in patients undergoing emergency colorectal surgery. Ann Saudi Med. 2023;43(6):364–72. https://doi.org/10.5144/0256-4947.2023.364
  17. Xu J, Zhou J, Hu J, Ren Q, Wang X, Shu Y. Development and validation of a machine learning model for survival risk stratification after esophageal cancer surgery. Front Oncol. 2022;12:1068198. https://doi.org/10.3389/fonc.2022.1068198
  18. Bilimoria KY, Liu Y, Paruch JL, Zhou L, Kmiecik TE, Ko CY, et al. Development and evaluation of the universal ACS NSQIP surgical risk calculator: a decision aid and informed consent tool for patients and surgeons. J Am Coll Surg. 2013;217(5):833–42.e1. https://doi.org/10.1016/j.jamcollsurg.2013.07.385
  19. Abikenari M, Jain B, Xu R, Jackson C, Huang J, Bettegowda C, et al. Bridging imaging and molecular biomarkers in trigeminal neuralgia: toward precision diagnosis and prognostication in neuropathic pain. Med Res Arch, [S.l.], v. 13, n. 5, May 2025. ISSN 2375-1924. Available at: https://esmed.org/MRA/mra/article/view/6605. Date accessed: 15 January 2026. https://doi.org/10.18103/mra.v13i5.6605
  20. Namavarian A, Gabinet-Equihua A, Deng Y, Khalid S, Ziai H, Deutsch K, et al. Length of stay prediction models for oral cancer surgery: machine learning, statistical and ACS-NSQIP. Laryngoscope. 2024;134(8):3664–72. https://doi.org/10.1002/lary.31443
  21. Kadomatsu Y, Emoto R, Kubo Y, Nakanishi K, Ueno H, Kato T, et al. Development of a machine learning-based risk model for postoperative complications of lung cancer surgery. Surg Today. 2024;54(12):1482–9. https://doi.org/10.1007/s00595-024-02878-y
  22. Cowling TE, Cromwell DA, Bellot A, Sharples LD, van der Meulen J. Logistic regression and machine learning predicted patient mortality from large sets of diagnosis codes comparably. J Clin Epidemiol. 2021;133:43–52. https://doi.org/10.1016/j.jclinepi.2020.12.018
  23. Klontzas ME, Ri M, Koltsakis E, Stenqvist E, Kalarakis G, Boström E, et al. Prediction of anastomotic leakage in esophageal cancer surgery: a multimodal machine learning model integrating imaging and clinical data. Acad Radiol. 2024;31(12):4878–85. https://doi.org/10.1016/j.acra.2024.06.026
  24. Chen H, Gou L, Fang Z, Dou Q, Chen H, Chen C, et al. Artificial intelligence-assisted real-time recognition of intra-abdominal metastasis during laparoscopic gastric cancer surgery. NPJ Digit Med. 2025;8(1):9. https://doi.org/10.1038/s41746-024-01372-6
  25. Barber EL, Garg R, Persenaire C, Simon M. Natural language processing with machine learning to predict outcomes after ovarian cancer surgery. Gynecol Oncol. 2021;160(1):182–6. https://doi.org/10.1016/j.ygyno.2020.10.004
  26. Rossi LA, Melstrom LG, Fong Y, Sun V. Predicting post-discharge cancer surgery complications via telemonitoring of patient-reported outcomes and patient-generated health data. J Surg Oncol. 2021;123(5):1345–52. https://doi.org/10.1002/jso.26413
  27. Yan YD, Wu XW, Li Y, Lin HW, Zhang ZT, Jia D, et al.; CRC-VTE investigators. Machine learning to predict venous thromboembolism after colorectal cancer surgery: a Chinese dynamic modelling study. Int J Surg. 2025; https://doi.org/10.1097/JS9.0000000000004036
  28. Ri M, Nunobe S, Narita T, Seto Y, Kawazoe Y, Ohe K, et al. Time-sequential prediction of postoperative complications after gastric cancer surgery using machine learning: a multicenter cohort study. Gastric Cancer. 2025;28(6):1273–81. https://doi.org/10.1007/s10120-025-01658-y
  29. Zeng S, Li L, Hu Y, Luo L, Fang Y. Machine learning approaches for the prediction of postoperative complication risk in liver resection patients. BMC Med Inform Decis Mak. 2021;21(1):371. https://doi.org/10.1186/s12911-021-01731-3
  30. Lin V, Tsouchnika A, Allakhverdiiev E, Rosen AW, Gögenur M, Clausen JS, et al. Training prediction models for individual risk assessment of postoperative complications after surgery for colorectal cancer. Tech Coloproctol. 2022;26(8):665–75. https://doi.org/10.1007/s10151-022-02624-x
  31. Lee S, Oh HJ, Yoo H, Kim CY. Machine learning insight: unveiling overlooked risk factors for postoperative complications in gastric cancer. Cancers (Basel). 2025;17(7):1225. https://doi.org/10.3390/cancers17071225
  32. Khan MF, Cahill RA. Peroperative personalised decision support and analytics for colon cancer surgery– Short report. Eur J Surg Oncol. 2021;47(2):477–9. https://doi.org/10.1016/j.ejso.2020.04.010
  33. Li R, Zhao Z, Yu J. Development and validation of a machine learning model for predicting early postoperative complications after radical gastrectomy. Front Oncol. 2025;15:1631260. https://doi.org/10.3389/fonc.2025.1631260
  34. Chen S, Deng T, Yang Q, Li J, Shen J, Luo X, et al. Development and validation of an explainable machine learning model for predicting postoperative pulmonary complications after lung cancer surgery: a machine learning study. EClinicalMedicine. 2025;86:103386. https://doi.org/10.1016/j.eclinm.2025.103386
  35. Yang X, Dou F, Tang G, Xiu R, Zhao X. Interpretable machine learning model for predicting anastomotic leak after esophageal cancer surgery via LightGBM. BMC Cancer. 2025;25(1):976. https://doi.org/10.1186/s12885-025-14387-3
  36. Hu J, Liu Q, He W, Wu J, Zhang D, Sun C, et al. Automated machine learning model for predicting anastomotic strictures after esophageal cancer surgery: a retrospective cohort study. Surg Endosc. 2025;39(6):3737–48. https://doi.org/10.1007/s00464-025-11759-5
  37. Bektaş M, Burchell GL, Bonjer HJ, van der Peet DL. Machine learning applications in upper gastrointestinal cancer surgery: a systematic review. Surg Endosc. 2023;37(1):75–89. https://doi.org/10.1007/s00464-022-09516-z
  38. Liu X, Cruz Rivera S, Moher D, Calvert MJ, Denniston AK, Chan AW, et al.; SPIRIT-AI and CONSORT-AI Working Group. Reporting guidelines for clinical trial reports for interventions involving artificial intelligence: the CONSORT-AI extension. Nat Med. 2020;26(9):1364–74. https://doi.org/10.1038/s41591-020-1034-x
  39. Rahman SA, Walker RC, Lloyd MA, Grace BL, van Boxel GI, Kingma BF, et al.; OCCAMS Consortium. Machine learning to predict early recurrence after oesophageal cancer surgery. Br J Surg. 2020;107(8):1042–52. https://doi.org/10.1002/bjs.11461
  40. Abikenari M, Schonfeld E, Choi J, Kim LH, Lim M. Revisiting glioblastoma classification through an immunological lens: a narrative review. Glioma. 2024;7(2)3–9. https://doi.org/10.4103/glioma.glioma_4_24
  41. Vargas LO, Himic V, Otaner F, Abikenari M, Chandar J, Govindarajan V, et al. Modulating the glioma microenvironment with laser interstitial thermal therapy: mechanisms and therapeutic implications. J Neurooncol. 2025;176(1):99. https://doi.org/10.1007/s11060-025-05305-5
  42. Abikenari M, Enayati I, Fountain DM, Leite MI. Navigating glioblastoma therapy: a narrative review of emerging immunotherapeutics and small-molecule inhibitors. Microbes Immun. 2025;2(4):132–43. https://doi.org/10.36922/mi.5075
  43. Van Calster B, McLernon DJ, van Smeden M, Wynants L, Steyerberg EW. Topic Group Evaluating diagnostic tests and prediction models of the STRATOS initiative. Calibration: the Achilles heel of predictive analytics. BMC Med. 2019;17(1):230. https://doi.org/10.1186/s12916-019-1466-7
  44. Vickers AJ, Elkin EB. Decision curve analysis: a novel method for evaluating prediction models. Med Decis Making. 2006;26(6):565–74. https://doi.org/10.1177/0272989X06295361
  45. Caldon LJ, Collins KA, Reed MW, Sivell S, Austoker J, Clements AM, et al.; BresDex Group. Clinicians’ concerns about decision support interventions for patients facing breast cancer surgery options: understanding the challenge of implementing shared decision-making. Health Expect. 2011;14(2):133–46. https://doi.org/10.1111/j.1369-7625.2010.00633.x
  46. Yen RW, Durand MA, Harris C, Cohen S, Ward A, O’Malley AJ, et al. Text-only and picture conversation aids both supported shared decision making for breast cancer surgery: analysis from a cluster randomized trial. Patient Educ Couns. 2020;103(11):2235–43. https://doi.org/10.1016/j.pec.2020.07.015
  47. Obermeyer Z, Powers B, Vogeli C, Mullainathan S. Dissecting racial bias in an algorithm used to manage the health of populations. Science. 2019;366(6464):447–53. https://doi.org/10.1126/science.aax2342
  48. Collins GS, Reitsma JB, Altman DG, Moons KGM. Transparent reporting of a multivariable prediction model for individual prognosis or diagnosis (TRIPOD): the TRIPOD statement. BMJ. 2015;350(4):g7594. https://doi.org/10.1136/bmj.g7594.
  49. Collins GS, Moons KG, Dhiman P, Riley RD, Beam AL, Van Calster B, et al. TRIPOD+AI statement: updated guidance for reporting clinical prediction models that use regression or machine learning methods. BMJ. 2024;385:e078378. https://doi.org/10.1136/bmj-2023-078378
  50. Moons KG, Damen JA, Kaul T, Hooft L, Andaur Navarro C, Dhiman P, et al. PROBAST+AI: an updated quality, risk of bias, and applicability assessment tool for prediction models using regression or artificial intelligence methods. BMJ. 2025;388:e082505. https://doi.org/10.1136/bmj-2024-082505
  51. Cruz Rivera S, Liu X, Chan AW, Denniston AK, Calvert MJ, Darzi A, et al.; SPIRIT-AI and CONSORT-AI Working Group; SPIRIT-AI and CONSORT-AI Steering Group; SPIRIT-AI and CONSORT-AI Consensus Group. Guidelines for clinical trial protocols for interventions involving artificial intelligence: the SPIRIT-AI extension. Nat Med. 2020;26(9):1351–63. https://doi.org/10.1038/s41591-020-1037-7
  52. Chan SY, Twohig P. Artificial intelligence in liver cancer surgery: predicting success before the first incision. World J Gastroenterol. 2025;31(16):107221. https://doi.org/10.3748/wjg.v31.i16.107221
  53. International Medical Device Regulators Forum (IMDRF). Software as a Medical Device (SaMD): key definitions. IMDRF/SaMD WG/N10FINAL:2013. Published December 18, 2013. Accessed February 25, 2026. https://www.imdrf.org/sites/default/files/docs/imdrf/final/technical/imdrf-tech-131209-samd-key-definitions-140901.pdf
  54. International Medical Device Regulators Forum (IMDRF). Software as a Medical Device (SaMD): Clinical Evaluation. IMDRF/SaMD WG/N41FINAL:2017. Published September 21, 2017. Accessed February 25, 2026. https://www.imdrf.org/sites/default/files/docs/imdrf/final/technical/imdrf-tech-170921-samd-n41-clinical-evaluation_1.pdf
  55. García-Granero Á, Jerí-McFarlane S, Ochogavía A, Gamundí-Cuesta M, Garcia-Granero E, González-Argenté FX. 3D reconstructions in rectal cancer. New tools for better diagnosis and surgical planning. Cir Esp (Engl Ed). 2025;103(9):800198. https://doi.org/10.1016/j.cireng.2025.800198
  56. US Food and Drug Administration. Software as a Medical Device (SaMD). Updated December 4, 2018. Accessed February 25, 2026. https://www.fda.gov/medical-devices/digital-health-center-excellence/software-medical-device-samd
  57. Checcucci E, De Cillis S, Porpiglia F. 3D-printed models and virtual reality as new tools for image-guided robot-assisted nephron-sparing surgery: a systematic review of the newest evidences. Curr Opin Urol. 2020;30(1):55–64. https://doi.org/10.1097/MOU.0000000000000686
  58. Chen X, Xu H, Qi Q, Sun C, Jin J, Zhao H, et al. AI-based chest CT semantic segmentation algorithm enables semi-automated lung cancer surgery planning by recognizing anatomical variants of pulmonary vessels. Front Oncol. 2022;12:1021084. https://doi.org/10.3389/fonc.2022.1021084
  59. Wang Z, Sun H, Li J, Chen J, Meng F, Li H, et al. Preoperative Prediction of Axillary Lymph Node Metastasis in Breast Cancer Using CNN Based on Multiparametric MRI. J Magn Reson Imaging. 2022 Sep;56(3):700–9. https://doi.org/10.1002/jmri.28082
  60. Tan Y, Feng LJ, Huang YH, Xue JW, Long LL, Feng ZB. A comprehensive radiopathological nomogram for the prediction of pathological staging in gastric cancer using CT-derived and WSI-based features. Transl Oncol. 2024;40:101864. https://doi.org/10.1016/j.tranon.2023.101864
  61. Yao Y, Cen X, Gan L, Jiang J, Wang M, Xu Y, et al. Automated esophageal cancer staging from free-text radiology reports: large language model evaluation study. JMIR Med Inform. 2025;13:e75556. https://doi.org/10.2196/75556
  62. Tang Z, Feng S, Liu Q, Ban Y, Zhang Y. Using pathomics-based model for predicting positive surgical margins in patients with esophageal squamous cell carcinoma: a comparative study of decision tree and nomogram. Int J Gen Med. 2024;17:5869–82. https://doi.org/10.2147/IJGM.S495296
  63. Awad MH, Sanchez M, Abikenari MA. The values work of restorative ventures: the role of founders’ embodied embeddedness with at-risk social groups. J Bus Ventur Insights. 2022;18:e00337. https://doi.org/10.1016/j.jbvi.2022.e00337
  64. Kaida S, Murakami Y, Masaki Y, et al. Artificial intelligence-driven 3-dimensional simulation system for enhanced preoperative planning in gastric cancer surgery: a retrospective validation study. J Gastrointest Surg. Published online December 2025:102295. https://doi.org/10.1016/j.gassur.2025.102295
  65. Hamabe A, Ishii M, Kamoda R, Sasuga S, Okuya K, Okita K, et al. Artificial intelligence-based technology to make a three-dimensional pelvic model for preoperative simulation of rectal cancer surgery using MRI. Ann Gastroenterol Surg. 2022;6(6):788–94. https://doi.org/10.1002/ags3.12574
  66. Aisu Y, Okada T, Itatani Y, Masuo A, Tani R, Fujimoto K, et al. Automatic segmentation of male pelvic floor soft tissue structures for anatomical simulation and morphological assessment in lower rectal cancer surgery. Tech Coloproctol. 2025;29(1):176. https://doi.org/10.1007/s10151-025-03218-z
  67. Wang F, Xiao C, Jia T, Pan L, Du F, Wang Z. Hepatobiliary surgery based on intelligent image segmentation technology. Open Life Sci. 2023 Aug;18(1):20220674. https://doi.org/10.1515/biol-2022-0674
  68. Hofmann FO, Heiliger C, Tschaidse T, Jarmusch S, Auhage LA, Aghamaliyev U, et al. Validation of body composition parameters extracted via deep learning-based segmentation from routine computed tomographies. Sci Rep. 2025;15(1):11909. https://doi.org/10.1038/s41598-025-96238-6
  69. Suthakaran R, Cao K, Arafat Y, Yeung J, Chan S, Master M, et al. Body composition assessment by artificial intelligence can be a predictive tool for short-term postoperative complications in Hartmann’s reversals. BMC Surg. 2024 Apr;24(1):111. https://doi.org/10.1186/s12893-024-02408-0
  70. Kim J, Han SH, Kim HI. Detection of sarcopenic obesity and prediction of long-term survival in patients with gastric cancer using preoperative computed tomography and machine learning. J Surg Oncol. 2021;124(8):1347–55. https://doi.org/10.1002/jso.26668
  71. Amparore D, Piramide F, De Cillis S, Verri P, Piana A, Pecoraro A, et al.; Renal Cancer Working Group of the Young Academic Urologists (YAU) and European Association of Urology (EAU). Robotic partial nephrectomy in 3D virtual reconstructions era: is the paradigm changed? World J Urol. 2022;40(3):659–70. https://doi.org/10.1007/s00345-022-03964-x
  72. Cheng H, Xu H, Peng B, Huang X, Hu Y, Zheng C, et al. Illuminating the future of precision cancer surgery with fluorescence imaging and artificial intelligence convergence. NPJ Precis Oncol. 2024;8(1):196. https://doi.org/10.1038/s41698-024-00699-3
  73. Iresjö BM, Smedh U, Engström C, Persson J, Mårtensson C, Lundholm K. A randomized translational study on protein and glucose metabolism in skeletal muscles evaluated by gene-ontology, following preoperative oral carbohydrate loading compared to overnight peripheral parenteral nutrition (PPN) before major cancer surgery. J Transl Med. 2024;22(1):675. https://doi.org/10.1186/s12967-024-05484-1
  74. van Oosterom MN, van Leeuwen SI, Mazzone E, Dell’Oglio P, Buckle T, van Beurden F, et al. Click-on fluorescence detectors: using robotic surgical instruments to characterize molecular tissue aspects. J Robot Surg. 2023;17(1):131–40. https://doi.org/10.1007/s11701-022-01382-0
  75. Müller D, Stier R, Straatman J, Babic B, Schiffmann L, Eckhoff J, et al. ICG-lymphknoten-mapping in der Tumorchirurgie des oberen Gastrointestinaltrakts. Chirurgie (Heidelb). 2022;93(10):925–33. https://doi.org/10.1007/s00104-022-01659-y
  76. Hoogendam JP, Hobbelink MG, Veldhuis WB, Verheijen RH, van Diest PJ, Zweemer RP. Preoperative sentinel node mapping with (99m)Tc-nanocolloid SPECT-CT significantly reduces the intraoperative sentinel node retrieval time in robot assisted laparoscopic cervical cancer surgery. Gynecol Oncol. 2013;129(2):389–94. https://doi.org/10.1016/j.ygyno.2013.02.020
  77. Alikarami M, Faraj TA, Hama NH, Hosseini AS, Habibi P, Samiei Mosleh I, et al. Artificial intelligence in advancing optical coherence tomography for disease detection and cancer diagnosis: A scoping review. Eur J Surg Oncol. 2025;51(9):110188. https://doi.org/10.1016/j.ejso.2025.110188
  78. Levy Y, Rempel D, Nguyen M, Yassine A, Sanati-Burns M, Salgia P, et al. The fusion of wide field optical coherence tomography and AI: advancing breast cancer surgical margin visualization. Life (Basel). 2023;13(12):2340. https://doi.org/10.3390/life13122340
  79. Pavone M, Innocenzi C, Carles E, Bizzarri N, Moro F, Ferrari FA, et al. Cutting edge microscopic intraoperative tissue assessment for guidance in oncologic surgery: a systematic review of the role of optical coherence tomography. Ann Surg Oncol. 2025;32(3): 2191–205. https://doi.org/10.1245/s10434-024-16632-8
  80. Abikenari M, Awad MH, Korouri S, Mohseni K, Abikenari D, Freichel R, et al. Reframing clinical AI evaluation in the era of generative models: toward multidimensional, stakeholder-informed, and safety-centric frameworks for real-world health care deployment. PJS. 2025;11:100089. https://doi.org/10.70389/PJS.100089
  81. Tanzim U, Khan IA, Abikenari M, Al-Deen RH, Miah L, Blaaza M, et al. Transforming patient-provider communication: the role of artificial intelligence in advancing health literacy – a comprehensive review. PJS. 2025;13:100095. https://doi.org/10.70389/PJS.100095
  82. Cheng Q, Dong Y (Hussein AF, editor). Da Vinci robot-assisted video image processing under artificial intelligence vision processing technology. Comput Math Methods Med. 2022. pp. 1–10. https://doi.org/10.1155/2022/2752444
  83. Ebigbo A, Mendel R, Scheppach MW, Probst A, Shahidi N, Prinz F, et al. Vessel and tissue recognition during third-space endoscopy using a deep learning algorithm. Gut. 2022;71(12):2388–90. https://doi.org/10.1136/gutjnl-2021-326470
  84. Checcucci E, Piana A, Volpi G, Piazzolla P, Amparore D, De Cillis S, et al. Three-dimensional automatic artificial intelligence driven augmented-reality selective biopsy during nerve-sparing robot-assisted radical prostatectomy: A feasibility and accuracy study. Asian J Urol. 2023;10(4):407–15. https://doi.org/10.1016/j.ajur.2023.08.001
  85. Park SH, Park HM, Baek KR, Ahn HM, Lee IY, Son GM. Artificial intelligence based real-time microcirculation analysis system for laparoscopic colorectal surgery. World J Gastroenterol. 2020;26(44):6945–62. https://doi.org/10.3748/wjg.v26.i44.6945
  86. Boland PA, Hardy NP, Moynihan A, McEntee PD, Loo C, Fenlon H, et al. Intraoperative near infrared functional imaging of rectal cancer using artificial intelligence methods – now and near future state of the art. Eur J Nucl Med Mol Imaging. 2024;51(10):3135–48. https://doi.org/10.1007/s00259-024-06731-9
  87. Checcucci E, Veccia A, Puliatti S, De Backer P, Piazza P, Kowalewski KF, et al. Metaverse in surgery – origins and future potential. Nat Rev Urol. 2026;23(1):50–63. https://doi.org/10.1038/s41585-024-00941-4
  88. Won NJ, Bartling M, La Macchia J, Markevich S, Holtshousen S, Jagota A, et al. Deep learning-enabled fluorescence imaging for surgical guidance: in silico training for oral cancer depth quantification. J Biomed Opt. 2025;30(S1 Suppl 1):S13706. https://doi.org/10.1117/1.JBO.30.S1.S13706
  89. Kurosawa H, Won NJ, Wunder JB, Patil S, Bartling M, Najjar E, et al. Deep learning-enabled fluorescence imaging for oral cancer margin classification in preclinical models. J Biomed Opt. 2025;30(S3 Suppl 3):S34109. https://doi.org/10.1117/1.JBO.30.S3.S34109
  90. Abosheisha M, Nasr E, Ali M, Tamanna R, Bin Halim S, Rakib Hasan M, et al. Near-infrared fluorescence imaging in general surgery: applications in vascularization, tumor margin detection, and biliary anatomy. Cureus. 2025;17(9):e92194. https://doi.org/10.7759/cureus.92194
  91. Müller DT, Schiffmann LM, Reisewitz A, Chon SH, Eckhoff JA, Babic B, et al. Mapping the lymphatic drainage pattern of esophageal cancer with near-infrared fluorescent imaging during robotic assisted minimally invasive Ivor Lewis esophagectomy (RAMIE) – First results of the prospective ESOMAP feasibility trial. Cancers (Basel). 2023;15(8):2247. https://doi.org/10.3390/cancers15082247
  92. Connolly L, Fooladgar F, Jamzad A, Kaufmann M, Syeda A, Ren K, et al. ImSpect: image-driven self-supervised learning for surgical margin evaluation with mass spectrometry. Int J Comput Assist Radiol Surg. 2024;19(6):1129–36. https://doi.org/10.1007/s11548-024-03106-1
  93. Connolly L, Jamzad A, Kaufmann M, Farquharson CE, Ren K, Rudan JF, et al. Combined mass spectrometry and histopathology imaging for perioperative tissue assessment in cancer surgery. J Imaging. 2021;7(10):203. https://doi.org/10.3390/jimaging7100203
  94. Boal M, Reali C, Duhoky R, Gill T, Khan J, Miskovic D, et al. Association of skill and errors with outcomes in robotic rectal cancer surgery. Surg Endosc. 2025. https://doi.org/10.1007/s00464-025-12393-x
  95. Quero G, Mascagni P, Kolbinger FR, Fiorillo C, De Sio D, Longo F, et al. Artificial intelligence in colorectal cancer surgery: present and future perspectives. Cancers (Basel). 2022;14(15):3803. https://doi.org/10.3390/cancers14153803
  96. Siddarth P, Abikenari M, Grzenda A, Cappelletti M, Oughli H, Liu C, et al. Inflammatory markers of geriatric depression response to Tai Chi or Health Education Adjunct Interventions. Am J Geriatr Psychiatry. 2023;31(1):22–32. https://doi.org/10.1016/j.jagp.2022.08.004
  97. Ajam Oughli H, Siddarth P, Lum M, Tang L, Ito B, Abikenari M, et al. Peripheral Alzheimer’s disease biomarkers are related to change in subjective memory in older women with cardiovascular risk factors in a trial of yoga vs. memory training: lien établi entre les biomarqueurs périphériques de la maladie d’Alzheimer et l’amélioration de la mémoire subjective chez les femmes âgées présentant des facteurs de risque cardiovasculaire dans le cadre d’un essai comparant le yoga à l’entraînement de la mémoire. Can J Psychiatry. 2025;0(0):7067437251343291. https://doi.org/10.1177/07067437251343291
  98. Basereh M, Abikenari MA, Sadeghzadeh S, et al. ConvNeXt-driven detection of Alzheimer’s disease: a benchmark study on expert-annotated alzaset mri dataset across anatomical planes. Diagnostics (Basel). 2025;15(23):2997. Published November 25, 2025. https://doi.org/10.3390/diagnostics15232997
  99. Luo X, Wang X, Eweje F, Zhang X, Yang S, Quinton R, et al. Ensemble learning of foundation models for precision oncology. arXiv preprint. 2025; arXiv:2508.16085.
  100. Abikenari M, Liu J, Ha JH, Annagiri S, Himic V, Medikonda R, et al. Emerging trends in cell-based therapies: contemporary advances and ethical considerations in translational neurosurgical oncology. J Neurooncol. 2025;175(1):1–20. https://doi.org/10.1007/s11060-025-05170-2
  101. Li Y, Wu X, Fang D, Luo Y. Informing immunotherapy with multi-omics driven machine learning. NPJ Digit Med. 2024;7(1):67. https://doi.org/10.1038/s41746-024-01043-6

Cite this article as:
Abikenari M., Enayati I, Mohseni K, Sanker V, Ablyazova F, Vargas L, Himic V, Chen Y, Baibussinov A, Polkampally S, Liu S, Poe J, Kim C, Davani D, Jain A, Freichel R, Abikenari D, Kerwan A and Agha R. Artificial Intelligence Across the Surgical Oncology Continuum: Decision Support, Operative Intelligence, and a Translation-First Roadmap. Premier Journal of Science 2026;20:100269

Export Test
Download an RIS file Download an RIS file
peer-reviewed
Screened by iThenticate - professional plagiarism prevention
Open Access


Premier Science
Publishing Science that inspires