Stop Holding Your Breath: CT-Informed Gaussian Splatting for Dynamic Bronchoscopy

2026-04-30Computer Vision and Pattern Recognition

Computer Vision and Pattern Recognition
AI summary

The authors address a problem in bronchoscopic navigation where breathing causes the airways to move, making it hard to match live videos to a pre-surgery CT scan. Normally, doctors ask patients to hold their breath to keep the anatomy still, but this is tricky and disrupts the procedure. The authors use pairs of CT scans taken during breathing to model how the airways move, allowing their system to track breathing and adjust the video alignment continuously without breath-holds. They test their method with a new simulation tool called RESPIRE and show it improves accuracy and speed compared to older methods.

bronchoscopic navigationCT scanrespiratory motionbreath-hold protocolrespiratory modelingimage registrationGaussian splattingendoscopic videosimulation
Authors
Andrea Dunn Beltran, Daniel Rho, Aarav Mehta, Xinqi Xiong, Raúl San José Estépar, Ron Alterovitz, Marc Niethammer, Roni Sengupta
Abstract
Bronchoscopic navigation relies on registering endoscopic video to a preoperative CT scan, but respiratory motion deforms the airway by 5-20 mm, creating CT-to-body divergence that limits localization accuracy. In practice, this is mitigated through breath-hold protocols, which attempt to match the intraoperative anatomy to a static CT, but are difficult to reproduce and disrupt clinical workflow. We propose to eliminate the need for breath-hold protocols by leveraging patient-specific respiratory modeling. Paired inhale-exhale CT scans, already acquired for planning, implicitly define the patient-specific deformation space of the breathing airway. By registering these scans, we reduce respiratory motion to a single scalar breathing phase per frame, constraining all reconstructions to anatomically observed configurations. We embed this representation within a mesh-anchored Gaussian splatting framework, where a lightweight estimator infers breathing phase directly from endoscopic RGB, enabling continuous, deformation-aware reconstruction throughout the respiratory cycle without breath-holds or external sensing. To enable quantitative evaluation, we introduce RESPIRE, a physically grounded bronchoscopy simulation pipeline with per-frame ground truth for geometry, pose, breathing phase, and deformation. Experiments on RESPIRE show that our approach achieves geometrically faithful reconstruction, over 20x faster training, and 1.22 mm target localization accuracy (within the 3mm clinically relevant tolerances) outperforming unconstrained single-CT baselines. Please check out our website for additional visuals: https://asdunnbe.github.io/RESPIRE/