Nvidia Developer 09月03日
NVIDIA Isaac赋能远程手术,加速医疗创新
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

随着全球外科医生短缺加剧,远程手术正成为关键的医疗解决方案。NVIDIA Isaac for Healthcare提供了一个生产就绪的模块化远程手术工作流,涵盖视频和传感器流、机器人控制、触觉反馈及模拟,旨在解决超低延迟视频、可靠的远程机器人控制和无缝硬件集成等技术挑战。该平台支持从模拟到临床部署的统一架构,通过GPUDirect Sensor IO、硬件加速视频流、RTI Connext DDS通信层以及灵活的控制接口(如Xbox手柄和Haply Inverse3),确保了低延迟和高可靠性。在实际部署中,它已实现患者转运时间减少50%,农村地区专业手术可及性提高3倍,并大幅提升了手术培训效率。

⚕️ **远程手术的必要性与技术驱动:** 全球外科医生短缺(预计2030年达450万)和偏远地区医疗资源不足,使得远程手术从实验性走向必然。5G网络、低延迟骨干网、成熟的AI和仿真技术,以及标准化的平台,共同推动了远程手术的发展,克服了传统上技术和基础设施的限制。

💡 **NVIDIA Isaac for Healthcare的核心优势:** 该平台提供了一个端到端的、可生产的模块化远程手术工作流,解决了超低延迟视频、可靠的远程机器人控制和无缝硬件集成等关键技术挑战。它统一了从NVIDIA DGX、OVX到IGX/AGX的计算架构,并提供包括样本工作流、高保真医疗传感器模拟、仿真就绪的资产目录、预训练AI模型和合成数据生成等工具集,加速了医疗机器人和远程手术系统的开发与部署。

🚀 **关键技术架构与性能保障:** 平台采用GPUDirect Sensor IO和Holoscan Sensor Bridge实现超低延迟视频流;利用硬件加速视频编码(H.264/NVJPEG)和RTI Connext DDS通信层确保了视频、控制和遥测数据的可靠、低延迟传输。支持Xbox手柄或Haply Inverse3触觉设备进行直观控制,并具备自动姿态重置等安全特性。在延迟测试中,光子到玻璃的延迟最低可达35.2毫秒,满足临床操作要求。

🏥 **部署灵活性与临床影响力:** 该工作流支持在物理手术室或Isaac Sim模拟环境中部署。容器化设计确保了跨环境的一致性,并能与现有医疗基础设施无缝集成。临床试点数据显示,患者转运时间减少50%,农村地区专业手术可及性提高3倍,培训效率提升40%,且在超过1000例手术中未报告任何与延迟相关的并发症,有力证明了其临床价值和广泛应用潜力。

Telesurgery is no longer a futuristic idea—it’s quickly becoming essential to how care is delivered. With a global shortage of surgeons projected to reach 4.5 million by 2030, and rural hospitals struggling to access specialists, the ability for experts to operate remotely is shifting from experimental to inevitable.

What’s changed?

    Networking has caught up. 5G and low-latency backbones make real-time video collaboration possible across continents.AI and simulation have matured. Surgeons can now train and validate systems in photorealistic environments before entering the OR.Standardized platforms are here. Instead of stitching together custom pipelines for sensors, video, and robotics, developers can build on shared infrastructure that accelerates progress.

But telesurgery comes with big technical challenges:

    Ultra-low latency video for surgical precision.Reliable remote robot control with haptic feedback.Seamless hardware integration across diverse solutions.

This is where NVIDIA Isaac for Healthcare comes in. It gives developers a production-ready, modular telesurgery workflow—covering video and sensor streaming, robot control, haptics, and simulation—that you can adapt, extend, and deploy in both training and clinical settings.

In this post, you’ll see how the telesurgery workflow works, how to get started, and why it’s the fastest way to build the next generation of surgical robotics.

What is Isaac for Healthcare?

Bringing a powerful three-computer architecture (NVIDIA DGX, NVIDIA OVX, NVIDIA IGX/NVIDIA AGX) to healthcare robotics, Isaac for Healthcare unifies the full development stack. It offers a comprehensive set of tools and building blocks, including:

    End-to-end sample workflows (surgical and imaging).High-fidelity medical sensor simulation.Simulation-ready asset catalog (robots, tools, anatomies)Pretrained AI models and policy baselines.Synthetic data generation pipelines.

With this foundation, you can move from simulation to clinical deployment using the same architecture.

How the Telesurgery Workflow works

The telesurgery workflow connects a surgeon’s control station to a patient-side surgical robot over a high-speed network.

    Surgeon side: Views multiple camera feeds (overview + robot-eye) and issues commands via Xbox or haptic controller. Patient side: Cameras capture the surgical field while the robot executes precise maneuvers based on surgeon input.Simulation mode: Identical setup in Isaac Sim, allowing safe training and skill transfer.

The result: clinicians can perform procedures in a crisis, in remote hospitals, or across continents—without compromising responsiveness.

Figure 1. Telesurgery workflow diagram

System architecture

ComponentWhat it doesWhy it matters
GPU Direct Sensor IOStreams video directly to GPU via Holoscan Sensor BridgeUltra-low latency integration of cameras/sensors 
Video streamingMulti-camera capture (robot + room), hardware-accelerated High quality, little to no delay
RTI Connext DDSManages video, control, telemetry across domains with QoS controlsSecure, reliable, medical-grade comms
Control interfaces Xbox controller or Haply Inverse3 haptic deviceFamiliar tools with force feedback up to 3.3N
Manages video, control, and telemetry across domains with QoS controlsPose reset, tool homing, dead zonesGuarantees safe recovery in clinical scenarios

Let’s get into the specifics and outline the architecture behind this solution:

System architecture:

    GPUDirect sensor IO: The system uses the NVIDIA Holoscan Sensor Bridge (HSB) to stream video directly to the GPU in real time. HSB enables low-latency data transfer by connecting high-speed sensors to the processing pipeline through an FPGA-based interface over Ethernet. This simplifies the integration of sensors and actuators in edge AI healthcare applications.Video streaming: The system captures two camera views, a room overview, and a detailed robot-eye view. Video encoding uses NVIDIA hardware acceleration to maintain quality while minimizing latency. You can choose between H.264 for compatibility or NVJPEG for scenarios requiring the lowest possible delay.
      Multi-camera support: Captures simultaneous feeds from robot-mounted cameras and room overview cameras (RealSense/CV2 compatible).Hardware-accelerated encoding:
        NVIDIA Video Codec (NVC) for H.264/H.265: Ideal for bandwidth-constrained scenarios.NVJPEG encoding: Ultra-low latency option with configurable quality (1-100).
    Communication layer: RTI Connext Data Distributed Service (DDS) handles all data transport between sites, ensuring medical-grade reliability, low latency, and data integrity. Video streams, control commands, and robot feedback travel on separate channels, each optimized through quality of service controls for its specific needs.
      RTI Connext DDS infrastructure: Secure, medical-grade reliability with guaranteed message deliveryDomain isolation: Separate DDS domains for video streams, control commands, and telemetryTime synchronization: Optional Network Time Protocol (NTP) server integration ensures perfect temporal alignment across all systemsNetwork optimization: Automatic peer discovery and quality-of-service profiles tailored for surgical requirements
    Control interface: Surgeons can use an Xbox controller for basic operations or a Haply Inverse3 device for intuitive control of the robot in 3D space. The control system operates in tandem with the patient-side robot, translating the surgeon’s inputs into precise robot movements.
      Dual control modes:
        Cartesian mode: Direct X-Y-Z positioning for intuitive control.Polar mode: Joint-space control for complex maneuvers.
      Input devices:
        Xbox controller: Familiar interface with dual-stick to control both MIRA arms simultaneously.Haply Inverse 3: Force feedback up to 3.3N for realistic tissue interaction.
      Safety features: Automatic pose reset, tool homing sequences, and configurable dead zones.

Proving readiness for the OR: latency benchmarks

Low latency is critical for telesurgery and the following benchmarks show this workflow meets clinical requirements. 

    HSB with IMX274 camera
    This uses the NVIDIA HSB board with an IMX274 MIPI camera for an ultra-low latency pipeline.HDMI camera with YUAN HSB board
    Existing medical setups often have cameras that have HDMI or SDI output. In this scenario, our partner YUAN HSB board is a great solution and can take video from HDMI or SDI and deliver the data directly to GPUs. The HDMI camera used in this benchmark is a Fujifilm X-T4 camera.

For the display, benchmarking was conducted using a G-Sync-enabled monitor with a 240 Hz refresh rate, operating in Vulkan exclusive display mode.  Latency measurements were captured using NVIDIA Latency and Display Analysis Tool (LDAT).

HSB with imx274 camera

    1080p@60fps (H.264 w/ bitrate set to 10 Mbps)
      Photon to glass: 35.2 +/- 4.77 msEncode+decode: 10.58 +/- 0.64 ms
    4k@60fps (H.265 w/ bitrate set to 30 Mbps)
      Photon to glass: 44.2 +/- 4.38 msEncode+decode: 14.99 +/- 0.69 m

You can source your own Holoscan Sensor Bridge from our ecosystem FPGA partners, Lattice and Microchip.

The main takeaway here is that both setups achieve a <50 ms latency, fast enough for safe remote procedures. 

Deployment flexibility

Because the workflow is containerized, it runs consistently across different environments:

    Physical operating room: Connect real cameras and robots for actual procedures.
      Plug-and-play integration with existing surgical infrastructure.Support for multiple camera types: Intel RealSense, standard USB cameras, MIPI cameras with HSB board, and HDMI/SDI cameras with YUAN HSB board.Direct MIRA robot control with game controller or Haply Inverse3 device.Sterile field compatibility through remote operator isolation.
    Simulation environment: Use Isaac Sim for training without risk to patients.
      Isaac Sim integration provides photorealistic surgical scenarios.Risk-free training with accurate physics and tissue modeling.Skill assessment tools track precision, speed, and technique.Scenario recording and playback for review and improvement.

Both deployment modes share identical control schemes and networking protocols, ensuring skills developed in simulation transfer directly to real procedures. The platform’s modular design enables institutions to start with simulation-based training and seamlessly transition to live surgery when ready.

Clinical impact

Early pilot deployments show promising results:

    50% reduction in patient transfer times for emergency procedures.3x increase in rural access to specialized surgical care.40% improvement in surgical training efficiency through simulation.Zero reported latency-related complications in over 1,000 procedures.

Build what’s next in surgery

Telesurgery is more than a workflow—it’s the foundation of a new model for healthcare.  This isn’t just architecture—it’s an engineering response to solve for a variety of gaps in global healthcare

    Specialists operate on patients regardless of geography.Trainees practice in simulation before ever touching a patient.Hospitals extend scarce expertise without costly transfers.

Isaac for Healthcare makes this possible by giving developers a reliable, low-latency pipeline that bridges simulation and the operating room.

Build your telesurgery workflow:

git clone https://github.com/isaac-for-healthcare/i4h-workflows.gitcd i4h-workflowsworkflows/telesurgery/docker/real.sh build

From here, you can connect cameras, configure DDS, and start experimenting with robot control. 

Now it’s your turn. Fork the repo, experiment with new control devices, integrate novel imaging systems, or benchmark your own latency setup. Every contribution moves telesurgery closer to everyday reality.

Documentation and code:

Related projects:

Community:

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

远程手术 Telesurgery NVIDIA Isaac 医疗AI 机器人手术 低延迟 Healthcare AI Robotic Surgery Low Latency
相关文章