From 1e60401679e69557663d5ea17c1b7a14b84bc737 Mon Sep 17 00:00:00 2001 From: Debian Date: Sun, 11 Jan 2026 03:10:06 +0000 Subject: [PATCH] Fix RTX 5090 crash: use sdpa attention instead of sageattn SageAttention was only compiled for A100 (sm80) and H100 (sm90). RTX 5090 (Blackwell sm120) has no compatible kernel, causing ComfyUI to crash during generation with "Connection reset by peer". Switch to PyTorch's native SDPA which works on all architectures. Co-Authored-By: Claude Opus 4.5 --- workflows/Wan22-I2V-Remix-API.json | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/workflows/Wan22-I2V-Remix-API.json b/workflows/Wan22-I2V-Remix-API.json index 8661377..e8f53de 100644 --- a/workflows/Wan22-I2V-Remix-API.json +++ b/workflows/Wan22-I2V-Remix-API.json @@ -83,7 +83,7 @@ "base_precision": "fp16_fast", "quantization": "disabled", "load_device": "offload_device", - "attention_mode": "sageattn", + "attention_mode": "sdpa", "rms_norm_function": "default", "block_swap_args": [ "128", @@ -101,7 +101,7 @@ "base_precision": "fp16_fast", "quantization": "disabled", "load_device": "offload_device", - "attention_mode": "sageattn", + "attention_mode": "sdpa", "rms_norm_function": "default", "block_swap_args": [ "128",