From f44050d777a1a470de176fc6765f2658d879d797 Mon Sep 17 00:00:00 2001 From: root <953994191@qq.com> Date: Fri, 1 May 2026 10:22:48 +0800 Subject: [PATCH] =?UTF-8?q?fix:=20=E5=89=8D=E7=AB=AF=E6=94=AF=E6=8C=81?= =?UTF-8?q?=E6=98=BE=E7=A4=BA=E5=85=9C=E5=BA=95=E8=BE=93=E5=87=BA=20(node:?= =?UTF-8?q?=20fallback)?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- frontend/src/components/chat_area.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/frontend/src/components/chat_area.py b/frontend/src/components/chat_area.py index 411e1cb..4711773 100644 --- a/frontend/src/components/chat_area.py +++ b/frontend/src/components/chat_area.py @@ -143,7 +143,7 @@ def _handle_ai_response(): # 1. 处理 LLM Token 流 (打字机效果) if event_type == "llm_token": # 确保只处理来自 LLM 的 token,避免将工具的输出作为 token 显示 - if event.get("node") == "llm_call": + if event.get("node") in ("llm_call", "fallback"): token = str(event.get("token", "")) reasoning_token = str(event.get("reasoning_token", ""))