diff --git a/frontend/src/components/chat_area.py b/frontend/src/components/chat_area.py index 4711773..207ad9a 100644 --- a/frontend/src/components/chat_area.py +++ b/frontend/src/components/chat_area.py @@ -143,7 +143,7 @@ def _handle_ai_response(): # 1. 处理 LLM Token 流 (打字机效果) if event_type == "llm_token": # 确保只处理来自 LLM 的 token,避免将工具的输出作为 token 显示 - if event.get("node") in ("llm_call", "fallback"): + if event.get("node") in ("llm_call", "fallback", "fast_path"): token = str(event.get("token", "")) reasoning_token = str(event.get("reasoning_token", ""))