From 2148efe65f38abcd441a91c54a7ed3ce263f4a6a Mon Sep 17 00:00:00 2001 From: kimtaewoong Date: Mon, 27 Oct 2025 22:43:22 +0900 Subject: [PATCH] feat: add reasoning_content field to LLMResultChunkDelta - Add optional reasoning_content field to support direct streaming of LLM reasoning/thinking content, separate from the main message content. --- python/dify_plugin/entities/model/llm.py | 1 + 1 file changed, 1 insertion(+) diff --git a/python/dify_plugin/entities/model/llm.py b/python/dify_plugin/entities/model/llm.py index fe4f7db2..0f8edf05 100644 --- a/python/dify_plugin/entities/model/llm.py +++ b/python/dify_plugin/entities/model/llm.py @@ -76,6 +76,7 @@ class LLMResultChunkDelta(BaseModel): index: int message: AssistantPromptMessage + reasoning_content: str | None = None usage: LLMUsage | None = None finish_reason: str | None = None