From eda7bca1c2f9ee8df78f43b58b3f8bc0ffed079a Mon Sep 17 00:00:00 2001 From: Li-8916 Date: Fri, 27 Feb 2026 11:34:38 +0000 Subject: [PATCH 1/2] Fix the issue where TTFT and TPOT have no data when running Kimi2.5 in a PD separation scenario. --- .../benchmark/models/api_models/vllm_custom_api_chat.py | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/ais_bench/benchmark/models/api_models/vllm_custom_api_chat.py b/ais_bench/benchmark/models/api_models/vllm_custom_api_chat.py index 1e792cfa..0fcf4e58 100644 --- a/ais_bench/benchmark/models/api_models/vllm_custom_api_chat.py +++ b/ais_bench/benchmark/models/api_models/vllm_custom_api_chat.py @@ -141,15 +141,15 @@ async def parse_stream_response(self, json_content, output): for item in json_content.get("choices", []): if item["delta"].get("content"): output.content += item["delta"]["content"] - if item["delta"].get("reasoning_content"): - output.reasoning_content += item["delta"]["reasoning_content"] + if item["delta"].get("reasoning_content") or item["delta"].get("reasoning"): + output.reasoning_content += item["delta"].get("reasoning_content") or item["delta"].get("reasoning") await self._parse_usage(json_content, output) async def parse_text_response(self, json_content, output): for item in json_content.get("choices", []): if content:=item["message"].get("content"): output.content += content - if reasoning_content:=item["message"].get("reasoning_content"): + if reasoning_content:=item["message"].get("reasoning_content") or item["message"].get("reasoning"): output.reasoning_content += reasoning_content await self._parse_usage(json_content, output) output.update_extra_details_data_from_text_response(json_content) From 0aa12e8c249638bb5ffc1483b90a5f26adcfa6f6 Mon Sep 17 00:00:00 2001 From: zhang GaoHua <73919261+GaoHuaZhang@users.noreply.github.com> Date: Wed, 11 Mar 2026 11:28:41 +0800 Subject: [PATCH 2/2] Apply suggestion from @gemini-code-assist[bot] Co-authored-by: gemini-code-assist[bot] <176961590+gemini-code-assist[bot]@users.noreply.github.com> --- ais_bench/benchmark/models/api_models/vllm_custom_api_chat.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/ais_bench/benchmark/models/api_models/vllm_custom_api_chat.py b/ais_bench/benchmark/models/api_models/vllm_custom_api_chat.py index 0fcf4e58..ca44e858 100644 --- a/ais_bench/benchmark/models/api_models/vllm_custom_api_chat.py +++ b/ais_bench/benchmark/models/api_models/vllm_custom_api_chat.py @@ -141,8 +141,8 @@ async def parse_stream_response(self, json_content, output): for item in json_content.get("choices", []): if item["delta"].get("content"): output.content += item["delta"]["content"] - if item["delta"].get("reasoning_content") or item["delta"].get("reasoning"): - output.reasoning_content += item["delta"].get("reasoning_content") or item["delta"].get("reasoning") + if reasoning := item["delta"].get("reasoning_content") or item["delta"].get("reasoning"): + output.reasoning_content += reasoning await self._parse_usage(json_content, output) async def parse_text_response(self, json_content, output):