Skip to content

Commit 04f124c

Browse files
kochj23claude
andcommitted
docs: Note tool calling loop fix in v6.3.0 changelog
Co-Authored-By: Claude Sonnet 4.6 (1M context) <noreply@anthropic.com>
1 parent 191886f commit 04f124c

1 file changed

Lines changed: 2 additions & 1 deletion

File tree

README.md

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -243,7 +243,7 @@ Being honest about limitations:
243243

244244
## Version History
245245

246-
### v6.3.0 (March 4, 2026) — Current
246+
### v6.3.0 build 7 (March 4, 2026) — Current
247247
- **Xcode Source Editor Extension** — 5 commands in Editor > MLX Code menu (Explain, Refactor, Generate Tests, Fix Issues, Ask). Communicates via shared App Group + `mlxcode://` URL scheme
248248
- **Native model downloads** — replaced Python downloader with `Hub.HubApi.snapshot()`. Python fully eliminated
249249
- **Syntax highlighting** — Swift, Python, JS/TS, Bash, JSON, Objective-C in all code blocks
@@ -253,6 +253,7 @@ Being honest about limitations:
253253
- **Tool call reliability** — JSON auto-repair, retry-on-failure loop, stricter system prompt rule
254254
- **Default temperature 0.2** — reduced from 0.7 to cut hallucinations in code analysis
255255
- **Jinja template fallback** — models with unsupported chat templates fall back to flat prompt format
256+
- **Fixed: agentic tool calling loop** — resolved "inference already in progress" error that occurred when the model called a tool and the follow-up generation failed. Root cause was `PythonService.terminate()` being a no-op with native MLX, causing `chatCompletion()` to run until `maxTokens` before returning. Stream now exits immediately when `</tool>` is detected.
256257

257258
### v6.2.0 (March 4, 2026)
258259
- Replaced Python subprocess daemon with native `mlx-swift-lm` framework for inference

0 commit comments

Comments
 (0)