diff --git a/.changeset/fix-default-model-tier.md b/.changeset/fix-default-model-tier.md deleted file mode 100644 index 5d583a42..00000000 --- a/.changeset/fix-default-model-tier.md +++ /dev/null @@ -1,11 +0,0 @@ ---- -"@perstack/runtime": patch -"@perstack/installer": patch -"perstack": patch ---- - -Fix: Preserve `defaultModelTier` when resolving experts from API - -The `toRuntimeExpert` function in both runtime and installer packages was dropping the `defaultModelTier` field when converting API response experts to runtime Expert objects. This caused all API-fetched experts to ignore their configured model tier and fall back to the provider's highest-tier model (e.g., claude-opus-4-6 instead of claude-haiku-4-5 for `defaultModelTier: "low"`). - -Also fixed `configExpertToExpert` in the installer package which had the same omission when converting TOML config experts. diff --git a/apps/perstack/CHANGELOG.md b/apps/perstack/CHANGELOG.md index fee03445..612b6115 100644 --- a/apps/perstack/CHANGELOG.md +++ b/apps/perstack/CHANGELOG.md @@ -1,5 +1,15 @@ # perstack +## 0.0.116 + +### Patch Changes + +- [#686](https://github.com/perstack-ai/perstack/pull/686) [`9ebcef3`](https://github.com/perstack-ai/perstack/commit/9ebcef36da30d64d8c61e10722725cae6d9537ac) Thanks [@FL4TLiN3](https://github.com/FL4TLiN3)! - Fix: Preserve `defaultModelTier` when resolving experts from API + + The `toRuntimeExpert` function in both runtime and installer packages was dropping the `defaultModelTier` field when converting API response experts to runtime Expert objects. This caused all API-fetched experts to ignore their configured model tier and fall back to the provider's highest-tier model (e.g., claude-opus-4-6 instead of claude-haiku-4-5 for `defaultModelTier: "low"`). + + Also fixed `configExpertToExpert` in the installer package which had the same omission when converting TOML config experts. + ## 0.0.115 ### Patch Changes diff --git a/apps/perstack/package.json b/apps/perstack/package.json index 899888a3..fa15f8dd 100644 --- a/apps/perstack/package.json +++ b/apps/perstack/package.json @@ -1,6 +1,6 @@ { "name": "perstack", - "version": "0.0.115", + "version": "0.0.116", "description": "PerStack CLI", "author": "Wintermute Technologies, Inc.", "license": "Apache-2.0", diff --git a/packages/installer/CHANGELOG.md b/packages/installer/CHANGELOG.md index f430de46..6c86c4a9 100644 --- a/packages/installer/CHANGELOG.md +++ b/packages/installer/CHANGELOG.md @@ -1,5 +1,15 @@ # @perstack/installer +## 0.0.35 + +### Patch Changes + +- [#686](https://github.com/perstack-ai/perstack/pull/686) [`9ebcef3`](https://github.com/perstack-ai/perstack/commit/9ebcef36da30d64d8c61e10722725cae6d9537ac) Thanks [@FL4TLiN3](https://github.com/FL4TLiN3)! - Fix: Preserve `defaultModelTier` when resolving experts from API + + The `toRuntimeExpert` function in both runtime and installer packages was dropping the `defaultModelTier` field when converting API response experts to runtime Expert objects. This caused all API-fetched experts to ignore their configured model tier and fall back to the provider's highest-tier model (e.g., claude-opus-4-6 instead of claude-haiku-4-5 for `defaultModelTier: "low"`). + + Also fixed `configExpertToExpert` in the installer package which had the same omission when converting TOML config experts. + ## 0.0.34 ### Patch Changes diff --git a/packages/installer/package.json b/packages/installer/package.json index d441a49a..945c3f8a 100644 --- a/packages/installer/package.json +++ b/packages/installer/package.json @@ -1,6 +1,6 @@ { "private": true, - "version": "0.0.34", + "version": "0.0.35", "name": "@perstack/installer", "description": "Perstack Installer - Generate lockfiles with tool definitions", "author": "Wintermute Technologies, Inc.", diff --git a/packages/runtime/CHANGELOG.md b/packages/runtime/CHANGELOG.md index 46284a3f..e9eba888 100644 --- a/packages/runtime/CHANGELOG.md +++ b/packages/runtime/CHANGELOG.md @@ -1,5 +1,15 @@ # @perstack/runtime +## 0.0.132 + +### Patch Changes + +- [#686](https://github.com/perstack-ai/perstack/pull/686) [`9ebcef3`](https://github.com/perstack-ai/perstack/commit/9ebcef36da30d64d8c61e10722725cae6d9537ac) Thanks [@FL4TLiN3](https://github.com/FL4TLiN3)! - Fix: Preserve `defaultModelTier` when resolving experts from API + + The `toRuntimeExpert` function in both runtime and installer packages was dropping the `defaultModelTier` field when converting API response experts to runtime Expert objects. This caused all API-fetched experts to ignore their configured model tier and fall back to the provider's highest-tier model (e.g., claude-opus-4-6 instead of claude-haiku-4-5 for `defaultModelTier: "low"`). + + Also fixed `configExpertToExpert` in the installer package which had the same omission when converting TOML config experts. + ## 0.0.131 ### Patch Changes diff --git a/packages/runtime/package.json b/packages/runtime/package.json index 422e4c7c..2eeb6373 100644 --- a/packages/runtime/package.json +++ b/packages/runtime/package.json @@ -1,6 +1,6 @@ { "name": "@perstack/runtime", - "version": "0.0.131", + "version": "0.0.132", "description": "Perstack Runtime", "author": "Wintermute Technologies, Inc.", "license": "Apache-2.0", diff --git a/packages/studio/CHANGELOG.md b/packages/studio/CHANGELOG.md index 317b7b8f..cae60093 100644 --- a/packages/studio/CHANGELOG.md +++ b/packages/studio/CHANGELOG.md @@ -1,5 +1,12 @@ # @perstack/studio +## 0.0.16 + +### Patch Changes + +- Updated dependencies [[`9ebcef3`](https://github.com/perstack-ai/perstack/commit/9ebcef36da30d64d8c61e10722725cae6d9537ac)]: + - @perstack/installer@0.0.35 + ## 0.0.15 ### Patch Changes diff --git a/packages/studio/package.json b/packages/studio/package.json index 12c3bca0..e51efb92 100644 --- a/packages/studio/package.json +++ b/packages/studio/package.json @@ -1,6 +1,6 @@ { "private": true, - "version": "0.0.15", + "version": "0.0.16", "name": "@perstack/studio", "description": "Perstack Studio - Expert management CLI handlers", "author": "Wintermute Technologies, Inc.",