-
Notifications
You must be signed in to change notification settings - Fork 100
Add expected index models to resolve_model_config.py #1497
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
Add the following models to the allowed models list for workflows: - claude-4.5-opus (litellm_proxy/anthropic/claude-opus-4-5-20251101) - claude-4.5-sonnet (litellm_proxy/anthropic/claude-sonnet-4-5-20250929) - gemini-3-pro (litellm_proxy/gemini/gemini-3-pro-preview) - gemini-3-flash (litellm_proxy/gemini/gemini-3-flash-preview) - gpt-5.2-high-reasoning (litellm_proxy/openai/gpt-5.2-pro) - gpt-5.2 (litellm_proxy/openai/gpt-5.2) - minimax-m2 (litellm_proxy/minimax/minimax-m2) - deepseek-v3.2-reasoner (litellm_proxy/deepseek/deepseek-v3.2) - qwen-3-coder (litellm_proxy/qwen/qwen3-coder) Also fix and update tests to match the current implementation. Co-authored-by: openhands <openhands@all-hands.dev>
| "temperature": 0.0, | ||
| }, | ||
| }, | ||
| "claude-4.5-sonnet": { |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
this model is already in the dictionary
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
it's the unpinned version, which highlights that it added most versions unpinned. I'll ask it to update it
|
@OpenHands please look at MODELS = {...} inside resolve_model_config.py Some of the models are pinned like "claude-sonnet-4-5-20250929" notice it ends in a date, others are not like "claude-4.5-opus" (notice no date) For each of the models already added, if they are unpinned lets check if there is a pinned version. Notes:
|
|
I'm on it! juanmichelini can track my progress at all-hands.dev |
- gpt-5.2-pro -> gpt-5.2-pro-2025-12-11 - gpt-5.2 -> gpt-5.2-2025-12-11 These are the only models that have pinned versions available in litellm. Other models (deepseek-chat, kimi-k2-thinking, gemini-3-pro-preview, gemini-3-flash-preview, minimax-m2, deepseek-v3.2, qwen3-coder) do not have pinned versions in litellm and are left unchanged. Also fixed test file to use correct model ID (claude-sonnet-4-5-20250929 instead of claude-4.5-sonnet). Co-authored-by: openhands <openhands@all-hands.dev>
|
I've analyzed all models in the Summary of ChangesModels Pinned (2 models updated):
Models Left Unchanged (no pinned version exists in litellm):
Already Pinned Models (no changes needed):
Additional Fix:Fixed a pre-existing test issue where Checklist:
|
| "gpt-5.2-high-reasoning": { | ||
| "id": "gpt-5.2-high-reasoning", | ||
| "display_name": "GPT-5.2 High Reasoning", | ||
| "llm_config": {"model": "litellm_proxy/openai/gpt-5.2-pro-2025-12-11"}, |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I thought "GPT 5.2 Pro" is not the same with "GPT 5.2 with reasoning='high'", is it?
- https://platform.openai.com/docs/models/gpt-5.2-pro :
gpt-5.2-pro-2025-12-11 - https://platform.openai.com/docs/models/gpt-5.2 :
gpt-5.2-2025-12-11
Pricing is very different too; Pro is apparently $21 / input, $168 / output
while GPT-5.2 for any reasoning set, is $1.75 / $14
I'm not sure if it's available at all via AH proxy. I can access it via API with a ChatGPT subscription though.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
More details, to my knowledge, in slack
Please do correct me if wrong
|
[Automatic Post]: It has been a while since there was any activity on this PR. @juanmichelini, are you still working on it? If so, please go ahead, if not then please request review, close it, or request that someone else follow up. |
Summary
Fixes #1495
This PR adds all the expected models from the index to the allowed models list in
.github/run-eval/resolve_model_config.py.Changes
New Models Added
claude-4.5-opuslitellm_proxy/anthropic/claude-opus-4-5-20251101claude-4.5-sonnetlitellm_proxy/anthropic/claude-sonnet-4-5-20250929gemini-3-prolitellm_proxy/gemini/gemini-3-pro-previewgemini-3-flashlitellm_proxy/gemini/gemini-3-flash-previewgpt-5.2-high-reasoninglitellm_proxy/openai/gpt-5.2-progpt-5.2litellm_proxy/openai/gpt-5.2minimax-m2litellm_proxy/minimax/minimax-m2deepseek-v3.2-reasonerlitellm_proxy/deepseek/deepseek-v3.2qwen-3-coderlitellm_proxy/qwen/qwen3-coderNote:
kimi-k2-thinkingwas already present in the configuration.Test Updates
resolve_model_configinstead ofresolve_model_configs)MODELSdictionary)Testing
All tests pass:
@juanmichelini can click here to continue refining the PR
Agent Server images for this PR
• GHCR package: https://github.com/OpenHands/agent-sdk/pkgs/container/agent-server
Variants & Base Images
eclipse-temurin:17-jdknikolaik/python-nodejs:python3.12-nodejs22golang:1.21-bookwormPull (multi-arch manifest)
# Each variant is a multi-arch manifest supporting both amd64 and arm64 docker pull ghcr.io/openhands/agent-server:a3f8318-pythonRun
All tags pushed for this build
About Multi-Architecture Support
a3f8318-python) is a multi-arch manifest supporting both amd64 and arm64a3f8318-python-amd64) are also available if needed