Skip to content

Conversation

@xlyk
Copy link
Contributor

@xlyk xlyk commented Feb 3, 2026

Update Gemini model capabilities with correct token limits from official Google Cloud documentation:

  • gemini_defaults(): context_window: 1,048,576 (1M), max_output_tokens: 65,536
  • gemini_2_0_flash_lite_defaults(): context_window: 1,048,576 (1M), max_output_tokens: 8,192
  • Updated gemini-2.0-flash-lite model to use gemini_2_0_flash_lite_defaults()

Correct GPT-5 context window from 1M to 400k tokens. Add support for Codex models with 192k context. Separate GPT-5 and o-series capabilities with distinct context windows (400k vs 200k). Update Gemini max output tokens from 8k to 65k, with special handling for Flash-Lite models.
@xlyk xlyk merged commit 0f7b716 into main Feb 3, 2026
3 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant