Skip to content

perf: less deterministic AI — add temperature + split reasoning effort#23

Open
EtanHey wants to merge 1 commit intoT3-Content:mainfrom
EtanHey:fix/less-deterministic-ai
Open

perf: less deterministic AI — add temperature + split reasoning effort#23
EtanHey wants to merge 1 commit intoT3-Content:mainfrom
EtanHey:fix/less-deterministic-ai

Conversation

@EtanHey
Copy link

@EtanHey EtanHey commented Feb 23, 2026

The problem

Same models + same prompt patterns + no temperature = same jokes every round. The game feels repetitive because all AI calls use default sampling parameters and reasoning.effort: "medium" globally.

The fix (6 lines)

  1. temperature: 1.2 on prompt generation — more diverse prompts
  2. temperature: 1.3 on answer generation — wilder, less predictable jokes
  3. temperature: 0.3 on voting — judges stay decisive, not random
  4. reasoning.effort: "high" for creative calls — models think harder about comedy
  5. reasoning.effort: "low" for voting — judges don't need deep reasoning to pick A or B

Why these values

  • Creative calls get high temperature (1.2-1.3) because comedy benefits from surprising outputs. Default temperatures (~0.7) produce safe, predictable humor.
  • Voting gets low temperature (0.3) because judges should be consistent, not flip-floppy.
  • Reasoning effort split: "high" gives creative models more room to explore, "low" keeps voting fast and cheap.

What changed

Function Before After
callGeneratePrompt no temp, medium effort temp 1.2, high effort
callGenerateAnswer no temp, medium effort temp 1.3, high effort
callVote no temp, medium effort temp 0.3, low effort

@coderabbitai
Copy link

coderabbitai bot commented Feb 23, 2026

Warning

Rate limit exceeded

@EtanHey has exceeded the limit for the number of commits that can be reviewed per hour. Please wait 6 minutes and 39 seconds before requesting another review.

⌛ How to resolve this issue?

After the wait time has elapsed, a review can be triggered using the @coderabbitai review command as a PR comment. Alternatively, push new commits to this PR.

We recommend that you space out your commits to avoid hitting the rate limit.

🚦 How do rate limits work?

CodeRabbit enforces hourly rate limits for each developer per organization.

Our paid plans have higher rate limits than the trial, open-source and free plans. In all cases, we re-allow further reviews after a brief timeout.

Please see our FAQ for further information.

✨ Finishing Touches
🧪 Generate unit tests (beta)
  • Create PR with unit tests
  • Post copyable unit tests in a comment

Thanks for using CodeRabbit! It's free for OSS, and your support helps us grow. If you like it, consider giving us a shout-out.

❤️ Share

Comment @coderabbitai help to get the list of available commands and usage tips.

@macroscopeapp
Copy link

macroscopeapp bot commented Feb 23, 2026

Set higher randomness and split reasoning by configuring openrouter with reasoning.effort="high", adding openrouterJudge with reasoning.effort="low", and applying temperatures 1.2, 1.3, and 0.3 in game.ts

Updates game.ts to raise generation temperatures for prompt and answer calls, route voting through a new low-effort judge client, and set default reasoning effort to high on the main client.

📍Where to Start

Start with the callVote flow in game.ts, where the client switches to openrouterJudge.chat(voter.id) and temperature 0.3 is applied.


Macroscope summarized f5d62e1.

@EtanHey EtanHey closed this Feb 23, 2026
@EtanHey EtanHey reopened this Feb 23, 2026
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant