From ffefcb8a044a422838c1d481e568baca3dbccefd Mon Sep 17 00:00:00 2001 From: Kenny Dizi Date: Tue, 11 Mar 2025 17:46:38 +0700 Subject: [PATCH 1/2] Fix default value for extended_thinking_max_output_tokens --- pr_agent/algo/ai_handlers/litellm_ai_handler.py | 2 +- pr_agent/settings/configuration.toml | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/pr_agent/algo/ai_handlers/litellm_ai_handler.py b/pr_agent/algo/ai_handlers/litellm_ai_handler.py index af9efc87..08810308 100644 --- a/pr_agent/algo/ai_handlers/litellm_ai_handler.py +++ b/pr_agent/algo/ai_handlers/litellm_ai_handler.py @@ -136,7 +136,7 @@ class LiteLLMAIHandler(BaseAiHandler): dict: Updated kwargs with extended thinking configuration """ extended_thinking_budget_tokens = get_settings().config.get("extended_thinking_budget_tokens", 2048) - extended_thinking_max_output_tokens = get_settings().config.get("extended_thinking_max_output_tokens", 2048) + extended_thinking_max_output_tokens = get_settings().config.get("extended_thinking_max_output_tokens", 4096) # Validate extended thinking parameters if not isinstance(extended_thinking_budget_tokens, int) or extended_thinking_budget_tokens <= 0: diff --git a/pr_agent/settings/configuration.toml b/pr_agent/settings/configuration.toml index 7f8daedf..8ea65e87 100644 --- a/pr_agent/settings/configuration.toml +++ b/pr_agent/settings/configuration.toml @@ -63,7 +63,7 @@ auto_approve_for_no_suggestions=false # If true, the PR will be auto-approved if # extended thinking for Claude reasoning models enable_claude_extended_thinking = false # Set to true to enable extended thinking feature extended_thinking_budget_tokens = 2048 -extended_thinking_max_output_tokens = 2048 +extended_thinking_max_output_tokens = 4096 [pr_reviewer] # /review # From 0acd5193cb6b46a1d804645d76cec52a07eee870 Mon Sep 17 00:00:00 2001 From: Kenny Dizi Date: Tue, 11 Mar 2025 17:46:53 +0700 Subject: [PATCH 2/2] Align it with document --- docs/docs/usage-guide/changing_a_model.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/docs/docs/usage-guide/changing_a_model.md b/docs/docs/usage-guide/changing_a_model.md index b0bde3fa..844cd6a5 100644 --- a/docs/docs/usage-guide/changing_a_model.md +++ b/docs/docs/usage-guide/changing_a_model.md @@ -240,4 +240,4 @@ With the OpenAI models that support reasoning effort (eg: o3-mini), you can spec [config] enable_claude_extended_thinking = false # Set to true to enable extended thinking feature extended_thinking_budget_tokens = 2048 -extended_thinking_max_output_tokens = 2048 +extended_thinking_max_output_tokens = 4096