From 97d6fb999a515ea6459c7474db886a593d4f325a Mon Sep 17 00:00:00 2001 From: tmokmss Date: Tue, 28 Nov 2023 20:58:57 +0900 Subject: [PATCH] set max_tokens_to_sample --- pr_agent/algo/ai_handler.py | 3 +++ 1 file changed, 3 insertions(+) diff --git a/pr_agent/algo/ai_handler.py b/pr_agent/algo/ai_handler.py index 6d873cd9..24273db6 100644 --- a/pr_agent/algo/ai_handler.py +++ b/pr_agent/algo/ai_handler.py @@ -63,6 +63,9 @@ class AiHandler: "VERTEXAI.VERTEX_LOCATION", None ) if get_settings().get("AWS.BEDROCK_REGION", None): + litellm.AmazonAnthropicConfig.max_tokens_to_sample = int(get_settings().get( + "AWS.CLAUDE_MAX_TOKENS_TO_SAMPLE", '2000' + )) self.aws_bedrock_client = boto3.client( service_name="bedrock-runtime", region_name=get_settings().aws.bedrock_region,