From 8d82cb2e049c049a2df5dd6e7217227966fdcf73 Mon Sep 17 00:00:00 2001 From: mrT23 Date: Sun, 15 Sep 2024 08:50:24 +0300 Subject: [PATCH] f string --- pr_agent/algo/ai_handlers/litellm_ai_handler.py | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/pr_agent/algo/ai_handlers/litellm_ai_handler.py b/pr_agent/algo/ai_handlers/litellm_ai_handler.py index ed59b5ab..43812386 100644 --- a/pr_agent/algo/ai_handlers/litellm_ai_handler.py +++ b/pr_agent/algo/ai_handlers/litellm_ai_handler.py @@ -215,13 +215,13 @@ class LiteLLMAIHandler(BaseAiHandler): response = await acompletion(**kwargs) except (openai.APIError, openai.APITimeoutError) as e: - get_logger().warning("Error during LLM inference: {e}") + get_logger().warning(f"Error during LLM inference: {e}") raise except (openai.RateLimitError) as e: - get_logger().error("Rate limit error during LLM inference: {e}") + get_logger().error(f"Rate limit error during LLM inference: {e}") raise except (Exception) as e: - get_logger().warning("Unknown error during LLM inference: {e}") + get_logger().warning(f"Unknown error during LLM inference: {e}") raise openai.APIError from e if response is None or len(response["choices"]) == 0: raise openai.APIError