Merge pull request #1473 from KennyDizi/main

Add support model deepseek/deepseek-reasoner
This commit is contained in:
Tal
2025-01-23 08:07:53 +02:00
committed by GitHub
2 changed files with 3 additions and 2 deletions

View File

@ -30,6 +30,7 @@ MAX_TOKENS = {
'claude-2': 100000,
'command-nightly': 4096,
'deepseek/deepseek-chat': 128000, # 128K, but may be limited by config.max_model_tokens
'deepseek/deepseek-reasoner': 64000, # 64K, but may be limited by config.max_model_tokens
'replicate/llama-2-70b-chat:2c1608e18606fad2812020dc541930f2d0495ce32eee50074220b87300bc16e1': 4096,
'meta-llama/Llama-2-7b-chat-hf': 4096,
'vertex_ai/codechat-bison': 6144,

View File

@ -200,10 +200,10 @@ class LiteLLMAIHandler(BaseAiHandler):
# Currently, model OpenAI o1 series does not support a separate system and user prompts
O1_MODEL_PREFIX = 'o1'
model_type = model.split('/')[-1] if '/' in model else model
if model_type.startswith(O1_MODEL_PREFIX):
if (model_type.startswith(O1_MODEL_PREFIX)) or ("deepseek-reasoner" in model):
user = f"{system}\n\n\n{user}"
system = ""
get_logger().info(f"Using O1 model, combining system and user prompts")
get_logger().info(f"Using model {model}, combining system and user prompts")
messages = [{"role": "user", "content": user}]
kwargs = {
"model": model,