2023-07-06 00:21:08 +03:00
|
|
|
MAX_TOKENS = {
|
2023-09-05 08:40:05 +03:00
|
|
|
'text-embedding-ada-002': 8000,
|
2023-07-06 00:21:08 +03:00
|
|
|
'gpt-3.5-turbo': 4000,
|
|
|
|
'gpt-3.5-turbo-0613': 4000,
|
|
|
|
'gpt-3.5-turbo-0301': 4000,
|
|
|
|
'gpt-3.5-turbo-16k': 16000,
|
|
|
|
'gpt-3.5-turbo-16k-0613': 16000,
|
|
|
|
'gpt-4': 8000,
|
|
|
|
'gpt-4-0613': 8000,
|
|
|
|
'gpt-4-32k': 32000,
|
2023-11-07 14:28:41 +02:00
|
|
|
'gpt-4-1106-preview': 128000, # 128K, but may be limited by config.max_model_tokens
|
2024-02-01 09:46:04 +02:00
|
|
|
'gpt-4-0125-preview': 128000, # 128K, but may be limited by config.max_model_tokens
|
2023-08-03 16:05:46 -07:00
|
|
|
'claude-instant-1': 100000,
|
|
|
|
'claude-2': 100000,
|
|
|
|
'command-nightly': 4096,
|
|
|
|
'replicate/llama-2-70b-chat:2c1608e18606fad2812020dc541930f2d0495ce32eee50074220b87300bc16e1': 4096,
|
2023-11-07 09:13:08 +00:00
|
|
|
'meta-llama/Llama-2-7b-chat-hf': 4096,
|
|
|
|
'vertex_ai/codechat-bison': 6144,
|
|
|
|
'vertex_ai/codechat-bison-32k': 32000,
|
|
|
|
'codechat-bison': 6144,
|
|
|
|
'codechat-bison-32k': 32000,
|
2023-11-28 20:11:40 +09:00
|
|
|
'anthropic.claude-v2': 100000,
|
|
|
|
'anthropic.claude-instant-v1': 100000,
|
|
|
|
'anthropic.claude-v1': 100000,
|
2023-07-06 00:21:08 +03:00
|
|
|
}
|