Throw descriptive error message if model is not in MAX_TOKENS array

This commit is contained in:
Phill Zarfos
2023-12-03 21:06:55 -05:00
parent 526ad00812
commit 317fec0536
2 changed files with 8 additions and 2 deletions

View File

@ -262,7 +262,7 @@ MAX_TOKENS = {
e.g.
MAX_TOKENS={
...,
"llama2": 4096
"ollama/llama2": 4096
}
@ -271,6 +271,8 @@ model = "ollama/llama2"
[ollama] # in .secrets.toml
api_base = ... # the base url for your huggingface inference endpoint
# e.g. if running Ollama locally, you may use:
api_base = "http://localhost:11434/"
```
**Inference Endpoints**