mirror of
https://github.com/qodo-ai/pr-agent.git
synced 2025-07-02 03:40:38 +08:00
adding documentation on how to call local hf models
This commit is contained in:
25
Usage.md
25
Usage.md
@ -169,6 +169,31 @@ in the configuration.toml
|
|||||||
|
|
||||||
#### Huggingface
|
#### Huggingface
|
||||||
|
|
||||||
|
**Local**
|
||||||
|
You can run Huggingface models locally through either [VLLM](https://docs.litellm.ai/docs/providers/vllm) or [Ollama](https://docs.litellm.ai/docs/providers/ollama)
|
||||||
|
|
||||||
|
E.g. to use a new Huggingface model locally via Ollama, set:
|
||||||
|
```
|
||||||
|
[__init__.py]
|
||||||
|
MAX_TOKENS = {
|
||||||
|
"model-name-on-ollama": <max_tokens>
|
||||||
|
}
|
||||||
|
e.g.
|
||||||
|
MAX_TOKENS={
|
||||||
|
...,
|
||||||
|
"llama2": 4096
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
[config] # in configuration.toml
|
||||||
|
model = "ollama/llama2"
|
||||||
|
|
||||||
|
[ollama] # in .secrets.toml
|
||||||
|
api_base = ... # the base url for your huggingface inference endpoint
|
||||||
|
```
|
||||||
|
|
||||||
|
**Inference Endpoints**
|
||||||
|
|
||||||
To use a new model with Huggingface Inference Endpoints, for example, set:
|
To use a new model with Huggingface Inference Endpoints, for example, set:
|
||||||
```
|
```
|
||||||
[__init__.py]
|
[__init__.py]
|
||||||
|
@ -29,6 +29,9 @@ key = "" # Optional, uncomment if you want to use Replicate. Acquire through htt
|
|||||||
key = "" # Optional, uncomment if you want to use Huggingface Inference API. Acquire through https://huggingface.co/docs/api-inference/quicktour
|
key = "" # Optional, uncomment if you want to use Huggingface Inference API. Acquire through https://huggingface.co/docs/api-inference/quicktour
|
||||||
api_base = "" # the base url for your huggingface inference endpoint
|
api_base = "" # the base url for your huggingface inference endpoint
|
||||||
|
|
||||||
|
[ollama]
|
||||||
|
api_base = "" # the base url for your huggingface inference endpoint
|
||||||
|
|
||||||
[github]
|
[github]
|
||||||
# ---- Set the following only for deployment type == "user"
|
# ---- Set the following only for deployment type == "user"
|
||||||
user_token = "" # A GitHub personal access token with 'repo' scope.
|
user_token = "" # A GitHub personal access token with 'repo' scope.
|
||||||
|
Reference in New Issue
Block a user