From c524492b8988a1a15b7209199eeda3911d003723 Mon Sep 17 00:00:00 2001 From: Icereed Date: Thu, 9 Jan 2025 15:34:30 +0100 Subject: [PATCH] docs: add optional OPENAI_BASE_URL configuration to README (#97) Closes #42 --- README.md | 2 ++ 1 file changed, 2 insertions(+) diff --git a/README.md b/README.md index f163ff6..124e551 100644 --- a/README.md +++ b/README.md @@ -90,6 +90,7 @@ services: LLM_PROVIDER: 'openai' # or 'ollama' LLM_MODEL: 'gpt-4o' # or 'llama2' OPENAI_API_KEY: 'your_openai_api_key' + # Optional - OPENAI_BASE_URL: 'https://litellm.yourinstallationof.it.com/v1' LLM_LANGUAGE: 'English' # Optional, default: English OLLAMA_HOST: 'http://host.docker.internal:11434' # If using Ollama VISION_LLM_PROVIDER: 'ollama' # (for OCR) - openai or ollama @@ -153,6 +154,7 @@ services: | `LLM_PROVIDER` | AI backend (`openai` or `ollama`). | Yes | | `LLM_MODEL` | AI model name, e.g. `gpt-4o`, `gpt-3.5-turbo`, `llama2`. | Yes | | `OPENAI_API_KEY` | OpenAI API key (required if using OpenAI). | Cond. | +| `OPENAI_BASE_URL` | OpenAI base URL (optional, if using a custom OpenAI compatible service like LiteLLM). | No | | `LLM_LANGUAGE` | Likely language for documents (e.g. `English`). Default: `English`. | No | | `OLLAMA_HOST` | Ollama server URL (e.g. `http://host.docker.internal:11434`). | No | | `VISION_LLM_PROVIDER` | AI backend for OCR (`openai` or `ollama`). | No |