From 67001d0ab1e1604f4dcb27f6f9b25624abb107f8 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Dominik=20Schr=C3=B6ter?= Date: Thu, 13 Mar 2025 23:33:36 +0100 Subject: [PATCH] fix(docs): update VISION_LLM_MODEL reference from gpt-4v to gpt-4o in README --- README.md | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/README.md b/README.md index effeabb..65652f4 100644 --- a/README.md +++ b/README.md @@ -120,7 +120,7 @@ services: # Option 1: LLM-based OCR OCR_PROVIDER: "llm" # Default OCR provider VISION_LLM_PROVIDER: "ollama" # openai or ollama - VISION_LLM_MODEL: "minicpm-v" # minicpm-v (ollama) or gpt-4v (openai) + VISION_LLM_MODEL: "minicpm-v" # minicpm-v (ollama) or gpt-4o (openai) OLLAMA_HOST: "http://host.docker.internal:11434" # If using Ollama # Option 2: Google Document AI @@ -191,7 +191,7 @@ paperless-gpt supports three different OCR providers, each with unique strengths ### 1. LLM-based OCR (Default) - **Key Features**: - - Uses vision-capable LLMs like GPT-4V or MiniCPM-V + - Uses vision-capable LLMs like gpt-4o or MiniCPM-V - High accuracy with complex layouts and difficult scans - Context-aware text recognition - Self-correcting capabilities for OCR errors @@ -203,7 +203,7 @@ paperless-gpt supports three different OCR providers, each with unique strengths ```yaml OCR_PROVIDER: "llm" VISION_LLM_PROVIDER: "openai" # or "ollama" - VISION_LLM_MODEL: "gpt-4v" # or "minicpm-v" + VISION_LLM_MODEL: "gpt-4o" # or "minicpm-v" ``` ### 2. Azure Document Intelligence