From c6a1d324981956f8cbb18f14efc7da8ce97c87aa Mon Sep 17 00:00:00 2001 From: jlightner Date: Fri, 3 Apr 2026 08:27:53 +0000 Subject: [PATCH] fix: route LLM calls through OpenWebUI tracked proxy for analytics Changed LLM_API_URL and LLM_FALLBACK_URL from /api/v1 to /api so requests hit OpenWebUI's /api/chat/completions (tracked) instead of /api/v1/chat/completions (passthrough with no analytics). --- .env.example | 5 +++-- 1 file changed, 3 insertions(+), 2 deletions(-) diff --git a/.env.example b/.env.example index 75e0d0b..a293346 100644 --- a/.env.example +++ b/.env.example @@ -10,10 +10,11 @@ POSTGRES_DB=chrysopedia REDIS_URL=redis://chrysopedia-redis:6379/0 # LLM endpoint (OpenAI-compatible — OpenWebUI on FYN DGX) -LLM_API_URL=https://chat.forgetyour.name/api/v1 +# Use /api (not /api/v1) so calls route through OpenWebUI's tracked proxy for analytics +LLM_API_URL=https://chat.forgetyour.name/api LLM_API_KEY=sk-changeme LLM_MODEL=fyn-llm-agent-chat -LLM_FALLBACK_URL=https://chat.forgetyour.name/api/v1 +LLM_FALLBACK_URL=https://chat.forgetyour.name/api LLM_FALLBACK_MODEL=fyn-llm-agent-chat # Per-stage LLM model overrides (optional — defaults to LLM_MODEL)