open-webui/backend
Sihyeon Jang 3ccbb46938 perf: fix cache key generation for model list caching
- Replace Request object with user.id in cache key for get_all_models
- Request objects are new instances per HTTP request, preventing cache hits
- Cache keys now use user.id ensuring proper cache functionality
- Affects both Ollama and OpenAI model list endpoints

Signed-off-by: Sihyeon Jang <sihyeon.jang@navercorp.com>
2025-09-03 05:17:41 +09:00
..
data refac: mv backend files to /open_webui dir 2024-09-04 16:54:48 +02:00
open_webui perf: fix cache key generation for model list caching 2025-09-03 05:17:41 +09:00
.dockerignore fix: litellm config issue 2024-02-24 22:35:11 -08:00
.gitignore refac 2024-09-06 04:59:20 +02:00
dev.sh feat: S3 vector support tested 2025-07-22 21:36:35 -06:00
requirements.txt Merge pull request #17093 from open-webui/dependabot/pip/backend/dev/argon2-cffi-25.1.0 2025-09-02 02:41:49 +04:00
start.sh fix: Resolve admin account creation on Hugging Face Spaces. 2025-08-31 14:33:48 +07:00
start_windows.bat fix: windows start script 2025-06-27 15:46:38 +04:00