Oh yes for Claude I use LiteLLM as a proxy to use it with OpenWebUI.
I'll try librechat too (never heard of it before) but I wonder if it has the same capabilities like voice and python tools. And ollama support (95% of my AI interactions are running locally)
I think that's probably the shim I was referring to - it has hardcoded context length, but it is either implemented incorrectly, Anthropic ignores it, or maybe it's on openwebui to manage the window and it just isn't? Not sure. I found it kept getting slow, so I was starting new conversations to work around that. Eventually I got suspicious and checked - I'd burned through almost $100 within a few hours.
LibreChat isn't as nice in some areas, but it's much more efficient in this regard.
wkat4242|1 year ago
I'll try librechat too (never heard of it before) but I wonder if it has the same capabilities like voice and python tools. And ollama support (95% of my AI interactions are running locally)
qqqult|1 year ago
emptiestplace|1 year ago
LibreChat isn't as nice in some areas, but it's much more efficient in this regard.