

2·
9 hours agoit’s come a long way in the last few majors. I’ve been using it since v7 and it’s been a struggle at times with users complaining about how glacial it could get. These days it’s pretty blazingly fast comparatively, I get no complaints.
I’ve set up OpenWebUI with the docker containers, which includes Ollama in API mode, and optionally Playwright if you want to add webscraping to your RAG queries. This gives you a ChatJippity format webpage that you can manage your models for Ollama, and add OpenAI usage as well if you want. You can manage all the users as well.
On top, then you have API support to your own Ollama instance, and you can also configure GPU usage for your local AI if available.
Honestly, it’s the easiest way to get local AI.