Any experiences with a self-hosted assistant like the modern Google Assistant? Looking for something LLM-powered that is smarter than older assistants that would just try to call 3rd party tools directly and miss or misunderstand requests half of the time.
I’d like integration with a mobile app to use it from the phone and while driving. I see Home Assistant has an Android Auto integration. Has anyone used this, or another similar option? Any blatant limitations?


I don’t think there’s a straightforward way like a HACS integration yet, but you can access Ollama from the web with open-webui and save the page to your homepage:
Just be warned, you’ll need a lot of resources depending on which model you choose and its parameter count (4B, 7B etc) – Gemma3 4B uses around 3GB storage, 0.5GB RAM and 4GB of VRAM to respond. It’s a compromise as I can’t get replacement RAM, and tends to be wildly inaccurate with large responses. The one I’d rather use, Dolphin-Mixtral 22B, takes 80GB storage and 17GB min RAM, the latter of which I can’t afford to take from my other services.
There’s an Ollama integration that adds it as a conversation agent.
ah, this puts it together and it’s exactly what I was looking for, thanks
And there’s another custom component, integrating all servers with an OpenAI-compatible API endpoint: https://github.com/jekalmin/extended_openai_conversation
Also https://github.com/acon96/home-llm/blob/90c0edc0907e4567eeca89a6f89b43c0b0b807eb/docs/Setup.md