this post was submitted on 23 Jul 2025
24 points (96.2% liked)

homeassistant

15603 readers
2 users here now

Home Assistant is open source home automation that puts local control and privacy first.
Powered by a worldwide community of tinkerers and DIY enthusiasts.

Home Assistant can be self-installed on ProxMox, Raspberry Pi, or even purchased pre-installed: Home Assistant: Installation

Discussion of Home-Assistant adjacent topics is absolutely fine, within reason.
If you're not sure, DM @GreatAlbatross@feddit.uk

founded 2 years ago
MODERATORS
 

What is everyone using for the LLM model for HA voice when selfhosting ollama? I've tried llama and qwen with varying degrees of understanding my commands. I'm currently on llama as it appears a little better. I just wanted to see if anyone found a better model.

Edit: as pointed out, this is more of a speech to text issue than llm model. I'm looking into the alternatives to whisper

you are viewing a single comment's thread
view the rest of the comments
[–] Rhaedas@fedia.io 1 points 1 week ago (5 children)

I don't use a HA so not familiar with the details of what's out there, but where are you having the problems? Is voice recognition fine and the model just isn't always following the directions precisely? If not, what are you using, whisper or something else? (I'm in search myself of a better voice to text that's local). Certainly by now there are local models fine tuned for being HA, which would work better than a general purpose that could drift or misunderstand common commands.

[–] smashing3606@feddit.online 1 points 1 week ago* (last edited 1 week ago) (4 children)

The issue is mainly voice recognition. Even if I pronounce stuff clear it thinks I've said something else.
Using whisper in HA.

[–] chaospatterns@lemmy.world 2 points 1 week ago* (last edited 1 week ago) (1 children)

That's not going to be fixed with a different LLM model though. I'm experiencing similar problems. If my stt is bad then, then the LLM just gets even more confused or requires a big model that doesnt run efficiently on my local GPU. won't trigger my custom automations because the tools don't consider custom automations phrases.

Speech2phrase improves accuracy for utterances that are basic like turn on X, or anything specified in an automation, but then struggles for other speech.

My next project is to implement a router that forwards the utterance to both speech2phrase and whisper and try to estimate which is correct.

[–] smashing3606@feddit.online 3 points 1 week ago* (last edited 1 week ago)

I was not aware of other alternatives to whisper. I will check them out.

load more comments (2 replies)
load more comments (2 replies)