this post was submitted on 14 Feb 2024
548 points (97.4% liked)

Technology

59605 readers
3501 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] CubitOom@infosec.pub 10 points 9 months ago* (last edited 9 months ago) (7 children)

So what is Home Assistant using for this?

If I were to build it myself I'd probably over complicate it by using multiple llm agents on a local server. Probably use whisper to do the speech to text and then Mistral fine tuned on the Rosetta code dataset to send the API calls to HA. However that wouldnt keep it from always listening to me and trying to interpret what I say into a command for HA. Is that just a prompting issue for whisper or would I need another agent to turn on whisper?

I could maybe get this to run without specialized hardware like a GPU but it would be better to have something for the llms to be a bit more responsive.

[–] redcalcium@lemmy.institute 7 points 9 months ago* (last edited 9 months ago) (6 children)

There is no LLM, it just used to recognize simple commands such as "turn on kitchen light". What the "conversation agent" can do is very limited, though you can extend it to recognize custom commands. It's not comparable to Google Assistant/Siri, let alone ChatGPT.

[–] 4am@lemm.ee 4 points 9 months ago (1 children)

I believe there is a ChatGPT integration in the works (optional, of course)

[–] Serinus@lemmy.world 3 points 9 months ago (2 children)

If it runs locally, that'll be awesome. I just hope it never decides to turn the heat up to 90F.

[–] Buddahriffic@lemmy.world 3 points 9 months ago (1 children)

Ideally IMO you'd want a system with safeties in place. Like acceptable temperature ranges or durations for the oven to be on to avoid situations where the software misinterprets a command in a dangerous way.

Something like this:

User: Set temperature to 19 degrees. (Yeah it's on the cold side even for Celsius, but not a crazy amount as room temperature is around 22 degrees)

Assistant: Setting temperature to 90 degrees. (Deadly in Celsius... Water boils at around 100 degrees, depending on pressure)

Assistant: 90 degrees is outside of the safe range defined by your configuration. Intrusion suspected. Deploying sentry guns.

[–] AA5B@lemmy.world 2 points 9 months ago

Good question - I have an allowed range configured on my thermostat but I don’t know if it applies to API calls or is just for the UI

[–] Saik0Shinigami@lemmy.saik0.com 2 points 9 months ago

There's plenty of local LLM options these days. It's entirely feasible to run it in house.

And if someone can do it... I would suspect that there'll be a HACS module up about 2 weeks ago...

load more comments (4 replies)
load more comments (4 replies)