this post was submitted on 21 Feb 2024
278 points (95.1% liked)

Technology

59589 readers
3332 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] QuadratureSurfer@lemmy.world 47 points 9 months ago (17 children)

Direct link to the GitHub repo:
https://github.com/nickbild/local_llm_assistant?tab=readme-ov-file

It's a small model by comparison. If you want something that's offline and actually closer to comparing to ChatGPT 3.5, you'll want the Mixtral 8x7B model instead (running on a beefy machine):

https://mistral.ai/news/mixtral-of-experts/

[–] Deceptichum@kbin.social 32 points 9 months ago (10 children)

Sick, I only need 90gb of VRAM!

[–] QuadratureSurfer@lemmy.world 15 points 9 months ago (2 children)

I've got it running with a 3090 and 32GB of RAM.

There are some models that let you run with hybrid system RAM and VRAM (it will just be slower than running it exclusively with VRAM).

[–] Deceptichum@kbin.social 16 points 9 months ago (1 children)

Yeah but damn does it get slow.

I always find it interesting how text is so much slower than image generation. I can do a 1024x1024 in probably 20s, but I get like 1 word a second with text.

[–] aBundleOfFerrets@sh.itjust.works 5 points 9 months ago

Languages are complex and, more importantly, much less forgiving to error

load more comments (7 replies)
load more comments (13 replies)