this post was submitted on 21 Feb 2024
89 points (90.1% liked)

Technology

64937 readers
4010 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] scottmeme@sh.itjust.works 14 points 1 year ago (2 children)

I'm just trying to get my hands on some faster hardware, https://groq.com has been able to do some crazy shit with their 500/tokens/sec on their LPUs

[–] Lmaydev@programming.dev 6 points 1 year ago* (last edited 1 year ago)

That is insanely fast! I figured we'd be getting "AI cards" at some point soon.

[–] Lojcs@lemm.ee 1 points 1 year ago (1 children)

What kind of a website is that? Super slow and doesn't work without web assembly. Do you really need that for a simple interface

[–] scottmeme@sh.itjust.works 2 points 1 year ago (3 children)

It's not about their frontend, they are running custom LPUs which can process LLM tokens at 500/sec which is insanely impressive.

For reference with a max size of 2k tokens, my dual xeon silver 4114 procs take 2-3 minutes.

[–] Lojcs@lemm.ee 1 points 1 year ago

No I got what you meant, but that site is weird if it's not doing anything on its own

[–] Amaltheamannen@lemmy.ml 1 points 1 year ago (1 children)

Isn't it those that cost $2000 per 250mb of memory?? Meaning you'd about 350 to load any half decent model.

[–] scottmeme@sh.itjust.works 2 points 1 year ago

Not sure how they are doing it, but it was actually $20k not $2k for 250mb of memory on the card. I suspect the models are probably cached in system memory.

[–] Finadil@lemmy.world 1 points 1 year ago

That with a fp16 model? Don't be scared to try even a 4 bit quantization, you'd be surprised at how little is lost and how much quicker it is.