this post was submitted on 28 Jan 2025
41 points (88.7% liked)

Linux

49393 readers
1748 users here now

From Wikipedia, the free encyclopedia

Linux is a family of open source Unix-like operating systems based on the Linux kernel, an operating system kernel first released on September 17, 1991 by Linus Torvalds. Linux is typically packaged in a Linux distribution (or distro for short).

Distributions include the Linux kernel and supporting system software and libraries, many of which are provided by the GNU Project. Many Linux distributions use the word "Linux" in their name, but the Free Software Foundation uses the name GNU/Linux to emphasize the importance of GNU software, causing some controversy.

Rules

Related Communities

Community icon by Alpár-Etele Méder, licensed under CC BY 3.0

founded 5 years ago
MODERATORS
 

Hello, I've been hearing a lot about this new DeepSeek LLM, and was wondering, would it be possible to get the 600+ billion parameter model running on my GPU? I've heard something about people have got it to run on their MacBooks. I have i7 4790K, 32GB DDR3, and 7900 XTX 24GB VRAM. I'm running Arch Linux, this computer is just for AI stuff really, not gaming as much. I did tried running the distilled 14B parameter model, but it didn't work for me, I was using GPT4All to run it. I'm thinking about getting one of the NVIDIA 5090s in the future. Thanks in advance!

you are viewing a single comment's thread
view the rest of the comments
[–] gay4dudes@sh.itjust.works 15 points 2 days ago (3 children)

I run the 32b Version on my 6700xt with an R9 3700x using ollama. It runs well but it gets a bit slower on complex problems. I once ran an 70b Llama model, but it took a long time to finish.

[–] ieatpwns@lemmy.world 3 points 2 days ago (1 children)

Hey not to side track ops post or your own but I’m new to the home llm space and I was wondering once you have the model set up is there a gui? And how do you input tasks for it to do?

[–] gay4dudes@sh.itjust.works 3 points 2 days ago

You can use the Terminal or something like AnythingLLM. It has a GUI and you can import pictures and Websites.

[–] Fisch@discuss.tchncs.de 2 points 2 days ago

I have the same GPU but I always run 7B/8B variants as exl2. Do you use GGUF to use your system RAM?

[–] MIXEDUNIVERS@discuss.tchncs.de 1 points 2 days ago (2 children)

i also have a 6700xt but i don't get ollama running on it. it only defaults to the cpu ryzen 5600 I plan to tackle this problem on a free weekend and now i have a new Reason for solving it.

[–] EddyBot@discuss.tchncs.de 2 points 1 day ago* (last edited 1 day ago)

on some Linux distros like Arch Linux you might need to install a ollama-rocm package too

[–] gay4dudes@sh.itjust.works 1 points 2 days ago

Well, I dont know what you are running, but on Debian or Fedora it automatically installed Drivers and picked the GPU. I had a Problem like this ones, where it had wrong Drivers (but it was in an NVIDIA GPU).