this post was submitted on 09 Aug 2024
259 points (93.6% liked)

Selfhosted

40347 readers
401 users here now

A place to share alternatives to popular online services that can be self-hosted without giving up privacy or locking you into a service you don't control.

Rules:

  1. Be civil: we're here to support and learn from one another. Insults won't be tolerated. Flame wars are frowned upon.

  2. No spam posting.

  3. Posts have to be centered around self-hosting. There are other communities for discussing hardware or home computing. If it's not obvious why your post topic revolves around selfhosting, please include details to make it clear.

  4. Don't duplicate the full text of your blog or github here. Just post the link for folks to click.

  5. Submission headline should match the article title (don’t cherry-pick information from the title to fit your agenda).

  6. No trolling.

Resources:

Any issues on the community? Report it using the report flag.

Questions? DM the mods!

founded 1 year ago
MODERATORS
 

I don't consider myself very technical. I've never taken a computer science course and don't know python. I've learned some things like Linux, the command line, docker and networking/pfSense because I value my privacy. My point is that anyone can do this, even if you aren't technical.

I tried both LM Studio and Ollama. I prefer Ollama. Then you download models and use them to have your own private, personal GPT. I access it both on my local machine through the command line but I also installed Open WebUI in a docker container so I can access it on any device on my local network (I don't expose services to the internet).

Having a private ai/gpt is pretty cool. You can download and test new models. And it is private. Yes, there are ethical concerns about how the model got the training. I'm not minimizing those concerns. But if you want your own AI/GPT assistant, give it a try. I set it up in a couple of hours, and as I said... I'm not even that technical.

you are viewing a single comment's thread
view the rest of the comments
[–] superglue@lemmy.dbzer0.com 5 points 3 months ago (5 children)

What kinds of specs do you need to run it well? I've got a laptop with a 3070.

[–] coffee_with_cream@sh.itjust.works -3 points 3 months ago* (last edited 3 months ago) (3 children)

You probably want 48gb of vram or more to run the good stuff. I recommend renting GPU time instead of using your own hardware, via AWS or other vendors - runpod.io is pretty good.

[–] NotMyOldRedditName@lemmy.world 6 points 3 months ago

Kinda defeats the purpose of doing it private and local.

I wouldn't trust any claims a 3rd party service makes with regards to being private.

[–] theterrasque@infosec.pub 4 points 3 months ago

Llama3 8b can be run at 6gb vram, and it's fairly competent. Gemma has a 9b I think, which would also be worth looking into.

[–] 31337@sh.itjust.works 3 points 3 months ago (1 children)

IDK, looks like 48GB cloud pricing would be 0.35/hr => $255/month. Used 3090s go for $700. Two 3090s would give you 48GB of VRAM, and cost $1400 (I'm assuming you can do "model-parallel" will Llama; never tried running an LLM, but it should be possible and work well). So, the break-even point would be <6 months. Hmm, but if Severless works well, that could be pretty cheap. Would probably take a few minutes to process and load a ~48GB model every cold start though?

[–] fhein@lemmy.world 1 points 3 months ago

Assuming they already own a PC, if someone buys two 3090 for it they'll probably also have to upgrade their PSU so that might be worth including in the budget. But it's definitely a relatively low cost way to get more VRAM, there are people who run 3 or 4 RTX3090 too.

load more comments (1 replies)