this post was submitted on 04 May 2024
289 points (95.6% liked)

Technology

59534 readers
3143 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 
  • Rabbit R1, AI gadget, runs on Android app, not requiring "very bespoke AOSP" firmware as claimed by Rabbit.
  • Rabbit R1 launcher app can run on existing Android phones, not needing system-level permissions for core functionality.
  • Rabbit R1 firmware analysis shows minimal modifications to standard AOSP, contradicting claims of custom hardware necessity by Rabbit.
you are viewing a single comment's thread
view the rest of the comments
[–] GlassHalfHopeful@lemmy.ca 14 points 6 months ago (14 children)

I've no skin in the game related to this device or software. I simply don't care. However, in terms of an AI assistant, I am curious if there is anything on the market, including this ~APK~ device, that is worth using. For someone who is privacy centric, advert avoidant, and security focused... does (or even can) such an app/tool exist?

[–] andrew0@lemmy.dbzer0.com 4 points 6 months ago (1 children)

What db2 already said. Microsoft just released Phi-3 mini, which could, allegedly, run locally on newer smartphones.

If I understood correctly, the Rabbit thingy just captures your information locally and then forwards it to their server. So, if you want more power, you could probably do the same by submitting the same info to a bigger open source model than Phi-3, like Llama 3, hosted on your homelab. I believe you can set it up with huggingface/gradio, which sort of provides an API that you could use.

That way, you don't need a shitty orange box, and can always get the latest open source models with a few lines of code. There are plenty of open source frameworks in the works at the moment, and I believe that we're not far off from having multi-modal LLMs running on homelab-level hardware (if you don't mind a bit of lag).

[–] GlassHalfHopeful@lemmy.ca 2 points 6 months ago (1 children)

huggingface! i found it once and never could remember again who hosted all those models. thank you!

i use a different device than homelab, but now I am curious what I may be able to achieve with my syno system. it's hw weak, so probably not a lot. but i would like to give it a go. if it's decent, i may consider another device for the purpose.

[–] andrew0@lemmy.dbzer0.com 3 points 6 months ago (1 children)

Good luck! You can try the huggingface-chat repo, or ollama with this web-ui. Both should be decent, as they have instructions to set up a docker container.

I believe the Llama 3 models are out there in a torrent somewhere, but I didn't dig to find it. For the 70B model, you'll probably need around 64GB of RAM available, but the 7B one should run fine with just 8GB. It will be somewhat slow though, compared to the ChatGPT experience. The self-attention mechanism can be parallelized, which is why you will see much better results on a GPU. According to some others that tested it, if you offload some stuff to RAM, you could see ~10-12 tokens per second on an RTX 3090 for certain 70B models. But more capable ones will be at less than 1 token per second, all depending on the context window you use.

If you don't have a GPU available, just give the Phi-3 model a try :D If you quantize it to 4 bits, it can apparently get 12 tokens per second on an iPhone haha. It should play nice with pooling information from a search engine, or a vector database like milvus, qdrant or chroma.

[–] GlassHalfHopeful@lemmy.ca 3 points 6 months ago

Thank you for all this! Much appreciated!

load more comments (12 replies)