this post was submitted on 21 Feb 2024
279 points (95.1% liked)

Technology

60440 readers
3489 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 2 years ago
MODERATORS
all 32 comments
sorted by: hot top controversial new old
[–] [email protected] 47 points 10 months ago (3 children)

Direct link to the GitHub repo:
https://github.com/nickbild/local_llm_assistant?tab=readme-ov-file

It's a small model by comparison. If you want something that's offline and actually closer to comparing to ChatGPT 3.5, you'll want the Mixtral 8x7B model instead (running on a beefy machine):

https://mistral.ai/news/mixtral-of-experts/

[–] [email protected] 32 points 10 months ago (2 children)

Sick, I only need 90gb of VRAM!

[–] [email protected] 15 points 10 months ago (1 children)

I've got it running with a 3090 and 32GB of RAM.

There are some models that let you run with hybrid system RAM and VRAM (it will just be slower than running it exclusively with VRAM).

[–] [email protected] 16 points 10 months ago (1 children)

Yeah but damn does it get slow.

I always find it interesting how text is so much slower than image generation. I can do a 1024x1024 in probably 20s, but I get like 1 word a second with text.

[–] aBundleOfFerrets 5 points 10 months ago

Languages are complex and, more importantly, much less forgiving to error

[–] [email protected] 10 points 10 months ago (1 children)

Nice! Thats a cool project, ill have to give it a try. I love the idea of self hosting local LLMs. Ive been playing around with: https://lmstudio.ai/ and it directly downloads from hugging face.

[–] [email protected] 2 points 10 months ago

There's also ollama which seems to be similar. Not sure if LMStudio is open source but ollama is.

[–] anticurrent 28 points 10 months ago (1 children)

Can we have smaller more domain specific models. that shouldn't require more than casual hardware. like a small model for coding, one for medicine, one for history, and so on. ???

[–] fruitycoder 14 points 10 months ago (1 children)

Check out hugging face! Honestly fine tunned models for specific domains seems very popular (if for nothing else because training smaller models is just easier!).

[–] [email protected] 8 points 10 months ago (1 children)

*cannot function correctly without T-Mobile speaker.

[–] [email protected] 3 points 10 months ago

I cannot function with T-Mobile internet, that is for sure. I'm moving to another ISP

[–] [email protected] 3 points 10 months ago

That’s gonna be a no from me dawg

[–] mindbleach 2 points 10 months ago (1 children)

This is a big part of why I'm not worried about this wave of AI.

It was all trained on consumer hardware. Lots of it, yes, at great expense... but brute force keeps ceding ground to smaller models built on that experience. Google went from a monolithic Go bot trained on historical games, to a much smaller Go bot trained by playing that bot and itself, to an even smaller bot that plays a wide variety of games. It's just matrix math and we know we're doing it badly. The endgame is running Not Hotdog on a Game Boy Camera.

On the other side, the fact you can run these on anything means we're never going to stop it. This fight is over. Fantasies about Bing and OpenAI preventing anyone from rendering Bad Things™ only push people toward local models. Higher adoption creates a virtuous circle of streamlining and empowerment for anyone getting into the technology. And since porn was the first thing all these billion-dollar companies tried stopping, well, guess what any rando with a high-end GPU can crank out.

... phrasing.

[–] mindbleach 1 points 10 months ago

Oh right, forgot to mention: democratization will destroy most markets for what these programs crank out. You can't sell ice to people with refrigerators.