this post was submitted on 30 Dec 2024
200 points (92.4% liked)

Technology

60239 readers
3405 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 2 years ago
MODERATORS
top 28 comments
sorted by: hot top controversial new old
[–] [email protected] 85 points 6 days ago

We need more people to code like this.

Everyone codes like it will be the only process running on their dev machine. Before you know it, you've opened a word processor, a mail client and a chat program. 8GB RAM is gone and the 3.2GHz 8-core CPU is idling at 35%.

[–] [email protected] 48 points 6 days ago

The story that this 260K parameter model generated (in the screenshot of their report):

Sleepy Joe said: "Hello, Spot. Do you want to be careful with me?" Spot told Spot, "Yes, I will step back!"
Spot replied, "I lost my broken broke in my cold rock. It is okay, you can't." Spoon and Spot went to the top of the rock and pulled his broken rock. Spot was sad because he was not l

It's still impressive that they got it running. I look forward to seeing if their BitNet architecture produces better results on the same hardware.

[–] [email protected] 37 points 6 days ago (2 children)

Any processor can run llms. The only issue is how fast, and how much ram it has access to. And you can trade the latter for disk space if you're willing to sacrifice even more speed.

If it can add, it can run any model

[–] [email protected] 6 points 5 days ago

Yes, and a big part of AI advancement is running it on leaner hardware, using less power, with more efficient models.

Not every team is working on building bigger with more resources. Showing off how much they can squeeze out of minimal hardware is an important piece of this.

[–] [email protected] 1 points 5 days ago* (last edited 5 days ago)

Yeah the Church-Turing thesis holds that you can run an LLM on a casio wrist watch (if for some reason you wanted to do that). I can't imagine this is exactly what you'd call 'good'...

[–] [email protected] 34 points 6 days ago (2 children)

Before we go, please note that EXO is still looking for help. If you also want to avoid the future of AI being locked into massive data centers owned by billionaires and megacorps and think you can contribute in some way, you could reach out.

Unfortunately Biden just signed an executive order allowing data centers to be built on federal land...

And Trump is 100% going to sell it off to the highest bidder and/or give a bunch to Elmo

[–] [email protected] 12 points 6 days ago

There's a huge bit of space between "more datacenter space" and "ah well, on-prem and self-host are dead". Like, this is a 2024-voter level of apathy.

[–] [email protected] -3 points 6 days ago (1 children)

Well......hold on a second. I was with you through most of that, but then you said Elmo at the end.

Maybe I'm stupid.....maybe there's another Elmo that makes WAAAAAY more logical sense. But what I'M envisioning is this army of AI Seaseme Street Elmo's harvesting your data, tracking/stalking you all day, and plotting the best time to come and tickle you.

I don't get it. What are you SAYING???

[–] [email protected] 21 points 6 days ago

Elongated Muskrat

[–] [email protected] 9 points 6 days ago (1 children)

This is like saying "my 30-year-old bike still works under very specific conditions"

[–] [email protected] 15 points 6 days ago (1 children)

You exist in a world where people overclock systems to eke a 3% more performance out of the metal, and somehow hammering some performance out of the software seems wasteful?

This kind of thinking seems to be a "slow? Throw more hardware at it" kind of mentality that I only see in ... Wait; you're a java programmer.

[–] sugar_in_your_tea 3 points 5 days ago

Or Python, don't leave us out.

[–] [email protected] 6 points 6 days ago (1 children)
[–] [email protected] 5 points 5 days ago

You can run AI on a smartwatch, it's just not going to be very intelligent. The fact that you can technically do it isn't necessarily very impressive.

[–] [email protected] 6 points 6 days ago (1 children)

If it can work in such a small space, think of what it can do in even a low-end android phone.

I don't need my phone to write me stories, but I would like it to notice when my flight is running late and either call up the customer support and book a new connection or get a refund (like Facebook's M was bizarrely adept at doing) or just let my contacts in the upcoming meeting know I'm a bit late.

If it searches my mail I'd like it to be on the device and never leverage a leaky, insecure cloud service.

With a trim setup like this, they've shown it's possible.

[–] [email protected] 9 points 6 days ago

Why would you need an LLM for that?

We have a standard, it's called RSS.

We have scripting. We also have visual scripting. That there's no customer tool for that ... is not customer's fault, but not a sign of some fundamental limitation either.

Customer support would, in fact, be more pleased with an e-mail from a template, and not a robot call (and it'll likely have robot detection and drop such calls anyway).

Informing your contacts is better done with a template too.

However, now when I think about it, if such a tool existed, it could use an LLM as a fallback in each case, where we don't have a better source of data about your flights, a fitting e-mail template, some point of that template lacking, or confusion in parsing the company page for support e-mail.

But that still would much rather be some "guesser" element in a visual script, one used when there's nothing more precise.

I think such a product could be more popular than just an LLM to which you say to do something and are never certain whether it's going to do a wildcard weirdness or it's going to be fine.

[–] [email protected] 4 points 5 days ago

What is this, Dr. Sbaitsbo?

[–] [email protected] 4 points 5 days ago

I installed gtp4all and tested all models available on it, it sucks ...

[–] [email protected] 5 points 6 days ago (1 children)

Let me know when someone gets an LLM to run on a Bendix G15. This Pentium stuff is way too new.

[–] [email protected] 5 points 6 days ago (1 children)

Me: Hmmmmmmmmm..............I only vaguely have an idea what's even being discussed here. They somehow got Windows 98 to run on a Llama. Probably picked a llama because if they used a camel it would BSOD.

[–] [email protected] 6 points 6 days ago

Username checks out

[–] mindbleach 2 points 6 days ago

Right: training is expensive. Neural networks are just matrix math. However much your hardware can do, a model can be trained to fit that hardware.

You could do Not Hotdog on a Game Boy Camera.

[–] [email protected] -4 points 6 days ago* (last edited 6 days ago) (1 children)

Imagine how much better it would run on a similar era version of redhat, gentoo, or beos.

They just proved that the hardware was perfectly capable, in the absolute garbage middle layer-the operating system is what matters about propelling the potential of the hardware forward into a usable form.

Many people may not remember, but there were a few Lins distributions around at the time. Certainly, they would have been able to make better use of the hardware had enough developers worked on it.

[–] Supernova1051 11 points 6 days ago (1 children)

but the hardware is not capable. it's running a miniscule custom 260k LLM and the "claim to fame" is that it wasn't slow. great? we already know tiny models are fast, they're just not as accurate and perform worse than larger models, all they did was make an even smaller than normal model. this is akin to getting Doom to run on anything with a CPU, while cool and impressive, it doesn't do much for anyone other than being an exercise in doing something because you can.